The Yuan requests your support! Our content will now be available free of charge for all registered subscribers, consistent with our mission to make AI a human commons accessible to all. We are therefore requesting donations from our readers so we may continue bringing you insightful reportage of this awesome technology that is sweeping the world. Donate now
Multimodal AI for Healthcare
By Kirk Borne  |  Dec 15, 2021
Multimodal AI for Healthcare
Image courtesy of and under license from Shutterstock.com
Multimodal artificial intelligence is the new kid on the AI block. It uses several data types, such as image, text, speech, and numerical data, and combines them with multiple intelligence processing algorithms to achieve higher performance. Data scientist and astrophysicist Dr Kirk Borne applies some context to multimodal AI’s ability to sense, understand, assimilate, synergize, and learn from multiple diverse inputs.

COLUMBIA, MARYLAND - Multimodal artificial intelligence derives its name from the multimodal datasets used to train and engage the artificial intelligence (AI). Those modes include images (still images and video), language (spoken and written words), and numerical data, primarily from databases, but also labeled values extracted from documents and images. 

Numerical data itself comes in different modes, e.g., analog data (non-digital readings on devices, perhaps recorded manually in logs), time series, tabular data, map data (points, lines, polygons), and chart data (distributions, histograms). 

Qualitative data, which can be any of those three modes just mentioned, also exists, but qualitative data itself is multimodal - collected and presented in different ways, such as words, multiple-choice responses to survey questions, sentiment, emotion, intent, and context.

In healthcare, a number of multimodal data collections can be used in AI applications, including medical imaging, patient photos, verbal and written notes from physicians, lab results, patient surveys, health diagnostics, electronic health records (EHR), electronic medical records (EMR), and much more. 

Emerging AI applications that require ingestion, processing, and decisioning on such multiple diverse data types will necessarily be more complex than unimodal (single data format) AI applications, such as a medical imaging diagnosis or a conversational AI chatbot for answering patients’ questions. Those complexity challenges are outweighed by the significant benefits that multimodal AI applications can bring to healthcare, including clinical care, diagnoses, treatments, and anomaly discovery. 

This article discusses multimodal AI, and includes a short tutorial, examples, and some healthcare applications.


Sensational Systems

The content herein is subject to copyright by The Yuan. All rights reserved. The content of the services is owned or licensed to The Yuan. Such content from The Yuan may be shared and reprinted but must clearly identify The Yuan as its original source. Content from a third-party copyright holder identified in the copyright notice contained in such third party’s content appearing in The Yuan must likewise be clearly labeled as such.
Continue reading
Sign up now to read this story for free.
- or -
Continue with Linkedin Continue with Google
Comments
Share your thoughts.
The Yuan wants to hear your voice. We welcome your on-topic commentary, critique, and expertise. All comments are moderated for civility.