Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More
MRI images are understandably complex and data-heavy.ย
Because of this, developers training large language models (LLMs) for MRI analysis have had to slice captured images into 2D. But this results in just an approximation of the original image, thus limiting the modelโs ability to analyze intricate anatomical structures. This creates challenges in complex cases involving brain tumors, skeletal disorders or cardiovascular diseases.ย
But GE Healthcare appears to have overcome this massive hurdle, introducing the industryโs first full-body 3D MRI research foundation model (FM) at this yearโs AWS re:Invent. For the first time, models can use full 3D images of the entire body.ย
GE Healthcareโs FM was built on AWS from the ground up โ there are very few models specifically designed for medical imaging like MRIs โ and is based on more than 173,000 images from over 19,000 studies. Developers say they have been able to train the model with five times less compute than previously required.ย
GE Healthcare has not yet commercialized the foundation model; it is still in an evolutionary research phase. An early evaluator, Mass General Brigham, is set to begin experimenting with it soon.ย
โOur vision is to put these models into the hands of technical teams working in healthcare systems, giving them powerful tools for developing research and clinical applications faster, and also more cost-effectively,โ GE HealthCare chief AI officer Parry Bhatia told VentureBeat.ย
Enabling real-time analysis of complex 3D MRI data
While this is a groundbreaking development, generative AI and LLMs are not new territory for the company. The team has been working with advanced technologies for more than 10 years, Bhatia explained.ย
One of its flagship products is AIR Recon DL, a deep learning-based reconstruction algorithm that allows radiologists to more quickly achieve crisp images. The algorithm removes noise from raw images and improves signal-to-noise ratio, cutting scan times by up to 50%. Since 2020, 34 million patients have been scanned with AIR Recon DL.ย
GE Healthcare began working on its MRI FM at the beginning of 2024. Because the model is multimodal, it can support image-to-text searching, link images and words, and segment and classify diseases. The goal is to give healthcare professionals more details in one scan than ever before, said Bhatia, leading to faster, more accurate diagnosis and treatment.
โThe model has significant potential to enable real-time analysis of 3D MRI data, which can improve medical procedures like biopsies, radiation therapy and robotic surgery,โ Dan Sheeran, GM for health care and life sciences at AWS, told VentureBeat.ย
Already, it has outperformed other publicly-available research models in tasks including classification of prostate cancer and Alzheimerโs disease. It has exhibited accuracy up to 30% in matching MRI scans with text descriptions in image retrieval โ which might not sound all that impressive, but itโs a big improvement over the 3% capability exhibited by similar models.ย
โIt has come to a stage where itโs giving some really robust results,โ said Bhatia. โThe implications are huge.โ
Doing more with (much less) data
The MRI process requires a few different types of datasets to support various techniques that map the human body, Bhatia explained.ย
Whatโs known as a T1-weighted imaging technique, for instance, highlights fatty tissue and decreases the signal of water, while T2-weighted imaging enhances water signals. The two methods are complementary and create a full picture of the brain to help clinicians detect abnormalities like tumors, trauma or cancer.ย
โMRI images come in all different shapes and sizes, similar to how you would have books in different formats and sizes, right?โ said Bhatia.ย
To overcome challenges presented by diverse datasets, developers introduced a โresize and adaptโ strategy so that the model could process and react to different variations. Also, data may be missing in some areas โ an image may be incomplete, for instance โ so they taught the model simply to ignore those instances.ย
โInstead of getting stuck, we taught the model to skip over the gaps and focus on what was available,โ said Bhatia. โThink of this as solving a puzzle with some missing pieces.โ
The developers also employed semi-supervised student-teacher learning, which is particularly helpful when there is limited data. With this method, two different neural networks are trained on both labeled and unlabeled data, with the teacher creating labels that help the student learn and predict future labels.ย
โWeโre now using a lot of these self-supervised technologies, which donโt require huge amounts of data or labels to train large models,โ said Bhatia. โIt reduces the dependencies, where you can learn more from these raw images than in the past.โ
This helps to ensure that the model performs well in hospitals with fewer resources, older machines and different kinds of datasets, Bhatia explained.ย
He also underscored the importance of the modelsโ multimodality. โA lot of technology in the past was unimodal,โ said Bhatia. โIt would look only into the image, into the text. But now theyโre becoming multi-modal, they can go from image to text, text to image, so that you can bring in a lot of things that were done with separate models in the past and really unify the workflow.โย
He emphasized that researchers only use datasets that they have rights to; GE Healthcare has partners who license de-identified data sets, and theyโre careful to adhere to compliance standards and policies.
Using AWS SageMaker to tackle computation, data challenges
Undoubtedly, there are many challenges when building such sophisticated models โ such as limited computational power for 3D images that are gigabytes in size.
โItโs a massive 3D volume of data,โ said Bhatia. โYou need to bring it into the memory of the model, which is a really complex problem.โ
To help overcome this, GE Healthcare built on Amazon SageMaker, which provides high-speed networking and distributed training capabilities across multiple GPUs, and leveraged Nvidia A100 and tensor core GPUs for large-scale training.ย
โBecause of the size of the data and the size of the models, they cannot send it into a single GPU,โ Bhatia explained. SageMaker allowed them to customize and scale operations across multiple GPUs that could interact with one another.ย
Developers also used Amazon FSx in Amazon S3 object storage, which allowed for faster reading and writing for datasets.ย
Bhatia pointed out that another challenge is cost optimization; with Amazonโs elastic compute cloud (EC2), developers were able to move unused or infrequently used data to lower-cost storage tiers.ย
โLeveraging Sagemaker for training these large models โ mainly for efficient, distributed training across multiple high-performance GPU clusters โ was one of the critical components that really helped us to move faster,โ said Bhatia.ย
He emphasized that all components were built from a data integrity and compliance perspective that took into account HIPAA and other regulatory regulations and frameworks.ย
Ultimately, โthese technologies can really streamline, help us innovate faster, as well as improve overall operational efficiencies by reducing the administrative load, and eventually drive better patient care โ because now youโre providing more personalized care.โ
Serving as a basis for other specialized fine-tuned models
While the model for now is specific to the MRI domain, researchers see great opportunities to expand into other areas of medicine.ย
Sheeran pointed out that, historically, AI in medical imaging has been constrained by the need to develop custom models for specific conditions in specific organs, requiring expert annotation for each image used in training.ย
But that approach is โinherently limitedโ due to the different ways diseases manifest across individuals, and introduces generalizability challenges.ย
โWhat we truly need is thousands of such models and the ability to rapidly create new ones as we encounter novel information,โ he said. High-quality labeled datasets for each model are also essential.ย
Now with generative AI, instead of training discrete models for each disease/organ combination, developers can pre-train a single foundation model that can serve as a basis for other specialized fine-tuned models downstream.ย
For instance, GE Healthcareโs model could be expanded into areas such as radiation therapy, where radiologists spend significant time manually marking organs that might be at risk. It could also help reduce scan time during x-rays and other procedures that currently require patients to sit still in a machine for extended periods, said Bhatia.ย
Sheeran marveled that โweโre not just expanding access to medical imaging data through cloud-based tools; weโre changing how that data can be utilized to drive AI advancements in healthcare.โ
source: https://venturebeat.com/ai/learn-how-ge-healthcare-used-aws-to-build-a-new-ai-model-that-interprets-mris/

