AWS GenAI in Health Care Impressions
Yesterday, I attended the AWS GenAI for Healthcare Summit in Lisbon, hosted by the Champalimaud Foundation
Here’s my impressions for the day.
Where is the Medical Sector using (Gen)AI
A non-exhaustive list:
-
Clinical Decision Support
-
AI Buddy in Diagnostics
-
Enhancing Imaging
For example, using AI to adjust X-ray imaging when there’s missing data, low resolution, cropping, or low contrast.
-
Data Storage Management
Some institutions have kept all their digital X-ray images. Since storage is so costly they started leveraging AI to discard 70% of their data and reconstruct when needed from the remaining 30% without information loss.
-
Reducing Invasive Examinations
Research suggests up to 5% of all cancers may be linked to exposure from CT scans. To avoid it, less precise scans can be used, making diagnosis harder. TheraPanacea trained a variational autoencoder on 250,000 CT scans to generate high-quality images from lower-quality scans. The demo results were impressive!
I found it fascinating that 3–5 are “human-out-of-the-loop” applications—offering scalable, software-like gains!
The most moving example, however, was a highly “human-in-the-loop,” non-scalable solution. In Africa, there’s a shortage of radiologists and widespread tuberculosis. To help, a new portable X-ray box with built-in AI models is being used. It’s transported village-to-village by motorbike. Everyone in the village can be scanned, and if the AI flags someone’s scan, they’re referred to a hospital for specialized care. The motorcyclist doesn’t need medical training—the X-ray box and AI do the work!
Challenges
Customized/fine-tuned foundational models for the medical subdomain is a must. That is, not only should models be specialized in health care: they should be specialized in the discipline within health care where they will be deployed. Why?
- Medical accuracy (goes without saying)
- Clinical relevance
- Direct access and quoting medical sources and evidence
- Healthcare context understanding (steering the model towards prioritizing clinical info—which is something the models should be able to do already, but I suspect that there’s a lot of improvised medical advice on the internet that a non-specialized model won’t be steered enough by prompting alone)
- A didactic approach in the replies geared towards what health professionals are used to. I cannot validate this, but I’ve heard that general models tends to be terser in their replies by default. I don’t know by which measure prompting can address it!
All these items contribute to trust, one of the big challenges in deploying AI in the medical sector. And with trust, adoption follows, and, from adoption, impact.
Which brings me to the observation that if AI doesn’t become operational, there’s no value. You could tell that health care professionals are not looking for the n-th pilot of PoC. They’re looking for stuff that makes their job easier, improve patients’ lives, and doesn’t require an Einstein to operate.
Closing thoughts
To close: everybody was screaming agents but there were no agentic demos.
Maybe next year?