Vivek Natarajan Profile picture
Jul 27, 2023 13 tweets 4 min read Read on X
Medicine is inherently multimodal.

Thrilled to share Med-PaLM M, the first demonstration of a generalist multimodal biomedical AI system with a stellar team @GoogleAI @GoogleDeepMind @GoogleHealth

Paper: https://t.co/ZgEtG0gXEsarxiv.org/pdf/2307.14334…
Med-PaLM M is a large multimodal generative model that flexibly encodes and interprets biomedical data spanning clinical language, medical imaging, genomics and more performing competently on a diverse array of tasks - all with the same set of model weights. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi Med-PaLM M is built by fine tuning and aligning PaLM-E - an embodied multimodal language model from @GoogleAI - to the biomedical domain using MultiMedBench, a newly curated open source biomedical benchmark. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI MutliMedBench spans 7 biomedical modalities and 14 diverse tasks such as medical QA, radiology report generation, genomic variant calling and more with over 1 milliion samples.

We hope the curation of MultiMedBench will further spur the development of generalist biomedical AI. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI Across all tasks in MultiMedBench, Med-PaLM M reaches performance competitive or exceeding SOTA, often exceeding specialist models by a wide margin.

Further, Med-PaLM M also significantly outperforms PaLM-E demonstrating the importance of biomedical fine tuning and alignment. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI A key intuition for building large scale generalist biomedical AI with language as a common grounding across tasks is the possibility of combinatorial generalization and positive task transfer.
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI Towards that, we find preliminary but exciting evidence that Med-PaLM M can generalize to novel medical tasks and concepts and perform zero-shot multimodal reasoning - all in a zero-shot fashion only through language based instructions and prompts.
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI For eg, we find that Med-PaLM M can accurately identify and describe TB in chest x-rays despite having never encountered presentations of the disease before in images - only through language based instructions and prompts. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI Finally, to understand the clinical applicability of Med-PaLM M, we conducted a radiologist evaluation of AI generated reports across model scales.

Clinically significant error rate for Med-PaLM M on par with radiologists from prior studies suggesting potential clinical utility. Image
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI Our paper has more experiments, details and insights.
Would love your feedback -

Zooming out, generalist biomedical AI is a dream many of us at @GoogleAI @GoogleHealth @GoogleDeepMind have been building towards for years.arxiv.org/pdf/2307.14334…
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI @GoogleHealth @GoogleDeepMind The possibilities with such generalist biomedical AI that can encode the biomedical universe is limitless with applications spanning the continuum of scientific biomedical discovery to care delivery.

The future of AI in medicine and bio is incredibly exciting!
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI @GoogleHealth @GoogleDeepMind Huge props to @taotu831 for driving this work with relentless energy and optimism together with stellar teammates @AziziShekoofeh @DannyDriess @HardyShakerman @peteflorence @thekaransinghal @alan_karthi @pichuan across @GoogleAI @GoogleDeepMind @GoogleHealth (1/2)
@taotu831 @AziziShekoofeh @HardyShakerman @peteflorence @DannyDriess @thekaransinghal @alan_karthi @GoogleAI @GoogleHealth @GoogleDeepMind @pichuan Mohamed Amin, Sara Mahdavi, Chris Semturs, Joelle Barral, @skornblith @acarroll_ATG @_basilM @RyutaroTanno @achowdhery @greg_corrado @ymatias and many more we have the privilege of working with (2/2)

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Vivek Natarajan

Vivek Natarajan Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @vivnat

May 17, 2023
Pleased to share our latest @GoogleAI @GoogleHealth pre-print on Med-PaLM 2 where we detail our progress towards physician expert-level medical question answering performance!

Link: arxiv.org/pdf/2305.09617… Image
Med-PaLM 2 builds on top of PaLM 2, the latest SOTA LLM from @GoogleAI

We further finetune / align the model to the requirements of the medical domain and introduce ensemble refinement as a simple new prompting strategy to improve the model’s medical reasoning capabilities. Image
Med-PaLM 2 approaches or exceeds SOTA on multiple medical question answering benchmarks.

In particular, it reaches a score of 86.5% on MEDQA (USMLE) dataset matching performance of expert test takers (>18% over Med-PaLM).

It also reaches a new SOTA of 81.8% on PubMedQA. Image
Read 10 tweets
Dec 27, 2022
Delighted to share our new @GoogleHealth @GoogleAI @Deepmind paper at the intersection of LLMs + health.

Our LLMs building on Flan-PaLM reach SOTA on multiple medical question answering datasets including 67.6% on MedQA USMLE (+17% over prior work).

arxiv.org/abs/2212.13138
In addition to objective metrics, we pilot a framework for clinician/layperson eval of answers revealing key gaps in Flan-PaLM responses.

To resolve this, we use instruction prompting tuning to further align LLMs to the medical domain and generate safe, helpful answers.
Med-PaLM performs encouragingly on several axes such as scientific and clinical precision, reading comprehension, recall of medical knowledge, medical reasoning and utility compared to Flan-PaLM.
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(