Skip to content
GCC AI Research

BiMediX2: Bio-Medical EXpert LMM for Diverse Medical Modalities

arXiv · · Significant research

Summary

MBZUAI releases BiMediX2, a bilingual (Arabic-English) Bio-Medical Large Multimodal Model, along with the BiMed-V dataset (1.6M samples) and BiMed-MBench evaluation benchmark. BiMediX2 supports multi-turn conversation in Arabic and English and handles diverse medical imaging modalities. The model achieves state-of-the-art results on medical LLM and LMM benchmarks, outperforming existing methods and GPT-4 in specific evaluations.

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

BiMediX: Bilingual Medical Mixture of Experts LLM

arXiv ·

MBZUAI researchers introduce BiMediX, a bilingual (English and Arabic) mixture of experts LLM for medical applications. The model is trained on BiMed1.3M, a new 1.3 million bilingual instruction dataset and outperforms existing models like Med42 and Jais-30B on medical benchmarks. Code and models are available on Github.

MedPromptX: Grounded Multimodal Prompting for Chest X-ray Diagnosis

arXiv ·

The paper introduces MedPromptX, a clinical decision support system using multimodal large language models (MLLMs), few-shot prompting (FP), and visual grounding (VG) for chest X-ray diagnosis, integrating imagery with EHR data. MedPromptX refines few-shot data dynamically for real-time adjustment to new patient scenarios and narrows the search area in X-ray images. The study introduces MedPromptX-VQA, a new visual question answering dataset, and demonstrates state-of-the-art performance with an 11% improvement in F1-score compared to baselines.

UniMed-CLIP: Towards a Unified Image-Text Pretraining Paradigm for Diverse Medical Imaging Modalities

arXiv ·

MBZUAI researchers introduce UniMed-CLIP, a unified Vision-Language Model (VLM) for diverse medical imaging modalities, trained on the new large-scale, open-source UniMed dataset. UniMed comprises over 5.3 million image-text pairs across six modalities: X-ray, CT, MRI, Ultrasound, Pathology, and Fundus, created using LLMs to transform classification datasets into image-text formats. UniMed-CLIP significantly outperforms existing generalist VLMs and matches modality-specific medical VLMs in zero-shot evaluations, improving over BiomedCLIP by +12.61 on average across 21 datasets while using 3x less training data.

MBZUAI’s bilingual healthcare model wins Meta award ahead of GITEX showcase

MBZUAI ·

MBZUAI's BiMediX2, a bilingual healthcare multi-modal model, won Meta's Llama Impact Innovation Award for its potential in solving healthcare accessibility challenges across the Middle East and Africa. Built using Llama 3.1, the model understands medical queries in both English and Arabic, interprets medical images, and is integrated as a chatbot on Telegram with speech functionality. The model was also presented at the AI for Sustainable Development Platform Launch Event and integrated into the UNDP for telemedicine. Why it matters: The model's bilingual capabilities and accessibility on low-cost devices and speech-based interaction have potential to improve healthcare access for marginalized populations in the region.