Skip to content
GCC AI Research

RP-SAM2: Refining Point Prompts for Stable Surgical Instrument Segmentation

arXiv · · Significant research

Summary

Researchers from MBZUAI introduced RP-SAM2, a method to improve surgical instrument segmentation by refining point prompts for more stable results. RP-SAM2 uses a novel shift block and compound loss function to reduce sensitivity to point prompt placement, improving segmentation accuracy in data-constrained settings. Experiments on the Cataract1k and CaDIS datasets show that RP-SAM2 enhances segmentation accuracy and reduces variance compared to SAM2, with code available on GitHub.

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

SALT: Parameter-Efficient Fine-Tuning via Singular Value Adaptation with Low-Rank Transformation

arXiv ·

Researchers introduce SALT, a parameter-efficient fine-tuning method for medical image segmentation that combines singular value adaptation with low-rank transformation. SALT selectively adapts influential singular values and complements this with a low-rank update for the remaining subspace. Experiments on five medical datasets show SALT outperforms state-of-the-art PEFT methods by 2-5% in Dice score with only 3.9% trainable parameters.

Adapting foundation models for medical image segmentation: a new approach presented at MICCAI

MBZUAI ·

MBZUAI researchers developed a method to adapt Meta's Segment Anything Model (SAM) for medical image segmentation, addressing its performance gap with natural images. Their approach improves SAM's accuracy without requiring extensive retraining or large medical image datasets. The research, led by Chao Qin, was nominated for the Best Paper Award at the MICCAI conference in Marrakesh. Why it matters: This offers a more efficient and effective way to leverage foundation models in specialized medical imaging applications, potentially improving diagnostic accuracy and reducing the need for large-scale, domain-specific training data.

UAE: Universal Anatomical Embedding on Multi-modality Medical Images

arXiv ·

Researchers propose a universal anatomical embedding (UAE) framework for medical image analysis to learn appearance, semantic, and cross-modality anatomical embeddings. UAE incorporates semantic embedding learning with prototypical contrastive loss, a fixed-point-based matching strategy, and an iterative approach for cross-modality embedding learning. The framework was evaluated on landmark detection, lesion tracking and CT-MRI registration tasks, outperforming existing state-of-the-art methods.

Image- and AI-guided robotics for minimally invasive surgery

MBZUAI ·

Researchers have developed robotic path-planning and control algorithms for minimally invasive surgery (MIS) that steer flexible needles, incorporating teleoperation and haptic feedback. An AI algorithm was designed to predict target motion due to respiratory movement, improving needle placement accuracy. GANs were used to generate synthetic images visualizing organ and tumor motion. Why it matters: This research demonstrates the potential of AI and robotics to enhance precision and adaptability in MIS, potentially reducing patient trauma and improving recovery times in the region and beyond.