Middle East AI

This Week arXiv

ConDiSR: Contrastive Disentanglement and Style Regularization for Single Domain Generalization

arXiv · · Significant research

Summary

This paper introduces a new Single Domain Generalization (SDG) method called ConDiSR for medical image classification, using channel-wise contrastive disentanglement and reconstruction-based style regularization. The method is evaluated on multicenter histopathology image classification, achieving a 1% improvement in average accuracy compared to state-of-the-art SDG baselines. Code is available at https://github.com/BioMedIA-MBZUAI/ConDiSR.

Keywords

Domain Generalization · Medical Image Classification · Contrastive Learning · Disentanglement · Histopathology

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

DGM-DR: Domain Generalization with Mutual Information Regularized Diabetic Retinopathy Classification

arXiv ·

This paper introduces a domain generalization (DG) method for Diabetic Retinopathy (DR) classification that maximizes mutual information using a large pretrained model. The method aims to address the challenge of domain shift in medical imaging caused by variations in data acquisition. Experiments on public datasets demonstrate that the proposed method outperforms state-of-the-art techniques, achieving a 5.25% improvement in average accuracy.

Towards Robust Multimodal Open-set Test-time Adaptation via Adaptive Entropy-aware Optimization

arXiv ·

This paper introduces Adaptive Entropy-aware Optimization (AEO), a new framework to tackle Multimodal Open-set Test-time Adaptation (MM-OSTTA). AEO uses Unknown-aware Adaptive Entropy Optimization (UAE) and Adaptive Modality Prediction Discrepancy Optimization (AMP) to distinguish unknown class samples during online adaptation by amplifying the entropy difference between known and unknown samples. The study establishes a new benchmark derived from existing datasets with five modalities and evaluates AEO's performance across various domain shift scenarios, demonstrating its effectiveness in long-term and continual MM-OSTTA settings.

OmniGen: Unified Multimodal Sensor Generation for Autonomous Driving

arXiv ·

The paper introduces OmniGen, a unified framework for generating aligned multimodal sensor data for autonomous driving using a shared Bird's Eye View (BEV) space. It uses a novel generalizable multimodal reconstruction method (UAE) to jointly decode LiDAR and multi-view camera data through volume rendering. The framework incorporates a Diffusion Transformer (DiT) with a ControlNet branch to enable controllable multimodal sensor generation, demonstrating good performance and multimodal consistency.

Domain Adaptable Fine-Tune Distillation Framework For Advancing Farm Surveillance

arXiv ·

The paper introduces a framework for camel farm monitoring using a combination of automated annotation and fine-tune distillation. The Unified Auto-Annotation framework uses GroundingDINO and SAM to automatically annotate surveillance video data. The Fine-Tune Distillation framework then fine-tunes student models like YOLOv8, transferring knowledge from a larger teacher model, using data from Al-Marmoom Camel Farm in Dubai.