Skip to content
GCC AI Research

Search

Results for "domain shift"

On Transferability of Machine Learning Models

MBZUAI ·

This article discusses domain shift in machine learning, where testing data differs from training data, and methods to mitigate it via domain adaptation and generalization. Domain adaptation uses labeled source data and unlabeled target data. Domain generalization uses labeled data from single or multiple source domains to generalize to unseen target domains. Why it matters: Research in mitigating domain shift enhances the robustness and applicability of AI models in diverse real-world scenarios.

Adapting to Distribution Shifts: Recent Advances in Importance Weighting Methods

MBZUAI ·

This article discusses distribution shifts in machine learning and the use of importance weighting methods to address them. Masashi Sugiyama from the University of Tokyo and RIKEN AIP presented recent advances in importance-based distribution shift adaptation methods. The talk covered joint importance-predictor estimation, dynamic importance weighting, and multistep class prior shift adaptation. Why it matters: Understanding and mitigating distribution shifts is crucial for deploying robust and reliable AI models in real-world scenarios within the GCC region and beyond.

DGM-DR: Domain Generalization with Mutual Information Regularized Diabetic Retinopathy Classification

arXiv ·

This paper introduces a domain generalization (DG) method for Diabetic Retinopathy (DR) classification that maximizes mutual information using a large pretrained model. The method aims to address the challenge of domain shift in medical imaging caused by variations in data acquisition. Experiments on public datasets demonstrate that the proposed method outperforms state-of-the-art techniques, achieving a 5.25% improvement in average accuracy.

Towards Robust Multimodal Open-set Test-time Adaptation via Adaptive Entropy-aware Optimization

arXiv ·

This paper introduces Adaptive Entropy-aware Optimization (AEO), a new framework to tackle Multimodal Open-set Test-time Adaptation (MM-OSTTA). AEO uses Unknown-aware Adaptive Entropy Optimization (UAE) and Adaptive Modality Prediction Discrepancy Optimization (AMP) to distinguish unknown class samples during online adaptation by amplifying the entropy difference between known and unknown samples. The study establishes a new benchmark derived from existing datasets with five modalities and evaluates AEO's performance across various domain shift scenarios, demonstrating its effectiveness in long-term and continual MM-OSTTA settings.

YaPO: Learnable Sparse Activation Steering Vectors for Domain Adaptation

arXiv ·

The paper introduces Yet another Policy Optimization (YaPO), a reference-free method for learning sparse steering vectors in the latent space of a Sparse Autoencoder (SAE) to steer LLMs. By optimizing sparse codes, YaPO produces disentangled, interpretable, and efficient steering directions. Experiments show YaPO converges faster, achieves stronger performance, exhibits improved training stability and preserves general knowledge compared to dense steering baselines.

ConDiSR: Contrastive Disentanglement and Style Regularization for Single Domain Generalization

arXiv ·

This paper introduces a new Single Domain Generalization (SDG) method called ConDiSR for medical image classification, using channel-wise contrastive disentanglement and reconstruction-based style regularization. The method is evaluated on multicenter histopathology image classification, achieving a 1% improvement in average accuracy compared to state-of-the-art SDG baselines. Code is available at https://github.com/BioMedIA-MBZUAI/ConDiSR.

Domain Adaptable Fine-Tune Distillation Framework For Advancing Farm Surveillance

arXiv ·

The paper introduces a framework for camel farm monitoring using a combination of automated annotation and fine-tune distillation. The Unified Auto-Annotation framework uses GroundingDINO and SAM to automatically annotate surveillance video data. The Fine-Tune Distillation framework then fine-tunes student models like YOLOv8, transferring knowledge from a larger teacher model, using data from Al-Marmoom Camel Farm in Dubai.

Closing the gap for real-world AI deployment in healthcare

MBZUAI ·

MBZUAI doctoral student Umaima Rahman is researching domain adaptation and generalization in deep learning for medical imaging to improve AI model performance across diverse hospitals and equipment. Her work focuses on building models that learn consistent features across different data sources to ensure reliability in various healthcare settings. Rahman emphasizes that generalization in healthcare AI is a necessity, especially in resource-limited settings, and aims to develop AI that assists clinicians rather than replaces them. Why it matters: This research addresses a critical challenge in deploying AI in healthcare, ensuring that models can be reliably used in diverse settings, particularly benefiting developing countries and improving global healthcare accessibility.