Skip to content
GCC AI Research

Search

Results for "MorphDiff"

Teaching AI to predict what cells will look like before running any experiments

MBZUAI ·

MBZUAI researchers have developed MorphDiff, a diffusion model that predicts cell morphology from gene expression data. MorphDiff uses the transcriptome to generate realistic post-perturbation images, either from scratch or by transforming a control image. The model combines a Morphology Variational Autoencoder (MVAE) with a Latent Diffusion Model, enabling both gene-to-image generation and image-to-image transformation. Why it matters: This could significantly accelerate drug discovery and biological research by allowing scientists to preview cellular changes before conducting experiments.

SemDiff: Generating Natural Unrestricted Adversarial Examples via Semantic Attributes Optimization in Diffusion Models

arXiv ·

This paper introduces SemDiff, a novel method for generating unrestricted adversarial examples (UAEs) by exploring the semantic latent space of diffusion models. SemDiff uses multi-attribute optimization to ensure attack success while preserving the naturalness and imperceptibility of generated UAEs. Experiments on high-resolution datasets demonstrate SemDiff's superior performance compared to state-of-the-art methods in attack success rate and imperceptibility, while also evading defenses.

When AI stops playing “spot the difference” and starts understanding changes in MRIs

MBZUAI ·

MBZUAI researchers presented DEFUSE-MS at MICCAI 2025, a novel AI system for analyzing changes in MRI scans of multiple sclerosis (MS) patients. DEFUSE-MS uses a deformation field-guided spatiotemporal graph-based framework to identify new lesions by reasoning about how the brain has changed. The model constructs graphs of small regions within baseline and follow-up MRIs, linking them across time with edges enriched with learned embeddings of the deformation field. Why it matters: DEFUSE-MS reframes the task from simple "spot the difference" to understanding structural changes, potentially improving the speed and accuracy of MS diagnosis and treatment monitoring.

Towards Unified and Lossless Latent Space for 3D Molecular Latent Diffusion Modeling

arXiv ·

The paper introduces UAE-3D, a multi-modal VAE for 3D molecule generation that compresses molecules into a unified latent space, maintaining near-zero reconstruction error. This approach simplifies latent diffusion modeling by eliminating the need to handle multi-modality and equivariance separately. Experiments on GEOM-Drugs and QM9 datasets show UAE-3D establishes new benchmarks in de novo and conditional 3D molecule generation, with significant improvements in efficiency and quality.

A Glass Bead Game of *-ology: Contemporary Computational Approaches to Linguistic Morphology, Typology and Social Psychology

MBZUAI ·

Ekaterina Vylomova from the University of Melbourne gave a talk on using NLP models to advance research in linguistic morphology, typology, and social psychology. The talk covered using models to study morphology, phonetic changes in words over time, and diachronic changes in language semantics. Vylomova presented the UniMorph project, a cross-lingual annotation schema and database with morphological paradigms for over 150 languages. Why it matters: This research demonstrates the potential of NLP to contribute to a deeper understanding of language evolution and structure, with applications in linguistic research and the study of social and cultural changes.

VENOM: Text-driven Unrestricted Adversarial Example Generation with Diffusion Models

arXiv ·

The paper introduces VENOM, a text-driven framework for generating high-quality unrestricted adversarial examples using diffusion models. VENOM unifies image content generation and adversarial synthesis into a single reverse diffusion process, enhancing both attack success rate and image quality. The framework incorporates an adaptive adversarial guidance strategy with momentum to ensure the generated adversarial examples align with the distribution of natural images.

Deep Surface Meshes

MBZUAI ·

Pascal Fua from EPFL presented an approach to implementing convolutional neural nets that output complex 3D surface meshes. The method overcomes limitations in converting implicit representations to explicit surface representations. Applications include single view reconstruction, physically-driven shape optimization, and bio-medical image segmentation. Why it matters: This research advances geometric deep learning by enabling end-to-end trainable models for 3D surface mesh generation, with potential impact on various applications in computer vision and biomedical imaging in the region.

Diffusion-BBO: Diffusion-Based Inverse Modeling for Online Black-Box Optimization

arXiv ·

This paper introduces Diffusion-BBO, a new online black-box optimization (BBO) framework that uses a conditional diffusion model as an inverse surrogate model. The framework employs an Uncertainty-aware Exploration (UaE) acquisition function to propose scores in the objective space for conditional sampling. The approach is shown theoretically to achieve a near-optimal solution and empirically outperforms existing online BBO baselines across 6 scientific discovery tasks.