Skip to content
GCC AI Research

Search

Results for "unsupervised learning"

Upsampling Autoencoder for Self-Supervised Point Cloud Learning

arXiv ·

This paper introduces a self-supervised learning method for point cloud analysis using an upsampling autoencoder (UAE). The model uses subsampling and an encoder-decoder architecture to reconstruct the original point cloud, learning both semantic and geometric information. Experiments show the UAE outperforms existing methods in shape classification, part segmentation, and point cloud upsampling tasks.

Self-Supervised Learning AI and AI for Molecular Biology

MBZUAI ·

Xiao Wang from Purdue University presented research on Adversarial Contrastive Learning (AdCo) and Cooperative-adversarial Contrastive Learning (CaCo) for improved self-supervised learning. He also discussed CryoREAD, a framework for building DNA/RNA structures from cryo-EM maps, and future work in deep learning for drug discovery. Wang's algorithms have impacted molecular biology, leading to new structure discoveries published in journals like Cell and Nature Microbiology. Why it matters: The research advances AI techniques for crucial tasks in molecular biology and drug discovery, with potential applications for institutions in the GCC region focused on healthcare and biotechnology.

Using child’s play for machine learning

MBZUAI ·

MBZUAI Professor Salman Khan is researching continuous, lifelong learning systems for computer vision, aiming to mimic human learning processes like curiosity and discovery. His work focuses on learning from limited data and adversarial robustness of deep neural networks. Khan, along with MBZUAI professors Fahad Khan and Rao Anwer, and partners from other universities, presented research at CVPR 2022. Why it matters: This research has the potential to significantly improve the ability of AI systems to understand and adapt to the real world, enabling more intelligent autonomous systems.

Computer vision: Teaching computers how to see the world

KAUST ·

KAUST's Visual Computing Center (VCC) is researching computer vision, image processing, and machine learning, with applications in self-driving cars, surveillance, and security. Professor Bernard Ghanem is working on teaching machines to understand visual data semantically, similar to how humans perceive the world. Self-driving cars use visual sensors to interpret traffic signals and detect obstacles, while computer vision also assists governments and corporations with security applications like facial recognition and detecting unattended luggage. Why it matters: Advancements in computer vision at KAUST can contribute to innovations in autonomous vehicles and enhance security measures in the region.

Unscented Autoencoder

arXiv ·

The paper introduces the Unscented Autoencoder (UAE), a novel deep generative model based on the Variational Autoencoder (VAE) framework. The UAE uses the Unscented Transform (UT) for a more informative posterior representation compared to the reparameterization trick in VAEs. It replaces Kullback-Leibler (KL) divergence with the Wasserstein distribution metric and demonstrates competitive performance in Fréchet Inception Distance (FID) scores.

Contrastive Pretraining for Echocardiography Segmentation with Limited Data

arXiv ·

This paper introduces a self-supervised contrastive learning method for segmenting the left ventricle in echocardiography images when limited labeled data is available. The approach uses contrastive pretraining to improve the performance of UNet and DeepLabV3 segmentation networks. Experiments on the EchoNet-Dynamic dataset show the method achieves a Dice score of 0.9252, outperforming existing approaches, with code available on Github.

Machine Learning Integration for Signal Processing

TII ·

Technology Innovation Institute's (TII) Directed Energy Research Center (DERC) is integrating machine learning (ML) techniques into signal processing to accelerate research. One project used convolutional neural networks to predict COVID-19 pneumonia from chest x-rays with 97.5% accuracy. DERC researchers also demonstrated that ML-based signal and image processing can retrieve up to 68% of text information from electromagnetic emanations. Why it matters: This adoption of ML for signal processing at TII highlights the potential for advanced AI techniques to enhance research and security applications in the UAE.

More than meets the eye: Identifying hidden causal variables with causal representation learning

MBZUAI ·

MBZUAI Professor Kun Zhang is developing machine learning techniques to identify hidden causal variables, which are underlying concepts driving cause-and-effect relationships. Zhang and colleagues from Carnegie Mellon University are presenting a new approach for this at ICML 2024. Their method, causal representation learning, assumes that measured variables are generated by unobserved latent variables. Why it matters: Uncovering hidden causal relationships can significantly advance understanding in various fields by revealing the underlying mechanisms driving observed phenomena.