Skip to content
GCC AI Research

Search

Results for "Autoencoder"

Upsampling Autoencoder for Self-Supervised Point Cloud Learning

arXiv ·

This paper introduces a self-supervised learning method for point cloud analysis using an upsampling autoencoder (UAE). The model uses subsampling and an encoder-decoder architecture to reconstruct the original point cloud, learning both semantic and geometric information. Experiments show the UAE outperforms existing methods in shape classification, part segmentation, and point cloud upsampling tasks.

Unscented Autoencoder

arXiv ·

The paper introduces the Unscented Autoencoder (UAE), a novel deep generative model based on the Variational Autoencoder (VAE) framework. The UAE uses the Unscented Transform (UT) for a more informative posterior representation compared to the reparameterization trick in VAEs. It replaces Kullback-Leibler (KL) divergence with the Wasserstein distribution metric and demonstrates competitive performance in Fréchet Inception Distance (FID) scores.

Mutually-Regularized Dual Collaborative Variational Auto-encoder for Recommendation Systems

arXiv ·

This paper introduces a mutually-regularized dual collaborative variational auto-encoder (MD-CVAE) for recommendation systems, addressing the limitations of user-oriented auto-encoders (UAEs) in handling sparse ratings and new items. MD-CVAE integrates item content and user ratings within a variational framework, regularizing UAE weights with item content to avoid non-optimal convergence. A symmetric inference strategy eliminates the need for retraining when introducing new items, enhancing efficiency in dynamic recommendation scenarios. Why it matters: The MD-CVAE approach offers a practical solution for improving recommendation accuracy and efficiency, especially in scenarios with data sparsity and frequent item updates, relevant to e-commerce and content platforms in the Middle East.

The Prism Hypothesis: Harmonizing Semantic and Pixel Representations via Unified Autoencoding

arXiv ·

The paper introduces the Prism Hypothesis, which posits a correspondence between an encoder's feature spectrum and its functional role, with semantic encoders capturing low-frequency components and pixel encoders retaining high-frequency information. Based on this, the authors propose Unified Autoencoding (UAE), a model that harmonizes semantic structure and pixel details using a frequency-band modulator. Experiments on ImageNet and MS-COCO demonstrate that UAE effectively unifies semantic abstraction and pixel-level fidelity, achieving state-of-the-art performance.