Skip to content
GCC AI Research

Search

Results for "Siamese neural networks"

MIT-QCRI Arabic Dialect Identification System for the 2017 Multi-Genre Broadcast Challenge

arXiv ·

This paper describes the MIT-QCRI team's Arabic Dialect Identification (ADI) system developed for the 2017 Multi-Genre Broadcast challenge (MGB-3). The system aims to distinguish between four major Arabic dialects and Modern Standard Arabic. The research explores Siamese neural network models and i-vector post-processing to handle dialect variability and domain mismatches, using both acoustic and linguistic features. Why it matters: The work contributes to the advancement of Arabic language processing, specifically in dialect identification, which is crucial for analyzing and understanding diverse Arabic speech content in media broadcasts.

Upsampling Autoencoder for Self-Supervised Point Cloud Learning

arXiv ·

This paper introduces a self-supervised learning method for point cloud analysis using an upsampling autoencoder (UAE). The model uses subsampling and an encoder-decoder architecture to reconstruct the original point cloud, learning both semantic and geometric information. Experiments show the UAE outperforms existing methods in shape classification, part segmentation, and point cloud upsampling tasks.

Enhancing Semantic Similarity Understanding in Arabic NLP with Nested Embedding Learning

arXiv ·

This paper introduces a nested embedding learning framework for Arabic NLP, utilizing Matryoshka Embedding Learning and multilingual models. The authors translated sentence similarity datasets into Arabic to enable comprehensive evaluation. Experiments on the Arabic Natural Language Inference dataset show Matryoshka embedding models outperform traditional models by 20-25% in capturing Arabic semantic nuances. Why it matters: This work advances Arabic NLP by providing a new method and evaluation benchmark for semantic similarity, which is crucial for tasks like information retrieval and text understanding.

UAE: Universal Anatomical Embedding on Multi-modality Medical Images

arXiv ·

Researchers propose a universal anatomical embedding (UAE) framework for medical image analysis to learn appearance, semantic, and cross-modality anatomical embeddings. UAE incorporates semantic embedding learning with prototypical contrastive loss, a fixed-point-based matching strategy, and an iterative approach for cross-modality embedding learning. The framework was evaluated on landmark detection, lesion tracking and CT-MRI registration tasks, outperforming existing state-of-the-art methods.

Deep Surface Meshes

MBZUAI ·

Pascal Fua from EPFL presented an approach to implementing convolutional neural nets that output complex 3D surface meshes. The method overcomes limitations in converting implicit representations to explicit surface representations. Applications include single view reconstruction, physically-driven shape optimization, and bio-medical image segmentation. Why it matters: This research advances geometric deep learning by enabling end-to-end trainable models for 3D surface mesh generation, with potential impact on various applications in computer vision and biomedical imaging in the region.

The Inception Team at NSURL-2019 Task 8: Semantic Question Similarity in Arabic

arXiv ·

The Inception Team presented a system for Semantic Question Similarity in Arabic as part of the NSURL 2019 Task 8. The system explores different methods for determining question similarity in Arabic. Their best result was an ensemble model using a pre-trained multilingual BERT model, achieving a 95.924% F1-Score and ranking first among nine participating teams. Why it matters: This demonstrates strong performance on a key Arabic NLP task, advancing the state-of-the-art in semantic understanding for the language.

The Prism Hypothesis: Harmonizing Semantic and Pixel Representations via Unified Autoencoding

arXiv ·

The paper introduces the Prism Hypothesis, which posits a correspondence between an encoder's feature spectrum and its functional role, with semantic encoders capturing low-frequency components and pixel encoders retaining high-frequency information. Based on this, the authors propose Unified Autoencoding (UAE), a model that harmonizes semantic structure and pixel details using a frequency-band modulator. Experiments on ImageNet and MS-COCO demonstrate that UAE effectively unifies semantic abstraction and pixel-level fidelity, achieving state-of-the-art performance.