MBZUAI researchers co-led a study published in Nature demonstrating that GluFormer, an AI foundation model trained on continuous glucose monitoring (CGM) data, more accurately predicts long-term diabetes and cardiovascular risk than current clinical standards. GluFormer, built on a transformer architecture and trained using NVIDIA AI infrastructure on over 10 million CGM measurements, forecasts individual health risks using short-term glucose dynamics. In a 12-year follow-up, the model captured 66% of new-onset diabetes cases and 69% of cardiovascular-death events in its highest-risk group, outperforming established CGM-derived metrics across 19 external cohorts. Why it matters: The development of GluFormer represents a significant advancement in personalized healthcare, enabling proactive and individualized health strategies through the analysis of dynamic glucose data.
Giovanni Puccetti from ISTI-CNR presented research on linguistic probing of language models like BERT and RoBERTa. The research investigates the ability of these models to encode linguistic properties, linking this ability to outlier parameters. Preliminary work on fine-tuning LLMs in Italian and detecting synthetic news generation was also presented. Why it matters: Understanding the inner workings and linguistic capabilities of LLMs is crucial for improving their reliability and adapting them to diverse languages like Arabic.
NYU and NYU Abu Dhabi researchers are working on user-centric gender rewriting in NLP, especially for Arabic. They are building an Arabic Parallel Gender Corpus and developing models for gender rewriting tasks. The work aims to address representational harms caused by NLP systems that don't account for user preferences regarding grammatical gender. Why it matters: This research promotes fairness and inclusivity in Arabic NLP by enabling systems to generate gender-specific outputs based on user preferences, mitigating biases present in training data.
This paper introduces a convolutional transformer model for classifying tomato maturity, along with a new UAE-sourced dataset, KUTomaData, for training segmentation and classification models. The model combines CNNs and transformers and was tested against two public datasets. Results showed state-of-the-art performance, outperforming existing methods by significant margins in mAP scores across all three datasets.