A new Bayesian matrix factorization approach is explored for performance prediction in multilingual NLP, aiming to reduce the experimental burden of evaluating various language combinations. The approach outperforms state-of-the-art methods in NLP benchmarks like machine translation and cross-lingual entity linking. It also avoids hyperparameter tuning and provides uncertainty estimates over predictions. Why it matters: Accurate performance prediction methods accelerate multilingual NLP research by reducing computational costs and improving experimental efficiency, especially valuable for Arabic NLP tasks.
The paper introduces a novel method for short-term, high-resolution traffic prediction, modeling it as a matrix completion problem solved via block-coordinate descent. An ensemble learning approach is used to capture periodic patterns and reduce training error. The method is validated using both simulated and real-world traffic data from Abu Dhabi, demonstrating superior performance compared to other algorithms.
MBZUAI Associate Professor Martin Takáč is working on high-performance computing and machine learning with applications in logistics, supply chain management, and other areas. His research focuses on using AI to improve precision and efficiency in tasks like predicting demand and optimizing delivery routes. Takáč's interests include imitative learning, predictive modeling, and reinforcement learning to enable AI to mimic human behavior and predict future outcomes. Why it matters: This research contributes to the development of more efficient and reliable AI systems that can be applied to a wide range of industries in the UAE and beyond.
Researchers studied user lifetime prediction in the location-based social network Jodel within Saudi Arabia, leveraging its disjoint communities. Machine learning models, particularly Random Forest, were trained to predict user lifetime as a regression and classification problem. A single countrywide model generalizes well and performs similarly to community-specific models.
This paper introduces a predictive analysis of Arabic court decisions, utilizing 10,813 real commercial court cases. The study evaluates LLaMA-7b, JAIS-13b, and GPT3.5-turbo models under zero-shot, one-shot, and fine-tuned training paradigms, also experimenting with summarization and translation. GPT-3.5 models significantly outperformed others, exceeding JAIS model performance by 50%, while also demonstrating the unreliability of most automated metrics. Why it matters: This research bridges computational linguistics and Arabic legal analytics, offering insights for enhancing judicial processes and legal strategies in the Arabic-speaking world.