InfiAgent is a new agent framework comparable to GPT4-Agent, developed by replicating Codex. It includes InfiCoder, an open-source model for text-to-code, code-to-code, and freeform code-related QA tasks. The framework focuses on data analysis and integrates an LLM with programming capabilities and a sandbox environment for executing Python code. Why it matters: This research demonstrates the potential for advancements in AI operating systems and highlights areas where current models like GPT-4V can be improved, contributing to the broader development of more capable and versatile AI agents.
Microsoft CEO Satya Nadella recently reassigned Mustafa Suleyman, co-founder of Inflection AI, to lead Microsoft AI, consolidating consumer AI efforts under one umbrella. Microsoft has also created a new organization, led by Rajesh Jha, focused on experiences and devices. Additionally, Microsoft is doubling down on its cloud computing and AI infrastructure investments. Why it matters: This restructuring signals Microsoft's intensified focus on integrating AI across its product ecosystem and solidifying its position as a leader in both AI software and infrastructure.
MBZUAI releases Bactrian-X, a multilingual parallel dataset of 3.4 million instruction-response pairs across 52 languages. They trained low-rank adaptation (LoRA) adapters using this dataset, creating lightweight, replaceable components for large language models. Experiments show the LoRA-based models outperform vanilla and existing instruction-tuned models in multilingual settings.
Inception, Cerebras, and MBZUAI have released Jais 2, a 70 billion parameter open-weight Arabic LLM. Jais 2 is trained on an Arabic-first dataset and features a redesigned architecture for stronger reasoning and fluency across Arabic dialects and English. It integrates a safety-first framework and demonstrates capabilities in understanding Arabic poetry, culture, and social media tone. Why it matters: Jais 2 addresses the historical underrepresentation of Arabic in AI by providing a culturally and linguistically faithful model, potentially accelerating innovation across the region.
MBZUAI has launched the Institute of Foundation Models (IFM) with a new Silicon Valley Lab in Sunnyvale, CA, joining existing facilities in Paris and Abu Dhabi. The launch event showcased PAN, a world model for simulating diverse realities with multimodal inputs. The IFM lab is also advancing K2-65B and JAIS AI systems. Why it matters: This expansion enhances MBZUAI's global presence and connects it with a critical AI ecosystem, supporting the UAE's economic diversification through advanced AI technologies.
Niket Tandon from the Allen Institute for AI presented a talk at MBZUAI on enabling large language models to focus on human needs and continuously learn from interactions. He proposed a memory architecture inspired by the theory of recursive reminding to guide models in avoiding past errors. The talk addressed who to ask, what to ask, when to ask and how to apply the obtained guidance. Why it matters: The research explores how to align LLMs with human feedback, a key challenge for practical and ethical AI deployment.
MBZUAI has appointed Professor Timothy Baldwin as Associate Provost and acting chair of its new NLP Department. Baldwin will focus on strengthening the curriculum and building a world-class faculty team. He previously spent 17 years at the University of Melbourne. Why it matters: The recruitment signals MBZUAI's commitment to becoming a leading center for NLP research and education in the region.
MBZUAI Assistant Professor Alham Fikri Aji is presenting research at EACL 2024 on efficient NLP for low-resource languages. The study uses knowledge distillation, transferring knowledge from a larger model (ChatGPT) to a smaller one using synthetic instruction data. The goal is to achieve similar performance with less computational resources, focusing on underrepresented languages. Why it matters: This work addresses the need for more accessible and inclusive NLP technologies, especially for languages lacking extensive datasets and computational resources.