A CMU professor and MBZUAI affiliated faculty presented research on how LLMs store and use knowledge learned during pre-training. The study used a synthetic biography dataset to show that LLMs may not effectively use memorized knowledge at inference time, even with zero training loss. Data augmentation during pre-training can force the model to store knowledge in specific token embeddings. Why it matters: The research highlights limitations in LLM knowledge manipulation and extraction, with implications for improving model architectures and training strategies for more effective knowledge utilization in Arabic LLMs.
A Caltech researcher presented at MBZUAI on memory representation and retrieval, contrasting AI and neuroscience approaches. Current AI retrieval systems like RAG retrieve via fine-tuning and embedding similarity, while the presenter argued for exploring retrieval via combinatorial object identity or spatial proximity. The research explores circuit-level retrieval via domain fine-tuned LLMs and distributed memory for image retrieval using semantic similarity. Why it matters: The work suggests structured databases and retrieval-focused training can allow smaller models to outperform larger general-purpose models, offering efficiency gains for AI development in the region.
MIT Professor Ahmed F. Ghoniem delivered a keynote at KAUST's Spring Enrichment Program discussing clean energy solutions for future cities. He emphasized a portfolio approach including electrochemical, solar thermochemical, and plasma technologies for renewable energy storage. Ghoniem highlighted the economic opportunities arising from clean energy technology deployment, R&D, and job creation. Why it matters: The focus on renewable energy and storage aligns with Saudi Arabia's Vision 2030 goals for sustainable urban development and diversification of the energy sector.
Akhil Arora from EPFL presented a framework for AI-assisted knowledge navigation, focusing on understanding and enhancing human navigation on Wikipedia. The framework includes methods for modeling navigation patterns, identifying knowledge gaps, and assessing their causal impact. He also discussed applications beyond Wikipedia, such as multimodal knowledge navigation assistants and multilingual knowledge gap mitigation. Why it matters: This research has the potential to improve information systems by making online knowledge more accessible and navigable, especially for platforms like Wikipedia that serve as critical resources for global knowledge sharing.
Prof. Chun Jason Xue from the City University of Hong Kong presented research on optimizing mobile memory and storage by analyzing mobile application characteristics, noting their differences from server applications. The research explores system software designs inherited from the Linux kernel and identifies optimization opportunities in mobile memory and storage management. Xue's work aims to enhance user experience on mobile devices through mobile application characterization, focusing on non-volatile and flash memories. Why it matters: Optimizing mobile systems based on the unique characteristics of mobile applications can significantly improve device performance and user experience in the region.
KAUST hosted the Frontiers in Energy Storage 2026 conference, emphasizing energy storage technologies for renewable energy. The conference highlighted electrochemical and chemical systems, including advanced batteries and hydrogen, as complementary layers for long-duration and industrial resilience. KAUST is developing energy-storage solutions relevant for the Kingdom and valuable to global partners, aiming to engineer solutions to withstand extreme environmental temperatures. Why it matters: This positions Saudi Arabia as a potential global exporter of resilient energy hardware, aligning with Saudi Vision 2030 goals in renewable energy.
Scimagine is a KAUST-based startup that provides a cloud-based platform for managing and storing experimental data for material scientists. The platform allows researchers to store, manage, and share their data, as well as create scientific visuals. It addresses the problem of experimental data being hidden in PDF files and not easily searchable. Why it matters: This platform improves data accessibility and collaboration in materials science research, potentially accelerating discovery and innovation in the field.