MBZUAI and Minerva Humanoids have launched a two-year research partnership to develop humanoid robotics for safety-critical industrial operations, initially focusing on the energy sector. The collaboration aims to create versatile, reasoning-based systems capable of operating with varying degrees of autonomy. MBZUAI students will be integrated into project teams to develop AI models and robotics systems. Why it matters: The partnership advances AI-driven robotics in safety-critical domains, supports the UAE's ambition to become a global hub, and develops local talent.
Yoshihiko Nakamura from the University of Tokyo discusses the computational challenges of humanoid robots, extending beyond sensing and control to understanding human movement, sensation, and relationships. The talk covers recent research on mechanical humanoid robots with a focus on actuators and computational problems related to human movements. Nakamura highlights the need for humanoid robots to interpret human actions and interactions for effective application. Why it matters: Addressing these computational challenges is crucial for developing more sophisticated and human-compatible robots for use in various human-centered applications within the region and globally.
MBZUAI's VP of Research, Professor Sami Haddadin, and his team at TUM have developed the 'Tree of Robots,' a new framework for categorizing robots based on capabilities and morphology rather than appearance or purpose. This framework uses a Process Database and Metrics Definitions to assess a robot's fitness for specific tasks, resulting in a fitness score and classification within the tree. The research appears in the March 2025 issue of Nature Machine Intelligence. Why it matters: This systematic approach could fundamentally change how we understand, compare, and develop robotic systems, enabling a deeper understanding of intelligent machines and their potential.
Mingyu Ding from UC Berkeley presented research on endowing robots with human-like commonsense and physical reasoning capabilities. The talk covered multimodal commonsense reasoning integrating vision, world models, and language-based task planners. It also discussed physical reasoning approaches for robots to infer dynamics and physical properties of objects. Why it matters: Enhancing robots with these capabilities can improve their ability to generalize across everyday tasks, leading to greater social benefits and impact.
This article previews a talk by Gül Varol from Ecole des Ponts ParisTech on bridging natural language and 3D human motions. The talk will cover text-to-motion synthesis using generative models and text-to-motion retrieval models based on the ACTOR, TEMOS, TMR, TEACH, and SINC papers. Varol's research interests include video representation learning, human motion synthesis, and sign languages. Why it matters: Research in this area could enable more intuitive human-computer interaction and new applications in areas like virtual reality and robotics.
Lorenzo Jamone from Queen Mary University of London presented on cognitive robotics, focusing on tactile exploration and manipulation by robots. The talk covered combining biology, engineering, and AI for advanced robotic systems. Jamone directs the CRISP group and has over 100 publications in cognitive robotics. Why it matters: This highlights the ongoing research into more sophisticated robotic systems that can interact with complex environments, an area crucial for future applications in manufacturing and human-robot collaboration in the GCC.
Professor Hava Siegelmann, a computer science expert, is researching lifelong learning AI, drawing inspiration from the brain's abstraction and generalization capabilities. The research aims to enable intelligent systems in satellites, robots, and medical devices to adapt and improve their expertise in real-time, even with limited communication and power. The goal is to develop AI systems applicable for far edge computing that can learn in runtime and handle unanticipated situations. Why it matters: This research could lead to more resilient and adaptable AI systems for critical applications in remote and resource-constrained environments, with potential benefits for various sectors in the Middle East.
Egor Zakharov from ETH Zurich AIT lab will present research on creating controllable and detailed 3D head avatars using data from consumer-grade devices. The presentation will cover high-fidelity image-based facial reconstruction/animation and video-based reconstruction of detailed structures like hairstyles. He will showcase integrating human-centric assets into virtual environments for real-time telepresence and entertainment. Why it matters: This research contributes to advancements in digital human modeling and telepresence, with applications in communication and gaming within the region.