Krishna Murthy, a postdoc at MIT, researches computational world models to enable robots to understand and operate effectively in the physical world. His work focuses on differentiable computing approaches for spatial perception and interfaces large image, language, and audio models with 3D scenes. Murthy envisions structured world models working with scaling-based approaches to create versatile robot perception and planning algorithms. Why it matters: This research could significantly advance robotics by enabling more sophisticated perception, reasoning, and action capabilities in embodied agents.
Gregory Chirikjian presented an overview of research on robot navigation in unstructured environments, using computer vision, sensor tech, ML, and motion planning. The methods use multi-modal observations from RGB cameras, 3D LiDAR, and robot odometry for scene perception, along with deep RL for planning. These methods have been integrated with wheeled, home, and legged robots and tested in crowded indoor scenes, home environments, and dense outdoor terrains. Why it matters: This research pushes the boundaries of robotics in complex environments, paving the way for more versatile and autonomous robots in the Middle East.
MBZUAI Professor Ivan Laptev is working to bridge the gap between data-driven AI systems and embodied agents (robots). He notes challenges in robotics including data scarcity, the need to generate new data through actions, and the requirement for real-time operation. Laptev aims to transfer innovations from computer vision to robotics, addressing these challenges to improve robots' ability to interpret and respond to the complexities of the real world. Why it matters: Overcoming these hurdles is crucial for advancing robotics and enabling robots to effectively interact with and navigate dynamic real-world environments.
Gregory Chirikjian presented an overview talk on applying probability, harmonic analysis, and geometry to robotics, emphasizing the need for robots to function beyond traditional industrial programming. He discussed a new approach where robots define affordances of objects, using simulation to 'imagine' object use and enabling reasoning about novel objects. Probabilistic methods on Lie-groups, initially developed for mobile robot state estimation, are now adapted for one-shot learning of affordances, with plans to integrate large language models. Why it matters: This research direction aims to enhance robot intelligence and adaptability, crucial for service robots in dynamic environments and aligning with broader goals of advanced AI integration in robotics.
Mingyu Ding from UC Berkeley presented research on endowing robots with human-like commonsense and physical reasoning capabilities. The talk covered multimodal commonsense reasoning integrating vision, world models, and language-based task planners. It also discussed physical reasoning approaches for robots to infer dynamics and physical properties of objects. Why it matters: Enhancing robots with these capabilities can improve their ability to generalize across everyday tasks, leading to greater social benefits and impact.