Skip to content
GCC AI Research

Search

Results for "robot navigation"

Robot Navigation in the Wild

MBZUAI ·

Gregory Chirikjian presented an overview of research on robot navigation in unstructured environments, using computer vision, sensor tech, ML, and motion planning. The methods use multi-modal observations from RGB cameras, 3D LiDAR, and robot odometry for scene perception, along with deep RL for planning. These methods have been integrated with wheeled, home, and legged robots and tested in crowded indoor scenes, home environments, and dense outdoor terrains. Why it matters: This research pushes the boundaries of robotics in complex environments, paving the way for more versatile and autonomous robots in the Middle East.

Human-Computer Conversational Vision-and-Language Navigation

MBZUAI ·

A presentation discusses the evolution of Vision-and-Language Navigation (VLN) from benchmarks like Room-to-Room (R2R). It highlights the role of Large Language Models (LLMs) such as GPT-4 in enabling more natural human-machine interactions. The presentation showcases work using LLMs to decode navigational instructions and improve robotic navigation. Why it matters: This research demonstrates the potential of merging vision, language, and robotics for advanced AI applications in navigation and human-computer interaction.

Language and Planning in Robotic Navigation: A Multilingual Evaluation of State-of-the-Art Models

arXiv ·

This paper introduces Arabic language integration into Vision-and-Language Navigation (VLN) in robotics, evaluating multilingual SLMs like GPT-4o mini, Llama 3 8B, Phi-3 14B, and Jais using the NavGPT framework. The study uses the R2R dataset to assess the impact of language on navigation reasoning through zero-shot sequential action prediction. Results show the framework enables high-level planning in both English and Arabic, though some models face challenges with Arabic due to reasoning limitations and parsing issues. Why it matters: This work highlights the need to improve language model planning and reasoning for effective navigation, especially to unlock the potential of Arabic-language models in real-world applications.

Learning Robot Super Autonomy

MBZUAI ·

Giuseppe Loianno from NYU presented research on creating "Super Autonomous" robots (USARC) that are Unmanned, Small, Agile, Resilient, and Collaborative. The research focuses on learning models, control, and navigation policies for single and collaborative robots operating in challenging environments. The talk highlighted the potential of these robots in logistics, reconnaissance, and other time-sensitive tasks. Why it matters: This points to growing research interest in advanced robotics in the region, especially given the focus on smart cities and automation.

Design and Deployment of an Autonomous Unmanned Ground Vehicle for Urban Firefighting Scenarios

arXiv ·

This paper presents the design and deployment of an autonomous unmanned ground vehicle (UGV) equipped with a robotic arm for urban firefighting. The UGV uses on-board sensors for navigation and a thermal camera for fire source identification, with a custom pump for fire suppression. The system was developed for the Mohamed Bin Zayed International Robotics Challenge (MBZIRC) 2020, where it achieved the highest score among UGV solutions and contributed to winning first place. Why it matters: This demonstrates the potential of autonomous robotics in addressing complex and dangerous real-world challenges like urban firefighting in the GCC region and beyond.

Structured World Models for Robots

MBZUAI ·

Krishna Murthy, a postdoc at MIT, researches computational world models to enable robots to understand and operate effectively in the physical world. His work focuses on differentiable computing approaches for spatial perception and interfaces large image, language, and audio models with 3D scenes. Murthy envisions structured world models working with scaling-based approaches to create versatile robot perception and planning algorithms. Why it matters: This research could significantly advance robotics by enabling more sophisticated perception, reasoning, and action capabilities in embodied agents.

Humanoid Robots and the Computational Problems Regarding the Human

MBZUAI ·

Yoshihiko Nakamura from the University of Tokyo discusses the computational challenges of humanoid robots, extending beyond sensing and control to understanding human movement, sensation, and relationships. The talk covers recent research on mechanical humanoid robots with a focus on actuators and computational problems related to human movements. Nakamura highlights the need for humanoid robots to interpret human actions and interactions for effective application. Why it matters: Addressing these computational challenges is crucial for developing more sophisticated and human-compatible robots for use in various human-centered applications within the region and globally.

Co-Modality Active sensing and Perception (C-MAP) in Autonomous Vehicles, Augmented Reality, Remote Environmental Monitoring, and Robotic Grasping

MBZUAI ·

Dezhen Song from Texas A&M University presented a talk on Co-Modality Active sensing and Perception (C-MAP) for robotics, covering sensor fusion for autonomous vehicles, augmented reality, and remote environmental monitoring. The talk highlighted lessons learned in sensor fusion using autonomous motorcycles and NASA Robonaut as examples. Recent works in robotic remote environment monitoring, especially focused on subsurface surface void and pipeline mapping were discussed. Why it matters: This research explores sensor fusion techniques to enhance robot perception, which could improve the robustness and capabilities of autonomous systems developed and deployed in the Middle East, particularly in challenging environments.