KAUST has launched self-driving shuttles on its campus, making it the first adopter of autonomous vehicles in Saudi Arabia. The pilot project utilizes vehicle technology from Local Motors and EasyMile. SAPTCO will operate the autonomous shuttles and manage operations with Saudi staff. Why it matters: This initiative advances Saudi Arabia's 2030 Vision and positions KAUST as a regional leader in smart city development and AI research.
KAUST researchers in the Image and Video Understanding Lab are applying machine learning to computer vision for automated navigation, including self-driving cars and UAVs. They tested their algorithms on KAUST roads, aiming to replicate the brain's efficiency in tasks like activity and object recognition. The team is also exploring the possibility of creative algorithms that can transfer skills without direct training. Why it matters: This research contributes to the advancement of autonomous systems and explores the fundamental questions of replicating human intelligence in machines within the GCC region.
KAUST, Intel, and Brightskies have launched REDD, a collaborative self-driving mobility platform, converting a conventional car into a self-driving vehicle with integrated AI software. Brightskies developed the self-driving system, powered by Intel® NUC platforms, utilizing their BrightDrive system. KAUST researchers will use the vehicle to test new techniques, leveraging real-world data to improve self-driving technologies. Why it matters: This partnership advances autonomous vehicle research in Saudi Arabia, aligning with the Kingdom's Vision 2030 by creating a platform for innovation and testing in a real-world environment.
Daniela Rus from MIT CSAIL discussed the role of AI in revolutionizing autonomous vehicles, emphasizing the need for risk evaluation, intent understanding, and adaptation to diverse driving styles. The talk highlighted integrating risk and behavior analysis in autonomous vehicle control systems. Social Value Orientation (SVO) can be incorporated into decision-making for self-driving vehicles. Why it matters: This research advances the development of safer and more adaptive autonomous vehicles, crucial for their successful deployment in diverse real-world driving scenarios within the GCC region and globally.
This paper introduces a minimalistic autonomous racing stack designed for high-speed time-trial racing, emphasizing rapid deployment and efficient system integration with minimal on-track testing. Validated on real speedways, the stack achieved a top speed of 206 km/h within just 11 hours of practice, covering 325 km. The system performance analysis includes tracking accuracy, vehicle dynamics, and safety considerations. Why it matters: This research offers insights for teams aiming to quickly develop and deploy autonomous racing stacks with limited track access, potentially accelerating innovation in autonomous vehicle technology within the A2RL and similar racing initiatives.
Giuseppe Loianno from NYU presented research on creating "Super Autonomous" robots (USARC) that are Unmanned, Small, Agile, Resilient, and Collaborative. The research focuses on learning models, control, and navigation policies for single and collaborative robots operating in challenging environments. The talk highlighted the potential of these robots in logistics, reconnaissance, and other time-sensitive tasks. Why it matters: This points to growing research interest in advanced robotics in the region, especially given the focus on smart cities and automation.
The paper introduces OmniGen, a unified framework for generating aligned multimodal sensor data for autonomous driving using a shared Bird's Eye View (BEV) space. It uses a novel generalizable multimodal reconstruction method (UAE) to jointly decode LiDAR and multi-view camera data through volume rendering. The framework incorporates a Diffusion Transformer (DiT) with a ControlNet branch to enable controllable multimodal sensor generation, demonstrating good performance and multimodal consistency.