Skip to content
GCC AI Research

Search

Results for "data silos"

Unlocking Decentralized AI and Vision: Overcoming Incentive Barriers, Orchestration Challenges, and Data Silos

MBZUAI ·

This article discusses the need for a decentralized approach to AI, especially in contexts where data and knowledge are distributed. It highlights five key technical challenges: privacy, verifiability, incentives, orchestration, and crowdUX. The author, Ramesh Raskar from MIT Media Lab, advocates for integrating privacy tech, distributed verifiable AI, data markets, orchestration, and crowd experience into the Web3 framework. Why it matters: Decentralized AI could unlock new possibilities for collaboration and problem-solving in the region, particularly in sectors like healthcare and logistics where data is often siloed.

Building Planetary-Scale Collaborative Intelligence

MBZUAI ·

Sai Praneeth Karimireddy from UC Berkeley presented a talk on building planetary-scale collaborative intelligence, highlighting the challenges of using distributed data in machine learning due to data silos and ethical-legal restrictions. He proposed collaborative systems like federated learning as a solution to bring together distributed data while respecting privacy. The talk addressed the need for efficiency, reliability, and management of divergent goals in these systems, suggesting the use of tools from optimization, statistics, and economics. Why it matters: Collaborative AI systems can unlock valuable distributed data in the region, especially in sensitive sectors like healthcare, while ensuring privacy and addressing ethical concerns.

Optimizing AI Systems through Cross-Layer Design: A Data-Centric Approach

MBZUAI ·

A Duke University professor presented a data-centric approach to optimizing AI systems by addressing the memory capacity and bandwidth bottleneck. The presentation covered collaborative optimization across algorithms, systems, architecture, and circuit layers. It also explored compute-in-memory as a solution for integrating computation and memory. Why it matters: Optimizing AI systems through a data-centric approach can improve efficiency and performance, critical for advancing AI applications in the region.

Bring Your Own Kernel! Constructing High-Performance Data Management Systems from Components

MBZUAI ·

Holger Pirk from Imperial College London is developing a novel approach to data management system composition called BOSS. The system uses a homoiconic representation of data and code and partial evaluation of queries by components, drawing inspiration from compiler-construction research. BOSS achieves a fully composable design that effectively combines different data models, hardware platforms, and processing engines, enabling features like GPU acceleration and generative data cleaning with minimal overhead. Why it matters: This research on composable database systems can broaden the applicability of data management techniques in the GCC region, enabling more flexible and efficient data processing for various applications.

Managing and Analyzing Big Traffic Data — An Uncertain Time Series Approach

MBZUAI ·

This article discusses the application of uncertain time series (UTS) approach to manage and analyze big traffic data for high-resolution vehicular transportation services. The study addresses challenges such as data sparseness, decision-making among multiple UTSs, and future forecasting with spatio-temporal correlations. Jilin Hui, previously a Research Associate at the Inception Institute of Artificial Intelligence (UAE), is applying this approach to solve problems related to increased congestion, greenhouse gas emissions, and reduced air quality in urban environments. Why it matters: The application of AI techniques to traffic management could significantly improve urban mobility and environmental sustainability in the GCC region and beyond.

Exploring science's fourth paradigm

KAUST ·

KAUST held a research conference on Computational and Statistical Interface to Big Data from March 19-21. The conference covered topics like data representation, visualization, parallel algorithms, and large-scale machine learning. Participants came from institutions including the American University of Sharjah, Aalborg University, and others to exchange ideas. Why it matters: The conference highlights KAUST's focus on promoting big data research and collaboration to address challenges and opportunities in various scientific fields within the Kingdom and globally.

Scalable Community Detection in Massive Networks Using Aggregated Relational Data

MBZUAI ·

A new mini-batch strategy using aggregated relational data is proposed to fit the mixed membership stochastic blockmodel (MMSB) to large networks. The method uses nodal information and stochastic gradients of bipartite graphs for scalable inference. The approach was applied to a citation network with over two million nodes and 25 million edges, capturing explainable structure. Why it matters: This research enables more efficient community detection in massive networks, which is crucial for analyzing complex relationships in various domains, but this article has no clear connection to the Middle East.

Enabling precision medicine with single cell omics and decentralized clinical studies

MBZUAI ·

Eduardo da Veiga Beltrame, bioinformatics lead at ImYoo (a Caltech spinout), presented on scalable methods for single-cell omics data analysis, including kallisto|bustools and scvi-tools. He highlighted their use in ImYoo's decentralized longitudinal study on Inflammatory Bowel Disease (IBD), where patients self-collect capillary blood samples. Beltrame also discussed his research on STEM education programs in Brazil as a visiting scholar at UC Berkeley. Why it matters: This highlights the growing trend of decentralized clinical studies leveraging advanced single-cell technologies for precision medicine, showcasing the potential of remote data collection and analysis in understanding complex diseases.