Saber Salehkaleybar from EPFL presented a talk on causal discovery, focusing on learning causal relationships from observational data and through interventions. He discussed an approximation algorithm for experiment design under budget constraints, with applications in gene-regulatory networks. The talk also covered improvements to reduce the computational complexity of experiment design algorithms. Why it matters: Causal AI systems can lead to more intelligent decision-making in various fields.
A new framework for constructing confidence sets for causal orderings within structural equation models (SEMs) is presented. It leverages a residual bootstrap procedure to test the goodness-of-fit of causal orderings, quantifying uncertainty in causal discovery. The method is computationally efficient and suitable for medium-sized problems while maintaining theoretical guarantees as the number of variables increases. Why it matters: This offers a new dimension of uncertainty quantification that enhances the robustness and reliability of causal inference in complex systems, but there is no indication of connection to the Middle East.
This article discusses a talk by Mingming Gong from the University of Melbourne at MBZUAI on bridging causality and machine learning. The talk focuses on using machine learning to discover causal structures from observational data, and leveraging causal structures to improve machine learning generalization and prediction in non-stationary environments. Gong's research explores theoretical foundations and computational innovations in causal structure learning from real-world data. Why it matters: This research direction is crucial for advancing AI systems that can reason about cause and effect, leading to more robust and reliable decision-making in complex environments.
MBZUAI researchers presented a study at ICML 2024 examining how data aggregation distorts causal discovery. The study argues that current methods are misled because real-world interactions happen at a micro level while observations are aggregated. Using the example of ice cream sales and temperature, they highlight how aggregation introduces "instantaneous causality" where time-lags exist. Why it matters: The research identifies a fundamental limitation in current causal discovery methods, potentially impacting disciplines relying on accurate causal inference from observational data.
MBZUAI Professor Kun Zhang is developing machine learning techniques to identify hidden causal variables, which are underlying concepts driving cause-and-effect relationships. Zhang and colleagues from Carnegie Mellon University are presenting a new approach for this at ICML 2024. Their method, causal representation learning, assumes that measured variables are generated by unobserved latent variables. Why it matters: Uncovering hidden causal relationships can significantly advance understanding in various fields by revealing the underlying mechanisms driving observed phenomena.