MBZUAI Professor Kun Zhang is developing machine learning techniques to identify hidden causal variables, which are underlying concepts driving cause-and-effect relationships. Zhang and colleagues from Carnegie Mellon University are presenting a new approach for this at ICML 2024. Their method, causal representation learning, assumes that measured variables are generated by unobserved latent variables. Why it matters: Uncovering hidden causal relationships can significantly advance understanding in various fields by revealing the underlying mechanisms driving observed phenomena.
Researchers from MBZUAI presented a new algorithm at ICLR 2024 that identifies causal relationships involving both observed and latent variables. The algorithm addresses limitations of existing methods that struggle with latent variables or assume observed variables don't directly influence latent variables. The proposed algorithm can accommodate both scenarios, offering a more generalizable approach to causal discovery. Why it matters: This research advances the development of AI systems that can analyze complex data and identify causal relationships, with potential applications in fields like medicine where understanding causality is crucial for developing treatments and preventative measures.
MBZUAI researchers presented a new machine learning method at ICLR for uncovering hidden variables from observed data. The method, called "complementary gains," combines two weak assumptions to provide identifiability guarantees. This approach aims to recover true latent variables reflecting real-world processes, while solving problems efficiently. Why it matters: The research advances disentangled representation learning by finding minimal assumptions necessary for identifiability, improving the applicability of AI models to real-world data.
Hassan Sajjad from Dalhousie University presented research on exploring the latent space of AI models to assess their safety and trustworthiness. He discussed use cases where analyzing latent space helps understand the robustness-generalization tradeoff in adversarial training and evaluate language comprehension. Sajjad's work aims to build better AI models and increase trust in their capabilities by looking at model internals. Why it matters: Intrinsic evaluation of model internals will become important to improving AI safety and robustness.
This article discusses approximating a high-dimensional distribution using Gaussian variational inference by minimizing Kullback-Leibler divergence. It builds upon previous research and approximates the minimizer using a Gaussian distribution with specific mean and variance. The study details approximation accuracy and applicability using efficient dimension, relevant for analyzing sampling schemes in optimization. Why it matters: This theoretical research can inform the development of more efficient and accurate AI algorithms, particularly in areas dealing with high-dimensional data such as machine learning and data analysis.
MBZUAI researchers presented a new causal representation learning framework at ICLR for identifying latent causal variables in multimodal biological data. The framework addresses the challenge of uncovering underlying causal factors from lab tests, genetic information, and medical images. The new approach can identify latent causal variables and their influence on observed biological outcomes across modalities. Why it matters: The model's ability to analyze causal mechanisms between modalities can lead to more complete insights in biomedical research.
A new mini-batch strategy using aggregated relational data is proposed to fit the mixed membership stochastic blockmodel (MMSB) to large networks. The method uses nodal information and stochastic gradients of bipartite graphs for scalable inference. The approach was applied to a citation network with over two million nodes and 25 million edges, capturing explainable structure. Why it matters: This research enables more efficient community detection in massive networks, which is crucial for analyzing complex relationships in various domains, but this article has no clear connection to the Middle East.
MBZUAI researchers have developed a new kernel-based method to identify dependence patterns in data, especially in small regions exhibiting 'rare dependence' where relationships between variables differ. The method uses sample importance reweighting, assigning more importance to regions with rare dependence. Tested on synthetic and real-world data, the algorithm successfully identified relations between variables even with rare dependence, outperforming traditional methods like HSIC. Why it matters: This advancement can improve data analysis in fields like public health, economics, genomics, and AI, enabling more accurate insights from complex observational data.