Kimon Fountoulakis from the University of Waterloo presented a talk on machine learning on graphs, covering node classification and algorithmic reasoning. The talk discussed the limitations and strengths of graph neural networks (GNNs). It also covered novel optimal architectures for node classification and the ability of looped GNNs to execute classical algorithms. Why it matters: Understanding GNN capabilities is crucial for advancing AI applications in areas like recommendation systems and drug discovery that rely on relational data.
The Qatar Computing Research Institute (QCRI) has introduced PDNS-Net, a large heterogeneous graph dataset for malicious domain classification, containing 447K nodes and 897K edges. It is significantly larger than existing heterogeneous graph datasets like IMDB and DBLP. Preliminary evaluations using graph neural networks indicate that further research is needed to improve model performance on large heterogeneous graphs. Why it matters: This dataset will enable researchers to develop and benchmark graph learning algorithms on a scale relevant to real-world cybersecurity applications, particularly for identifying and mitigating malicious online activity.
KAUST Associate Professor Xiangliang Zhang leads the Machine Intelligence and Knowledge Engineering (MINE) group, focusing on machine learning and data mining algorithms for AI applications. The MINE group researches complex graph data to profile nodes, predict links, detect computing communities, and understand their connections. Zhang's team also works on graph alignment and recommender systems. Why it matters: This research contributes to advancing machine learning techniques at a leading GCC institution, potentially impacting various AI applications in the region.
This talk explores modern machine learning through high-dimensional statistics, using random matrix theory to analyze learning models. The speaker, Denny Wu from University of Toronto and the Vector Institute, presents two examples: hyperparameter selection in overparameterized models and gradient-based representation learning in neural networks. The analysis reveals insights such as the possibility of negative optimal ridge penalty and the advantages of feature learning over random features. Why it matters: This research provides a deeper theoretical understanding of deep learning phenomena, with potential implications for optimizing training and improving model performance in the region.
Mausam, head of Yardi School of AI at IIT Delhi and affiliate professor at University of Washington, will discuss Neuro-Symbolic AI. The talk will cover recent research threads with applications in NLP, probabilistic decision-making, and constraint satisfaction. Mausam's research explores neuro-symbolic machine learning, computer vision for radiology, NLP for robotics, multilingual NLP, and intelligent information systems. Why it matters: Neuro-Symbolic AI is gaining importance as it combines the strengths of neural and symbolic approaches, potentially leading to more robust and explainable AI systems.