Skip to content
GCC AI Research

Uncertainty Estimation: Can your neural network provide confidence for its predictions?

MBZUAI · Notable

Summary

Dr. Maxim Panov from TII Abu Dhabi will give a talk on uncertainty estimation in neural networks, covering model calibration, ensemble methods, and Bayesian approaches. The talk will focus on efficient single-network methods for quantifying prediction confidence, without requiring ensembles or major training changes. Panov's background includes experience at Skolkovo Institute of Science and Technology and DATADVANCE Company. Why it matters: Improving uncertainty estimation is crucial for deploying reliable AI systems in critical applications across the GCC region.

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

Separating fact from fiction with uncertainty quantification

MBZUAI ·

MBZUAI's Maxim Panov is developing uncertainty quantification methods to improve the reliability of language models. His work focuses on providing insights into the confidence level of machine learning models' predictions, especially in scenarios where accuracy is critical, such as medicine. Panov is working on post-processing techniques that can be applied to already-trained models. Why it matters: This research aims to address the issue of "hallucinations" in language models, enhancing their trustworthiness and applicability in sensitive domains within the region and globally.

Uncertainty Modeling of Emerging Device-based Computing-in-Memory Neural Accelerators with Application to Neural Architecture Search

arXiv ·

This paper analyzes the impact of device uncertainties on deep neural networks (DNNs) in emerging device-based Computing-in-memory (CiM) systems. The authors propose UAE, an uncertainty-aware Neural Architecture Search scheme, to identify DNN models robust to these uncertainties. The goal is to mitigate accuracy drops when deploying trained models on real-world platforms.

Deep Ensembles Work, But Are They Necessary?

MBZUAI ·

A recent study questions the necessity of deep ensembles, which improve accuracy and match larger models. The study demonstrates that ensemble diversity does not meaningfully improve uncertainty quantification on out-of-distribution data. It also reveals that the out-of-distribution performance of ensembles is strongly determined by their in-distribution performance. Why it matters: The findings suggest that larger, single neural networks can replicate the benefits of deep ensembles, potentially simplifying model deployment and reducing computational costs in the region.

Distribution-Free Conformal Joint Prediction Regions for Neural Marked Temporal Point Processes

MBZUAI ·

A presentation will demonstrate the construction of well-calibrated, distribution-free neural Temporal Point Process (TPP) models from multiple event sequences using conformal prediction. The method builds a distribution-free joint prediction region for event arrival time and type with a finite-sample coverage guarantee. The refined method is based on the highest density regions, derived from the joint predictive density of event arrival time and type to address the challenge of creating a joint prediction region for a bivariate response that includes both continuous and discrete data types. Why it matters: This research from a KAUST postdoc improves uncertainty quantification in neural TPPs, which are crucial for modeling continuous-time event sequences, with applications in various fields, by providing more reliable prediction regions.