Skip to content
GCC AI Research

A Unified Deep Model of Learning from both Data and Queries for Cardinality Estimation

arXiv · · Significant research

Summary

This paper introduces a unified deep autoregressive model (UAE) for cardinality estimation that learns joint data distributions from both data and query workloads. It uses differentiable progressive sampling with the Gumbel-Softmax trick to incorporate supervised query information into the deep autoregressive model. Experiments show UAE achieves better accuracy and efficiency compared to state-of-the-art methods.

Get the weekly digest

Top AI stories from the GCC region, every week.

Related

Duet: efficient and scalable hybriD neUral rElation undersTanding

arXiv ·

The paper introduces Duet, a hybrid neural relation understanding method for cardinality estimation. Duet addresses limitations of existing learned methods, such as high costs and scalability issues, by incorporating predicate information into an autoregressive model. Experiments demonstrate Duet's efficiency, accuracy, and scalability, even outperforming GPU-based methods on CPU.

Distribution-Free Conformal Joint Prediction Regions for Neural Marked Temporal Point Processes

MBZUAI ·

A presentation will demonstrate the construction of well-calibrated, distribution-free neural Temporal Point Process (TPP) models from multiple event sequences using conformal prediction. The method builds a distribution-free joint prediction region for event arrival time and type with a finite-sample coverage guarantee. The refined method is based on the highest density regions, derived from the joint predictive density of event arrival time and type to address the challenge of creating a joint prediction region for a bivariate response that includes both continuous and discrete data types. Why it matters: This research from a KAUST postdoc improves uncertainty quantification in neural TPPs, which are crucial for modeling continuous-time event sequences, with applications in various fields, by providing more reliable prediction regions.

Explainable Fact Checking for Statistical and Property Claims

MBZUAI ·

EURECOM researchers developed data-driven verification methods using structured datasets to assess statistical and property claims. The approach translates text claims into SQL queries on relational databases for statistical claims. For property claims, they use knowledge graphs to verify claims and generate explanations. Why it matters: The methods aim to support fact-checkers by efficiently labeling claims with interpretable explanations, potentially combating misinformation in the region and beyond.

Learning to act in noisy contexts using deep proxy learning

MBZUAI ·

Researchers are exploring methods for evaluating the outcome of actions using off-policy observations where the context is noisy or anonymized. They employ proxy causal learning, using two noisy views of the context to recover the average causal effect of an action without explicitly modeling the hidden context. The implementation uses learned neural net representations for both action and context, and demonstrates outperformance compared to an autoencoder-based alternative. Why it matters: This research addresses a key challenge in applying AI in real-world scenarios where data privacy or bandwidth limitations necessitate working with noisy or anonymized data.