Aleatoric and Epistemic Uncertainty
in Statistics and Machine Learning 

   Tutorial at the
European Conference on Machine Learning and Principles
and Practice of Knowledge Discovery in Databases (ECML PKDD) 


Without any doubt, the notion of uncertainty is of major importance in machine learning and constitutes a key element of modern machine learning methodology. In recent years, it has gained in importance due to the increasing relevance of machine learning for practical applications, many of which are coming with safety requirements. In this regard, new problems and challenges have been identified by machine learning scholars, many of which call for novel methodological developments. Indeed, while uncertainty has a long tradition in statistics, and a broad range of useful concepts for representing and quantifying uncertainty have been developed on the basis of probability theory, recent research has gone beyond traditional approaches and also leverages more general formalism and uncertainty calculi.

In the statistics literature, two inherently different sources of uncertainty are commonly distinguished, referred to as aleatoric and epistemic. While the former refers to variability due to inherently random effects, the latter is uncertainty caused by a lack of knowledge and hence relates to the epistemic state of an agent. Thus, epistemic uncertainty can in principle be reduced on the basis of additional information, while aleatoric uncertainty is non-reducible. The distinction between different types of uncertainty and their quantification has also been adopted in the recent ML literature, and various methods for quantifying aleatoric and epistemic uncertainty have been proposed. In the context of supervised learning, the focus is typically on predictive uncertainty, i.e., the learner’s uncertainty in the outcome given a query instance for which a prediction is sought. The aleatoric part of this uncertainty is due to the supposedly stochastic nature of the dependence between instances and outcomes, e.g. due to a lack of informative features or noisy class annotations. Therefore, the “ground-truth” is a conditional probability distribution, i.e., each outcome has a certain probability to occur. Even with full knowledge about the underlying data-generating process, the outcome cannot be predicted with certainty.

Obviously, the learner does not have full knowledge of this ground-truth. Instead, it produces a “guess” on the basis of the sample data provided for training. Broadly speaking, epistemic uncertainty is uncertainty about the true probability and hence the discrepancy between the ground-truth and the learner's guess. This (second-order) uncertainty can be captured and represented in different ways. One approach is to train a probabilistic predictor in a more or less standard way, and to quantify the uncertainty of that predictor in a kind of post hoc manner. An example is the estimation of epistemic uncertainty in terms of the mutual information between the target variable and the model parameters.

Another idea is to estimate uncertainty in a more direct way, and to let the learner itself predict, not only the target variable, but also its own uncertainty about the prediction. For example, instead of predicting a probability distribution, the learner may predict a second-order distribution (distribution of distributions) or a set of distributions. Its epistemic uncertainty is then represented by the “peakedness” of the former or the size of the latter.  

This tutorial aims to provide an overview of uncertainty quantification in machine learning, a topic that has received increasing attention in the recent past. Starting with a recapitulation of classical statistical concepts, we specifically focus on novel approaches for distinguishing and representing so-called aleatoric and epistemic uncertainty. By the end of the tutorial, attendees will have a comprehensive understanding of the fundamental concepts and recent advances in this field.

TARGET AUDIENCE

With the tutorial we aim to attract both researchers that are already active in one of the above domains, as well as researchers with little or no prior experience in uncertainty quantification. We will only assume a general familiarity with well-known machine learning techniques for classification and regression (neural networks, probabilistic models, risk minimization, etc.). Therefore, the tutorial is of interest to scholars from diverse subfields of machine learning and with different backgrounds.


PRESENTERS

Viktor Bengs

LMU Munich, Germany

Eyke Hüllermeier

LMU Munich, Germany

Willem Waegeman

Ghent University, Belgium

SCHEDULE

This tutorial will be held in-person at ECML-PKDD 2023 at the Officine Grandi Riparazioni on September 22nd 2023 in the morning. The program schedule will be as follows.


Introduction


by Eyke Hüllermeier

Basic statistical concepts

Frequentist inference

Bayesian inference

by Viktor Bengs

First-order uncertainty representation

Representations based on probability distributions

Representations based on sets

Extensions for regression 

by Willem Waegeman

Second-order uncertainty representation

Aleatoric and epistemic uncertainty 

Bayesian inference 

Direct uncertainty prediction  

Uncertainty quantification 

by Eyke Hüllermeier

SLIDES

Slides are available: here


PRACTICAL INFORMATION

For registration, see the ECML-PKDD 2023 website.


VENUE (September 22nd, Room  PoliTo Room 7T)

Officine Grandi Riparazioni, Turin, Italy

Map View