Doctoral thesis (Dissertations and theses)
Distribution-Free and Calibrated Predictive Uncertainty in Probabilistic Machine Learning
Dheur, Victor
2025
 

Files


Full Text
Victor_Dheur_thesis.pdf
Author postprint (25.18 MB) Creative Commons License - Attribution
Download

All documents in ORBi UMONS are protected by a user license.

Send to



Details



Keywords :
Uncertainty Quantification; Probabilistic Calibration; Generative Modeling; Conformal Prediction; Multi-Output Regression
Abstract :
[en] Machine learning models are increasingly deployed in high-stakes domains such as healthcare and autonomous systems, where decisions carry significant risks. Probabilistic machine learning is valuable in these settings, as it quantifies predictive uncertainty, notably by generating probabilistic predictions. We focus on regression, where the goal is to predict one or more continuous outputs given a set of inputs. In this context, we consider two main forms of uncertainty representation: predictive distributions, which assign probabilities to possible output values, and prediction sets, which are designed to contain the true output with a pre-specified probability. For these predictions to be reliable and informative, they must be calibrated and sharp, i.e., statistically consistent with observed data and concentrated around the true value. In this thesis, we develop distribution-free regression methods to produce calibrated and sharp probabilistic predictions using neural network models. We consider both single-output and the less-explored multi-output regression settings. Specifically, we develop and study recalibration, regularization, and conformal prediction (CP) methods. The first adjusts predictions after model training, the second augments the training objective, and the last produces prediction sets with finite-sample coverage guarantees. For single-output regression, we conduct a large-scale experimental study to provide a comprehensive comparison of these methods. The results reveal that post-hoc approaches consistently achieve superior calibration. We explain this finding by establishing a formal link between recalibration and CP, showing that recalibration also benefits from finite-sample coverage guarantees. However, the separate training and recalibration steps typically lead to degraded negative log-likelihood. To address this issue, we develop an end-to-end training procedure that incorporates the recalibration objective directly into learning, resulting in improved negative log-likelihood while maintaining calibration. For multi-output regression, we conduct a comparative study of CP methods and introduce new classes of approaches that offer novel trade-offs between sharpness, compatibility with generative models, and computational efficiency. A key challenge in CP is achieving conditional coverage, which ensures that coverage guarantees hold for specific inputs rather than only on average. To address this, we propose a method that improves conditional coverage using conditional quantile regression, thereby avoiding the need to estimate full conditional distributions. Finally, for tasks requiring a full predictive density, we introduce a recalibration technique that operates in the latent space of invertible generative models such as conditional normalizing flows. This approach yields an explicit, calibrated multivariate probability density function. Collectively, these contributions advance the theory and practice of uncertainty quantification in machine learning, facilitating the development of more reliable predictive systems across diverse applications.
Disciplines :
Computer science
Author, co-author :
Dheur, Victor  ;  Université de Mons - UMONS > Faculté des Sciences > Service Big Data and Machine Learning
Language :
English
Title :
Distribution-Free and Calibrated Predictive Uncertainty in Probabilistic Machine Learning
Defense date :
2025
Institution :
UMONS - University of Mons [Faculté des Sciences], Mons, Belgium
Degree :
Doctorat en Sciences
Promotor :
Ben Taieb, Souhaib  ;  Université de Mons - UMONS > Faculté des Sciences > Service Big Data and Machine Learning
Dupont, Stéphane  ;  Université de Mons - UMONS > Faculté des Sciences > Service d'Intelligence Artificielle
President :
Wijsen, Jef  ;  Université de Mons - UMONS > Faculté des Sciences > Service des Systèmes d'information
Secretary :
Geurts, Pierre;  ULiège - University of Liège > Department of Electrical engineering and Computer Science > Systems and Modeling
Jury member :
Sesia, Matteo;  University of Southern California > Department of Data Sciences and Operations
Research unit :
S861 - Big Data and Machine Learning
Research institute :
Infortech
Complexys
Funders :
F.R.S.-FNRS - Fund for Scientific Research
Available on ORBi UMONS :
since 16 December 2025

Statistics


Number of views
18 (8 by UMONS)
Number of downloads
7 (3 by UMONS)

Bibliography


Similar publications



Contact ORBi UMONS