Calibrated Bayesian Uncertainty in Random Feature Regression via Optimal Posterior Tempering

Sayan Mukherjee1, Samuel I. Berchuck2
1Center for Scalable Data Analysis and Artificial Intelligence Universität Leipzig Leipzig 04105
2Department of Biostatistics & Bioinformatics Duke University Durham, NC 27705
DOI: https://doi.org/10.71448/bcds2561-2
Published: 30/03/2025
Cite this article as: Sayan Mukherjee, Samuel I. Berchuck. Calibrated Bayesian Uncertainty in Random Feature Regression via Optimal Posterior Tempering. Bulletin of Computer and Data Sciences, Volume 6 Issue 1. Page: 16-38.

Abstract

Bayesian posterior predictive intervals are a central tool for uncertainty quantification in regression. However, in highly overparameterized models—including random feature regression and wide neural networks—recent work has shown that posterior predictive variances can be badly misaligned with the frequentist generalization error of the corresponding point estimator, especially near interpolation thresholds and in low-noise regimes. This misalignment contributes to the “cold posterior effect”, where artificially concentrating the posterior improves empirical performance. In this paper, we propose a principled framework for calibrated Bayesian random feature regression based on posterior tempering. We consider a family of generalized posteriors indexed by a temperature parameter T > 0 that smoothly interpolates between over-dispersed and under-dispersed uncertainty estimates by rescaling the effective regularization in Bayesian random feature regression. Using high-dimensional random matrix asymptotics for random feature models, we derive deterministic limits for the test risk and posterior predictive variance as functions of the feature and sample aspect ratios, signal-to-noise ratio, and temperature. We then define calibration-optimal and risk-optimal temperatures that, respectively, align posterior predictive variance with frequentist prediction error and minimize test risk. Our analysis shows that in overparameterized, low-noise regimes, both optimal temperatures are strictly smaller than one, thus providing a theoretical explanation for the cold posterior effect in random feature models. We complement our asymptotic results with finite-sample simulations that demonstrate substantial improvements in uncertainty calibration and competitive or superior test risk across a wide range of regimes. Finally, we propose practical, data-driven procedures to estimate the temperature from a single dataset, making calibrated Bayesian random feature regression a viable tool for uncertainty-aware prediction in overparameterized systems.

Keywords: Bayesian random feature regression, posterior tempering, uncertainty calibration, cold posterior effect, high-dimensional asymptotics

Abstract

Bayesian posterior predictive intervals are a central tool for uncertainty quantification in regression. However, in highly overparameterized models—including random feature regression and wide neural networks—recent work has shown that posterior predictive variances can be badly misaligned with the frequentist generalization error of the corresponding point estimator, especially near interpolation thresholds and in low-noise regimes. This misalignment contributes to the “cold posterior effect”, where artificially concentrating the posterior improves empirical performance. In this paper, we propose a principled framework for calibrated Bayesian random feature regression based on posterior tempering. We consider a family of generalized posteriors indexed by a temperature parameter T > 0 that smoothly interpolates between over-dispersed and under-dispersed uncertainty estimates by rescaling the effective regularization in Bayesian random feature regression. Using high-dimensional random matrix asymptotics for random feature models, we derive deterministic limits for the test risk and posterior predictive variance as functions of the feature and sample aspect ratios, signal-to-noise ratio, and temperature. We then define calibration-optimal and risk-optimal temperatures that, respectively, align posterior predictive variance with frequentist prediction error and minimize test risk. Our analysis shows that in overparameterized, low-noise regimes, both optimal temperatures are strictly smaller than one, thus providing a theoretical explanation for the cold posterior effect in random feature models. We complement our asymptotic results with finite-sample simulations that demonstrate substantial improvements in uncertainty calibration and competitive or superior test risk across a wide range of regimes. Finally, we propose practical, data-driven procedures to estimate the temperature from a single dataset, making calibrated Bayesian random feature regression a viable tool for uncertainty-aware prediction in overparameterized systems.

Keywords: Bayesian random feature regression, posterior tempering, uncertainty calibration, cold posterior effect, high-dimensional asymptotics
Sayan Mukherjee
Center for Scalable Data Analysis and Artificial Intelligence Universität Leipzig Leipzig 04105
Samuel I. Berchuck
Department of Biostatistics & Bioinformatics Duke University Durham, NC 27705

DOI

Cite this article as:

Sayan Mukherjee, Samuel I. Berchuck. Calibrated Bayesian Uncertainty in Random Feature Regression via Optimal Posterior Tempering. Bulletin of Computer and Data Sciences, Volume 6 Issue 1. Page: 16-38.

Publication history

Copyright © 2025 Sayan Mukherjee, Samuel I. Berchuck. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Browse Advance Search