Surrogate-based quantification of policy uncertainty in generative flow networks

📅 2025-10-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Generative flow networks rely on approximate reward functions to construct high-reward objects, but noisy data induce epistemic uncertainty in reward estimation, compromising policy reliability. To address this, we propose a surrogate-model-based method for quantifying policy uncertainty: we innovatively employ Polynomial Chaos Expansion (PCE) to establish an analytical mapping between low-dimensional reward parameters and the action distribution; this is combined with lightweight model ensembling and Monte Carlo sampling to efficiently characterize policy sensitivity to reward uncertainty. Evaluated on discrete/continuous grid worlds, symbolic regression, and Bayesian structure learning tasks, our approach achieves high-accuracy, interpretable policy uncertainty estimation while substantially reducing computational cost. It establishes a novel paradigm for trustworthy generative flow modeling.

Technology Category

Application Category

📝 Abstract
Generative flow networks are able to sample, via sequential construction, high-reward, complex objects according to a reward function. However, such reward functions are often estimated approximately from noisy data, leading to epistemic uncertainty in the learnt policy. We present an approach to quantify this uncertainty by constructing a surrogate model composed of a polynomial chaos expansion, fit on a small ensemble of trained flow networks. This model learns the relationship between reward functions, parametrised in a low-dimensional space, and the probability distributions over actions at each step along a trajectory of the flow network. The surrogate model can then be used for inexpensive Monte Carlo sampling to estimate the uncertainty in the policy given uncertain rewards. We illustrate the performance of our approach on a discrete and continuous grid-world, symbolic regression, and a Bayesian structure learning task.
Problem

Research questions and friction points this paper is trying to address.

Quantifying epistemic uncertainty in generative flow network policies
Modeling policy uncertainty from noisy reward function estimations
Estimating policy uncertainty via inexpensive Monte Carlo sampling
Innovation

Methods, ideas, or system contributions that make the work stand out.

Polynomial chaos expansion models policy uncertainty
Surrogate model maps rewards to action distributions
Monte Carlo sampling estimates policy uncertainty efficiently
🔎 Similar Papers
No similar papers found.
R
Ramón Nartallo-Kaluarachchi
IBM Research Europe, Daresbury, United Kingdom
R
Robert Manson-Sawko
IBM Research Europe, Daresbury, United Kingdom
Shashanka Ubaru
Shashanka Ubaru
IBM Research
Numerical Linear AlgebraMachine LearningQuantum Algorithms
D
Dongsung Huh
MIT-IBM Watson AI Lab, Massachusetts, USA
M
Małgorzata J. Zimoń
IBM Research Europe, Daresbury, United Kingdom
Lior Horesh
Lior Horesh
IBM TJ Watson Research Center | Columbia University
Inverse ProblemsNumerical OptimizationNumerical AnalysisAI
Yoshua Bengio
Yoshua Bengio
Professor of computer science, University of Montreal, Mila, IVADO, CIFAR
Machine learningdeep learningartificial intelligence