🤖 AI Summary
This paper addresses functional output regression—a neglected yet critical supervised learning problem involving predictions of structured outputs such as spectra or probability density functions. We propose a novel Bayesian framework integrating full kernelization within a reproducing kernel Hilbert space (RKHS). Our key contribution is the first analytical derivation of the posterior predictive distribution over functional outputs in RKHS, enabled by a multitask-style functional covariance model that captures intrinsic correlations among output functions. The method retains model simplicity while efficiently handling high-dimensional nonlinearities. It supports principled uncertainty quantification and yields closed-form functional predictive distributions. Empirical evaluation on synthetic benchmarks and a materials science task—predicting electronic density of states—demonstrates substantial improvements in accuracy and generalization over baselines, confirming the approach’s robustness and practical utility.
📝 Abstract
In supervised learning, the output variable to be predicted is often represented as a function, such as a spectrum or probability distribution. Despite its importance, functional output regression remains relatively unexplored. In this study, we propose a novel functional output regression model based on kernel methods. Unlike conventional approaches that independently train regressors with scalar outputs for each measurement point of the output function, our method leverages the covariance structure within the function values, akin to multitask learning, leading to enhanced learning efficiency and improved prediction accuracy. Compared with existing nonlinear function-on-scalar models in statistical functional data analysis, our model effectively handles high-dimensional nonlinearity while maintaining a simple model structure. Furthermore, the fully kernel-based formulation allows the model to be expressed within the framework of reproducing kernel Hilbert space (RKHS), providing an analytic form for parameter estimation and a solid foundation for further theoretical analysis. The proposed model delivers a functional output predictive distribution derived analytically from a Bayesian perspective, enabling the quantification of uncertainty in the predicted function. We demonstrate the model's enhanced prediction performance through experiments on artificial datasets and density of states prediction tasks in materials science.