Integrating Predictive and Generative Capabilities by Latent Space Design via the DKL-VAE Model

📅 2025-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the decoupling of prediction and generation capabilities in generative models for scientific discovery, this paper proposes a VAE-DKL synergistic framework: deep kernel learning (DKL) is embedded into the latent space of a variational autoencoder (VAE), enabling structured modeling and property-guided optimization of latent variables via Gaussian process (GP) regression. This approach achieves, for the first time, end-to-end joint training of generation and property prediction. On the QM9 dataset, it attains a prediction error of 0.12 eV for enthalpy—surpassing a standalone VAE+GP baseline—and enables controllable generation of novel molecular structures outside the training set that satisfy target property constraints. The core innovation lies in the co-optimization mechanism within the latent space: it preserves the VAE’s high-fidelity sampling capability while endowing it with interpretable and intervenable property prediction functionality, thereby significantly enhancing efficiency in materials inverse design.

Technology Category

Application Category

📝 Abstract
We introduce a Deep Kernel Learning Variational Autoencoder (VAE-DKL) framework that integrates the generative power of a Variational Autoencoder (VAE) with the predictive nature of Deep Kernel Learning (DKL). The VAE learns a latent representation of high-dimensional data, enabling the generation of novel structures, while DKL refines this latent space by structuring it in alignment with target properties through Gaussian Process (GP) regression. This approach preserves the generative capabilities of the VAE while enhancing its latent space for GP-based property prediction. We evaluate the framework on two datasets: a structured card dataset with predefined variational factors and the QM9 molecular dataset, where enthalpy serves as the target function for optimization. The model demonstrates high-precision property prediction and enables the generation of novel out-of-training subset structures with desired characteristics. The VAE-DKL framework offers a promising approach for high-throughput material discovery and molecular design, balancing structured latent space organization with generative flexibility.
Problem

Research questions and friction points this paper is trying to address.

Integrates VAE generative power with DKL predictive capabilities
Enhances latent space for Gaussian Process-based property prediction
Enables high-throughput material discovery and molecular design
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates VAE generative power with DKL predictive nature
Refines latent space using Gaussian Process regression
Enables high-precision property prediction and novel structure generation
🔎 Similar Papers
No similar papers found.
Boris Slautin
Boris Slautin
Postdoctoral researcher, University of Duisburg-Essen
Solid-State PhysicsScanning Probe MicroscopyAutomated ExperimentationAI for Material Discovery
U
Utkarsh Pratiush
Department of Materials Science and Engineering, University of Tennessee, Knoxville, TN 37996, USA
D
D. Lupascu
Institute for Materials Science and Center for Nanointegration Duisburg-Essen (CENIDE), University of Duisburg-Essen, Essen, 45141, Germany
M
M. Ziatdinov
Pacific Northwest National Laboratory, Richland, WA 99354, USA; Department of Materials Science and Engineering, University of Tennessee, Knoxville, TN 37996, USA
Sergei V. Kalinin
Sergei V. Kalinin
Weston Fulton Chair Professor, UT Knoxville. Chief Scientist, AI/ML for Physical Sciences, PNNL
AI4Materialsautomated experimentelectron microscopySPMatomic fabrication