🤖 AI Summary
Modeling parameterized quantum circuits—both near-term and fault-tolerant—becomes intractable at high qubit counts due to exponential state-space growth and computational bottlenecks in gradient-based optimization.
Method: We propose a hardware-agnostic, scalable differentiable quantum state-vector simulation framework built on PyTorch, featuring a distributed tensor computation architecture that integrates automatic differentiation, memory optimization, and accelerator-agnostic parallelization for heterogeneous hardware co-simulation.
Contribution/Results: Our framework achieves, for the first time, end-to-end gradient training of parameterized quantum circuits with up to 1000 qubits while preserving full-state fidelity. It significantly improves simulation throughput and parallel scalability for large-scale circuits. Experiments demonstrate superior training stability and cross-hardware portability across representative variational quantum algorithms—including VQE and QAOA—establishing a unified simulation infrastructure bridging near-term noisy devices and fault-tolerant quantum computing.
📝 Abstract
TorchQuantumDistributed (tqd) is a PyTorch-based [Paszke et al., 2019] library for accelerator-agnostic differentiable quantum state vector simulation at scale. This enables studying the behavior of learnable parameterized near-term and fault- tolerant quantum circuits with high qubit counts.