🤖 AI Summary
High-dimensional engineering design optimization is challenged by the coupled effects of aleatory and epistemic uncertainties. Method: This paper proposes a multi-level surrogate modeling framework based on hierarchical orthogonal decomposition, adaptively constructing and updating multiple non-intrusive Kriging surrogates. It integrates uncertainty-sensitive data filtering, hierarchical sampling, and orthogonal projection to achieve efficient, high-fidelity mapping of large-scale design spaces. Contribution/Results: The method significantly enhances scalability and statistical robustness. On benchmark analytical test problems, it improves both computational efficiency and optimization accuracy by two to three orders of magnitude over state-of-the-art approaches, while incurring minimal computational resource overhead. It establishes a new paradigm for uncertainty-driven high-dimensional engineering optimization—balancing economic feasibility, predictive accuracy, and practical deployability.
📝 Abstract
Engineering design involves demanding models encompassing many decision variables and uncontrollable parameters. In addition, unavoidable aleatoric and epistemic uncertainties can be very impactful and add further complexity. The state-of-the-art adopts two steps, uncertainty quantification and design optimization, to optimize systems under uncertainty by means of robust or stochastic metrics. However, conventional scenario-based, surrogate-assisted, and mathematical programming methods are not sufficiently scalable to be affordable and precise in large and complex cases. Here, a multi-level approach is proposed to accurately optimize resource-intensive, high-dimensional, and complex engineering problems under uncertainty with minimal resources. A non-intrusive, fast-scaling, Kriging-based surrogate is developed to map the combined design/parameter domain efficiently. Multiple surrogates are adaptively updated by hierarchical and orthogonal decomposition to leverage the fewer and most uncertainty-informed data. The proposed method is statistically compared to the state-of-the-art via an analytical testbed and is shown to be concurrently faster and more accurate by orders of magnitude.