Optimal Configuration of API Resources in Cloud Native Computing

📅 2025-12-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the lack of systematic optimization for CPU and memory resource allocation during the Release phase of cloud-native DevOps. We propose the first pre-deployment offline performance optimization framework for microservices—distinct from mainstream auto-scaling research focused on the Ops phase. Our approach performs fine-grained resource configuration tuning *before* deployment, thereby mitigating auto-scaling failures caused by suboptimal memory provisioning. Methodologically, we integrate Bayesian optimization, statistical experimental design, and a goal-directed factor screening strategy to balance sampling cost and approximation accuracy. Extensive evaluation on the TeaStore benchmark demonstrates that our pre-deployment optimization significantly improves memory suitability and API-level resource utilization. Moreover, it empirically validates the necessity and context-dependent applicability of factor screening under varying optimization objectives.

Technology Category

Application Category

📝 Abstract
This paper presents how an existing framework for offline performance optimization can be applied to microservice applications during the Release phase of the DevOps life cycle. Optimization of resource allocation configuration parameters for CPU and memory during the Release phase remains a largely unexplored problem as most research has focused on intelligent scheduling and autoscaling of microservices during the Ops stage of the DevOps cycle. Yet horizontal auto-scaling of containers, based on CPU usage for instance, may still leave these containers with an inappropriately allocated amount of memory, if no upfront fine-tuning of both resources is applied before the Deployment phase. We evaluate the performance optimization framework using the TeaStore microservice application and statistically compare different optimization algorithms, supporting informed decisions about their trade-offs between sampling cost and distance to the optimal resource configuration. This shows that upfront factor screening, for reducing the search space, is helpful when the goal is to find the optimal resource configuration with an affordable sampling budget. When the goal is to statistically compare different algorithms, screening must also be applied to make data collection of all data points in the search space feasible. If the goal is to find a near-optimal configuration, however, it is better to run bayesian optimization without screening.
Problem

Research questions and friction points this paper is trying to address.

Optimize CPU and memory resource allocation for microservices pre-deployment
Address unexplored resource configuration in DevOps Release phase
Compare optimization algorithms for cost-effective near-optimal configurations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Applies offline performance optimization to DevOps Release phase
Uses factor screening to reduce search space for resource configuration
Compares Bayesian optimization with screening for near-optimal results
🔎 Similar Papers
No similar papers found.