Machine learning for predictive resource scaling of microservices on kubernetes platforms

Research output: Chapter in Book/Report/Conference proceedingConference contribution

3 Citations (Scopus)
313 Downloads (Pure)

Abstract

Resource scaling is the process of adjusting the amount of resources allocated to a system or a service according to the changing demand. For microservices, resource scaling can be done at different levels, such as the container, the pod, or the cluster. However, the current approaches for resource scaling are not good enough because they rely on reactive or rule-based methods that do not account for the dynamic and complex nature of microservices. These methods often lead to over-provisioning or under-provisioning of resources, both affecting the quality of service and the cost efficiency. To address these issues, this work focuses on testing multiple machine learning approaches to optimize the pod dimensioning problem for Kubernetes platforms through predicting resource requirements for an upscaled number of users. The proposed approach aims to address the limitations of the standard Horizontal Pod Autoscaler (HPA), which often results in resource wastage or suboptimal performance. The results were promising and demonstrated high precision and performance of multiple ML models to accurately forecast future resource needs.
Original languageEnglish
Title of host publicationProceedings of the IEEE/ACM 16th International Conference on Utility and Cloud Computing
Number of pages8
DOIs
Publication statusPublished - 04 Dec 2023

Fingerprint

Dive into the research topics of 'Machine learning for predictive resource scaling of microservices on kubernetes platforms'. Together they form a unique fingerprint.

Cite this