VM Scaling and Load Balancing via Cost Optimal MDP Solution

IEEE Transactions on Cloud Computing(2022)

引用 8|浏览0
暂无评分
摘要
Dynamic resource allocation mechanism is an essential building block in contemporary cloud computing environment, enabling the support of the large variability of incoming requests from an enormous number of applications utilizing such cloud infrastructure. In this article, we devise a dynamic resource allocation mechanism that optimizes the application’s profit under the set of costs and revenues while maintaining performance constraints. Specifically, we devise a decision-maker (DM) agent which formulates the joint admission control, scaling and load balancing problem as a stochastic process solvable by Markov decision process (MDP) which provides the optimal policy. Accordingly, at each time instance, the DM can determine based on the system’s current state, on the set of requirements and on the set of costs, whether to add or release a VM (scale-out or scale-in, respectively), whether to admit or reject an upcoming task and if admitting it, which VM to allocate it to. We explore the value function structure and provide insights with respect to the optimal policies produced from it. To address scalability issues of the detailed MDP solution we provide an alternative solution by abstract MDP which consolidates multiple system states into a single abstract state, hence can cope with much larger systems at the expense of slight performance degradation. To demonstrate the feasibility of the suggested scheme, we designed and implemented it, alongside with two traditional auto-scalers, on the Amazon Web Services (AWS) infrastructure. We ran numerous MATLAB simulations and AWS-based experiments which provided insights and demonstrated superiority against the traditional policies we compared with.
更多
查看译文
关键词
Cloud computing,markov decision processes,load balancing,auto-scaling
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要