IDEAS home Printed from https://ideas.repec.org/a/gam/jftint/v14y2022i12p368-d996917.html
   My bibliography  Save this article

Holistic Utility Satisfaction in Cloud Data Centre Network Using Reinforcement Learning

Author

Listed:
  • Pejman Goudarzi

    (ICT Research Institute (ITRC), Tehran 14155-3961, Iran
    These authors contributed equally to this work.)

  • Mehdi Hosseinpour

    (ICT Research Institute (ITRC), Tehran 14155-3961, Iran
    These authors contributed equally to this work.)

  • Roham Goudarzi

    (Faculty of Science, University of British Columbia, Vancouver, BC V6T 1Z4, Canada
    These authors contributed equally to this work.)

  • Jaime Lloret

    (Department of Communications, Polytechnic University of Valencia, Camino de Vera, 46022 Valencia, Spain
    These authors contributed equally to this work.)

Abstract

Cloud computing leads to efficient resource allocation for network users. In order to achieve efficient allocation, many research activities have been conducted so far. Some researchers focus on classical optimisation theory techniques (such as multi-objective optimisation, evolutionary optimisation, game theory, etc.) to satisfy network providers and network users’ service-level agreement (SLA) requirements. Normally, in a cloud data centre network (CDCN), it is difficult to jointly satisfy both the cloud provider and cloud customer’ utilities, and this leads to complex combinatorial problems, which are usually NP-hard. Recently, machine learning and artificial intelligence techniques have received much attention from the networking community because of their capability to solve complicated networking problems. In the current work, at first, the holistic utility satisfaction for the cloud data centre provider and customers is formulated as a reinforcement learning (RL) problem with a specific reward function, which is a convex summation of users’ utility functions and cloud provider’s utility. The user utility functions are modelled as a function of cloud virtualised resources (such as storage, CPU, RAM), connection bandwidth, and also, the network-based expected packet loss and round-trip time factors associated with the cloud users. The cloud provider utility function is modelled as a function of resource prices and energy dissipation costs. Afterwards, a Q-learning implementation of the mentioned RL algorithm is introduced, which is able to converge to the optimal solution in an online and fast manner. The simulation results exhibit the enhanced convergence speed and computational complexity properties of the proposed method in comparison with similar approaches from the joint cloud customer/provider utility satisfaction perspective. To evaluate the scalability property of the proposed method, the results are also repeated for different cloud user population scenarios (small, medium, and large).

Suggested Citation

  • Pejman Goudarzi & Mehdi Hosseinpour & Roham Goudarzi & Jaime Lloret, 2022. "Holistic Utility Satisfaction in Cloud Data Centre Network Using Reinforcement Learning," Future Internet, MDPI, vol. 14(12), pages 1-21, December.
  • Handle: RePEc:gam:jftint:v:14:y:2022:i:12:p:368-:d:996917
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1999-5903/14/12/368/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1999-5903/14/12/368/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Xin Xu & Huiqun Yu, 2014. "A Game Theory Approach to Fair and Efficient Resource Allocation in Cloud Computing," Mathematical Problems in Engineering, Hindawi, vol. 2014, pages 1-14, April.
    2. Šemrov, D. & Marsetič, R. & Žura, M. & Todorovski, L. & Srdic, A., 2016. "Reinforcement learning approach for train rescheduling on a single-track railway," Transportation Research Part B: Methodological, Elsevier, vol. 86(C), pages 250-267.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Wenxing Wu & Jing Xun & Jiateng Yin & Shibo He & Haifeng Song & Zicong Zhao & Shicong Hao, 2023. "An Integrated Method for Reducing Arrival Interval by Optimizing Train Operation and Route Setting," Mathematics, MDPI, vol. 11(20), pages 1-20, October.
    2. Wang, Xuekai & D’Ariano, Andrea & Su, Shuai & Tang, Tao, 2023. "Cooperative train control during the power supply shortage in metro system: A multi-agent reinforcement learning approach," Transportation Research Part B: Methodological, Elsevier, vol. 170(C), pages 244-278.
    3. Ding, Yida & Wandelt, Sebastian & Wu, Guohua & Xu, Yifan & Sun, Xiaoqian, 2023. "Towards efficient airline disruption recovery with reinforcement learning," Transportation Research Part E: Logistics and Transportation Review, Elsevier, vol. 179(C).
    4. Guang Yang & Feng Zhang & Cheng Gong & Shiwen Zhang, 2019. "Application of a Deep Deterministic Policy Gradient Algorithm for Energy-Aimed Timetable Rescheduling Problem," Energies, MDPI, vol. 12(18), pages 1-19, September.
    5. Zhang, Lang & He, Deqiang & He, Yan & Liu, Bin & Chen, Yanjun & Shan, Sheng, 2022. "Real-time energy saving optimization method for urban rail transit train timetable under delay condition," Energy, Elsevier, vol. 258(C).
    6. Sairong Peng & Xin Yang & Hongwei Wang & Hairong Dong & Bin Ning & Haichuan Tang & Zhipeng Ying & Ruijun Tang, 2019. "Dispatching High-Speed Rail Trains via Utilizing the Reverse Direction Track: Adaptive Rescheduling Strategies and Application," Sustainability, MDPI, vol. 11(8), pages 1-20, April.
    7. Yan, Dongyang & Li, Keping & Zhu, Qiaozhen & Liu, Yanyan, 2023. "A railway accident prevention method based on reinforcement learning – Active preventive strategy by multi-modal data," Reliability Engineering and System Safety, Elsevier, vol. 234(C).
    8. Li, Wenqing & Ni, Shaoquan, 2022. "Train timetabling with the general learning environment and multi-agent deep reinforcement learning," Transportation Research Part B: Methodological, Elsevier, vol. 157(C), pages 230-251.
    9. Zhang, Wenyu & Chen, Qian & Yan, Jianyong & Zhang, Shuai & Xu, Jiyuan, 2021. "A novel asynchronous deep reinforcement learning model with adaptive early forecasting method and reward incentive mechanism for short-term load forecasting," Energy, Elsevier, vol. 236(C).
    10. Allan M. C. Bretas & Alexandre Mendes & Martin Jackson & Riley Clement & Claudio Sanhueza & Stephan Chalup, 2023. "A decentralised multi-agent system for rail freight traffic management," Annals of Operations Research, Springer, vol. 320(2), pages 631-661, January.
    11. Jonas Harbering & Abhiram Ranade & Marie Schmidt & Oliver Sinnen, 2019. "Complexity, bounds and dynamic programming algorithms for single track train scheduling," Annals of Operations Research, Springer, vol. 273(1), pages 479-500, February.
    12. Gkiotsalitis, K. & Cats, O., 2021. "At-stop control measures in public transport: Literature review and research agenda," Transportation Research Part E: Logistics and Transportation Review, Elsevier, vol. 145(C).
    13. Ying, Cheng-shuo & Chow, Andy H.F. & Nguyen, Hoa T.M. & Chin, Kwai-Sang, 2022. "Multi-agent deep reinforcement learning for adaptive coordinated metro service operations with flexible train composition," Transportation Research Part B: Methodological, Elsevier, vol. 161(C), pages 36-59.
    14. Liu, Hui & Yu, Chengqing & Wu, Haiping & Duan, Zhu & Yan, Guangxi, 2020. "A new hybrid ensemble deep reinforcement learning model for wind speed short term forecasting," Energy, Elsevier, vol. 202(C).
    15. Kang, Liujiang & Li, Hao & Sun, Huijun & Wu, Jianjun & Cao, Zhiguang & Buhigiro, Nsabimana, 2021. "First train timetabling and bus service bridging in intermodal bus-and-train transit networks," Transportation Research Part B: Methodological, Elsevier, vol. 149(C), pages 443-462.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:14:y:2022:i:12:p:368-:d:996917. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.