IDEAS home Printed from https://ideas.repec.org/a/eee/appene/v323y2022ics0306261922008571.html
   My bibliography  Save this article

A reinforcement and imitation learning method for pricing strategy of electricity retailer with customers’ flexibility

Author

Listed:
  • Zhang, Yang
  • Yang, Qingyu
  • Li, Donghe
  • An, Dou

Abstract

The effective pricing of retail broker in competitive electricity market constitutes a key problem toward four goals: (1) the maximization of the broker’s economic benefits; (2) the balance between customers’ energy supply and demand; (3) the realization of the energy supply and demand flexibility potential of customers; (4) the constraint that prevents the retail prices from too high or too low. Unfortunately, few studies can achieve four goals simultaneously. Moreover, the complicated electricity trading environment with continuous states and actions also increases the difficulty of learning optimal pricing strategy. To solve these problems, a reinforcement and imitation learning approach is proposed to develop the optimal pricing strategy of retail broker in this paper. Specifically, the proposed approach consists of a demand prediction method to predict customers’ energy demand and supply volume, a self-generated expert knowledge imitation learning mechanism to instruct the agent to imitate given expert policy with generated expert knowledge, and an action policy learning method. Different from existing studies, our approach achieves all four goals and exploits the generated transition tuples fully to learn a more effective pricing strategy. The proposed scheme is verified by experiments using real-world market data, the experimental results illustrate our proposed approach gains 9.71%, 3.32%, and 15.94% higher economic profits than three state-of-the-art pricing strategies, respectively. Meanwhile, the total needed computation time for our method to learn an effectiveness pricing strategy is only 4102 s. The results show that our method gains the highest economic profits for the broker with acceptable computation cost. Moreover, the changing curves of customers’ consumption/production habits demonstrate that the proposed method could achieve demand/supply response of customers.

Suggested Citation

  • Zhang, Yang & Yang, Qingyu & Li, Donghe & An, Dou, 2022. "A reinforcement and imitation learning method for pricing strategy of electricity retailer with customers’ flexibility," Applied Energy, Elsevier, vol. 323(C).
  • Handle: RePEc:eee:appene:v:323:y:2022:i:c:s0306261922008571
    DOI: 10.1016/j.apenergy.2022.119543
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0306261922008571
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.apenergy.2022.119543?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Nojavan, Sayyad & Zare, Kazem & Mohammadi-Ivatloo, Behnam, 2017. "Optimal stochastic energy management of retailer based on selling price determination under smart grid environment in the presence of demand response program," Applied Energy, Elsevier, vol. 187(C), pages 449-464.
    2. Feihu Hu & Xuan Feng & Hui Cao, 2018. "A Short-Term Decision Model for Electricity Retailers: Electricity Procurement and Time-of-Use Pricing," Energies, MDPI, vol. 11(12), pages 1-18, November.
    3. Lu, Renzhi & Hong, Seung Ho, 2019. "Incentive-based demand response for smart grid with reinforcement learning and deep neural network," Applied Energy, Elsevier, vol. 236(C), pages 937-949.
    4. Saeian, Hosein & Niknam, Taher & Zare, Mohsen & Aghaei, Jamshid, 2022. "Coordinated optimal bidding strategies methods of aggregated microgrids: A game theory-based demand side management under an electricity market environment," Energy, Elsevier, vol. 245(C).
    5. Volodymyr Mnih & Koray Kavukcuoglu & David Silver & Andrei A. Rusu & Joel Veness & Marc G. Bellemare & Alex Graves & Martin Riedmiller & Andreas K. Fidjeland & Georg Ostrovski & Stig Petersen & Charle, 2015. "Human-level control through deep reinforcement learning," Nature, Nature, vol. 518(7540), pages 529-533, February.
    6. Tuballa, Maria Lorena & Abundo, Michael Lochinvar, 2016. "A review of the development of Smart Grid technologies," Renewable and Sustainable Energy Reviews, Elsevier, vol. 59(C), pages 710-725.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Siying Xu & Gaoyu Zhang & Xianzhi Yuan, 2024. "An Enterprise Multi-agent Model with Game Q-Learning Based on a Single Decision Factor," Computational Economics, Springer;Society for Computational Economics, vol. 64(4), pages 2523-2562, October.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Razzak, Abdur & Islam, Md. Tariqul & Roy, Palash & Razzaque, Md. Abdur & Hassan, Md. Rafiul & Hassan, Mohammad Mehedi, 2024. "Leveraging Deep Q-Learning to maximize consumer quality of experience in smart grid," Energy, Elsevier, vol. 290(C).
    2. Omar Al-Ani & Sanjoy Das, 2022. "Reinforcement Learning: Theory and Applications in HEMS," Energies, MDPI, vol. 15(17), pages 1-37, September.
    3. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    4. Qi, Chunyang & Zhu, Yiwen & Song, Chuanxue & Yan, Guangfu & Xiao, Feng & Da wang, & Zhang, Xu & Cao, Jingwei & Song, Shixin, 2022. "Hierarchical reinforcement learning based energy management strategy for hybrid electric vehicle," Energy, Elsevier, vol. 238(PA).
    5. Mohseni, Soheil & Brent, Alan C. & Kelly, Scott & Browne, Will N., 2022. "Demand response-integrated investment and operational planning of renewable and sustainable energy systems considering forecast uncertainties: A systematic review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 158(C).
    6. Park, Keonwoo & Moon, Ilkyeong, 2022. "Multi-agent deep reinforcement learning approach for EV charging scheduling in a smart grid," Applied Energy, Elsevier, vol. 328(C).
    7. Antonopoulos, Ioannis & Robu, Valentin & Couraud, Benoit & Kirli, Desen & Norbu, Sonam & Kiprakis, Aristides & Flynn, David & Elizondo-Gonzalez, Sergio & Wattam, Steve, 2020. "Artificial intelligence and machine learning approaches to energy demand-side response: A systematic review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 130(C).
    8. Deng, Tingting & Yan, Wenzhou & Nojavan, Sayyad & Jermsittiparsert, Kittisak, 2020. "Risk evaluation and retail electricity pricing using downside risk constraints method," Energy, Elsevier, vol. 192(C).
    9. Ma, Siyu & Liu, Hui & Wang, Ni & Huang, Lidong & Goh, Hui Hwang, 2023. "Incentive-based demand response under incomplete information based on the deep deterministic policy gradient," Applied Energy, Elsevier, vol. 351(C).
    10. Lu, Renzhi & Hong, Seung Ho & Zhang, Xiongfeng, 2018. "A Dynamic pricing demand response algorithm for smart grid: Reinforcement learning approach," Applied Energy, Elsevier, vol. 220(C), pages 220-230.
    11. Dadashi, Mojtaba & Haghifam, Sara & Zare, Kazem & Haghifam, Mahmoud-Reza & Abapour, Mehdi, 2020. "Short-term scheduling of electricity retailers in the presence of Demand Response Aggregators: A two-stage stochastic Bi-Level programming approach," Energy, Elsevier, vol. 205(C).
    12. Xie, Jiahan & Ajagekar, Akshay & You, Fengqi, 2023. "Multi-Agent attention-based deep reinforcement learning for demand response in grid-responsive buildings," Applied Energy, Elsevier, vol. 342(C).
    13. Han, Gwangwoo & Joo, Hong-Jin & Lim, Hee-Won & An, Young-Sub & Lee, Wang-Je & Lee, Kyoung-Ho, 2023. "Data-driven heat pump operation strategy using rainbow deep reinforcement learning for significant reduction of electricity cost," Energy, Elsevier, vol. 270(C).
    14. Perera, A.T.D. & Kamalaruban, Parameswaran, 2021. "Applications of reinforcement learning in energy systems," Renewable and Sustainable Energy Reviews, Elsevier, vol. 137(C).
    15. Ning Zhang & Nien-Che Yang & Jian-Hong Liu, 2021. "Optimal Time-of-Use Electricity Price for a Microgrid System Considering Profit of Power Company and Demand Users," Energies, MDPI, vol. 14(19), pages 1-13, October.
    16. Guo, Chenyu & Wang, Xin & Zheng, Yihui & Zhang, Feng, 2022. "Real-time optimal energy management of microgrid with uncertainties based on deep reinforcement learning," Energy, Elsevier, vol. 238(PC).
    17. Dinh, Huy Truong & Lee, Kyu-haeng & Kim, Daehee, 2022. "Supervised-learning-based hour-ahead demand response for a behavior-based home energy management system approximating MILP optimization," Applied Energy, Elsevier, vol. 321(C).
    18. Lu, Renzhi & Li, Yi-Chang & Li, Yuting & Jiang, Junhui & Ding, Yuemin, 2020. "Multi-agent deep reinforcement learning based demand response for discrete manufacturing systems energy management," Applied Energy, Elsevier, vol. 276(C).
    19. Lu, Renzhi & Hong, Seung Ho, 2019. "Incentive-based demand response for smart grid with reinforcement learning and deep neural network," Applied Energy, Elsevier, vol. 236(C), pages 937-949.
    20. Tulika Saha & Sriparna Saha & Pushpak Bhattacharyya, 2020. "Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning," PLOS ONE, Public Library of Science, vol. 15(7), pages 1-28, July.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:appene:v:323:y:2022:i:c:s0306261922008571. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/wps/find/journaldescription.cws_home/405891/description#description .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.