IDEAS home Printed from https://ideas.repec.org/a/eee/appene/v329y2023ics0306261922014970.html
   My bibliography  Save this article

Incremental incentive mechanism design for diversified consumers in demand response

Author

Listed:
  • Liu, Di
  • Qin, Zhaoming
  • Hua, Haochen
  • Ding, Yi
  • Cao, Junwei

Abstract

Demand response has been proven to be an effective way to improve energy utilization efficiency. It is notable that the diversified characteristics of residential consumers, which many greatly affect its performance in demand response, have not been fully considered in existing incentive mechanisms. In this paper, an incremental incentive mechanism for incentive-based demand response (IBDR) is proposed, in which consumers obtain different incentives according to the increment of response, so that the incentive can follow the change of consumers' marginal cost. We theoretically illustrate that the proposed incremental incentive mechanism can effectively improve the profit of load service entity (LSE), as well as the benefit of highly flexible consumers, compared with other existing incentive mechanism. In practice, LSE's bidding strategy in the day ahead market is affected by the intraday IBDR strategy that cannot be known in advance. In order to solve the bidding problem with incomplete information in the day ahead market, we propose an asynchronous double-interaction deep reinforcement learning (DRL) algorithm to maximize LSE’s cumulative profit of multiple time slots throughout the day. Numerical simulation results show that the proposed mechanism can improve the consumers' response depth while reducing the unit incentive cost, and the proposed DRL algorithm has relatively stable and satisfactory performance even in highly uncertain environment.

Suggested Citation

  • Liu, Di & Qin, Zhaoming & Hua, Haochen & Ding, Yi & Cao, Junwei, 2023. "Incremental incentive mechanism design for diversified consumers in demand response," Applied Energy, Elsevier, vol. 329(C).
  • Handle: RePEc:eee:appene:v:329:y:2023:i:c:s0306261922014970
    DOI: 10.1016/j.apenergy.2022.120240
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0306261922014970
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.apenergy.2022.120240?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Wen, Lulu & Zhou, Kaile & Li, Jun & Wang, Shanyong, 2020. "Modified deep learning and reinforcement learning for an incentive-based demand response model," Energy, Elsevier, vol. 205(C).
    2. Zhao, Liyuan & Yang, Ting & Li, Wei & Zomaya, Albert Y., 2022. "Deep reinforcement learning-based joint load scheduling for household multi-energy system," Applied Energy, Elsevier, vol. 324(C).
    3. Hua, Haochen & Qin, Yuchao & Hao, Chuantong & Cao, Junwei, 2019. "Optimal energy management strategies for energy Internet via deep reinforcement learning approach," Applied Energy, Elsevier, vol. 239(C), pages 598-609.
    4. Wang, Zhaohua & Li, Hao & Deng, Nana & Cheng, Kaiwei & Lu, Bin & Zhang, Bin & Wang, Bo, 2020. "How to effectively implement an incentive-based residential electricity demand response policy? Experience from large-scale trials and matching questionnaires," Energy Policy, Elsevier, vol. 141(C).
    5. Lin, Jin & Dong, Jun & Dou, Xihao & Liu, Yao & Yang, Peiwen & Ma, Tongtao, 2022. "Psychological insights for incentive-based demand response incorporating battery energy storage systems: A two-loop Stackelberg game approach," Energy, Elsevier, vol. 239(PC).
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Kanakadhurga, Dharmaraj & Prabaharan, Natarajan, 2024. "Smart home energy management using demand response with uncertainty analysis of electric vehicle in the presence of renewable energy sources," Applied Energy, Elsevier, vol. 364(C).
    2. Wenhui Zhao & Zilin Wu & Bo Zhou & Jiaoqian Gao, 2024. "Wind and PV Power Consumption Strategy Based on Demand Response: A Model for Assessing User Response Potential Considering Differentiated Incentives," Sustainability, MDPI, vol. 16(8), pages 1-22, April.
    3. Wang, Tonghe & Hua, Haochen & Shi, Tianying & Wang, Rui & Sun, Yizhong & Naidoo, Pathmanathan, 2024. "A bi-level dispatch optimization of multi-microgrid considering green electricity consumption willingness under renewable portfolio standard policy," Applied Energy, Elsevier, vol. 356(C).
    4. Hui Wang & Yao Xu, 2024. "Optimized Decision-Making for Multi-Market Green Power Transactions of Electricity Retailers under Demand-Side Response: The Chinese Market Case Study," Energies, MDPI, vol. 17(11), pages 1-15, May.
    5. Gao, Hongchao & Jin, Tai & Feng, Cheng & Li, Chuyi & Chen, Qixin & Kang, Chongqing, 2024. "Review of virtual power plant operations: Resource coordination and multidimensional interaction," Applied Energy, Elsevier, vol. 357(C).

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Lin, Jin & Dong, Jun & Liu, Dongran & Zhang, Yaoyu & Ma, Tongtao, 2022. "From peak shedding to low-carbon transitions: Customer psychological factors in demand response," Energy, Elsevier, vol. 238(PA).
    2. Sabarathinam Srinivasan & Suresh Kumarasamy & Zacharias E. Andreadakis & Pedro G. Lind, 2023. "Artificial Intelligence and Mathematical Models of Power Grids Driven by Renewable Energy Sources: A Survey," Energies, MDPI, vol. 16(14), pages 1-56, July.
    3. Zhu, Ziqing & Hu, Ze & Chan, Ka Wing & Bu, Siqi & Zhou, Bin & Xia, Shiwei, 2023. "Reinforcement learning in deregulated energy market: A comprehensive review," Applied Energy, Elsevier, vol. 329(C).
    4. Seongwoo Lee & Joonho Seon & Byungsun Hwang & Soohyun Kim & Youngghyu Sun & Jinyoung Kim, 2024. "Recent Trends and Issues of Energy Management Systems Using Machine Learning," Energies, MDPI, vol. 17(3), pages 1-24, January.
    5. Hui Wang & Yao Xu, 2024. "Optimized Decision-Making for Multi-Market Green Power Transactions of Electricity Retailers under Demand-Side Response: The Chinese Market Case Study," Energies, MDPI, vol. 17(11), pages 1-15, May.
    6. Yang, Ting & Zhao, Liyuan & Li, Wei & Zomaya, Albert Y., 2021. "Dynamic energy dispatch strategy for integrated energy system based on improved deep reinforcement learning," Energy, Elsevier, vol. 235(C).
    7. Wang, Yi & Qiu, Dawei & Sun, Mingyang & Strbac, Goran & Gao, Zhiwei, 2023. "Secure energy management of multi-energy microgrid: A physical-informed safe reinforcement learning approach," Applied Energy, Elsevier, vol. 335(C).
    8. Zhu, Jiaoyiling & Hu, Weihao & Xu, Xiao & Liu, Haoming & Pan, Li & Fan, Haoyang & Zhang, Zhenyuan & Chen, Zhe, 2022. "Optimal scheduling of a wind energy dominated distribution network via a deep reinforcement learning approach," Renewable Energy, Elsevier, vol. 201(P1), pages 792-801.
    9. Zhang, Yijie & Ma, Tao & Elia Campana, Pietro & Yamaguchi, Yohei & Dai, Yanjun, 2020. "A techno-economic sizing method for grid-connected household photovoltaic battery systems," Applied Energy, Elsevier, vol. 269(C).
    10. Ahmad, Tanveer & Chen, Huanxin, 2019. "Deep learning for multi-scale smart energy forecasting," Energy, Elsevier, vol. 175(C), pages 98-112.
    11. Ahmad, Tanveer & Madonski, Rafal & Zhang, Dongdong & Huang, Chao & Mujeeb, Asad, 2022. "Data-driven probabilistic machine learning in sustainable smart energy/smart energy systems: Key developments, challenges, and future research opportunities in the context of smart grid paradigm," Renewable and Sustainable Energy Reviews, Elsevier, vol. 160(C).
    12. Zeyue Sun & Mohsen Eskandari & Chaoran Zheng & Ming Li, 2022. "Handling Computation Hardness and Time Complexity Issue of Battery Energy Storage Scheduling in Microgrids by Deep Reinforcement Learning," Energies, MDPI, vol. 16(1), pages 1-20, December.
    13. Kandasamy, Jeevitha & Ramachandran, Rajeswari & Veerasamy, Veerapandiyan & Irudayaraj, Andrew Xavier Raj, 2024. "Distributed leader-follower based adaptive consensus control for networked microgrids," Applied Energy, Elsevier, vol. 353(PA).
    14. Qi Huang & Aihua Jiang & Yu Zeng & Jianan Xu, 2022. "Community Flexible Load Dispatching Model Based on Herd Mentality," Energies, MDPI, vol. 15(13), pages 1-18, June.
    15. Ioanna-M. Chatzigeorgiou & Christos Diou & Kyriakos C. Chatzidimitriou & Georgios T. Andreou, 2021. "Demand Response Alert Service Based on Appliance Modeling," Energies, MDPI, vol. 14(10), pages 1-15, May.
    16. Wang, Zhaohua & Pham, Thi Le Hoa & Sun, Kaining & Wang, Bo & Bui, Quocviet & Hashemizadeh, Ali, 2022. "The moderating role of financial development in the renewable energy consumption - CO2 emissions linkage: The case study of Next-11 countries," Energy, Elsevier, vol. 254(PB).
    17. Wang, Bo & Deng, Nana & Li, Haoxiang & Zhao, Wenhui & Liu, Jie & Wang, Zhaohua, 2021. "Effect and mechanism of monetary incentives and moral suasion on residential peak-hour electricity usage," Technological Forecasting and Social Change, Elsevier, vol. 169(C).
    18. Fathy, Ahmed, 2023. "Bald eagle search optimizer-based energy management strategy for microgrid with renewable sources and electric vehicles," Applied Energy, Elsevier, vol. 334(C).
    19. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    20. Qi, Chunyang & Zhu, Yiwen & Song, Chuanxue & Yan, Guangfu & Xiao, Feng & Da wang, & Zhang, Xu & Cao, Jingwei & Song, Shixin, 2022. "Hierarchical reinforcement learning based energy management strategy for hybrid electric vehicle," Energy, Elsevier, vol. 238(PA).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:appene:v:329:y:2023:i:c:s0306261922014970. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/wps/find/journaldescription.cws_home/405891/description#description .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.