IDEAS home Printed from https://ideas.repec.org/a/eee/energy/v238y2022ipbs0360544221019733.html
   My bibliography  Save this article

Deep reinforcement learning based direct torque control strategy for distributed drive electric vehicles considering active safety and energy saving performance

Author

Listed:
  • Wei, Hongqian
  • Zhang, Nan
  • Liang, Jun
  • Ai, Qiang
  • Zhao, Wenqiang
  • Huang, Tianyi
  • Zhang, Youtong

Abstract

Distributed drive electric vehicles are regarded as a broadly promising transportation tool owing to their convenience and maneuverability. However, reasonable and efficient allocation of torque demand to four wheels is a challenging task. In this paper, a deep reinforcement learning-based torque distribution strategy is proposed to guarantee the active safety and energy conservation. The torque distribution task is explicitly formulated as a Markov decision process, in which the vehicle dynamic characteristics can be approximated. The actor-critic networks are utilized to approximate the action value and policy functions for a better control effect. To guarantee continuous torque output and further stabilize the learning process, a twin delayed deep deterministic policy gradient algorithm is deployed. The motor efficiency is incorporated into the cumulative reward to reduce the energy consumption. The results of double lane change demonstrate that the proposed strategy results in better handling stability performance. In addition, it can improve the vehicle transient response and eliminate the static deviation in the step steering maneuver test. For typical steering maneuvers, the proposed direct torque distribution strategy significantly improves the average motor efficiency and reduces the energy loss by 5.25%–10.51%. Finally, a hardware-in-loop experiment was implemented to validate the real-time executability of the proposed torque distribution strategy. This study provides a foundation for the practical application of intelligent safety control algorithms in future vehicles.

Suggested Citation

  • Wei, Hongqian & Zhang, Nan & Liang, Jun & Ai, Qiang & Zhao, Wenqiang & Huang, Tianyi & Zhang, Youtong, 2022. "Deep reinforcement learning based direct torque control strategy for distributed drive electric vehicles considering active safety and energy saving performance," Energy, Elsevier, vol. 238(PB).
  • Handle: RePEc:eee:energy:v:238:y:2022:i:pb:s0360544221019733
    DOI: 10.1016/j.energy.2021.121725
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S0360544221019733
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.energy.2021.121725?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Ouyang, Danhua & Zhou, Shen & Ou, Xunmin, 2021. "The total cost of electric vehicle ownership: A consumer-oriented study of China's post-subsidy era," Energy Policy, Elsevier, vol. 149(C).
    2. Li, Zhenhe & Khajepour, Amir & Song, Jinchun, 2019. "A comprehensive review of the key technologies for pure electric vehicles," Energy, Elsevier, vol. 182(C), pages 824-839.
    3. Zou, Yuan & Liu, Teng & Liu, Dexing & Sun, Fengchun, 2016. "Reinforcement learning-based real-time energy management for a hybrid tracked vehicle," Applied Energy, Elsevier, vol. 171(C), pages 372-382.
    4. Zhou, Quan & Li, Ji & Shuai, Bin & Williams, Huw & He, Yinglong & Li, Ziyang & Xu, Hongming & Yan, Fuwu, 2019. "Multi-step reinforcement learning for model-free predictive energy management of an electrified off-highway vehicle," Applied Energy, Elsevier, vol. 255(C).
    5. Wu, Jingda & He, Hongwen & Peng, Jiankun & Li, Yuecheng & Li, Zhanjiang, 2018. "Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus," Applied Energy, Elsevier, vol. 222(C), pages 799-811.
    6. Long, Zoe & Axsen, Jonn & Miller, Inger & Kormos, Christine, 2019. "What does Tesla mean to car buyers? Exploring the role of automotive brand in perceptions of battery electric vehicles," Transportation Research Part A: Policy and Practice, Elsevier, vol. 129(C), pages 185-204.
    7. Volodymyr Mnih & Koray Kavukcuoglu & David Silver & Andrei A. Rusu & Joel Veness & Marc G. Bellemare & Alex Graves & Martin Riedmiller & Andreas K. Fidjeland & Georg Ostrovski & Stig Petersen & Charle, 2015. "Human-level control through deep reinforcement learning," Nature, Nature, vol. 518(7540), pages 529-533, February.
    8. Han, Zhongliang & Xu, Nan & Chen, Hong & Huang, Yanjun & Zhao, Bin, 2018. "Energy-efficient control of electric vehicles based on linear quadratic regulator and phase plane analysis," Applied Energy, Elsevier, vol. 213(C), pages 639-657.
    9. Hu, Xiao & Wang, Ping & Hu, Yunfeng & Chen, Hong, 2020. "A stability-guaranteed and energy-conserving torque distribution strategy for electric vehicles under extreme conditions," Applied Energy, Elsevier, vol. 259(C).
    10. Zhenpo Wang & Yachao Wang & Lei Zhang & Mingchun Liu, 2017. "Vehicle Stability Enhancement through Hierarchical Control for a Four-Wheel-Independently-Actuated Electric Vehicle," Energies, MDPI, vol. 10(7), pages 1-18, July.
    11. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    12. Zhang, LiPeng & Liu, Wei & Qi, BingNan, 2020. "Energy optimization of multi-mode coupling drive plug-in hybrid electric vehicles based on speed prediction," Energy, Elsevier, vol. 206(C).
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Sun, Binbin & Li, Bo & Xing, Jilei & Yu, Xiao & Xie, Mengxue & Hu, Zihao, 2024. "Analysis of the influence of electric flywheel and electromechanical flywheel on electric vehicle economy," Energy, Elsevier, vol. 295(C).
    2. Marouane Adnane & Ahmed Khoumsi & João Pedro F. Trovão, 2023. "Efficient Management of Energy Consumption of Electric Vehicles Using Machine Learning—A Systematic and Comprehensive Survey," Energies, MDPI, vol. 16(13), pages 1-39, June.
    3. Louback, Eduardo & Biswas, Atriya & Machado, Fabricio & Emadi, Ali, 2024. "A review of the design process of energy management systems for dual-motor battery electric vehicles," Renewable and Sustainable Energy Reviews, Elsevier, vol. 193(C).
    4. Fuwu Yan & Jinhai Wang & Changqing Du & Min Hua, 2022. "Multi-Objective Energy Management Strategy for Hybrid Electric Vehicles Based on TD3 with Non-Parametric Reward Function," Energies, MDPI, vol. 16(1), pages 1-17, December.
    5. Deping Wang & Changyang Guan & Junnian Wang & Haisheng Wang & Zhenhao Zhang & Dachang Guo & Fang Yang, 2023. "Review of Energy-Saving Technologies for Electric Vehicles, from the Perspective of Driving Energy Management," Sustainability, MDPI, vol. 15(9), pages 1-17, May.
    6. Wei, Hongqian & Zhang, Youtong & Wang, Yongzhen & Hua, Weiqi & Jing, Rui & Zhou, Yue, 2022. "Planning integrated energy systems coupling V2G as a flexible storage," Energy, Elsevier, vol. 239(PB).
    7. Fan, Likang & Wang, Yufei & Wei, Hongqian & Zhang, Youtong & Zheng, Pengyu & Huang, Tianyi & Li, Wei, 2022. "A GA-based online real-time optimized energy management strategy for plug-in hybrid electric vehicles," Energy, Elsevier, vol. 241(C).

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Daniel Egan & Qilun Zhu & Robert Prucka, 2023. "A Review of Reinforcement Learning-Based Powertrain Controllers: Effects of Agent Selection for Mixed-Continuity Control and Reward Formulation," Energies, MDPI, vol. 16(8), pages 1-31, April.
    2. Dong, Peng & Zhao, Junwei & Liu, Xuewu & Wu, Jian & Xu, Xiangyang & Liu, Yanfang & Wang, Shuhan & Guo, Wei, 2022. "Practical application of energy management strategy for hybrid electric vehicles based on intelligent and connected technologies: Development stages, challenges, and future trends," Renewable and Sustainable Energy Reviews, Elsevier, vol. 170(C).
    3. Hu, Dong & Xie, Hui & Song, Kang & Zhang, Yuanyuan & Yan, Long, 2023. "An apprenticeship-reinforcement learning scheme based on expert demonstrations for energy management strategy of hybrid electric vehicles," Applied Energy, Elsevier, vol. 342(C).
    4. Perera, A.T.D. & Kamalaruban, Parameswaran, 2021. "Applications of reinforcement learning in energy systems," Renewable and Sustainable Energy Reviews, Elsevier, vol. 137(C).
    5. Yang, Ningkang & Han, Lijin & Xiang, Changle & Liu, Hui & Li, Xunmin, 2021. "An indirect reinforcement learning based real-time energy management strategy via high-order Markov Chain model for a hybrid electric vehicle," Energy, Elsevier, vol. 236(C).
    6. Feng, Zhiyan & Zhang, Qingang & Zhang, Yiming & Fei, Liangyu & Jiang, Fei & Zhao, Shengdun, 2024. "Practicability analysis of online deep reinforcement learning towards energy management strategy of 4WD-BEVs driven by dual-motor in-wheel motors," Energy, Elsevier, vol. 290(C).
    7. Lian, Renzong & Peng, Jiankun & Wu, Yuankai & Tan, Huachun & Zhang, Hailong, 2020. "Rule-interposing deep reinforcement learning based energy management strategy for power-split hybrid electric vehicle," Energy, Elsevier, vol. 197(C).
    8. Christian Montaleza & Paul Arévalo & Jimmy Gallegos & Francisco Jurado, 2024. "Enhancing Energy Management Strategies for Extended-Range Electric Vehicles through Deep Q-Learning and Continuous State Representation," Energies, MDPI, vol. 17(2), pages 1-21, January.
    9. Deng, Huifan & Zhao, Youqun & Feng, Shilin & Wang, Qiuwei & Zhang, Chenxi & Lin, Fen, 2021. "Torque vectoring algorithm based on mechanical elastic electric wheels with consideration of the stability and economy," Energy, Elsevier, vol. 219(C).
    10. Yang, Ningkang & Ruan, Shumin & Han, Lijin & Liu, Hui & Guo, Lingxiong & Xiang, Changle, 2023. "Reinforcement learning-based real-time intelligent energy management for hybrid electric vehicles in a model predictive control framework," Energy, Elsevier, vol. 270(C).
    11. Alessia Musa & Pier Giuseppe Anselma & Giovanni Belingardi & Daniela Anna Misul, 2023. "Energy Management in Hybrid Electric Vehicles: A Q-Learning Solution for Enhanced Drivability and Energy Efficiency," Energies, MDPI, vol. 17(1), pages 1-20, December.
    12. Xu, Bin & Rathod, Dhruvang & Zhang, Darui & Yebi, Adamu & Zhang, Xueyu & Li, Xiaoya & Filipi, Zoran, 2020. "Parametric study on reinforcement learning optimized energy management strategy for a hybrid electric vehicle," Applied Energy, Elsevier, vol. 259(C).
    13. Chen, Zheng & Hu, Hengjie & Wu, Yitao & Zhang, Yuanjian & Li, Guang & Liu, Yonggang, 2020. "Stochastic model predictive control for energy management of power-split plug-in hybrid electric vehicles based on reinforcement learning," Energy, Elsevier, vol. 211(C).
    14. Wu, Yuankai & Tan, Huachun & Peng, Jiankun & Zhang, Hailong & He, Hongwen, 2019. "Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus," Applied Energy, Elsevier, vol. 247(C), pages 454-466.
    15. Zhou, Jianhao & Xue, Siwu & Xue, Yuan & Liao, Yuhui & Liu, Jun & Zhao, Wanzhong, 2021. "A novel energy management strategy of hybrid electric vehicle via an improved TD3 deep reinforcement learning," Energy, Elsevier, vol. 224(C).
    16. Wei, Hongqian & Ai, Qiang & Zhao, Wenqiang & Zhang, Youtong, 2022. "Modelling and experimental validation of an EV torque distribution strategy towards active safety and energy efficiency," Energy, Elsevier, vol. 239(PA).
    17. Wang, Hanchen & Ye, Yiming & Zhang, Jiangfeng & Xu, Bin, 2023. "A comparative study of 13 deep reinforcement learning based energy management methods for a hybrid electric vehicle," Energy, Elsevier, vol. 266(C).
    18. Liu, Teng & Tan, Wenhao & Tang, Xiaolin & Zhang, Jinwei & Xing, Yang & Cao, Dongpu, 2021. "Driving conditions-driven energy management strategies for hybrid electric vehicles: A review," Renewable and Sustainable Energy Reviews, Elsevier, vol. 151(C).
    19. Ye, Yiming & Wang, Hanchen & Xu, Bin & Zhang, Jiangfeng, 2023. "An imitation learning-based energy management strategy for electric vehicles considering battery aging," Energy, Elsevier, vol. 283(C).
    20. Matteo Acquarone & Claudio Maino & Daniela Misul & Ezio Spessa & Antonio Mastropietro & Luca Sorrentino & Enrico Busto, 2023. "Influence of the Reward Function on the Selection of Reinforcement Learning Agents for Hybrid Electric Vehicles Real-Time Control," Energies, MDPI, vol. 16(6), pages 1-22, March.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:energy:v:238:y:2022:i:pb:s0360544221019733. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/energy .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.