Human-leading automated truck platooning has been an effective technique to improve traffic capacity and fuel economy and eliminate uncertainties of the traffic environment. Aiming for a tradeoff between the dynamic response of car following and energy-efficient platooning control, a predictive information multiagent soft actor–critic (PI-MASAC) control framework is proposed for a human-leading automated heavy-duty-truck platoon. In this framework, predictive information of environmental dynamics is modeled as the state representation of a deep reinforcement learning algorithm to address the uncertainties of a partially observable environment. In the truck model, the impact of intraplatoon aerodynamic interactions is modeled, which is used to design a constant spacing policy for platooning control. We demonstrate the effectiveness of our approach by testing the human-leading truck platoon under multiple scenarios compared to proximal policy optimization, an intelligent driver model, and linear-based cooperative adaptive cruise control. Our results show that the PI-MASAC learns a novel car-following strategy of peak shaving and valley filling and therefore significantly enhances energy savings by reducing high-intensity accelerations and decelerations. In addition, the PI-MASAC demonstrates its adaptability to various initial scenarios and exhibits good generalization to a larger platoon size.
详细内容请参阅论文原文
If you find our work is useful in your research, please consider citing:
@ARTICLE{10273625,
author={Lian, Renzong and Li, Zhiheng and Wen, Boxuan and Wei, Junqing and Zhang, Jiawei and Li, Li},
journal={IEEE Intelligent Transportation Systems Magazine},
title={Multiagent Deep Reinforcement Learning for Automated Truck Platooning Control},
year={2023},
volume={},
number={},
pages={2-17},
doi={10.1109/MITS.2023.3319091}}