DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Implications of stop-and-go traffic on training learning-based car-following control

Journal Article · · Transportation Research Part C: Emerging Technologies

Learning-based car-following control (LCC) of connected and autonomous vehicles (CAVs) is gaining significant attention with the advancement of computing power and data accessibility. While the flexibility and large model capacity of model-free architecture enable LCC to potentially outperform the model-based car-following (CF) model in improving traffic efficiency and mitigating congestion, the generalizability of LCC for traffic conditions different from the training environment/dataset is not well-understood. Herein, this study seeks to explore the impact of stop-and-go traffic in the training dataset on the generalizability of LCC. It uses the characteristics of lead vehicle trajectories to describe stop-and-go traffic, and links the theory of identifiability (i.e., obtaining a unique parameter estimation result using sensor measurements) to the generalizability of behavior cloning (BC) and policy-based deep reinforcement learning (DRL). Correspondingly, the study shows theoretically that: (i) stop-and-go traffic can enable the property of identifiability and enhance the control performance of BC-based LCC in different traffic conditions; (ii) stop-and-go traffic is not necessary for DRL-based LCC to generalize to different traffic conditions; (iii) DRL-based LCC trained with only constant-speed lead vehicle trajectories (not sufficient to ensure identifiability) can be generalized to different traffic conditions; and (iv) stop-and-go traffic increases variance in the training dataset, which improves the convergence of parameter estimation while negatively impacting the convergence of DRL to the optimal control policy. Numerical experiments validate the above findings, illustrating that BC-based LCC entails comprehensive training datasets for generalizing to different traffic conditions, while DRL-based LCC can achieve generalization with simple free-flow traffic training environments. This further suggests DRL as a more promising and cost-effective LCC approach to reduce operational costs, mitigate traffic congestion, and enhance safety and mobility, which can accelerate the deployment and acceptance of CAVs.

Research Organization:
Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States)
Sponsoring Organization:
USDOE
Grant/Contract Number:
AC05-00OR22725
OSTI ID:
2438666
Journal Information:
Transportation Research Part C: Emerging Technologies, Journal Name: Transportation Research Part C: Emerging Technologies Vol. 168; ISSN 0968-090X
Publisher:
ElsevierCopyright Statement
Country of Publication:
United States
Language:
English

References (20)

System Identification book December 1999
Proactive longitudinal control to preclude disruptive lane changes of human-driven vehicles in mixed-flow traffic journal July 2023
Physics-informed deep reinforcement learning-based integrated two-dimensional car-following control strategy for connected automated vehicles journal June 2023
Characterization of Traffic Oscillation Propagation under Nonlinear Car-Following Laws journal January 2011
Measurement and estimation of traffic oscillation properties journal January 2010
Trajectory data reconstruction and simulation-based validation against macroscopic traffic patterns journal October 2015
Robust local and string stability for a decentralized car following control strategy for connected automated vehicles journal July 2019
Distributed model predictive control approach for cooperative car-following with guaranteed local and string stability journal October 2019
A real-time deployable model predictive control-based cooperative platooning approach for connected and autonomous vehicles journal October 2019
Influence of connected and autonomous vehicles on traffic flow stability and throughput journal October 2016
A physics-informed deep learning paradigm for car-following models journal September 2021
Connected automated vehicle cooperative control with a deep reinforcement learning approach in a mixed traffic environment journal December 2021
Congestion-mitigating MPC design for adaptive cruise control based on Newell’s car following model: History outperforms prediction journal September 2022
Dense reinforcement learning for safety validation of autonomous vehicles journal March 2023
Stability-Certified Reinforcement Learning: A Control-Theoretic Perspective journal January 2020
Proactive Longitudinal Control of Connected and Autonomous Vehicles with Lane-Change Assistance for Human-Driven Vehicles conference September 2021
A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients journal November 2012
Smooth-Switching Control-Based Cooperative Adaptive Cruise Control by Considering Dynamic Information Flow Topology journal March 2020
Liquid Time-constant Networks journal May 2021
Gershgorin's Theorem and the Zeros of Polynomials journal March 1965