This paper proposes a deep reinforcement learning (DRL)-based distributed longitudinal control strategy for connected and automated vehicles (CAVs) under communication failure to stabilize traffic oscillations. Specifically, the signalinterference-plus-noise ratio-based vehicle-to-vehicle communication is incorporated into the DRL training environment to reproduce the realistic communication and time-space varying information flow topologies (IFTs). A dynamic information fusion mechanism is designed to smooth the high-jerk control signal caused by the dynamic IFTs. Based on that, each CAV controlled by the DRL-based agent was developed to receive the real-time downstream CAVs' state information and take longitudinal actions to achieve the equilibrium consensus in the multi-agent system. Simulated experiments are conducted to tune the communication adjustment mechanism and further validate the control performance, oscillation dampening performance and generalization capability of our proposed algorithm.
INTRODUCTIONTraffic oscillations, known as the stop-and-go phenomenon (X. Li et al., 2010), contribute to traffic flow instability, traffic unsafety, and energy inefficiency. Connected and automated vehicles (CAVs), equipped with advanced communication and automation capability, have great potential to alleviate traffic oscillations to enhance the traffic flow performance through adaptive cruise control (ACC; Marsden et al., 2001) and cooperative ACC (CACC;Arem et al., 2006).