Towards Safe and Robust Autonomous Vehicle Platooning:

A Data-Model-Knowledge Hybrid-Driven Self-Organizing Cooperative Control Framework

View on GitHub

Abstract

In the emerging hybrid traffic flow environment, which includes both human-driven vehicles (HDVs) and autonomous vehicles (AVs), ensuring safe and robust decision-making and control is crucial for the effective operation of self-driving formations. Current systems for cooperative adaptive cruise control and lane changing are inadequate in responding to real-world emergency situations, limiting the potential of autonomous vehicle platooning technology. To address the aforementioned challenges, we propose a Twin-World Safety-Enhanced Data-Model-Knowledge Hybrid-Driven autonomous vehicle platooning Cooperative Control Framework. Within this framework, a deep reinforcement learning formation decision model integrating traffic priors is designed, and a twin-world deduction model based on safety priority judgment is proposed. Subsequently, an optimal control-based multi-scenario decision-control right adaptive switching mechanism is designed to achieve adaptive switching between data-driven and model-driven methods. Through simulation experiments and hardware-in-the-loop tests, our algorithm has demonstrated excellent performance in terms of safety, robustness, and flexibility.

Images

The overview of the proposed data-model-knowledge hybrid-driven framework for self-organizing autonomous vehicle platooning.

The overview of the proposed data-model-knowledge hybrid-driven framework for self-organizing autonomous vehicle platooning.

Illustration of the considered traffic scenario and the solution. CAVs (Red) and HDVs (Green) are both considered.

Illustration of the considered traffic scenario and the solution. CAVs (Red) and HDVs (Green) are both considered.

Illustration of the designed headway maintaining reward, which considers both the single-vehicle rewards and the multi-vehicle rewards.

Illustration of the designed headway maintaining reward, which considers both the single-vehicle rewards and the multi-vehicle rewards.

Videos

SIL experiment

The model has been trained in a multitude of scenarios, and the following illustration demonstrates the model's performance in SIL with the selected training results from three typical high-risk scenarios.

The risk factors encountered are: human interference, traffic accidents and fluctuations in traffic flow.

HIL experiment

We build a HIL experimental platform and conduct HIL experiments to further test and verify the model's safety and robustness.

In the first scenario, the platoon encounters a severe static obstruction: a multi-vehicle collision blocking multiple lanes.

Our model resolves this complex situation through adaptive decision-making, avoiding secondary collisions by allowing the first vehicle to initiate a lane change and the other vehicles to follow once adequate space is available.

In the second scenario, the platoon faces a dynamic obstruction: the leading vehicle decelerates suddenly at 5 m/s² due to a breakdown.

Our model effectively coordinates lane changes and vehicle positioning, leading to a smooth reformation of the platoon with optimal efficiency.