Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
GPT-5.1
GPT-5.1 130 tok/s
Gemini 3.0 Pro 29 tok/s Pro
Gemini 2.5 Flash 145 tok/s Pro
Kimi K2 191 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

Enhancing System-Level Safety in Mixed-Autonomy Platoon via Safe Reinforcement Learning (2401.11148v2)

Published 20 Jan 2024 in eess.SY and cs.SY

Abstract: Connected and automated vehicles (CAVs) have recently gained prominence in traffic research due to advances in communication technology and autonomous driving. Various longitudinal control strategies for CAVs have been developed to enhance traffic efficiency, stability, and safety in mixed-autonomy scenarios. Deep reinforcement learning (DRL) is one promising strategy for mixed-autonomy platoon control, thanks to its capability of managing complex scenarios in real time after sufficient offline training. However, there are three research gaps for DRL-based mixed-autonomy platoon control: (i) the lack of theoretical collision-free guarantees, (ii) the widely adopted but impractical assumption of skilled and rational drivers who will not collide with preceding vehicles, and (iii) the strong assumption of a known human driver model. To address these research gaps, we propose a safe DRL-based controller that can provide a system-level safety guarantee for mixed-autonomy platoon control. First, we combine control barrier function (CBF)-based safety constraints and DRL via a quadratic programming (QP)-based differentiable neural network layer to provide theoretical safety guarantees. Second, we incorporate system-level safety constraints into our proposed method to account for the safety of both CAVs and the following HDVs to address the potential collisions due to irrational human driving behavior. Third, we devise a learning-based system identification approach to estimate the unknown human car-following behavior in the real system. Simulation results demonstrate that our proposed method effectively ensures CAV safety and improves HDV safety in mixed platoon environments while simultaneously enhancing traffic capacity and string stability.

Citations (6)

Summary

  • The paper introduces a safe RL controller that integrates CBFs to ensure collision-free, system-level safety for CAVs in mixed traffic.
  • It employs a learning-based system identification method to adapt to unpredictable human-driven vehicle behaviors.
  • Simulation results show improved traffic efficiency and expanded safety regions under high-disturbance conditions.

Enhancing System-Level Safety in Mixed-Autonomy Platoon via Safe Reinforcement Learning

Introduction

The paper "Enhancing System-Level Safety in Mixed-Autonomy Platoon via Safe Reinforcement Learning" (2401.11148) investigates the challenges of controlling connected and automated vehicles (CAVs) in mixed-autonomy traffic environments using Deep Reinforcement Learning (DRL). It addresses three significant gaps in current DRL research for CAVs: the lack of collision-free guarantees, the assumption of rational driver behavior, and the neglect of unknown human driver models. The paper presents a safe DRL-based approach integrating Control Barrier Functions (CBFs) for system-level safety guarantees and proposes a learning-based system identification method for human driver behavior.

System Design and Methodology

Mixed-Autonomy Traffic Modeling

The framework considers a single CAV in a platoon of human-driven vehicles (HDVs), assuming realistic scenarios where CAV penetration remains low. The dynamics of CAVs are governed by standard acceleration models, while HDVs are simulated through car-following laws, notably the Optimal Velocity Model (OVM), with the actual parameters unknown to the CAVs. This setup enables the CAV to learn and adapt its behavior based on observed traffic patterns using a reinforcement learning framework.

Reinforcement Learning and Safety Considerations

PPO is utilized as the main RL algorithm, formulated to improve traffic efficiency, stability, and ensure safety. The innovation lies in a differentiable neural network embedding quadratic programming layers to ensure CBF-based safety constraints are met. This ensures that actions chosen by the RL agent remain within safe operational boundaries, accounting for both immediate and system-level safety constraints. Figure 1

Figure 1: Overview of the proposed controller for CAVs in mixed-autonomy traffic framework, the dotted line represents backpropagation of the differentiable QP.

Learning-Based System Identification

The approach employs a hybrid system identification strategy combining linear models with neural networks to approximate and update unknown HDV behavior in real-time. This method enables CAVs to dynamically adapt to evolving traffic conditions without pre-assumed driver models, reducing system vulnerabilities to unexpected HDV actions.

Implementation and Results

Simulation Setup

Two primary scenarios were simulated: (1) where an HDV makes an emergency deceleration, and (2) where an HDV behind the CAV accelerates unexpectedly. Training incorporated realistic disturbances, engineering CAV responses to unmodeled, erratic human behaviors.

Performance Analysis

The proposed method effectively expanded safety regions and decreased collision risks under high-disturbance conditions (Figure 2), outperforming standard DRL implementations by a significant margin. The impacts on traffic efficiency and system-level stability were quantified, showing marked improvements over traditional approaches, particularly in mitigating adverse effects from disregard for HDV interactions. Figure 2

Figure 2

Figure 2: Safety-guaranteed regions associated with two specific scenarios: (a) the deceleration disturbance of the preceding vehicle and (b) the acceleration disturbance of the following vehicle. The horizontal axis denotes the magnitude of the disturbance signal (acceleration/deceleration), and the vertical axis is the duration of the disturbance signal. The dark blue regions denote the safety region of the LCC with PPO controller without the safety layer, and the light blue regions illustrate the expanded safety region achieved by implementing the safe RL controller for LCC (i.e., with safety layer), while the white regions indicate the unsafe region.

Comparative Performance

The integrated framework improved robustness against disturbances and better handled uncertainty, particularly in mixed braking and acceleration scenarios. Regular DRL methods failed in maintaining non-collision conditions, necessitating the integration of CBF for guaranteed operational safety.

Conclusion and Future Work

The paper introduces an advanced safe RL-based controller for utilizing CBFs in reinforcing CAV safety within mixed-traffic platoons. By incorporating system-level considerations and adopting a learning-based dynamic identification approach, the proposed framework exhibits substantial improvements in handling stochastic traffic dynamics and enhancing cooperative vehicular operations. Future work may explore multi-agent RL for collaborative strategies among CAVs, adaptive to varying traffic densities and real-world complexities like urban settings and diverse environmental conditions.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.