Reliable and Efficient Lane Changing Behaviour for Connected Autonomous Vehicle through Deep Reinforcement Learning

The establishment of future intelligent transport systems is dependable on the reliable and seamless function of Connected and Autonomous Vehicles (CAV). Reinforcement learning (RL), which allows autonomous vehicles (AVs) to learn an ideal driving strategy through constant contact with the environment, plays a significant part in the decision-making process of autonomous driving (AD). The networking of CAV is advantageous since it allows for the transmission of traffic-related data to vehicles via Vehicle-to-External (V2X) communication. Recognition and anticipation of driving behaviour are critical for avoiding collisions because they can provide useful information to other drivers and vehicles. The fundamental challenge in developing CAV is the construction of an autonomous controller that can effectively perform close real-time control selections, such as a fast acceleration while merging onto a highway and rapid speed adjustments in stop-and-go traffic congestion. CAV driving behaviours can be considerably improved by utilizing shared information, resulting in more accountable, intelligent, and efficient driving. In the present work, a deep reinforcement learning approach is proposed that integrates the information gathered through connectivity capabilities and sensing from neighbour automobiles in the vicinity of CAV. The fused information is used for providing safe and cooperative lane-changing behaviour. The deployment of an algorithm in CAV is expected to improve the transportation safety of CAV driving behaviours.

Language

  • English

Media Info

Subject/Index Terms

Filing Info

  • Accession Number: 01875512
  • Record Type: Publication
  • Files: TRIS
  • Created Date: Mar 13 2023 12:53PM