Abstract
Keywords
Introduction
Traffic congestion and safety have long been the focus of traffic management authorities and drivers. With the continuous increase of car ownership around the world, traffic congestion becomes more frequent, and the number of traffic accidents increases significantly. Therefore, to effectively alleviate congestion and reduce the occurrence of vehicle accidents is an urgent problem to be solved.
Vehicle-to-infrastructure (V2I) technology can effectively improve vehicle safety and reduce unnecessary parking, 1 reduce fuel consumption and exhaust emissions of vehicles at intersections. 2 5G network can greatly improve the efficiency of data collection, storage, and processing in vehicle infrastructure integration, making this technology to have a broader application scenario.
An automated vehicle (AV) can achieve safe and efficient driving by receiving control commands from controllers in real time through wireless information interaction function, which helps the application of auto-driving technology at intersections. It is expected that by 2035, the AV may occupy about 25% of new car market. Research shows that online controlled transportation can significantly reduce traffic accidents and improve road traffic efficiency. 3 However, as the AV is introduced, there will be a period of mixed traffic flow of AV and human-driving vehicle (HV), then gradually evolve into an environment full of AV. Therefore, it is important to study the intersection control problem in the case of mixed traffic flow of AV and HV.
For the control of unsignalized intersection, the current research mainly focuses on designing vehicle control algorithm to control the vehicle movement process. 4 In the study of micro-control at unsignalized intersections, the method of optimizing the model or setting safety constraints in the simulation model to improve the safety of vehicles is adopted by most of the studies, while there are few studies on the method of directly coordinating vehicle driving behavior in real time to avoid collision.
Reinforcement learning has the advantages of autonomous learning by agents, and is appropriate for solving decision-making problems under the setting of high-dimensional state space and high-dimensional action space. 5 Over the years, deep reinforcement learning (DRL) has evolved into a more mature research framework. In traffic control, 6 automatic driving, 7 and other fields, DRL has obtained some meaningful research results.
AV control at unsignalized intersections refers to the optimization of driving strategies of AV using micro-control methods, so as to achieve the goals of minimizing the collision and delay of vehicles within the control range of unsignalized intersections. Therefore, focusing on the control of unsignalized intersection, this article proposes an improved Double Dueling Deep Q network method, that is, 3DQN method, to reduce the number of vehicle collisions and improve the traffic efficiency.
Thus, considering the mixed situation of AVs and HVs, this article designs a connected and automated vehicle (CAV) control method based on DRL method for unsignalized intersections with V2I technology. The approach proposed in this article can provide a new control method for CAV to drive in an unsignalized intersection. The main contributions of this article are as follows:
Aiming at the micro-control problem of left-turning CAV at an unsignalized intersection, a control method based on an improved 3DQN method is presented in this article. The left-turn CAV approaching, inside, and leaving an intersection is integrated into the control method.
In the proposed 3DQN method, in which double deep Q network and dueling deep Q network are combined, important features are extracted with convolutional neural network and an advantageous strategy is obtained based on historical and existing features with Long Short-Term Memory (LSTM) network. A new multi-step reward and penalty method is proposed to solve the problem where there is relatively little experience in vehicle collisions and vehicle passing. At the same time, the sparse reward problem is alleviated using positive and negative reward experience replay buffer (PNRERB).
To solve the problems studied in this article, a central control method and multi-agent control method of unsignalized intersection are designed based on the improved 3DQN method. When constructing the DRL model, considering the particularity of the problem, the state of the model is processed. In addition, this article uses micro-simulation software VISSIM to build a virtual environment and uses it as an agent-learning environment. The left-turn and straight-line vehicle in the simulation environment are used as the learning agent, so that the agent can learn independently in the virtual environment. Finally, to verify the effectiveness of agent training in different environments, this article analyzes the effect of unsignalized intersections for different traffic flow levels at different market penetration rates of 0, 0.2, 0.4, 0.6, and 0.8.
The remainder of this article is organized as follows. The section “Literature review” summarizes the current state of research on control and DRL for CAVs at unsignalized intersections. The section “Problem formulation” describes the control problem in V2I environments and the assumptions studied in this article, along with the state space, behavior space, and reward function used in this article for DRL. The section “Improved double dueling deep Q network 3DQN” discusses the improved 3DQN method with Convolutional Neutral Network and LSTM (3DQN-CNN-LSTM) method proposed in this article, including multi-step reward and penalty methods, positive and negative sampling experience pools, and so on. The section “Experiment” describes the simulation experiments conducted and analyzes the experimental results. Section “Conclusion” concludes the article and identifies future research directions.
Literature review
Unsignalized intersection control of CAV
The aim of intelligent CAV control at unsignalized intersection is to alleviate vehicle collision and reduce delay by optimizing the driving strategy of CAV. To achieve a safe and efficient access of vehicles in unsignalized intersections, researchers mainly design vehicle control algorithm that control the movement of vehicles. A simulation experiment was designed by Shao. 8 In the environment of possible conflict, by optimizing the driving strategy of vehicles, and studying the influence of sight distance conditions and warning conditions on collision avoidance of drivers, a collision avoidance model in no control intersection was designed. The results indicate that the safety of vehicles can be improved by their model, but more complicated situation of straight and left turn hybrid driving is not considered. A distributed and collective intelligence framework is proposed by Kalantari et al. 9 to provide navigation for vehicles in intersection. Through the simulation of intersection, it is proved that the proposed framework can reduce collisions and vehicle travel time at the same time. However, the paper did not analyze the optimization effect of control methods under different market penetration. To solve the problem of vehicle collision at the unsignalized intersection, an improved reinforcement learning method was proposed by Isele et al. 10 It is verified that the proposed method can effectively ensure the safety of vehicles at unsignalized intersection through simulation experiment. However, the situation of fully autonomous vehicle is only considered through the research.
A two-car collision judgment model based on time to collision (TTC) is established by Duan et al. 11 to reduce collisions within the range of unsignalized intersection. The current state of the vehicle can be adjusted by the proposed model through estimating the possible collision area in advance. With the results of simulation, it is shown that the model can reduce the collisions effectively.
Wu et al. 12 have proposed a multi-agent learning method decentralized coordination learning of autonomous intersection management (DCL-AIM) to minimize the intersection delay under the condition of no collision constraints, which was verified by simulation experiments. From the experimental results, it is shown that the proposed method can reduce vehicle delay effectively. The safety constraints are considered to avoid the vehicle collision. In the paper by Li et al., 13 two kinds of unsignalized intersection control methods, namely priority-based method and discrete forward rolling optimal control (DFROC) method, are proposed through coordinating the driving behavior of vehicles within the control range of intersection. The safe driving of vehicles was ensured, and the traffic efficiency of vehicles was improved. It is found that two kinds of unsignalized intersection control methods can reduce vehicle delay through simulation.
In addition, to reduce vehicle collision at unsignalized intersections, a lot of methods have been proposed. Xu et al. 14 proposed a distributed no collision cooperation method, which coordinated the driving strategies of multiple vehicles at the same time to realize the vehicle no collision passing. Through numerical simulation experiments, it was found that the proposed method can realize vehicles passing through the intersection without any collision. Moreau et al. 15 proposed a Bayes curve optimization method. The problem of avoiding collision with obstacles was transformed into an optimization problem. Lagrange method and gradient method were used to solve the problem. It was shown that the method can reduce the occurrence of vehicle unsafe behavior. However, for unsignalized intersections, in most of the studies, avoiding vehicle collision was considered as the optimization objective, and collision avoidance constraints were set as the constraints in the model. In addition, in the current research, the vehicle collision optimization problem was considered in the case of fully CAVs. However, the mixed driving of AV and HV was considered rarely.
In conclusion, the weakness of current research is as follows:
Most of the current research in unsignalized intersection control problem has a limitation. In simulation experiments, only the situation that all vehicles are combined with CAV is considered while the situation with mixed AV and HV is not considered often. However, the latter situation might be the main traffic state in the future.
Vehicle efficiency, which is described as the total throughput of vehicles in an intersection within a given period of time, can be regarded as a key evaluation indicator for unsignalized intersections control. However, in much of the research, fuel consumption and exhaust emission were considered as the optimization objectives, while the vehicle efficiency is considered relatively less. In addition, setting of safety constraints in the optimization model or simulation model is adopted by most researchers to improve the safety of vehicles, while directly using real-time coordination of vehicle driving behavior to achieve collision avoidance is relatively lacking.
In view of the above problems, aiming at the control problem of unsignalized intersection, this article proposes a reinforcement learning model to study the control problem of unsignalized intersection. Therefore, the methods of DRL will be reviewed below.
DRL
Reinforcement learning 16 is favored by researchers because of the advantages of agent autonomous learning. To solve practical problems more effectively, the advantages of deep learning, such as priority in solving the multidimensional state space and multidimensional action space decision problem, are fully utilized. The application of reinforcement learning has been developed to a new stage, which is known as DRL. Through years of study by researchers, a relatively mature framework of DRL has been proposed, with research findings in many fields, such as robotics, 17 traffic control, 18 automatic driving, 19 and so on.
In view of different characteristics of actual problems, various improvements in DRL have been proposed by researchers to solve practical problems more efficiently. In addition, when solving practical problems, it is necessary to consider the specific situation and choose the most appropriate method. DRL methods can be divided into four categories: (1) Value function-based DRL (VF-DRL) method, (2) Policy gradient-based DRL (PG-DRL) method, (3) Value function and policy gradient-based DRL (VFPG-DRL) method, and (4) Multi-agent DRL (MADRL) method. Since this article is mainly based on DQN, a classical value function-based reinforcement learning method, the VF-DRL method will be highlighted among the four methods, while the other methods will be briefly described.
Q learning method is one of the VF-DRL methods. 20 The agent interacts with the environment, stores the historical data in the Q-table, and updates the Q-table to learn the optimal strategy. In addition, Q-learning can be applied to practical problems. For instance, to optimize the traffic flow of highways, the problem was described as a Markov problem in Walraven et al., 21 and Q-learning was used to explore the optimal strategy, that is, the maximum driving speed on the expressway. Meanwhile, the effect of the method was verified through simulation. The outcomes of simulation show that the strategy learned by Q-learning can greatly reduce traffic congestion in high traffic demand. However, one weakness of Q-learning is that a Q-table related to state and action is necessary to be established, and the size of state and action space depends on the computer memory, which may affect the calculation speed of the algorithm.
To further develop the reinforcement learning method, in 2013, the deep learning method was combined with the reinforcement learning method by Google deep mind team, and the DRL structure was proposed, which is named as deep Q network (DQN). 22 In 2015, the improved DQN algorithm was proposed and successfully published in Nature magazine. 23 DQN has also obtained some achievements in other practical problems. To solve the problem of obstacle avoidance of underactuated unmanned ships under unknown environment interference, the obstacle avoidance algorithm was designed by Cheng and Zhang 24 with DQN structure. The researchers found that the method based on reinforcement learning can avoid obstacles more accurately. Certainly, there are some problems of DQN method in application, such as overestimation and other weaknesses. Therefore, some improved methods for DQN were proposed by Van et al., 25 which can alleviate the overestimation problem of DQN method. The concept of advantage function was introduced by Wang et al. 26 in DQN, and the competitive DQN was proposed. A priority based DQN was designed by Schaul et al. 27 for priority sampling of important experience. After these studies, DQN series methods are gradually applied to practical problems. Based on the actual driving data constructed, the bidirectional DQN method is applied by Zhang et al. 28 to control the vehicle speed. The experimental results are compared with the traditional DQN, which shows that the accuracy of the estimation value and the quality of the strategy are greatly improved. The real-time GPS data is used by Zeng et al. 29 to replace the neural network in DQN with recurrent neural network. A new deep recurrent Q network (DRQN) is designed and applied to solve traffic light control problem at unsignalized intersection. The simulation shows that the effect of DRQN is better than DQN. Based on the above studies, it is clear that there is an excellent performance of VF-DRL method on discrete action space problems. However, many of the practical problems are the problems with continuous action and state space, which are high dimensional space problems. Therefore, VF-DRL has some limitations in solving more extensive practical problems.
In addition to the VF-DRL method, PG-DRL, VFPG-DRL, and MADRL methods also exist. These methods may outperform VF-DRL in some specific scenarios. PG-DRL method can be used to solve the continuous action space problem, and the convergence speed is relatively fast. However, the method is easy to converge to the local optimal solution and difficult to evaluate the policy. VFPG-DRL, a DRL method based on value function and strategy gradient, combines the advantages of VF-DRL and PG-DRL methods. However, there are some problems, such as low sampling efficiency and overestimation of evaluation network. Therefore, in practical applications, it is necessary to regulate and improve the algorithm according to the actual problem. In MADRL method, the agent can obtain the state information and reward information from the environment based on interactions among the agents and environment, and adjust the strategy of the agents based on the acquired information, thus learning the optimal strategy.
Problem formulation
Through the V2I environment, the information of the vehicles can be obtained by the Road side unit (RSU) within the control range in real time, and the information is transmitted to the control center. The current driving strategy of the CAV is calculated by the control center according to the vehicle information to avoid collision with other vehicles and accelerate the vehicle traffic efficiency. The yellow dot denotes the collision point where the vehicle may have a collision. In addition, in order to describe the state of vehicles within the control range more accurately, this article discretizes the road network of unsignalized intersections at the interval of length M, and Figure 1(b) shows the discrete diagram of the road network.

The road network structure diagram: (a) road network structure diagram in unsignalized intersection and (b) the discrete diagram of the road network.
In Figure 1, Agent 1, 2, 3, and 4 were used to obtain optimal driving strategies for vehicles in the straight and left-turn directions, respectively. For example, Agent 1 controls the vehicles going straight from Lane 2 to Lane 17 and Lane 18, and the vehicles turning left from Lane 1 to Lane 20. In Figure 1(a), at the intersection, left turn and through vehicles are more likely to collide, that is, there are more collision points. In order to focus on such complex traffic environment, right turn is not considered in this article. However, the possible collision situations with right turning vehicles in other directions may also be considered. As long as the through vehicle and right turning vehicle are designed in the same lane and the model proposed in this article is trained, the situation including right turning vehicles can be obtained.
This article is based on the following basic assumptions:
Within the control range of unsignalized intersection, vehicles are not allowed to overtake or change lanes.
In the control range of unsignalized intersection, when receiving the control information, the vehicle drivers are completely subject to the control command of the controller.
Within the control range of unsignalized intersection, the delay of information communication between vehicles and road, vehicles and vehicles are acceptable, and there is no packet loss.
Problem description
State space
In this article, the position and speed of vehicles in the control range of unsignalized intersection are considered in the state space, and the state is determined according to the road network discrete diagram in Figure 1
In equations (1) to (3),
Action space
This article aims to reduce the number of vehicle collisions in the control range of unsignalized intersection, and improve the traffic efficiency of vehicles through the control of vehicle acceleration and deceleration behavior
According to equation (4),
Reward function
The objective of this article is to reduce the number of vehicle collisions in the intersection and improve the traffic efficiency of vehicles at the same time. In this article, at the unsignalized intersection, the first problem to be solved is vehicle collision. Therefore, this article sets the penalty when a vehicle collision occurs as the maximum penalty, as shown in equation (5a). And the vehicle passing the intersection safely is given a larger reward, as shown in equation (5b). In other cases, in order to ensure the vehicle passing the intersection as soon as possible, a fixed reward is given to the vehicle when it is driving at normal speed, while a penalty of the same value is given when the vehicle speed is small, as shown in equations (5c) and (5d). The reward function formula, equation (5) in the original text, is as follows
where
The principle of the reward function is as follows. In order to reduce the number of collisions of control vehicles at unsignalized intersections, if the control vehicles collide, a large penalty value will be given, which is related to the speed of collision. When the vehicle passes through the unsignalized intersection successfully, a larger positive reward value is given, which is also related to the speed of the vehicle passing through the detection point. In other cases, in order to make the vehicles stay in the control range of unsignalized intersection, when the control vehicle speed is less than the threshold, a smaller penalty can be given, otherwise a smaller reward can be given. Based on the reward function, it can avoid collision and ensure the vehicles to pass through the intersection as soon as possible.
Improved double dueling deep Q network 3DQN
3DQN-CNN-LSTM
Figure 2 shows the structure diagram of the improved double dueling deep Q network (3DQN-CNN-LSTM) model proposed in this article, which is different from the structure diagram of the double dueling deep Q network (3DQN) in the work by Gong et al. 30 In this article, CNN are used to extract important features. Meanwhile, LSTM networks are used to select the best strategy combined with memory information and current information. In addition, the positive and negative reward experience buffer pool method, 26 multi-step reward, and penalty method 15 are used in this method to speed up convergence.

The structure diagram of the improved double dueling deep Q network, 3DQN-CNN-LSTM.
Multi-step reward and penalty
The less experience of vehicle collision and vehicle passing leads to the decrease of learning efficiency of the algorithm. Therefore, the article proposes a multi-step reward and penalty method based on the error correction method of failure experience proposed by Zhang et al.
28
The experience of collision or possible collision, and the experience of vehicle successfully passing or promoting vehicle passing are mainly increased through the method. The main idea is that if the vehicle collides or successfully passes the intersection at time
However, in some cases, in order to prevent the possibility of collision, the behavior of deceleration or braking is taken by the agent, which is not supposed to be punished. Therefore, the algorithm is designed to determine whether the agent takes the behavior of deceleration or braking. If the agent takes actions to avoid collision, equation (6) is not used to calculate the reward value from time
According to equation (6), the reward value at time
Positive and negative sampling experience pool
At present, a large number of researchers have applied Deep Deterministic Policy Gradient (DDPG) to solve many problems of continuous action space. 31 However, in the application of DDPG algorithm, there are still some problems. For example, when sampling in the experience playback buffer pool, the selected historical experience is randomly selected. Therefore, it is difficult to balance the proportion of positive reward and negative reward experience, which leads to the poor stability of the algorithm.
In the original DDPG, 32 the experience replays buffer (ERB) is mixed with positive and negative experiences. In order to solve the problem of vehicle delay caused by parking within the control range of unsignalized intersection, the historical experience is divided into positive reward experience and negative reward experience, the positive and negative experience are stored in the PNRERB, respectively, in this article. Similar to the original DDPG method, the sizes of the two cache pools are initialized first in this article, and the historical experience are replaced with the new one after the cache pool is full of experience.
In addition, small batch learning method is used to train DRL network in this article. Therefore, let the agent interact with the environment, collect a certain amount of historical experience, and then extract experience from the cache pool to train the neural network. In order to improve the sampling performance of the algorithm, the experience pool is divided into two experience pools, which are the experience pool with reward value greater than 0, that is, positive reward experience pool, and the experience pool with reward value less than or equal to 0, that is, negative reward experience pool, respectively. The sampling ratio from positive and negative experience pool is set as 3:1.
Central control method and multi-agent control method based on DRL
At present, intersection control can be divided into central control and distributed control. 5 Based on the idea of these two control methods, this article proposes a central control method and a multi-agent control method for unsignalized intersection based on improved DRL. In these methods, the implementation of multi-agent control method is based on the distributed control of multi-agent reinforcement learning method. The structure diagram and pseudo code of these two control methods are introduced in the following section.
Central control method
The central control method is to set up an agent in the central control system, and the agent controls all the CAV in the intersection control area, as shown in Figure 3. Table 1 gives the pseudo code of the central control method.

Interaction diagram of central control method.
The central control method.
Multi-agent control method
Multi-agent control method is mainly to set up multiple control agents in the central control system, which will control all the CAV in their respective control areas of the intersection.
In this article, the CAV of the four entrances in southeast and northwest directions are regarded as the control vehicles of the four agents. Each agent controls the whole process from the CAV entering the control area to the vehicles leaving the control area.
Figure 4 shows the interaction structure diagram of multi-agent control with the environment. From Figure 4, it can be seen that multiple agents interact with the VISSIM environment separately. Taking the four agents in this article as an example, each agent obtains the state information of its own control direction within the control range of the intersection from the VISSIM environment, and unites with the states of the other three directions to constitute the combined state. Each agent makes the action of the current control direction of the CAV according to the joint state. Finally, by sending its action to the VISSIM environment, each agent gets the immediate reward under the current state-action and the state of the next moment.

Multi-agent control method and VISSIM interaction diagram.
Table 2 shows the pseudo code for the multi-agent control method.
Multi-agent control method.
In this article, the main difference between Algorithms 1 and 2 manly lies in the Lines 14 to 16 of the pseudo code. It can be seen from Tables 1 and 2 that the central control method in the pseudo code Line 14 mainly uses the central control agent to select the corresponding action according to the state in time
Experiment
The optimization effect of different DRL methods has been verified through the simulation experiment of vehicle control at unsignalized intersection, which is presented in this section. First, the simulation platform and parameters of the unsignalized intersection are described. Second, the control methods based on DRL are discussed. Specifically, the structure and parameter setting of each method is explained. Third, the scheme in this article is outlined and finally, the result based on the DRL control method is analyzed.
Simulation platform and parameter setting
A virtual road network is built in the experiment under the VISSIM simulation environment. The road network structure is shown in Figure 5. The cycle length in the signal control intersection is 80 s, the green time of east–west and north–south straight line is 37 s, the yellow time is 3 s, the green time of east–west and north–south straight line is 22 s and the green time of left-turning is 8 s, and the yellow time is 3 s.

VISSIM simulation road network.
The differences between AV and HV are mainly explained from the following two aspects: (1) In terms of the concept defined, AV refers to the intelligent connected vehicle. HV refers to traditional manned vehicle. AV mainly simulates the actual intelligent vehicle behavior, and HV mainly simulates the actual human driving vehicle behavior; (2) In terms of the driving strategy, the optimal driving strategy of AV are obtained according to the DRL algorithm proposed in this article, while the driving strategy of HV are obtained according to the built-in model of VISSIM simulation system. HVs are used to simulate the actual driving behavior of human beings. It receives the driving strategy of the built-in model of VISSIM simulation system and does not receive the control of DRL agent. DRL agent is mainly used to control AVs driving behavior.
The simulation interval in VISSIM is 1 s, that is, 1 frame/1 simulation step. This article mainly adopts VISSIM and Python interactive simulation. The DRL algorithm is realized in Python environment. At time
The results of the two classes of methods are compared in this section. One class is mainly based on DQN methods, which contains Deep Q Network method (DQN-NN), 26 Deep Q Network method based on CNN (DQN-CNN), 27 Double Deep Q Network method based on CNN (Double-DQN-CNN), 29 and Dueling Deep Q Network method based on CNN (Dueling-DQN-CNN). 27 The other class is mainly based on 3DQN method, which contains Double Dueling Deep Q Network method based on CNN (3DQN-CNN), 14 Multi Agent of Double Dueling Deep Q Network method based on CNN (multi-3DQN-CNN), 14 and the two methods proposed in this article, that is, Double Dueling Deep Q Network method based on CNN and LSTM (3DQN-CNN-LSTM), and Multi Agent of Double Dueling Deep Q Network method based on CNN and LSTM (multi-3DQN-CNN-LSTM). The network structure of multi-3DQN-CNN-LSTM method used in the article is the same as 3DQN-CNN-LSTM. In addition, multi-agent method is also adopted in multi-3DQN-CNN-LSTM.
In order to verify the optimization result of the vehicle performance index of the proposed method for the actual intersection, the mixed driving of straight vehicles, and left turning vehicles in the unsignalized intersection is considered in this article. The experimental scheme is given in Table 3.
Experimental scheme.
EL: entrance lane.
In the scheme, there are straight vehicles and left turning vehicles driving in the unsignalized intersection. Table 3 shows the distribution of traffic flow at each entrance lane of the scheme. When the vehicles in the unsignalized intersection are straight vehicles and left turning vehicles, the total flow is 800 vehicles/h, 1680 vehicles/h, and 2560 vehicles/h, respectively. The first column in Table 3 is the total flow. The second to ninth columns show the flow of each inlet. According to Figure 1, Inlets 1, 4, 7, and 10 are left turn lanes, and Inlets 2, 5, 8, and 11 are straight lanes.
According to Figure 1, Inlets 3, 6, 9, and 12 are reserved lanes for these right-turning vehicles. Meanwhile, Lanes 13 to 20 are outlets, and these lanes are controlled by the agents in the corresponding directions. In the simulated environment of the intersection, there are three inlets but only two outlets in each direction.
Experimental result
In this article, the success rate of vehicles successfully passing through the intersection 12 is taken as the evaluation index to test the effect of the method based on DRL. The calculation formula of vehicle success rate is given in equation (7)
where
The calculation formula of average travel time (ATT) is
where
The two key evaluation metrics in this article are the average vehicle travel time and the total throughput of the vehicle. Equation (8) shows the correlation between these two metrics
Equation (9) represents the ATT of all the vehicles, from entering the control area to leaving the control area of the intersection. Here,
In the following section, the success rate, ATT, and vehicle trajectory experimental results of vehicles passing through the intersection under this scheme will be analyzed, respectively.
The success rate
Figure 6 shows the success rate of vehicles passing through the intersection when the total flow of the intersection is 800 vehicles/h based on DQN method and 3DQN method. It can be seen from Figure 6 that under the five permeability levels (20%, 40%, 60%, 80%, and 100%), the success rate based on DQN method and 3DQN method is greater than that when the permeability is 0%, and when the permeability is 100%, the success rate based on 3DQN method can be higher than that based on DQN method. This is also shown at the intersection of left turn and straight traffic. In addition, in the 3DQN-based method, under the same permeability, the success rates of 3DQN -CNN-LSTM method and 3DQN-CNN method are higher than those of multi-3DQN-CNN-LSTM method and multi-3DQN-CNN method, respectively, indicating that the optimization effect of central control method is better than that of multi-agent control method.

Success rate of vehicles passing through the intersection when the intersection flow is 800 vehicles/h.
The percentage of success rate under two methods when traffic flow is equal to 800 vehicles/h, 1680 vehicles/h, and 2560 vehicles/h is shown as Table 4. The values highlighted in gray in Table 4 indicate the optimal values under the same penetration rate. In Table 4, M1 is DQN-NN method, M2 is DQN-CNN method, M3 is Double-DQN-CNN method, M4 is Dueling-DQN-CNN method, M5 is 3DQN-CNN method, M6 is multi-3DQN-CNN method, M7 is 3DQN-CNN-LSTM method, M8 is multi-3DQN-LSTM method, and MPR is market penetration rate. According to Table 4, under three different traffic flow, the percentage of success is more than 70% based on DQN method and 3DQN method. However, under the same traffic flow, the percentage of success based on the two methods is rising gradually with the increase of penetration rate. However, under the same penetration rate, the percentage of success is decreasing with the increase of traffic flow. Therefore, comprehensively, under three different traffic flows, the percentage of success can be optimized effectively based on DQN method and 3DQN method. Within the same traffic flow, the higher the penetration rate, the better the optimization effect of two methods, and within the same penetration rate, the greater the traffic flow, the worse the optimization effect of the two methods. Within the same traffic flow and penetration rate, the percentage of success of 3DQN method is higher than DQN method. Based on the DQN method, the percentage of success of DQN-NN method, DQN-CNN method, Double-DQN-CNN method, and Dueling-DQN-CNN is similar. Within the method based on DQN, there is a little difference between 3DQN-CNN method, multi-3DQN-CNN method, 3DQN-CNN-LSTM method, and multi-3DQN-LSTM method when the traffic flow is 800 vehicles/h and 1680 vehicles/h, respectively. Under the same penetration rate, the percentage of success of 3DQN-CNN-LSTM and multi-3DQN-LSTM are higher than 3DQN-CNN and multi-3DQN-CNN, respectively, when the traffic flow is equal to 2560 vehicles/h. In addition, under three different traffic flow and the same penetration, the optimization effect of 3DQN-CNN-LSTM is the best (within the method based on 3DQN, the highest percentage of success is obtained by 3DQN-CNN-LSTM method). However, under the same penetration rate and traffic flow, within the method based on 3DQN, the percentage of success of 3DQN-CNN-LSTM method and 3DQN-CNN method are higher than multi-3DQN-LSTM method and multi-3DQN-CNN method, respectively. In a summary, the percentage of success of 3DQN-CNN-LSTM is the highest, which means there is a better performance of central control method than the multi-agent control method in the optimization of percentage of success. Therefore, the 3DQN-CNN-LSTM method proposed in this article is the best for optimizing.
Vehicle success rate under different methods when the flow is 800 vehicles/h, 1680 vehicles/h, and 2560 vehicles/h.
3DQN: Double Dueling Deep Q Network; MPR: market penetration rate; M1: DQN-NN method; M2: DQN-CNN method; M3: Double-DQN-CNN method; M4: Dueling-DQN-CNN method; M5: 3DQN-CNN method; M6: Multi-3DQN-CNN method; M7: 3DQN-CNN-LSTM method; M8: Multi-3DQN-LSTM method.
ATT
Table 5 shows the ATT of vehicles and the number of vehicles successfully passing through the intersection obtained based on DQN-based method and 3DQN-based method when the total flow is 2560 vehicles/h. The values in bold font in Table 5 represent the optimal values under the same permeability. It can be seen from Table 5 that the ATT obtained by DQN-based method is less than that obtained by 3DQN-based method, but the number of vehicles successfully passing through the intersection obtained by 3DQN-based method is greater than that obtained by DQN-based method. The signal control method has the highest total throughput at 0% penetration. However, its ATT is about 3 times that of the other methods. It can be considered that the signal control method has the highest security, but its combined efficiency is not as good as the other reinforcement learning methods. When the penetration rate is less than or equal to 100%, the number of vehicles passing through the intersection successfully obtained by DQN-NN method increases very little, perhaps because DQN-NN method cannot correctly judge the driving strategies of other vehicles, resulting in wrong driving decisions. When the penetration rate is 100%, the number of vehicles successfully passing through the intersection under 3DQN-NN-LSTM method is more than that of the signal control method. Similarly, the ATT is shorter than the signal control method, and the ATT gain can reach 69%. At the same time, under the same permeability, considering the ATT and the number of vehicles successfully passing through the intersection, the effect of 3DQN-NN-LSTM method is better than that of DQN method, 3DQN-CNN method, multi-3DQN-CNN method, and multi-3DQN-CNN-LSTM method. This shows that even when the traffic flow is large, 3DQN-NN-LSTM method can still ensure the safe and rapid passage of vehicles. At the same time, it also shows that the optimization effect of central control method is better than that of multi-agent control method.
Number of vehicles successfully passing through the intersection when the flow is 2560 vehicles/h.
DQN-NN: Deep Q Network method; CNN: convolutional neural networks; 3DQN: double dueling deep Q network; LSTM: long short-term memory.
Figure 7 shows the ATT of vehicles obtained by DQN-based method and 3DQN-based method with a total flow of 2560 vehicles/h. It can be seen from Figure 7 that (1) The ATT under signal control method is nearly 3 times larger compared to other DQN-based method and 3DQN-based method when the permeability is larger than 0%; (2) The ATT under 3DQN-CNN method and multi-3DQN-CNN method is higher than that of DQN method, 3DQN-CNN-LSTM method, and multi-3DQN-CNN-LSTM method.

The average travel time under the two methods when the flow is 2560 vehicles/h.
To further verify the experimental results, different permeability environments with traffic flow of 800 vehicles/h and 1680 vehicles/h are also investigated in this article, and the experimental results are shown in Tables 6 and 7, respectively. According to Table 6, the total throughput under the 3DQN-CNN-LSTM method gradually increases with the increase of penetration rate, but is always lower than the total throughput under signal control method. According to Table 7, the total throughput of the 3DQN-CNN-LSTM method is higher than the total throughput under signal control at a penetration rate of 100%. In addition, the ATT required for the model with DRL method is consistently smaller than that of the signal-controlled approach. Therefore, it can be concluded that the 3DQN-CNN-LSTM method proposed in this article will outperform the signal control method under each evaluation indicator in the case of high penetration rate and high throughput.
Number of vehicles successfully passing through the intersection when the flow is 800 vehicles/h.
DQN-NN: Deep Q Network method; CNN: convolutional neural networks; 3DQN: double dueling deep Q network; LSTM: long short-term memory.
Number of vehicles successfully passing through the intersection when the flow is 1680 vehicles/h.
DQN-NN: deep Q network method; CNN: convolutional neural networks; 3DQN: double dueling deep Q network; LSTM: long short-term memory.
To sum up, in general, the DQN-based method and the 3DQN-based method can effectively reduce the number of vehicle collisions at intersections under different permeability and improve the success rate of vehicles passing through intersections. At the same time, compared with the signal control method, the ATT based on DQN method and 3DQN method is shorter. Specifically, (1) Under different traffic flows, with the increase of the market penetration of CAV, the success rate of vehicles passing through the intersection based on DQN method and 3DQN method increases gradually. However, when the total flow at the intersection is slightly larger (greater than 1200 vehicles/h), the increase based on 3DQN method is higher than that of DQN method. In addition, based on 3DQN method, 3DQN-CNN-LSTM method performs better under any intersection flow. When the permeability is 100%, the success rate of 3DQN-CNN-LSTM method can reach 99%; (2) Compared with the signal control method, the ATT based on DQN method and 3DQN method is shorter. When the permeability is less than or equal to 80%, the number of vehicles successfully passing through the intersection based on DQN method and 3DQN method is less than that of the signal control method. When the permeability is 100%, the number of vehicles successfully passing through the intersection based on 3DQN method is close to that of signal control method, and the number of vehicles under 3DQN-CNN-LSTM method is higher than that of signal control method under some total flow (such as 2560 vehicles/h); (3) No matter the convergence of the algorithm or the reduction of the number of vehicle collisions, the optimization effects of 3DQN-CNN and 3DQN-CNN-LSTM methods are better than multi-3DQN-CNN and multi-3DQN-CNN-LSTM methods, respectively, which shows that the central control method proposed in this article is better than multi-agent control method.
In summary, the collision times and travel time of vehicles can be effectively reduced based on DQN method and 3DQN method. The 3DQN-CNN-LSTM method proposed in this article is the most excellent method among DQN-NN, DQN-CNN, double-DQN-CNN, dueling-DQN-CNN, 3DQN-CNN, multi-3DQN-CNN, 3DQN-CNN-LSTM, and multi-3DQN-CNN-LSTM methods, which can reduce the number of vehicle collisions effectively, and reduce the ATT of vehicles between 47% and 71% no matter under low traffic flow or high traffic flow.
The evaluation indicators in this article are mainly the ATT and the total throughput. In this article, the strengths and weaknesses of the model are analyzed mainly based on these two indicators. The signal control method has the highest total throughput when the penetration rate is 0%. However, its ATT is about 3 times higher than the other methods. Therefore, we believe that although the signal control method has the highest security, its overall efficiency is inferior to the 3DQN-CNN-LSTM method proposed in this article.
The goal of this article is to optimize the traffic efficiency on the premise of avoiding the collision of vehicles at the intersection. Therefore, if the vehicle collides, a greater penalty will be given. Therefore, in order to ensure that there is no collision, the vehicle may reduce the speed, thereby reducing the traffic efficiency. When the penetration rate is lower than 100%, in order not to collide with other vehicles, HV may not deliberately reduce the speed on the simulation road; however, CAV may accelerate or decelerate to reduce the collision. Therefore, in order to ensure safety, some traffic efficiency may be lost. However, when the penetration rate is set as 100%, all vehicles are connected vehicles, and all vehicles themselves will take the behavior of avoiding collision, so the traffic efficiency is further guaranteed in a relatively safe environment.
The success rate and ATT under complex traffic environment
The reason why the right turn movement is not considered in above discussion is that the purpose of this article is to focus on the unsignalized intersection control under the objective of reducing vehicle collision and delay as much as possible and considering the mixed traffic of HV and AV. In the two-way four lane unsignalized intersection environment tested in this article, the setting of collision points in the road network diagram of unsignalized intersection is shown in Figure 1. It can be seen from Figure 1 that the collision between motor vehicles mainly exists in the left turn and through traffic flow, and the right turn traffic flow has little impact on the traffic flow. In addition, this article does not consider the mixed traffic flow in the presence of both pedestrians and non-motor vehicles. Therefore, right turning vehicles may be ignored. However, if the mixed traffic flow with both pedestrians and non-motor vehicles is considered in the research, the right turn traffic flow must be considered.
We also analyzed and discussed the experimental results of the success rate of vehicles passing through the intersection and the ATT of vehicles under different traffic flows, comprehensively considering the through traffic flow, left turn traffic flow, and right turn traffic flow in the next part.
Figure 8 shows the success rate of vehicles passing through the intersection under the comprehensive consideration of straight, left turn, and right turn traffic flows. In Figure 8, the total flow is 2000 vehicles/h. Different methods, such as DQN-NN, DQN-CNN, double-DQN-CNN, dueling-DQN-CNN, 3DQN-CNN, multi-3DQN-CNN, 3DQN-CNN-LSTM, and multi-3DQN-CNN-LSTM are tested. It can be seen from Figure 8 that with the increase of the penetration rate of connected vehicles, the success rate of DQN-NN, DQN-CNN, double DQN-CNN, dueling DQN-CNN, 3DQN-CNN, multi-3DQN-CNN, 3DQN-CNN-LSTM, and multi-3DQN-CNN-LSTM methods has increased to some extent. The 3DQN-CNN-LSTM method can get the best results. When the penetration rate is 100%, the success rate can reach 99%, while 3DQN-CNN and multi-CNN-LSTM methods and 3DQN-CNN and multi-3DQN-LSTM are better than DQN-NN, DQN-CNN, double DQN-CNN, and dueling DQN-CNN, that is, they have a higher success rate.

Success rate of vehicles passing through the intersection when the total flow at the intersection is 2000 vehicles/h.
Figure 9 shows the ATT of vehicles under the comprehensive consideration of straight, left turn, and right turn traffic flows. Table 8 shows the ATT of vehicles and the number of vehicles successfully passing through the intersection under the comprehensive consideration of straight, left turn, and right turn traffic flows. In Figure 9 and Table 3, the total flow is 2000 vehicles/h. Different methods, such as DQN-NN, DQN-CNN, double-DQN-CNN, dueling-DQN-CNN, 3DQN-CNN, multi-3DQN-CNN, 3DQN-CNN-LSTM, and multi-3DQN-CNN-LSTM are tested. It can be seen from Figure 9 and Table 8 that the ATT optimized by 3DQN-CNN and multi-3DQN-CNN is longer, indicating that although 3DQN-CNN and multi-3DQN-CNN can increase vehicle safety, they also increase vehicle travel time. Compared with the 3DQN-CNN-LSTM method and the multi-3DQN-CNN-LSTM method, the 3DQN-CNN method and the multi-3DQN-CNN method not only have a longer ATT but also have a relatively small number of vehicles successfully passing through the intersection.

Average travel time of each method when the flow is 2000 vehicles/h.
Number of vehicles successfully passing through the intersection when the flow is 2000 vehicles/h.
DQN-NN: Deep Q Network method; CNN: convolutional neural networks; 3DQN: double dueling deep Q network; LSTM: long short-term memory.
Vehicle trajectory
The 3DQN-CNN-LSTM method is adopted as control method. Through the above analysis, it is clear that there is a better performance of 3DQN-CNN-LSTM method, thus the vehicle trajectory of 3DQN-CNN-LSTM method is mainly analyzed in this section. When the penetration is 0%, the trajectory is the vehicle trajectory obtained under the signal control method, in which the network CAV trajectory is represented by red line and the HV trajectory is represented by blue line. The trajectories in this section are those of vehicles passing through the intersection successfully. The trajectories of vehicles in collision are removed in this section.
The spatiotemporal trajectory of vehicles in Lane 5 at a total traffic flow of 1680 vehicles/h is shown as Figure 10. According to Figure 10(b)–(f), there is an increment in the vehicle trajectory with the increase of penetration. Compared with Figure 10(a), the vehicle trajectory in Figure 10(b)–(f) is smoother, which means that the optimization of the vehicle trajectory based on 3DQN-CNN-LSTM method is excellent at the unsignalized intersection with through and left-turning vehicles.

Vehicle trajectory of Lane 5 when the total flow at the intersection is 1680 vehicles/h: (a) penetration rate = 0%, (b) penetration rate = 20%, (c) penetration rate = 40%, (d) penetration rate = 60%, (e) penetration rate = 80%, and (f) penetration rate = 100%.
As can be seen from Figure 10, compared with the signal control method, the vehicle trajectory optimized by 3DQN-CNN-LSTM method has less parking waiting, and most of them are relatively smoother. Therefore, 3DQN-CNN-LSTM method can effectively alleviate the phenomenon of vehicle parking queue. Under the same traffic flow, with the increase of the penetration of CAV, the vehicle trajectory gradually increases, indicating that with the increase of the penetration of CAV, the number of vehicle collisions is less.
Conclusion
In order to reduce vehicle collisions and improve traffic efficiency, the unsignalized intersection is researched and the methods based on DQN and 3DQN are designed to solve this problem. The results show that the percentage of success based on DQN method and 3DQN method increases with the increase of the penetration rate of CAV under the same traffic flow. In addition, under the same traffic flow and penetration rate, the optimization effect based on 3DQN method is better than that of the DQN method, and the success rate is up to 99%. At the same time, compared with the signal control method, the ATT of vehicles passing through the intersection based on DQN method and 3DQN method has been greatly reduced, and the ATT is in the range of 18%–72%. In addition, the optimization effects of 3DQN-CNN and 3DQN-CNN-LSTM methods perform better than those of multi-3DQN-CNN and multi-3DQN-CNN-LSTM methods in terms of vehicle collisions. It shows that the optimization effect of central control method is better than that of multi-agent control method.
An improved DRL method for vehicle control at urban road intersections has been proposed in this article. Although the methods proposed have some results for vehicle control problems, there is still room for improvement. For example, the vehicle type studied in this article is car, and the possibility of other vehicle types coexisting has not been considered. In the future, more vehicle types and effective MADRL will be considered.
