网刊加载中。。。

使用Chrome浏览器效果最佳,继续浏览,你可能不会看到最佳的展示效果,

确定继续浏览么?

复制成功,请在其他浏览器进行阅读

Precise Decision-Making Learning for Automated Vehicles in Lane-Change Scenario Based on Parameter Description  PDF

  • ZHANG Yuxiang
  • HE Ganglei
  • LI Xin
  • LIU Qifang
  • CONG Yanfeng
  • WANG Yuhai
State Key Laboratory of Automotive Simulation and Control, Jilin University, Changchun 130022, China.

CLC: U471.1

Updated:2022-08-11

DOI:10.11908/j.issn.0253-374x.22787

  • Full Text
  • Figs & Tabs
  • References
  • Authors
  • About
CN CITE
OUTLINE

Abstract

To promote safety and fully consider human drivers' acceptance, precise decision-making is realized for automated vehicles under the lane-change scenario in this paper. More specifically, automated vehicles not only decide to change lanes or not but also decide specific microcosmic behaviors, such as lane-change time and expected acceleration. Thus, precise decisions for lane-change are described with three parameters and learned by reinforcement learning. The rationality of such parameter-based precise decisions is shown in two aspects. First, different values of decision parameters will notably influence the planned trajectory, which means other microcosmic behaviors will be a significant uncertainty when they are not precisely decided in the decision-making layer. Secondly, based on the analysis of real traffic data, NGSIM, changeable lane-change time, and expected acceleration are revealed in lane-change behaviors, which is seldom explicitly considered in the decision-making layer of current researches. The decision parameters that include lane-change time and expected acceleration are learned with kernel-based least-squares policy iteration reinforcement learning (KLSPI). Safety, current driver's willingness, and average human driving style are considered in the reward function. Simulation results demonstrate that using reinforcement learning (RL) to learn decision parameters can realize more precise decisions, promote safety performance, and imitate human drivers' behaviors in the lane-change scenario.

Automatic control will be fully realized from the decision-making layer to the planning layer in automated vehicles

1-2. In a higher-performance autopilot driving system, the human driver's willingness and diverse driving preference should also be taken into consideration to improve the acceptance in either general or individual human drivers3-5. However, the human's decision process is much more complicated to be captured with a mathematical model6 and has no obvious multi-layer hierarchical framework as same as autopilot driving system7. Meanwhile, human behavior is influenced by driving style and skills8-9, which is different in both the decision-making and planning layers for different drivers.

In previous researches, learning-based methods have obtained great attention to learn human-like driving behaviors

10. Beyond numerous learning-based methods, reinforcement learning (RL), as a model-free method for sequential control, is widely researched in decision-making11. Generally speaking, RL learns with reward signal, which can not only be physical measurements about safety constraints, like the distance between the host vehicle and surrounding vehicles but also can be the correspondence with human's decision.

As it is hoped that the autopilot driving system can ensure safety performance as well as respect human willingness, using parameters that can be some physical measurements to describe decisions helps to make more precise decisions

12-13 and realize such performance for the following reasons. Regarding vehicle's motion, despite macroscopical behavior, such as lane change or not, microcosmic behaviors, such as lane-change time and expected acceleration, will also influence a vehicle's trajectory. Specific microcosmic behaviors actually will fix the trajectory under a smaller range. If macroscopical behaviors are only used to describe a decision, decisions will be much more conservative to keep safe and different from human drivers. Meanwhile, different drivers and the driving situations will reflect microcosmic behaviors. In the driving process, a driver will make decisions with driving experience naturally to predict the influence of these microcosmic behaviors, which will lead to some specific or intelligent behaviors, like accelerate to cut in or reduce the time for lane-change when the distance gap is not sufficient, especially in heavy traffic14.

The precise decision-making method based on parameter description in the lane-change scenario is investigated in this paper. The overall framework is shown in Fig.1, which mainly focuses on the RL-based decision-making module and the model predictive control (MPC)-based trajectory planning module. The first contribution is to analyze the rationality of parameter-based precise decisions framework in two parts. First, the influence on trajectory under different lane-change time and expected acceleration are shown basing on the designed trajectory planning controller. Second, a comparison is made to show different decision parameters between real driving profiles. The trajectory planning controller used here is proposed in reference [

15] and will be extended in this research to realize precise decision-making. The public dataset of vehicle trajectories used here is from NGSIM, a program developed by the US Federal Highway Administration9. The second contribution is to design the learning-based precise decision-making method. The lane-change decision-making problem is modeled as a Markov decision process (MDP) in the parameter decision framework. The reward function considers the human driver's willingness and safety situation. Thus, the learning promotes the safety performance on the potential dangerous driving scenario and imitates actual human drivers' behavior in NGSIM. The kernel-based least-squares policy iteration reinforcement learning (KLSPI) that is proposed in reference [16] is used to solve this problem. In the simulation, the driving scenarios are recognized and divided into three sample sets for learning and evaluation. By training, the parameter-based precise decisions are realized, and simulation results are also shown.

Fig.1  Diagram of framework

1 Analysis of potential performance with parameter decision framework

The trajectory planning controller is described in Fig. 1. Based on this controller, parallel simulations with different decision parameters are conducted to show its influence on the planned trajectory. The real traffic data, NGSIM,are analyzed to find out the range of decision parameters that fit the trajectory planning controller and potential promoted performance in the decision layer.

1.1 Optimization trajectory planning controller

The trajectory planning controller is designed with nonlinear model predictive control in the parameter decision framework

15. The main intention is to simplify the implementation of driving behavior by only constraining terminal states to fit with the road segment. For example, on a straight road, as shown in Fig.2, at the end time step in lane-change and lane-keeping, the vehicle's lateral position should be on the center of the target lane and will stay on the lane (the same heading angle with the target lane, zero yaw rate, and lateral acceleration). Meanwhile, the sequence of control inputs is directly optimized to obtain a feasible trajectory without a reference polynomial function or B-spline trajectory.

Fig.2  Diagram of the straight-road scenario

The nonlinear motion control model x˙=f(x,u) is established as:

f(x,u)=vxcosφ-vysinφvxsinφ+vycosφωralfIzFyf-lrIzFyr1MFyf+1MFyr-vxωr (1)

where: the vector of states is x=[X,Y,φ,vx,vy,ωr]X and Y are the positions, φ is the heading angle, vx and vy are the longitudinal and lateral velocity, ωr is the yaw rate; the vector of control inputs is u=[a,δf]δf is the steering-wheel angle, the simple longitudinal dynamics v˙x=a is considered to simplify the motion control model; M is the mass of the vehicle; Iz is the moment of inertia of the vehicle around the z-axis; lf and lr are the distances from the center of gravity (CoG) to the front and rear axles, respectively. The linear tire model is considered, and the tire slip angle of the front wheel αf and rear-wheel αr can be linearized with the small slip angle, the lateral tire force Fyf and Fyr on the front and each rear tires is written as:

Fyf-Cf(vy+lfwrvx-δf),Fyr-Crvy-lrwrvx, (2)

where: Cf and Cr are the cornering stiffness values of the front and rear tires, respectively.

The trajectory planning problem in the straight lane-change scenario at decision parameters (YT,T,atar) is formulated after discretization as

15

mink=0T-1||u(k)||Q2+||atar-a(k)||P2      +k=0T-2||u(k+1)-u(k)||R2s.t. x(k+1)=x(k)+f(x(k),u(k))Δt     uminu(k)umax     vy(T)=0, ωr(T)=0     φ(T)=0, Y(T)=YT (3)

where:Q, P, R are the weighting matrices; YT{0, L,-L} is the lateral offset to the center of target lane, which is different in lane-keeping and lane-change; L is the distance between two neighboring lanes. The lane-change time T decides the predictive horizon. As have illustrated before, the terminal equality equation ensures that the lane-change behavior is finished in the predictive horizon without a reference trajectory.

1.2 Influence of different decision parameters

Decisions in the lane-change scenario can be precisely described as lane-keeping or lane-change to the right/ left lane in fast/ moderate/ mild mode with acceleration/maintained velocity/ deceleration. The examples represented with decision parameters are listed in Tab.1. Such a description of the decision can provide more precise and diverse behaviors for decision-making and be more human-like.

Tab.1  Examples of decision description for lane—change at parameter representation
DescriptionYT/mT/s
Lane keeping 0 1
Right lane—change in fast mode -L 3
Right lane—change in mild mode -L 5
Left lane—change in moderate mode L 4

The action atar is not listed, while its value will decide whether to accelerate, maintain velocity or decelerate.

Then, the influence on the planned trajectory is shown with different decision parameters through parallel simulations. In the parallel simulations, the left lane-change (YT=L) is implemented with changeable decision parameters T and atar at two initial longitudinal velocities vx, which involves different trajectory optimization problems as shown in Eq.(3). These optimization problems are solved by sequential quadratic programming (SQP) algorithm using MATLAB.

The trajectories with different action durations T{T1,T2,T3}, expected accelerations atar{A1,A2,A3}, and initial longitudinal velocities vx{V1,V2} are shown in Fig.3. Here, the initial longitudinal velocity V1=10 m/s and V2=20 m/s. The action duration T1=3 s, T2=4 s, and T3=5 s. The expected acceleration A1=-0.5 m/s2A2=0, and A3=0.5 m/s2.

Fig.3  Trajectories at different values of decision parameters

As shown in Fig.3, there is a certain distance gap between the trajectories with different microcosmic decisions, which may cause absolutely different situations when these microcosmic decisions are not considered in the decision layer.

1.3 Analysis of range of decision parameters in NGSIM

Firstly, the parameter-based decisions and the corresponding trajectories after optimization are compared with real traffic trajectories from NGSIM to verify the rationality of parameter-based decision-making. Three typical examples (C1C2C3) are shown in Fig.4. The start time step and end time step of lane-change in each trajectory are recognized with the method in reference [

9], which is set as the action duration T. The expected acceleration atar with an interval of 0.1 is calculated to fit the actual trajectory. The optimized trajectory is compared with the actual trajectory from NGSIM, in which the optimized trajectory has a good fitting performance compared with the actual trajectory. Meanwhile, action duration T and expected acceleration atar are listed in the right side of Fig.4. It can be seen that different lane-change maneuvers are executed with different drivers, which is a potential behavior that can be considered in the decision layer.

Fig.4  Comparison of actual trajectories from NGSIM (R in legend) and optimized trajectory with optimization trajectory planning controller (C in legend)

Meanwhile, the ranges of these decision parameters are analyzed with this real traffic dataset. The lane-change trajectories from the NGSIM are selected and labeled with the method in reference [

9]. The statistic values in the whole data are listed in Tab.2. The ranges of expected acceleration and action duration in NGSIM are used as a reference in the parameter-based decision-making problem in the next section.

Tab.2  Statistic results in NGSIM
Action durationTmin/sTmax/sT¯/s
T/s 2.5 7.5 4.5
Expected acceleration atar,min atar,max a¯tar
atar/(m/s2 -1.1 2.5 0.2

1.4 Driving scenarios analysis in NGSIM

In free traffic flow, there are eight potential surrounding positions around the host vehicle, which is numbered as shown in Fig.5. The ranges of different positions are given in Tab.3. For each labeled host vehicle, the vehicle that is in the range of positions P1P8 will be added into its driving scenario. The ranges of expected acceleration and action duration in NGSIM are used as a reference in the parameter-based decision-making problem in the next section.

Fig.5  Diagram of host vehicle and its surrounding vehicles

Tab.3  Range of different positions
PositionRangePositionRange
P4 TH n/s[0,3] P3P5 dn/m[15,60]
P2P6 dn/m[-15,15] P1P7P8 dn/m[-15,-60]

dn and THn=dn/vh are the distance in the lane direction and timeheadway between the host vehicle and the surrounding vehicles in the position i. vhis the velocity of the host vehicle.

In these driving scenarios, the relative distance between the host vehicle and the surrounding vehicles are calculated on its current lane and target lane when the host vehicle is changing the lane. 15 driving scenarios that the minimal distance is beyond 4 m are assumed as a potential emergency driving situation and can be promoted with a better decision while the driver's intention is also fully respected. Thus, the start time step of lane-change, the action duration for lane-change, and the average acceleration are recorded in these driving scenarios. The learning is done on the driving scenarios to obtain better safety situations without violating the driver's willingness. The driver's willingness includes the intention of changing lane, the action duration, and the average acceleration of lane-change.

2 MDP modeling and RL algorithm

The RL based parameter lane-change decision-making problem is established, and the kernel-based least squares policy iteration algorithm (KLSPI) that is proposed in reference [

16] is applied to learn the lane-change decision parameters.

2.1 MDP modeling

The decision process for driving is modeled as a MDP, which contains the design of state space, action space, and reward function. The trajectory planning controller changes the state of the host vehicle with the action selected in the decision layer.

In the design of state, to depict each of the vehicle in the potential positions P1P8, the relative velocity Δvn(k)=vn(k)-vh(k), the acceleration an(k), the relative distance in the lane direction dn(k), and the intention of the surrounding vehicles In(k) are considered. The intention of the surrounding vehicles In(k) is calculated using the method from reference [

9]. The subscript n indicates the position of surrounding vehicles while the subscript h indicates the host vehicle. Thus, the state vector can be expressed as:

s(k)=(Δv1(k),a1(k),d1(k),I1(k)P1,           ...,Δv8(k),a8(k),d8(k),I8(k)P8)T (4)

In the parameter-based decision framework, the decision is described with parameters, which is the lateral offset, the time of lane-change and expected acceleration, respectively. The action vector is expressed as:

a(k)=(YL(k),T(k),atar(k))T (5)

where: YL(k){-L,0,L} is the target lateral offset, L is the distance between two neighboring lanes; T(k)[3,6] is the time of lane-change or T(k)=1 is the time of lane keeping; atar(k)[-1,2] is the expected acceleration.

The reward function is designed with the consideration of safety rs, the intention of diver rr, and the consistency with all drivers rc, which can be expressed as:

r(k)=ksrs(k)+krrr(k)+kcrc(k) (6)

The safety reward rs evaluates the safety situation compared with the action taken rs,e and the original situation in NGSIM rs,o, which is the driver's willingness considered in this paper. The safety reward rs is:

rs=rs,o-rs,e (7)

The relative distance di between the host vehicle and the surrounding vehicles on its current lane and target lane when the host vehicle is changing the lane is calculated during the lane change process. Assuming there are nev surrounding vehicles, the incremental equation of action taken rs,e and the original situation in NGSIM rs,o are calculated in the same way and expressed as:

rs,e/o=rs,e/o+di-de   for di<de, i=1,2,...,nevrs,e/o-10   for di<dc, i=1,2,...,nevrs,e/o    else (8)

where: de=4 is the emergency distance, dc=2 is the collision distance. The reward for the intention of diver rr and the consistency with all drivers rc can be expressed as:

rr=-kr,1(T-To)2-kr,2(atar-atar,o)2 (9)
rc=-kc,1(T-Tc)2-kc,2(atar-atar,c)2 (10)

Here: To and Tc are the time for lane-change with the current driver and the average of all drivers, respectively; atar,o and atar,c are the average acceleration with the current driver and the average of all drivers.

2.2 KLSPI algorithm

The reinforcement learning algorithm KLSPI that is proposed in reference [

16] is used to solve this lane-change decision-making problem. The detailed of this algorithm is not discussed in this paper, only some changes that can better fit this algorithm to the problem and give the pseudocode in Algorithm 1 are explained. The main equations in Algorithm 1 are from reference [16].

Algorithm 1 KLSPI for lane-change decision-making:

1) Collect sample set

{s(k),a(k),r(k),s'(k), k=1,2,...,n}

with random policy and trajectory planning controller.

2) Sparsification: Initialize empty dictionary Dic={}.

For k=1 to ndo

Assuming that the current dictionary has t features, for feature mk) calculate

c=W-1w(m(k))ξ=wkk-wT(m(k))c (11)

where:

[W]i,j=κ(m(i),m(j))wkk=κ(m(k),m(k))
w(m(k))=[κ(m(1),m(k)),...,κ(m(t),m(k))]T

If ξ<μdo Dic=Dicm(k)

else continue

end if

end for

3) Policy Iteration: Random initialize A,b

Loop for j=1 to maximum iteration nm

For k=1 to n

Compute A,b,α with

A=A+w(m(k))[w(m(k))-γw(m'(k))]Tb=b+w(m(k))r(k)α=(A)-1b (12)

end for

For k=1 to n

Implement policy improvement with

π'=argmaxaj'Q˜(s',aj') (13)

end for

Until (aj'-aj-1')2<μe

4) Output α

In KLSPI, first, all training samples are collected. Secondly, the sparsification procedure is done to obtain features in this sample set that are not evident linear correlation to form a dictionary for function approximation, which is calculated withby using Eq.(11) and decided by using threshold parameter μ. Thirdly, policy iteration is carried outconducted until the termination threshold μ. Thirdly, policy iteration is carried out until the termination threshold μe is satisfied. These iteration equations can be directly found in reference [

16]. Finally, the policy is output with the weight vector αof estimated state-action value function Q˜(s,a).

The radial basis function (RBF) network is used as the function approximation to approximate the state-action value function, which can be expressed as:

Q˜(s(k),a(k))=j=1tαjκ(m(j),m(k)) (14)

Feature representation m(k) is customized, which combines the state vector s(k) and the action vector a(k)=[YT,T, atar] and can be expressed as:

m(k)=(h(1)s(k),h(2)s(k),h(3)s(k),T,atar)T (15)

here h is the activation vector and can be expressed as:

h=(1,0,0)for YT=-L(0,1,0)for YT=0(0,0,1)for YT=L (16)

The weight vector kφ is manually set to normalize the feature vector with a different range and measure state and action differently. The kernel function can be described as:

κ(m(k),m(j))=e-(kφ(m(k)-m(j)))22σ2 (17)

3 Simulation results

The driving scenarios in NGSIM are obtained and divided into the training set,the test set, and the cross-validation set. Sample sets are generated from the training set. The learning is implemented as illustrated in Sections 2 and 3, and influential parameters are decided. Finally, simulation results in the test set and the cross-validation set and promoted performance are verified.

3.1 Sample sets and learning process analysis

In NGSIM, 254 driving scenarios whose host vehicle execute lane-change are selected and are randomly divided into the training set, the test set, and the cross-validation set, which has 214, 30, 10 driving scenarios, respectively. In the training set, the time step that the host vehicle changes lane is found and three decision time steps that before and behind this time step are also considered whose time interval is 0.2 s. We use tThese time steps rather than the whole time are used to sample, because the behavior of other environments could be assumed to be maintained in this short time interval. Lane-change or lane-keeping decisions are both simulated in these decision points to collect sample sets in these training scenarios. Finally, 10 327 samples are obtained.

In the learning process, the threshold parameter μ will be compared with a feature linear correlation with features in the current dictionary to decides whether it will be added to the dictionary, which will be used to approximate the action-state value function. Thus, the threshold parameter μ will influence the dimension of the dictionary and function approximation. The dimension of the dictionary is 572, 170, and 91, respectively, when μ=0.01, 0.06, and 0.11. The average estimated error in the training set and test set are shown in Fig.6. Eventually, μ= 0.01 is chosen because of its better performance in both the training set and the test set.

Fig.6  Average estimated error in the training set and test set

The error in each sample is calculated using ee=γ Q˜(s',a')+r -Q˜(s,a)

3.2 Performance validation

First, an emergency scenario whose minimal distance between the host vehicle and the surrounding vehicles in the whole lane-change process is only 2.5 m in the test set is simulated. As the performance shown in Fig.7, the control policy after learning tends to change lane late with acceleration and a faster mode to obtain a safer driving situation. Thus the minimal distance in the initial stage is considerable, and the minimal distance in the whole lane-change process has increased to 3.7 m, which promotes the safety performance and retains the driver's intention to change lane rather than keep the lane. A better-performing behavior but also acceptable for still executing the human driver's willingness is realized.

Fig.7  Results for an emergency scenario in test set

In the cross-validation set, a common scenario is simulated, and the performance is shown in Fig.8. During diving, the control policy after learning change lane in a shorter time, which maintains the performance in both safety, velocity control, and finishing lane-change task. The maintained performance in the cross-validation set verifies the generalization performance of the learning results.

Fig.8  Results for a common scenario in cross-validation set

4 Conclusions and future works

Microcosmic behaviors, such as lane-change time and expected acceleration, in precise decision-making not only influence trajectory but also differ in drivers, which are shown in NGSIM but seldom considered in the decision layer. As emergency driving scenarios exist in NGSIM, a learning-based parameter decision-making method for automated vehicles to learn the precise decision-making has been investigated, which can balance safety and human driving willingness. The lane-change time and expected acceleration are added to the action space. Safety, the current driver's willingness, and the average human driving style are considered in the reward function. After training by KLSPI with driving scenarios in NGSIM, precise decision-making is realized. Safety performance is promoted in an emergency lane-change scenario of the test set, which indicates a better-performing behavior and is acceptable for still executing the human driver's willingness. Safety performance is maintained in the cross-validation set, which verifies its generalization performance of the learning results.

In the further, other deep reinforcement learning methods will be explored in more complex and changeable driving scenarios.

Reference

1

PADEN BČÁP MYONG S Zet al. A survey of motion planning and control techniques for self-driving urban vehicles[J]. IEEE Transactions on Intelligent Vehicles201611): 33. [Baidu Scholar] 

2

LI X HSUN Z PCAO D Pet al. Real-time trajectory planning for autonomous urban driving: Framework, algorithms, and verifications[J]. IEEE/ASME Transactions on mechatronics2015212): 740. [Baidu Scholar] 

3

GUO C ZKIDONO KTERASHIMA Ret al. Toward human-like behavior generation in urban environment based on Markov decision process with hybrid potential maps[C]// 2018 IEEE Intelligent Vehicles Symposium (IV). ChangshuIEEE20182209. [Baidu Scholar] 

4

CHU H QGUO L LYAN Y Jet al. Self-learning optimal cruise control based on individual car-following style[J], IEEE Transactions on Intelligent Transportation Systems2020991. [Baidu Scholar] 

5

GINDELE TBRECHTEL SDILLMANN Ret al. Learning driver behavior models from traffic observations for decision making and planning[J]. IEEE Intelligent Transportation Systems Magazine201571): 69. [Baidu Scholar] 

6

MARTINEZ C MHEUCKE MWANG F Yet al. Driving style recognition for intelligent vehicle control and advanced driver assistance: A survey[J]. IEEE Transactions on Intelligent Transportation Systems2018193): 666. [Baidu Scholar] 

7

GONZÁLEZ DPÉREZ JMILANÉS Vet al. A review of motion planning techniques for automated vehicles[J]. IEEE Trans. Intelligent Transportation Systems, 2016174): 1135. [Baidu Scholar] 

8

VALLON CERCAN ZCARVALHO Aet al. A machine learning approach for personalized autonomous lane change initiation and control[C]// 2017 IEEE Intelligent Vehicles Symposium (IV). Los AngelesIEEE20171590. [Baidu Scholar] 

9

HE G LLI XLYU Yet al. Probabilistic intention prediction and trajectory generation based on dynamic bayesian networks[C]// 2019 Chinese Automation Congress (CAC)HangzhouIEEE20192646. [Baidu Scholar] 

10

TAN Y VELLIOTT M RFLANNAGAN C A Cet al. Development of a real-time prediction model of driver behavior at intersections using kinematic time series data[J]. Accident Analysis & Prevention2017106428. [Baidu Scholar] 

11

YOU C XLU J BFILEV Det al. Highway traffic modeling and decision making for autonomous vehicle using reinforcement learning[C]// 2018 IEEE Intelligent Vehicles Symposium (IV). ChangshuIEEE20181227. [Baidu Scholar] 

12

SHALEV-SHWARTZ SSHAMMAH SSHASHUA Aet al. On a formal model of safe and scalable self-driving cars [DB/OL]. arXiv: 1708.063742017. https://doi.org/10.48550/arXiv.1708.06374. [Baidu Scholar] 

13

ZHANG Y XGAO B ZGUO L Let al. Adaptive decision-making for automated vehicles under roundabout scenarios using optimization embedded reinforcement learning[J]. IEEE Transactions on Neural Networks and Learning Systems2020991. [Baidu Scholar] 

14

ARIKERE AYANG DKLOMP Met al. Integrated evasive manoeuvre assist for collision mitigation with oncoming vehicles[J]. Vehicle System Dynamics20185610): 1. [Baidu Scholar] 

15

ZHANG Y XGAO B ZGUO L Let al. A novel trajectory planning method for automated vehicles under parameter decision framework[J]. IEEE Access2019788264. [Baidu Scholar] 

16

XU XHU D WLU X Cet al. Kernel-based least squares policy iteration for reinforcement learning[J]. IEEE Transactions on Neural Networks2007184): 973. [Baidu Scholar]