Next Article in Journal
A Grid-Based Approach for Measuring Similarities of Taxi Trajectories
Next Article in Special Issue
A Review on Vehicle Classification and Potential Use of Smart Vehicle-Assisted Techniques
Previous Article in Journal
Analysis for Joint Delay-Power Tradeoff with Buffer/Channel-Aware and Its FPGA Implementation in Wireless Sensor Networks
Previous Article in Special Issue
Robust Stereo Visual Inertial Navigation System Based on Multi-Stage Outlier Removal in Dynamic Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lane Position Detection Based on Long Short-Term Memory (LSTM)

1
College of Automotive Engineering, Chongqing University, Chongqing 400044, China
2
School of Information, Zhejiang University of Finance Economics, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(11), 3115; https://doi.org/10.3390/s20113115
Submission received: 17 April 2020 / Revised: 12 May 2020 / Accepted: 28 May 2020 / Published: 31 May 2020

Abstract

:
Accurate detection of lane lines is of great significance for improving vehicle driving safety. In our previous research, by improving the horizontal and vertical density of the detection grid in the YOLO v3 (You Only Look Once, the 3th version) model, the obtained lane line (LL) algorithm, YOLO v3 (S × 2S), has high accuracy. However, like the traditional LL detection algorithms, they do not use spatial information and have low detection accuracy under occlusion, deformation, worn, poor lighting, and other non-ideal environmental conditions. After studying the spatial information between LLs and learning the distribution law of LLs, an LL prediction model based on long short-term memory (LSTM) and recursive neural network (RcNN) was established; the method can predict the future LL position by using historical LL position information. Moreover, by combining the LL information predicted with YOLO v3 (S × 2S) detection results using Dempster Shafer (D-S) evidence theory, the LL detection accuracy can be improved effectively, and the uncertainty of this system be reduced correspondingly. The results show that the accuracy of LL detection can be significantly improved in rainy, snowy weather, and obstacle scenes.

1. Introduction

In order to avoid road accidents and ensure the safety of driving, in recent years, the research and development of the advanced driving assistance system (ADAS), which provides help for drivers, is booming all over the world [1,2,3,4]. Lane departure warning and lane keeping, lane change and forward collision warning, adaptive cruise control, and blind spot detection systems all belong to the category of ADAS. Among the many assisted driving technologies, LL detection is the core part of these systems [5,6,7]. Therefore, it is of great theoretical value and practical significance to study LL detection.
At present, LL detection is still a difficult research topic. In the real world, the structure is complex. Different countries have different traffic laws and regulations, and LL signs are different. Even in the same country, due to the differences in landform, the types and ways LL signs are used between cities are different. The differences of these types of markings are should be overcome by the algorithm itself. In addition, the LL is easily corroded and blurred with time, making it difficult for some algorithms based on computer vision for detection [8,9,10,11,12]. Furthermore, weather is also an important factor affecting the accuracy of the LL detection algorithm. When snow and rainstorms occur, LL is easily blocked, and visual detection of LL signs cannot be implemented at all [13]. Some other rule-based methods are needed to solve LL in similar scenes.
In the current LL detection algorithms, there is little research on the spatial distribution law of LL. However, LL has a clear distribution law in the structure, and the same type of LL sequence exists in the front of LL with high probability. Moreover, the distribution law is not affected by weather, the visual field, or obstruction. This study intended to establish an LL prediction model based on Long Short-Term Memory (LSTM) [14] and recurrent neural network by analyzing the distribution relationship of LL on the road surface. Only part of the known LL coordinate information is needed to predict the location of other LLs in complex scenarios.
The grid size of the bounding box is S × S in the YOLO v3 [15]; however, this density of grid cannot be applied for the detection of such a small size and large aspect ratio targets, like LLs. In our previous research [16], by improving the horizontal and vertical density of the bounding box in YOLO v3, the obtained LL detection algorithm YOLO v3 (S × 2S) improved the accuracy of LL detection. However, when the LL is occluded, the accuracy of the algorithm is poor. This research is a continuation of the problem of LL detection. On the basis of YOLO v3 (S × 2S), we further studied LL detection in complex scenes. The main contributions are as follows:
  • The use of a YOLO v3 (S × 2S)-based LL recognition model to detect LL position information and construct an LL position sequence;
  • Based on the historical position information of the detected LLs, a prediction model (ALSTM-RcNN) combined with angle information and a recurrent neural network, respectively, is presented.
  • To ensure the accuracy of the final LL detection results, the D-S evidence theory is used to integrate the output of the detection model and the prediction model to obtain the optimal position information.

2. Related Work

To reduce the influence of environmental factors on the detection algorithm, researchers use schemes that combine cameras, radars, high-precision maps, and so on to improve the system’s lane detection performance under complex operating conditions. Shin et al. [17] proposed an LL detection algorithm using a combination of cameras and Light Detection and Ranging (LiDAR), which used images and radar information to improve the LL recognition capabilities in harsh environments. Rose et al. [18] integrated cameras, Global Positioning System (GPS), and LiDAR to estimate the distance between the vehicle and lateral LL. De Paula et al. [19] proposed a lane detection algorithm using LiDAR, where a local grid reflection map around the vehicle was constructed using radar, and the LL was extracted from map information. Jung et al. [20] proposed a real-time model for lane detection using LiDAR data, which classified drivable areas, identified road signs, performed lane detection, and updated the corresponding parameters in real time. In combination with the GPS system, the detected road parameters were integrated to build a lane-level digital map. Because the reflectivity of different roads could cause large errors in laser measurements, it is difficult to use lasers for real vehicle applications of the LL detection system. High-precision maps can improve the accuracy of LL detection to a certain extent, but there are still some difficulties in mapping, which restrict the research and development of LL detection algorithms based on these sensors.
However, the sensor fusion algorithms, which depend on LiDAR, high-precision maps, and GPS, are complex, and the cost is higher than that of vision-based algorithms. Thus, the popularization and application of ADAS is blocked to a certain extent. Therefore, this study proposes an algorithm based on machine vision to study the LL high-precision detection in complex scenes in a low-cost way. At present, the lane detection method using visual information mainly uses three categories: Color-based, model-based, and learning-based methods.
The feature-based method does not need to consider the shape information of the road. Generally, according to the color, edge, or gradient change of LL and other feature information, the lane marking and the non-LL area on the road surface are separated by clustering or segmentation. Li et al. [21] delineated the region of interest in the captured image, eliminating noise using image graying and median filtering, while a spike neural network was used to detect the edge of the LL and the line itself was obtained using the Hough transform. Madrid et al. [22] proposed a lane departure warning system suitable for mobile devices, which used the Hough transform and image blurring to reduce the amount of calculations substantially and meet the real-time requirements. Gaikwad et al. [23] grayed the image and used a piecewise linear stretching function to improve the contrast of the interested region, achieving LL detection using the Hough transform. Yenİaydin et al. [24] proposed a robust lane detection algorithm using machine vision. However, grayscale images are greatly affected by some factors, such as shadows and the light intensity, and their use in lane detection is susceptible to environmental interference. After the image is converted to grayscale, color information is lost, resulting in the low robustness of the detection algorithm.
Detection methods based on color features can extract the color information in the road image effectively and obtain the LL information. Mammeri et al. [25] proposed an LL detection architecture, which combined the maximally stable extremal regions (MSER) algorithm and the Hough transform. By using MSER in a color image to determine the region of interest, and applying a three-level refined optimization algorithm to enhance the MSER results and filter out irrelevant information, such as trees and vehicles, the color and shape of LLs were matched. Nugraha et al. [26] converted the RGB color mode of the image into LUV and LAB color spaces, respectively, and extracted the yellow and white LLs on the road surface, achieving two-color LL detection after filtering. Comparing the aforementioned works with the LL detection algorithm using only grayscale images, we conclude that, to a certain extent, robustness is improved due to the use of color information. However, this grayscale information cannot be used to accurately perceive the changes in shape feature information and image size, so it is very difficult to detect local features.
In recent years, methods based on the convolutional neural network (CNN) [27,28,29] have also been gradually used for LL detection. Kim et al. [29] considered the global image information and transformed the detection of LLs into the task of region segmentation. Lee et al. [30] proposed an end-to-end trainable vanishing point guided network (VPGNet), which can obtain the position of the LL in the image by training a large number of pictures. Zhang et al. [16] proposed an LL detection algorithm based on the YOLO v3 using K-means clustering algorithm to optimize the network parameters. Kim et al. [31] proposed an LL detection scheme combining CNN and the random sample consensus (RANSAC) algorithm. Firstly, the image was input into the CNN network for denoising and enhancement after edge detection, and then the RANSAC algorithm was used to fit the LL. He et al. [32] input perspective and overhead images to CNN for training, respectively. The input of the network includes the image set of LL-positive samples (target LL) and the image set of LL-negative samples (road surface interference markers), and with the help of positive and negative labels, the network was trained. In order to solve the problem that the deep learning model needs a large number of label images, Zhang [28] et al. proposed a method using real scene images to synthesize label image sets for detecting lateral LL. The detection accuracy based on CNN is higher, because it uses a moving window to traverse the entire image, and the extracted features are the most abundant, but it also brings a huge amount of calculation, which greatly reduces the efficiency.
The length, width, and direction of LL have strong regularity, and the characteristics of serialization and structure association are obvious. The recurrent neural network (RNN) [33] is a kind of neural network that captures the dynamic information in serialized data through the periodic connection of hidden layer nodes. Ye et al. [34] proposed an LL detection method, which combines the CNN and long-term memory neural network. Li et al. [35] developed a multitask deep convolution network to detect the existence of the target and its geometric attributes (position and direction). Zou et al. [36] combined CNN and RNN, first extracted the feature information of each frame by CNN, and then input RNN for feature learning and LL prediction. Li [35] and Zou et al. [36] verified that RNN can successfully detect LLs in the case of missing LLs. Although RNN can improve the detection accuracy of LL by using continuous multi-frame information, the number of frames input into the network and the sampling interval between frames will affect the detection results, and multi-frame data is bound to increase the processing time. In this study, by using the spatial distribution law of LLs, a method of LL detection using LSTM and RNN is proposed. Only part of the known LL coordinate information is needed to estimate the position of other LLs.

3. Lane Line Prediction Algorithm

3.1. Recurrent Neural Network

A recurrent neural network (RcNN) can construct a network structure recursively from the bottom to the top according to a given directed acyclic graph. In order to solve the problem that LLs are difficult to detect in the case of turns, we used the RcNN to represent lane line sequences with certain rules. By traversing the nodes in the directed acyclic graph topology, the vector of a child node is used to represent the vector of the learnt parent node. To unify and simplify the network structure, the directed acyclic graph was transformed to obtain a form of binary tree, as shown in Figure 1.
All nodes in the recurrent neural network have input data. For the ith node, the system state equation is:
h ( i ) = f ( U T h c ( i ) + W T X + b ) ,
where h ( i ) ,   h c ( i ) represent the system state of the current node and its parent node, respectively; X denotes the input data of this node; f represents the activation function of the encapsulated feedforward neural network; and U, W, and b represent the weight coefficients. The recursive neural network can learn the distribution characteristics of LLs with a larger span by combining the characteristics of adjacent LLs. In the learning process, an LL sequence is formed according to the distribution relation of LLs, which not only retains LL feature information but also combines the hierarchical structure characteristics of LLs.

3.2. Lane Line Prediction Model Considering Angle Information

The recurrent neural network has obtained a good result for LL prediction on straight roads in previous works, but for curved lanes, it is difficult to accurately predict its position using only frame coordinate information. To solve this problem, the angle information must be taken into account. When constructing a deep neural network to train the lane detection model, the coordinate information of the LL should be calculated, and we should also consider whether there is an angle shift in the direction of the LL. This is because this shift is closely related to the direction of the front and rear LLs. In view of this structural phenomenon, an angle LSTM-RcNN (ALSTM-RcNN) based on an angular transfer model was constructed, as shown in Figure 2. The LSTM-RcNN unit has a binary tree structure composed of two input gates, one output gate, one memory unit, and two forgetting units. When calculating the vector of a parent node, the position and direction information of the two child nodes are taken into account in the ALSTM-RcNN model, which is proposed in this paper.
Similar to LSTM, the memory unit of ALSTM-RcNN is also composed of an input gate, forget gate, output gate, and memory cell. However, since ALSTM-RcNN is based on a tree structure, there is more than one input of LSTM-RcNN at time t (for binary tree, there are two inputs), so ALSTM-RcNN has two input gates and forget for each LSTM unit gate, as shown in Figure 2, and an ALSTM-RcNN cell based on the binary tree structure is composed of two input gates, one output gate, one memory cell, and two forget cells. The sign ‘~’ in Figure 2 represents the activation function ‘tanh’.
An LL steering angle example is shown in Figure 3, which shows a discontinuous LL composed of multiple LL blocks. The original image is mapped into a bird’s eye image and excludes irrelevant information. A, B, and C are the center coordinate points of the bounding boxes of the three LL blocks in the map. The cosine theorem can be used to determine the angle transfer α of CB to AB.
For driving safety, the angle of the LL is usually limited to a range to avoid sudden large-angle steering corrections. Considering that the method should try to include as many cases as possible, the range of α was limited to [−15°, 15°], with an interval of 2°, so that an angle transfer label set was obtained as A = {−15, −13, …, 0, …, 13, 15}. We defined the direction label of the left child of the current node as α l , that of the right child as α r , and the angle transfer vector as α * l r . The vector of the current node is represented by the position and direction information of its two children. When training this model, a sequence containing the LL coordinates and the steering angle is input to the ALSTM-RcNN network, and the distribution rule of LLs is learned through the training model. Let the child nodes be m and n, respectively, while the parent nodes are denoted using p. h t 1 m , h t 1 n represent the vectors of the child nodes in the hidden layer; c t 1 m ,   c t 1 n represent the memory vectors of the child nodes; P and Q are coefficient matrices; and b * is the bias term. The LL prediction model considering the angle information can be expressed as:
i 1 = δ ( P i 1 m h t 1 m + P i 1 n h t 1 n + Q i 1 m c t 1 m + Q i 1 n c t 1 n + b i 1 ) ,
i 2 = δ ( P i 2 m h t 1 m + P i 2 n h t 1 n + Q i 2 m c t 1 m + Q i 2 n c t 1 n + b i 2 ) ,
f 1 = δ ( P f 1 m h t 1 m + P f 1 n h t 1 n + Q f 1 m c t 1 m + Q f 1 n c t 1 n + b f 1 ) ,
f 2 = δ ( P f 2 m h t 1 m + P f 2 n h t 1 n + Q f 2 m c t 1 m + Q f 2 n c t 1 n + b f 2 ) ,
c t = f 1 c t 1 m + f 2 c t 1 n + tanh ( P c x h t 1 x i 1 + P c y h t 1 y i 2 + b c ) ,
o = δ ( P o m t 1 m + P o n t 1 n + Q m c t 1 m + Q n c t 1 n + Q n c t + b o ) ,
where i 1 , i 2 indicate the input gate; f 1 , f 2 denote the forgetting gate; o indicates the parent node output gate; c t indicates the parent node’s forgetting gate; and h t p indicates the output of the parent node.

4. Integration of Lane Detection Results Based on D-S

D-S evidence theory [37] is a kind of imprecise reasoning theory proposed by Dempster Shafer, which satisfies the weaker condition than Bayesian probability theory. It has unique advantages in uncertain information representation and multi-source evidence fusion. The basic principle is the uncertain information describing the system is transformed into evidence, and then the D-S evidence combination rule is used to fuse the multi-source uncertain evidence to form the mass function value of the identification framework Θ. Therefore, in this study, for the sake of improving the detection accuracy of the LL and avoiding the low precision caused by the false and weak detection of certain types of detectors (YOLO v3(S × 2S) and ALSTM-RcNN), the D-S evidence theory is used to fuse the output results of the two detectors and optimize the detection result. For the sake of improving the detection accuracy of the LL and avoiding the low precision caused by the false and weak detection of certain types of detectors, the D-S evidence theory was used to fuse the output results of the two detectors and optimize the detection result. The LL detection framework using the D-S evidence theory [30] is shown in Figure 4. The image collected by the camera is detected by the YOLO v3(S × 2S) model, and the lane position of LL is obtained. By constructing the LL sequence, the training of the prediction model is completed, and then the D-S evidence theory is used to integrate the output results of the YOLO v3(S × 2S) and the prediction model to obtain the optimal LL position information in complex scenes.
In D-S evidence theory, if is the problem to be discerned, all possible and mutually incompatible decision results form a non-empty set Θ, which is the recognition framework of [34]. In the present study, we determined whether there is an LL in a certain area, so the recognition framework can be expressed as Θ = { ω l a n e , ω n o l a n e } . We defined the set function as m: 2 Θ [ 0 , 1 ] , which represents the basic assignment probability in Θ, satisfying the following equation:
{ m ( ) = 0 A Θ m ( A ) = 1 ,
where m(A) represents the trust degree of the focus element A and represents the empty set. In terms of two independent pieces of evidence, it was assumed that m1 and m2 are the distribution functions of the corresponding basic probabilities in Θ, and the focal elements are A 1 , , A M and B 1 , , B M , respectively. According to the D-S evidence theory, we can synthesize the equation of the basic probability distribution function as follows:
{ m ( ) = 0 m ( C ) = A i B j = C m 1 ( A i ) m 2 ( B j ) 1 K , C ,
where K = A B = m 1 ( A i ) m 2 ( B j ) represents the uncertainty factor and reflects the degree of conflict between different evidence. The basic probability assignment (BPA) given by m is called the orthogonal sum of m1 and m2, which is m 1 m 2 . If K = 1, then m 1 m 2 does not exist and the basic amplitude function cannot be synthesized.
In order to better manage the evidence conflict problem in DS theory, reliability is introduced by assigning a weight coefficient to the output of the two detectors. Reliability can reflect the quality of the fused information quantitatively, as information with high credibility has a higher weight and conversely for low credibility. This approach can enhance the advantages of the credible information selected, and improve the stability and reliability of the system during analysis. The performance of YOLO v3 (S × 2S) and ALSTM-RcNN are tested on the test set and the average accuracy rate is used as the credibility, recorded as R1 and R2, respectively. A1, …, Ak and B1, …, Bk are observed and assigned 1 R * to the recognition frame U as unknown information. According to D-S theory, the combination rule is as follows:
m ( C ) = { i , j C [ R 1 m 1 ( A i ) , ( 1 R 1 ) ] [ R 2 m 2 ( B j ) , ( 1 R 2 ) ] 1 K ,   C U , C 0                                                                                                           ,   C = ,
where K = i , j A i B j = [ R 1 m 1 ( A i ) ] [ R 2 m 2 ( B j ) ] , C = [ A i , ( 1 R 1 ) ] [ B i , ( 1 R 2 ) ] .
The decision-making based on the minimum risk is to find the minimum one in the decision set as the optimal decision-making. According to the calculation process of LL detection and the detection results of YOLO and LSTM-RcNN, the decision-making based on the basic probability assignment was adopted in this paper. The rules were as follows:
Suppose A 1 , A 2 U ,
If
{ m ( A 1 ) = m a x { m ( A i ) , A i Θ } m ( A 2 ) = m a x { m ( A i ) , A i Θ   &   A i A 1 } ,
and:
{ m ( A 1 ) m ( A 2 ) > ε 1 m ( φ ) < ε 2 m ( A 1 ) > m ( Θ ) ,
Then, A1 is the judgment result, where ε 1 and ε 2 represent preset thresholds, while A1 represents the result of the judgment.

5. Result Analysis and Discussion

The computer operating system used for testing was Ubuntu 14.04, the CPU was i7-7600K, with 16Gb of RAM and an NVIDIA GTX1080. We used Euro Truck Simulator 2’s photo gallery for training and testing, as it contains a wealth of traffic scenes, including a collection of night, rain, and fog images.

5.1. Model Training and Result Analysis

5.1.1. Model Training Based on Transfer Learning

The transfer learning method [38,39] was used to train on the Euro Truck Simulator 2 (ETS2) traffic scene image set using the YOLO v3 (S × 2S) LL detection model. As shown in Figure 5, photo galleries including different weather conditions, such as night, rainy, and foggy images, were used. Caltech and KITTI traffic scene pictures were the source domains [40]. The YOLO v3 (S × 2S) LL detection model was trained using these two data sets. The task of transfer learning was to identify LLs in ETS2 traffic scenes, and the traffic scenes were the target areas.
ETS2 has many traffic scenarios, which basically cover all weather types. In this study, sunny, rainy, and snowy weather was selected. The road types were mainly urban roads, highways, tunnels, and bridges. In order to improve the training efficiency, the screen recording software Camtasia Studio 8 was used to record the game process as a video file at 36 frames per second. The online training method was still used during training. The pictures collected from ETS2 traffic scenes were input to the YOLO v3 (S × 2S) model. The total time of the training video was 16 h and 26 min. The training parameters were initialized. The initial parameter values for training were set as follows: learning_rate = 0.001, batchsize = 128, momentum = 0.9. During the training process, the model’s detection capability was gradually enhanced, so we set a growth factor ζ = 2e−5 for T0. The training process is shown in Figure 6.
The picture of the game scene was input to the YOLO v3 (S × 2S) model. Initially, the detection was not ideal, so the detection results needed to be pre-processed so that they could be used as standard label pictures for model training. Meanwhile, using only the condition Tξ to confirm whether an LL is recognized will cause some LLs to be missed. During model training in the Caltech and KITTI scenes, because the predicted values with low confidence were probably non-LLs, the pixel value of the bounding box area of the confidence value of the detection result in the range of T/4 < ξ < T was set to 0. Due to the lower interference in the ETS2 traffic scene images, and to accelerate the training speed, the threshold of the confidence value was increased from the original T/4 to T/3, namely, the LL boundary box in the range of T/3 < ξ < T was expanded to (x, y, w + 2δ, h + 2δ), and the pixels in this area were set to 0. For bounding boxes whose confidence value satisfied Tξ, they were expanded to (x, y, w + 2δ, h + 2δ), and then an adaptive edge detection algorithm based on Canny [41] was used to quickly relocate the detected LLs.
The detection results of the YOLO V3 (S × 2S) LL recognition model in ETS2 traffic scene images are shown in Figure 7. In Figure 7a,d,g, the model achieves better recognition results in scenes with good lighting conditions. Figure 7e,h show the results of the test conditions at night, where images of the road surface containing clear LLs are accurately identified under the illumination of the headlights. In Figure 7c, the YOLO v3 (S× 2S) LL detection model cannot predict the position of the LL using machine vision due to vehicle occlusion. In Figure 7f, despite the interference of the zebra crossing, the obtained model can distinguish the LL accurately, so the trained model is proved to have strong robustness. Figure 7i is a game scene in rainy weather. As the raindrops hit the windshield, the spray was generated, which caused the LL in the picture to be deformed. In the process of moving the wiper, part of the LL was also blocked, so some LLs could not be detected accurately.
To evaluate the performance of the YOLO v3 (S × 2S) LL recognition model, manual labeling was used to label the pictures in the Euro Truck Simulator 2 traffic scenes, and a standard data set used for testing was obtained and classified according to weather conditions, as shown in Table 1.
As shown in Table 1, the accuracy on the ETS2 traffic scene images reached 91% on sunny and cloudy scenes, exceeding the accuracy rates obtained on KITTI and Caltech due to the game scenes, in which the LLs were clearly marked and there was almost no interference from other obstacles. The accuracy of the sunny and cloudy scenes is the same, which indicates that the model’s detection of the LL is almost unaffected by the light intensity. In the rainy scenes, due to the interference of raindrops on the front windshield and the swing of the wipers, the shape of the LL in the image is deformed, which reduces the accuracy of LL recognition.

5.1.2. Model Training for ALSTM-RcNNs

The analysis in the previous section showed that the LL detection model of LSTM-RcNN does not consider the angle information of the LL under steering conditions, which leads to a low LL prediction accuracy. In the process of constructing the ALSTM-RcNN model, the angle of each LL module is fully considered through the training of the network model, and the angle and coordinate information of the LLs are output using the softmax function. The YOLO v3 (S × 2S) LL detection model was used to detect LLs on the road surface and remove poorly performing pictures. We relocated the detected LLs using the Canny-based adaptive edge detection algorithm, and the steering angle of the LL was obtained using the cosine algorithm, so the parameters of each LL were (xi, yi, wi, hi, αi), where (xi, yi) are the center point coordinates of the bounding box, and wi and hi represent the width and height of the bounding box. To determine the images sets accurately identified, the LL area was expanded to (xi, yi, wi +2δ, hi +). Using the adaptive Canny edge detection algorithm, the detected LL was relocated to obtain accurate coordinate information. After the preprocessing was completed, the optimized LL detection results were input into the ALSTM for model training. The training process is shown in Figure 8. The parameters of each LL were input into the ALSTM-RcNN model, and the Adam optimization method was used for training. The initial values of the training parameters were set as follows: learning_rate = 0.006, batchsize = 128. Through the feedback training network, the corresponding angle was output through the softmax layer, so that the coordinate information of the LL was obtained and the prediction of the LL in the case of the curve was realized.
The LL detection results using the ALSTM-RcNN model are shown in Figure 9. In nighttime (Figure 9a) and shadow scenes (Figure 9b), LLs were detected accurately. In Figure 9c, the LLs obtained were severely deformed due to the water droplets hitting the front windshield; however, the ALSTM-RcNN model still predicted them accurately.
The accuracy rate in each scenario is shown in Table 2, and is higher than the LSTM-RCNN model, with the MAP reaching 70%. The algorithm can realize LL detection under steering conditions. Although accurate detection cannot be performed when there is a change in the steering angle, the predicted LL coordinate information is more similar to the standard data set, thus the accuracy of traffic line identification is improved.

5.2. Results Analysis of Lane Line Detection

The ALSTM-RcNN uses the known LL position information and considers the angular variation of the LL in the curve situation, where the LL prediction is performed in the case of limited vision. The MAP of the LL detection model based on YOLO v3 (S × 2S) reaches 90%, which has a high detection accuracy. However, it is difficult to detect the LL when the LL is deformed due to occlusions or raindrops.
In order to comprehensively consider the identification results of two kinds of LL detectors and make the output LL coordinate information reach the optimum, D-S evidence theory was adopted to integrate the detection results of YOLO v3(S × 2S) and ALSTM-RcNN to obtain the final output accurate LL position. The identification framework of D-S evidence theory is Θ = { ω l a n e , ω n o l a n e } , which is divided into two cases, depending on whether an LL exists or not. If an LL exists, the boundary box of the LL will be output. The YOLO detector detects the position of the LL using visual information, and the output value is (ξYOLO, x, y, h, w). ALSTM-RcNN predicts the position center coordinates of the LL using existing LL position information, and ultimately outputs the maximum coordinates (x, y) of the confidence value ξRcNN using the softmax function. Considering the confidence value as the LL detection accuracy, the probability of LL presence is ξYOLO and ξRcNN, respectively, and the probability of LL absence is 1−ξYOLO v3(S × 2S) and 1−ξALSTM-RcNN, respectively. To obtain the reliability of the two detectors, we used the images of the ETS2 image collection as test data to determine the reliability of YOLO v3 (S × 2S) and ALSTM-RcNN, as shown in Table 3. The accuracy of the ALSTM-RcNN algorithm by itself is significantly lower than that of the YOLO v3 (S × 2S) algorithm. However, since it can predict the LL position information in the case of rainy days and obstacles, after D-S evidence theory fusion, the MAP value of LL recognition is increased to 93.74%, reaching a high recognition level.
The LL detection results using YOLO v3 (S × 2S), ALSTM-RcNN, and D-S fusion are shown in Figure 10. The YOLO v3 (S × 2S) model can detect the position of the LL without obstacle occlusion accurately, while the ALSTM-RcNN algorithm predicts the future LL sequence using the detected LL, making the recognition accuracy of the line slightly higher. After fusing the D-S evidence theory, the accurate LL position is output, as shown in Figure 10a. Part of the LL in Figure 10b is blocked by other vehicles, and the LL in Figure 10c is affected by raindrops. The LL obtained using the camera was severely deformed. The YOLO v3 (S × 2S) model can only detect visible LLs and not occluded and severely deformed LLs. The ALSTM-RcNN algorithm uses information on existing LLs to predict LL presence. Finally, the D-S evidence theory fuses the results of the two detectors and outputs accurate LL position information.
To test the accuracy of the lane recognition algorithm, a set of system devices was designed using the real vehicle. The camera was placed at the upper center position of the windshield using a suction cup to collect the traffic images in front of the vehicle. The roads in the experiment were highways, urban streets, and tunnels marked with obvious LLs, and also included vehicle congestion and unobstructed scenes. Because the designed algorithm used machine vision, the brightness had a great influence on the detection results, so the test scene in this study contained the road conditions under different brightness. The experiment was carried out in Chongqing. The layout of the vehicle sensors and driving route is shown in Figure 11. Moreover, Caltech was introduced to analyze the detection results of LLs and verify the effectiveness of the algorithm by comparison.
In order to illustrate the effectiveness of the algorithm, the average accuracy (mAP) of the detected LL bounding box and the detection speed were used as evaluation criteria. The algorithm in this paper was compared with the detection results of current LL detection algorithms, such as Fast-RCNN [42], Faster-RCNN [43], Sliding window + CNN [44], SSD [45], and Context + RCNN [46]). The results are shown in the Table 4. Since the algorithm in this study adopted the method of combining YOLO v3(S × 2S) and ALSTM-RcNN, it has strong robustness to the detection of LLs in complex scenes, but because of its use of the RNN network, the detection speed is slower.
In addition, due to the complexity of road conditions in the real scene, the accuracy of LL detection is lower than that in Caltech. However, the algorithm in this study adopts the method of fusion of YOLO v3(S × 2S) and ALSTM-RcNN, which can predict the location of the lane line in complex scenes, such as LL wear and occlusion. Therefore, compared with the single YOLO v3(S × 2S) method, the accuracy of LL detection is greatly improved. However, in the Caltech data set, the LLs in the scenes of Cordoval and Washington are relatively clear, and only YOLO v3(S × 2S) can achieve better detection results. The algorithm in this study does not improve the accuracy much. In summary, the LL algorithm proposed in this paper combining YOLO v3(S × 2S) and ALSTM-RcNN has strong robustness in harsh scenarios.
The LL bounding boxes detected by the algorithm in this study are independent of each other, but in the actual LL detection, a continuous line is needed. Therefore, the K-means and RANSAC algorithm were combined, and each individual LL bounding box was fitted into a curve. The final LL detection results are shown in Figure 12. In cloudy and sunny scenes, when the road conditions are good and there are no shadows or obstructions, the LL can be accurately detected. While in the night and tunnel scenes, the light source in the picture is mainly the lights and streetlights, which are darker than in the daytime, and the image quality collected by the camera at night is poor. The method in this study can predict the blurred and obscured LLs through the detected LLs, and output the complete LL results, which has strong robustness.

6. Conclusions

To solve the problem that the LLs are difficult to detect when occluded or deformed, the distribution of LLs and an LL detection method based on sequence information were studied. The LL prediction model ALSTM-RcNN with angle polarity was proposed. The tests showed that ALSTM-RcNN can predict the position of the LL accurately in both straight and curved situations. To improve the accuracy of LL detection and decrease the large errors in the recognition algorithm of single LLs, we used D-S evidence theory to integrate the output of the LL detection model YOLO v3 (S × 2S) and LL prediction model ALSTM-RcNN. The results showed that the detection accuracy of the LL after D-S fusion is significantly higher than the single detector.
However, though the experiments showed that the research method in this paper has high detection accuracy, due to the complexity of the algorithm, it cannot meet the real-time requirements. In future research, the multi-sensor fusion method will be considered to enable LL detection’s removal of visual restrictions. In complex scenes, not only accuracy but also real-time LL detection should be guaranteed.

Author Contributions

X.Z. and W.Y. conceived and designed the method and performed experiments; Q.L. and D.S. reviewed and edited this manuscript; P.X. and Y.H. carried out part of the experimental verification. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Chongqing Natural Science Foundation, grant number No. cstc2018jcyjAX0468.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sun, X.; Cai, Y.; Wang, S.; Xu, X.; Chen, L.; Long, C. Optimal control of intelligent vehicle longitudinal dynamics via hybrid model predictive control. Robot. Auton. Syst. 2019, 112, 190–200. [Google Scholar] [CrossRef]
  2. Tang, X.; Zhang, D.; Liu, T.; Khajepour, A.; Yu, H.; Wang, H. Research on the energy control of a dual-motor hybrid vehicle during engine start-stop process. Energy 2019, 166, 1181–1193. [Google Scholar] [CrossRef]
  3. Brinkley, J.; Posadas, B.; Sherman, I.; Daily, S.; Gilbert, J.E. An open road evaluation of a self-driving Vehicle human–machine interface designed for visually impaired users. Int. J. Hum. Comput. Int. 2019, 35, 1018–1032. [Google Scholar] [CrossRef]
  4. Li, L.; Wang, X.; Shao, Y.; Li, W.; Song, F. Entropic pressure between fluctuating membranes in multilayer systems. Sci. China Phys. Mech. Astron. 2018, 61, 128711. [Google Scholar] [CrossRef]
  5. Zhang, X.; Yang, W.; Zhang, Y.; Liu, J.; Zhou, S. An Improved Algorithm for Image Synthesis based on Gradient and Database. In Proceedings of the 3rd International Conference on Robotics, Control and Automation (CACRE), Chengdu, China, 19–22 July 2018; pp. 164–168. [Google Scholar]
  6. Butakov, V.; Ioannou, P. Personalized driver/vehicle lane change models for ADAS. IEEE Trans. Veh. Technol. 2014, 64, 4422–4431. [Google Scholar] [CrossRef]
  7. Baskar, L.; Hellendoorn, J.; De Schutter, B.; Papp, Z. Traffic control and intelligent vehicle highway systems: A survey. IET Intell. Transp. Syst. 2011, 5, 38–52. [Google Scholar] [CrossRef] [Green Version]
  8. Lee, H.; Tomizuka, M. Adaptive vehicle traction force control for intelligent vehicle highway systems (IVHSs). IEEE Trans. Ind. Electron. 2003, 50, 37–47. [Google Scholar]
  9. Tang, X.; Yang, W.; Hu, X.; Zhang, D. A novel simplified model for torsional vibration analysis of a series-parallel hybrid electric vehicle. Mech. Syst. Signal Process. 2017, 85, 329–338. [Google Scholar] [CrossRef]
  10. Zhang, X.; Yang, W.; Tang, X.; He, Z. Estimation of the lateral distance between vehicle and lanes using convolutional neural network and vehicle dynamics. Appl. Sci. 2018, 8, 2508. [Google Scholar] [CrossRef] [Green Version]
  11. Wang, R.; Ye, Q.; Cai, Y.; Wang, Y.; Xu, X.; Meng, X.; Long, C. Analyzing the influence of automatic steering system on the trajectory tracking accuracy of intelligent vehicle. Adv. Eng. Softw. 2018, 121, 188–196. [Google Scholar] [CrossRef]
  12. Deng, G.; Wu, Y. Double Lane Line Edge Detection Method based on Constraint Conditions Hough Transform. In Proceedings of the 2018 17th International Symposium on Distributed Computing and Applications for Business Engineering and Science (DCABES), Wuxi, China, 19–23 October 2018; pp. 107–110. [Google Scholar]
  13. Tang, X.; Hu, X.; Yang, W.; Yu, H. Novel torsional vibration modeling and assessment of a power-split hybrid electric vehicle equipped with a dual-mass flywheel. IEEE Trans. Veh. Technol. 2018, 67, 1990–2000. [Google Scholar] [CrossRef]
  14. Gao, L.; Guo, Z.; Zhang, H.; Xu, X.; Shen, H.T. Video captioning with attention-based LSTM and semantic consistency. IEEE Trans. Multimed. 2017, 19, 2045–2055. [Google Scholar] [CrossRef]
  15. Redmon, J.; Farhadi, A. Yolov3: An Incremental Improvement. Available online: https://arxiv.org/abs/1804.02767 (accessed on 8 April 2018).
  16. Zhang, X.; Yang, W.; Tang, X.; Liu, J. A fast learning method for accurate and robust lane detection using two-stage feature extraction with YOLO v3. Sensors 2018, 18, 4308. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Shin, S.; Shim, I.; Kweon, I.S. Combinatorial Approach for Lane Detection using Image and LIDAR Reflectance. In Proceedings of the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Goyang, Korea, 28–30 October 2015; pp. 485–487. [Google Scholar]
  18. Rose, C.; Britt, J.; Allen, J.; Bevly, D. An integrated vehicle navigation system utilizing lane-detection and lateral position estimation systems in difficult environments for GPS. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2615–2629. [Google Scholar] [CrossRef]
  19. De Paula Veronese, L.; Ismail, A.; Narayan, V.; Schulze, M. An Accurate and Computational Efficient System for Detecting and Classifying Ego and Sides Lanes using LiDAR. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1476–1483. [Google Scholar]
  20. Jung, J.; Bae, S.-H. Real-time road lane detection in urban areas using LiDAR data. Electronic 2018, 7, 276. [Google Scholar] [CrossRef] [Green Version]
  21. Li, X.; Wu, Q.; Kou, Y.; Hou, L.; Yang, H. Lane Detection Based on Spiking Neural Network and Hough Transform. In Proceedings of the 2015 8th International Congress on Image and Signal Processing (CISP), Shenyang, China, 14–16 October 2015; pp. 626–630. [Google Scholar]
  22. Madrid, N.; Hurtik, P. Lane departure warning for mobile devices based on a fuzzy representation of images. Fuzzy Sets Syst. 2016, 291, 144–159. [Google Scholar] [CrossRef]
  23. Gaikwad, V.; Lokhande, S. Lane departure identification for advanced driver assistance. IEEE Trans. Intell. Transp. Syst. 2014, 16, 1–9. [Google Scholar] [CrossRef]
  24. Yenİaydin, Y.; Schmidt, K.W. A Lane Detection Algorithm Based on Reliable Lane Markings. In Proceedings of the 2018 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018; pp. 1–4. [Google Scholar]
  25. Mammeri, A.; Boukerche, A.; Tang, Z. A real-time lane marking localization, tracking and communication system. Comput. Commun. 2016, 73, 132–143. [Google Scholar] [CrossRef]
  26. Nugraha, B.T.; Su, S.F. Towards Self-Driving Car using Convolutional Neural network and Road Lane Detector. In Proceedings of the 2017 2nd International Conference on Automation, Cognitive Science, Optics, Micro Electro-Mechanical System, and Information Technology (ICACOMIT), Jakarta, Indonesia, 23–24 October 2017; pp. 65–69. [Google Scholar]
  27. Cao, Y.; Chen, Y.; Khosla, D. Spiking deep convolutional neural networks for energy-efficient object Recognition. Int. J. Comput. Vis. 2014, 113, 54–66. [Google Scholar] [CrossRef]
  28. Zhang, X.; Yang, W.; Tang, X.; Wang, Y. Lateral distance detection model based on convolutional neural network. IET Intell. Transp. Syst. 2019, 13, 31–39. [Google Scholar] [CrossRef]
  29. Kim, J.; Kim, J.; Jang, G.-J.; Lee, M. Fast learning method for convolutional neural networks using extreme learning machine and its application to lane detection. Neural Netw. 2017, 87, 109–121. [Google Scholar] [CrossRef] [PubMed]
  30. Lee, S.; Kim, J.; Yoon, J.S.; Shin, S.; Bailo, O.; Kim, N.; Lee, T.-H.; Hong, H.S.; Han, S.-H.; Kweon, I.S. VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1947–1955. [Google Scholar]
  31. Kim, J.; Lee, M. Robust Lane Detection Based on Convolutional Neural Network and Random Sample Consensus. In Proceedings of the International Conference on Neural Information Processing (ICNIP 2017), Montreal Convention Center, Montreal, QC, Canada, 12–13 December 2014; pp. 454–461. [Google Scholar]
  32. He, B.; Ai, R.; Yan, Y.; Lang, X. Accurate and Robust Lane Detection based on Dual-View Convolutional Neutral Network. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gotenburg, Sweden, 19–22 June 2016; pp. 1041–1046. [Google Scholar]
  33. Mishra, A.K.; Desai, V. Drought forecasting using feed-forward recursive neural network. Ecol. Model. 2006, 198, 127–138. [Google Scholar] [CrossRef]
  34. Ye, Y.Y.; Hao, X.L.; Chen, H.J. Lane detection method based on lane structural analysis and CNNs. IET Intell. Transp. Syst. 2018, 12, 513–520. [Google Scholar] [CrossRef]
  35. Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep neural network for structural prediction and lane detection in traffic scene. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 690–703. [Google Scholar] [CrossRef] [Green Version]
  36. Zou, Q.; Jiang, H.; Dai, Q.; Yue, Y.; Chen, L.; Wang, Q. Robust lane detection from continuous driving scenes using deep neural networks. IEEE Trans. Veh. Technol. 2020, 69, 41–54. [Google Scholar] [CrossRef] [Green Version]
  37. Ayoub, N.; Gao, Z.; Chen, B.; Jian, M. A synthetic fusion rule for salient region detection under the framework of DS-evidence theory. Symmetry 2018, 10, 183. [Google Scholar] [CrossRef] [Green Version]
  38. Shin, H.-C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M.; Hoo-Chang, S. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Singh, J.; Hanson, J.; Paliwal, K.K.; Zhou, Y. RNA secondary structure prediction using an ensemble of two-dimensional deep neural networks and transfer learning. Nat. Commun. 2019, 10, 1–13. [Google Scholar] [CrossRef] [Green Version]
  40. Narote, S.P.; Bhujbal, P.N.; Narote, A.S.; Dhane, D. A review of recent advances in lane detection and departure warning system. Pattern Recognit. 2018, 73, 216–234. [Google Scholar] [CrossRef]
  41. Son, J.; Yoo, H.; Kim, S.; Sohn, K. Real-time illumination invariant lane detection for lane departure warning system. Expert Syst. Appl. 2015, 42, 1816–1824. [Google Scholar] [CrossRef]
  42. Zhang, X.; Zhuang, Y.; Wang, W.; Pedrycz, W. Online feature transformation learning for cross-domain object category recognition. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 1–15. [Google Scholar] [CrossRef] [PubMed]
  43. Pan, J.; Sayrol, E.; Giro-I-Nieto, X.; McGuinness, K.; O’Connor, N.E. Shallow and Deep Convolutional Networks for Saliency Prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 59–606. [Google Scholar]
  44. Ahmed, E.; Clark, A.; Mohay, G. A novel sliding window based change detection algorithm for asymmetric traffic. In Proceedings of the 2008 IFIP International Conference on Network and Parallel Computing, Shanghai, China, 18–21 October 2008; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2008; pp. 168–175. [Google Scholar]
  45. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot Multibox Detector. In Proceedings of the 14th European Conference on Computer Vision (ECCV) 2016, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  46. Tian, Y.; Gelernter, J.; Wang, X.; Chen, W.; Gao, J.; Zhang, Y.; Li, X. Lane marking detection via deep convolutional neural network. Neurocomputing 2018, 280, 46–55. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Structure of recurrent neural network (RcNN).
Figure 1. Structure of recurrent neural network (RcNN).
Sensors 20 03115 g001
Figure 2. Network structure of Angle LSTM-RcNN (ALSTM-RcNN).
Figure 2. Network structure of Angle LSTM-RcNN (ALSTM-RcNN).
Sensors 20 03115 g002
Figure 3. The angle of lane lines (LLs).
Figure 3. The angle of lane lines (LLs).
Sensors 20 03115 g003
Figure 4. LL detection framework.
Figure 4. LL detection framework.
Sensors 20 03115 g004
Figure 5. Three traffic scene images.
Figure 5. Three traffic scene images.
Sensors 20 03115 g005
Figure 6. Training process of lane detection model based on the improved You Only Look Once (the third version) (YOLO V3(S × 2S)).
Figure 6. Training process of lane detection model based on the improved You Only Look Once (the third version) (YOLO V3(S × 2S)).
Sensors 20 03115 g006
Figure 7. The results detected obtained using YOLO v3(S × 2S) in Euro Truck Simulator 2 (ETS2). The (a,d,g) are the detection result in scenes with good lighting conditions. The (b,e,h) are in the night. The (c,f,i) are in complex scenes such as vehicle occlusion, the interference of the zebra crossing, and rainy weather.
Figure 7. The results detected obtained using YOLO v3(S × 2S) in Euro Truck Simulator 2 (ETS2). The (a,d,g) are the detection result in scenes with good lighting conditions. The (b,e,h) are in the night. The (c,f,i) are in complex scenes such as vehicle occlusion, the interference of the zebra crossing, and rainy weather.
Sensors 20 03115 g007
Figure 8. Training flow of the ALSTM-RcNN.
Figure 8. Training flow of the ALSTM-RcNN.
Sensors 20 03115 g008
Figure 9. ALSTM-RcNN detection results on ETS2. The (ac) are the lane detection result in the nighttime, shadow, and rain scenes, respectively.
Figure 9. ALSTM-RcNN detection results on ETS2. The (ac) are the lane detection result in the nighttime, shadow, and rain scenes, respectively.
Sensors 20 03115 g009
Figure 10. Detection results of YOLO v3 (S × 2S), ALSTM-RcNN, and Dempster Shafer (D-S) evidence theory. The (a) are the lane detection result in the scenes without obstacle occlusion; (b) are the result in the scenes with obstacle occlusion; (c) are in the rain scene.
Figure 10. Detection results of YOLO v3 (S × 2S), ALSTM-RcNN, and Dempster Shafer (D-S) evidence theory. The (a) are the lane detection result in the scenes without obstacle occlusion; (b) are the result in the scenes with obstacle occlusion; (c) are in the rain scene.
Sensors 20 03115 g010
Figure 11. The layout of the camera and the driving route. The (a) is the layout of camera; (b) is the driving route.
Figure 11. The layout of the camera and the driving route. The (a) is the layout of camera; (b) is the driving route.
Sensors 20 03115 g011
Figure 12. The LL detection result in real scenes. The LLs were detected in three scenarios, where (a) is on a sunny day, (b) is a cloudy day, and (c) is the detection result in tunnel and night scenes.
Figure 12. The LL detection result in real scenes. The LLs were detected in three scenarios, where (a) is on a sunny day, (b) is a cloudy day, and (c) is the detection result in tunnel and night scenes.
Sensors 20 03115 g012
Table 1. Distribution and mean average precision (MAP) of label images.
Table 1. Distribution and mean average precision (MAP) of label images.
WeatherNumberMAP
sunny40091.7%
cloudy40091.5%
rainy30086.4%
Table 2. MAP of ALSTM-RcNN.
Table 2. MAP of ALSTM-RcNN.
WeatherMAP
Sunny69.7%
Cloudy70.4%
Rainy68.4%
Table 3. The confidence value of ALSTM-RcNN on ETS2.
Table 3. The confidence value of ALSTM-RcNN on ETS2.
AlgorithmConfidence Value
YOLO v3 (S × 2S)90.7%
ALSTM-RcNN69.4%
D-S evidence theory93.74%
Table 4. The detection accuracy and speed of algorithms on real scene images.
Table 4. The detection accuracy and speed of algorithms on real scene images.
AlgorithmmAPTime/ms
Fast RCNN45.71%2271
Faster RCNN53.37%122
Sliding window + CNN64.82%79,000
SSD70.26%29.3
Context + RCNN73.43%197
Yolo v3(S × 2S)83.94%25.2
The proposed algorithm85.45%124

Share and Cite

MDPI and ACS Style

Yang, W.; Zhang, X.; Lei, Q.; Shen, D.; Xiao, P.; Huang, Y. Lane Position Detection Based on Long Short-Term Memory (LSTM). Sensors 2020, 20, 3115. https://doi.org/10.3390/s20113115

AMA Style

Yang W, Zhang X, Lei Q, Shen D, Xiao P, Huang Y. Lane Position Detection Based on Long Short-Term Memory (LSTM). Sensors. 2020; 20(11):3115. https://doi.org/10.3390/s20113115

Chicago/Turabian Style

Yang, Wei, Xiang Zhang, Qian Lei, Dengye Shen, Ping Xiao, and Yu Huang. 2020. "Lane Position Detection Based on Long Short-Term Memory (LSTM)" Sensors 20, no. 11: 3115. https://doi.org/10.3390/s20113115

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop