Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review

Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review applied sciences Review Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review Monirul Islam Pavel , Siok Yee Tan * and Azizi Abdullah Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia; pavel@ieee.org (M.I.P.); azizia@ukm.edu.my (A.A.) * Correspondence: esther@ukm.edu.my Abstract: In the past decade, autonomous vehicle systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. However, the AVS is still far away from mass production because of the high cost of sensor fusion and a lack of combination of top-tier solutions to tackle uncertainty on roads. To reduce sensor dependency and to increase manufacturing along with enhancing research, deep learning-based approaches could be the best alternative for developing practical AVS. With this vision, in this systematic review paper, we broadly discussed the literature of deep learning for AVS from the past decade for real-life implementation in core fields. The systematic review on AVS implementing deep learning is categorized into several modules that cover activities including perception analysis (vehicle detection, traffic signs and light identification, pedestrian detection, lane and curve detection, road object localization, traffic scene analysis), decision making, end-to-end controlling and prediction, path and motion planning and augmented reality-based HUD, analyzing research works from 2011 to 2021 that focus on RGB camera vision. The literature is also analyzed for final representative outcomes as visualization in augmented reality-based head-up display (AR-HUD) with categories such as early warning, road markings for improved navigation and enhanced safety with overlapping on vehicles and pedestrians Citation: Pavel, M.I.; Tan, S.Y.; in extreme visual conditions to reduce collisions. The contribution of the literature review includes Abdullah, A. Vision-Based detailed analysis of current state-of-the-art deep learning methods that only rely on RGB camera Autonomous Vehicle Systems Based on Deep Learning: A Systematic vision rather than complex sensor fusion. It is expected to offer a pathway for the rapid development Literature Review. Appl. Sci. 2022, 12, of cost-efficient and more secure practical autonomous vehicle systems. 6831. https://doi.org/10.3390/ app12146831 Keywords: autonomous controlling; deep learning; decision making; intelligent vehicle; perception; self-driving Academic Editor: Seong-Ik Han Received: 11 May 2022 Accepted: 20 June 2022 Published: 6 July 2022 1. Introduction Publisher’s Note: MDPI stays neutral Recently, the autonomous vehicle system (AVS) has become one of the most trend- with regard to jurisdictional claims in ing research domains that focus on driverless intelligent transport for better safety and published maps and institutional affil- reliability on roads [1]. One of the main motives for enhancing AVS developments is iations. its ability to overcome human driving mistakes, including distraction, discomfort and lack of experience, that cause nearly 94% of accidents, according to a statistical survey by the National Highway Traffic Safety Administration (NHTSA) [2]. In addition, almost 50 million people are severely injured by road collisions, and over 1.25 million people Copyright: © 2022 by the authors. worldwide are killed annually in highway accidents. The possible reasons for these injuries Licensee MDPI, Basel, Switzerland. may derive from less emphasis on educating drivers with behavior guidance and poorly This article is an open access article developed drivers’ training procedures, fatigue while driving, visual complexities, that distributed under the terms and is, human error, which can be potentially solved by adopting highly efficient self-driving conditions of the Creative Commons vehicles [3,4]. The NHTSA and the U.S. Department of Transportation formed the SAE Attribution (CC BY) license (https:// International levels of driving automation, identifying autonomous vehicles (AV) from creativecommons.org/licenses/by/ 0 0 ‘level 0 to the ‘level 5 [5], where levels 3 to 5 are considered to be fully AV. However, as of 4.0/). Appl. Sci. 2022, 12, 6831. https://doi.org/10.3390/app12146831 https://www.mdpi.com/journal/applsci Appl. Sci. 2022, 12, 6831 2 of 51 2019, the manufacturing of level 1 to 3 vehicle systems has been achieved but level 4 vehicle systems are in the testing phase [6]. Moreover, it is highly anticipated that autonomous vehicles will be employed to support people in need of mobility as well as reduce the costs and times of transport systems and provide assistance to people who cannot drive [7,8]. In the past couple of years, not only the autonomous driving academic institutions but also giant tech companies like Google, Baidu, Uber and Nvidia have shown great interest [9–11] and vehicle manufacturing companies such as Toyota, BMW and Tesla are already working on launching AVSes within the first half of this decade [12]. Although different sensors such as radar, lidar, geodimetric, computer views, Kinect and GPS are used by conventional AVS to perceive the environment [13–17], it is indeed expensive to equip vehicles with these sensors and the high costs of these sensors are often limited to on-road vehicles [18]. Table 1 shows a comparison of three major vision sensors based on a total of nine factors. While the concept of driverless vehicles has existed for decades, the exorbitant costs have inhibited development for large-scale deployment [19]. To resolve this issue and build a system that is cost efficient with high accuracy, deep learning applied vision-based systems are becoming more popular where RGB vision is used as the only camera sensor. The recent developments in this field of deep learning have accelerated the potential of profound learning applications for the solution of complex real-world challenges [20]. Table 1. Comparison of vision sensors. VS VR FoV Cost PT DA AAD FE LLP AWP Camera High High Low Medium Medium High High Medium Medium Lidar High Medium High Medium High Medium Medium High Medium Radar Medium Low Medium High High Low Low High Low VS = Vision Sensor, VR = Visibility Range, FoV = Field of View, PT = Processing Time, DA = Distance Accuracy, AAD = AI Algorithm Deployment, FE = Feature Engineering, LLP = Low-Light Performance, AWP = All-Weather Performance. In this systematic review paper, a broad discussion and survey of the implementation of deep learning are applied to aspects of AVS such as vehicle detection (VD), traffic signs and light identification (TSL), pedestrian detection (PD), lane detection and tracking (LDT), traffic scene analysis (TSA), decision making (DM), end-to-end controlling and prediction (E2EP), path and motion planning (PMP) and augmented reality-based HUD (ARH) ana- lyzing research articles from 2011 to 2021 research articles on deep learning-applied AVS to reduce the dependency on sensor fusion and the high cost of manufacturing and to enhance the focus on developing a level 5 autonomous driving vehicle. We represent and thoroughly discuss the best deep learning algorithms for each domain, provide solutions to their limitations and analyze their performance for increasing practical implementation concepts. Moreover, this systematic review explored the most complete and predominate domains compared to other surveys from [21–33] (shown in Table 2) that indicates its impact on AVS implementing deep learning where the review article covered all aspects of the human–machine interface (HMI). The overall contribution of the research is set out below: Analyzed recent solution of state-of-the-art deep learning algorithms for cost-efficient AVS using RGB camera. Detailed literature review covering major domains and most subcategories to decrease vision sensor complexities. Discussed the key advantages and disadvantages of deep learning methods applied to AVS. Appl. Sci. 2022, 12, 6831 3 of 51 Table 2. Comparison of existing studies. Survey Coverage Ref. Year VD LRCD PD TSL E2EC TSA PMP DM ARH HMI [21] 2019 4 X 4 X X X 4 4 X 4 [22] 2020 X X X 4 X 4 4 4 X X [23] 2016 X X X X 4 X 4 X X X [24] 2020 X X X X 4 X 4 X X 4 [25] 2018 X X X X 4 X 4 X X X [26] 2018 X X X X 4 X 4 X X X [27] 2021 X X X 4 4 X 4 4 X 4 [28] 2020 4 4 4 4 4 4 X X X 4 [29] 2018 X X 4 X X X X X X X [30] 2020 4 4 4 X X X 4 X X X [31] 2020 X 4 4 4 4 4 4 X X X [32] 2021 X 4 X X X X 4 4 X X [33] 2020 4 X X X 4 4 4 X X 4 Ours 2022 4 4 4 4 4 4 4 4 4 X 2. Methodology 2.1. Review Planning The study is based on a systematic review methodology, an approach for analyzing and evaluating accessible studies related to a particular issue of current research where the core three phases are preparing the review, conducting the review, and creating a report that summarizes the review. In this systematic review paper, the researchers have included 142 papers containing deep learning and belonging to a different domain of AVS. To finalize the papers, we initially focused on the entire domain of autonomous driving, then we restricted our search to the usage of deep learning in AVS. Only papers with full text in English from renowned journals, conferences and book chapters that were published between 2011 and 2021 were selected. Due to an increase in the scope of advanced autonomous transportation, we finally limited our search to the vision-based application of deep learning in AVS, and the rest were rejected. We also took the most complete edition to avoid dealing with duplication. The key plan and protocol of the review includes source of data, searching criteria and procedure, research questions, data selection and data extraction. 2.2. Sources of Data Research papers were gathered from various famous research databases to incorporate specific field and research questions. Irrelevant research papers that could not address or endorse our research questions were dismissed. To achieve a broad coverage for the literature review, we used the following databases as our key resources: Web of Science, Scopus, IEEE Xplorer, ScienceDirect, MDPI, Springer, Wiley Library and ACM. 2.3. Research Questions Research questions were formed to refine the survey and maintain the aim of the topic. The following research questions are answered throughout the discussion in the different sections of the paper. How does deep learning reduce sensor dependency? How are on-road objects detected and localized? What decision-making processes are solved for AVS? How does deep learning contribute to end-to-end controlling and path planning? How should final outcomes be represented in AR-HUD? Appl. Sci. 2022, 12, 6831 4 of 51 2.4. Searching Criteria To find research papers according to the methodology, a pattern was followed to gather suitable papers which were mostly necessary for our study. We adopted a Boolean searching method with multiple AND, OR in the advance search options of each data source. During the search for the relevant papers, we selected “Autonomous Driving” and “Autonomous Vehicle” or “Intelligent Vehicle” or “Self-Driving” and “Deep Learning” as the main phrases. For a further refined search, various keywords were included to obtain the desired research papers according to our aim in this review. The following queries were developed based on Boolean operations: ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Object) AND ([Vehicle] OR [Pedestrians] [Traffic Sign] AND [Traffic Light])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self-Driving) AND (Deep Learning) AND ([Traffic Scene] OR [Localization] OR [Segmentation])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Lane) AND ([Track] OR [Shift] OR [Segmentation])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Control) AND ([Steering] OR [Motion])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND ([Deep Learning] OR [Deep Reinforcement Learning]) AND (Decision Making) AND ([Uncertainty] OR [Lane Keeping] OR [Overtaking] OR [Braking] OR [Acceleration])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND ([Augmented Reality] AND [Head Up Display] OR [HUD])) 2.5. Searching and Extraction Procedure The selection procedure for choosing papers includes four core iteration filtering processes. As the aim of the study is to discuss implementation of deep learning and comprehensive literature searches to analyze the frameworks and system designs, first, a total of 760 papers were selected from eight data sources based on the queries mentioned in the searching criteria (Section 2.4). Web of Science had the highest 151 and ACM had the lowest 40 papers. Then, the selected papers had to be processed through an eligibility stage where 209 duplicated papers were eliminated at first. Furthermore, 121 papers were screened out during abstract scanning and 276 papers were chosen after full text reading. In the next iteration, studies containing domains of deep learning in relation to AVS were selected where all the papers were published between 2011 and 2021. The final dataset contains a total of 142 papers that covers the literature on the implementation of deep learning methods for AVS. The structure of the whole selection process is presented in Figure 1. Table 3 presents the final calculation for the selection of 142 papers according to these steps and based on the most relatable topics and in-depth analysis. 2.6. Analysis of Publication by Year Out of 142 final papers for review, the studies published between 2011 and 2021 were selected. The year 2019 had the highest number of selected research papers, with 31, which is 17.9% of the total, and 2011 had the lowest number of papers (2). The distribution of publication is visualized in Figure 2. Appl. Sci. 2022, 12, x FOR PEER REVIEW 5 of 52 Springer 96 53 21 Wily Library 71 44 8 Appl. Sci. 2022, 12, 6831 5 of 51 MDPI 112 62 21 Total 760 430 142 Figure 1. Framework for searching the literature and making the selection. Figure 1. Framework for searching the literature and making the selection. 2.6. Analysis of Publication by Year Table 3. Paper selection from multiple sources. Out of 142 final papers for review, the studies published between 2011 and 2021 were Scheme 151 Primary Candidate Selected selected. The year 2019 had the highest number of selected research papers, with 31, which Web of Science 151 72 24 is 17.9% of the total, and 2011 had the lowest number of papers (2). The distribution of IEEE Xplore 124 69 22 publication is visualized in Figure 2. Scopus 72 47 20 ScienceDirect 94 51 19 ACM 40 32 7 Springer 96 53 21 Wily Library 71 44 8 MDPI 112 62 21 Total 760 430 142 Appl. Sci. 2022, 12, x FOR PEER REVIEW 6 of 52 Appl. Sci. 2022, 12, x FOR PEER REVIEW 6 of 52 Appl. Sci. 2022, 12, 6831 6 of 51 Figure 2. Distribution of studies in terms of year of publication (2011–2021). 2.Figure 7. Analy 2. Distribution sis of Publicat of studies ion by in Country terms of year of publication (2011–2021). Figure 2. Distribution of studies in terms of year of publication (2011–2021). Among the 142 selected papers for the literature review, 56 countries contributed to 2.7. Analysis of Publication by Country autonomous vehicle system development. Figure 3 shows the top 10 countries and the 2.7. Analy Among sis the of P 142 ublicat selected ion bpapers y Country for the literature review, 56 countries contributed to number of papers they contributed before the final selection. The graphical representation autonomous vehicle system development. Figure 3 shows the top 10 countries and the Among the 142 selected papers for the literature review, 56 countries contributed to shows that China made the largest contribution, with 34 papers, and the USA contributed number of papers they contributed before the final selection. The graphical representation autonomous vehicle system development. Figure 3 shows the top 10 countries and the 21 shows papers that , wh China ich wa made s th the e secon largest d la contribution, rgest. with 34 papers, and the USA contributed number of papers they contributed before the final selection. The graphical representation 21 papers, which was the second largest. shows that China made the largest contribution, with 34 papers, and the USA contributed 21 papers, which was the second largest. Figure 3. Distribution of studies over top 15 countries of first authors. Figure 3. Distribution of studies over top 15 countries of first authors. 2.8. Analysis of Publication Domains 2.8. Analy Thes142 is of final Publicat papers ion D wer om e ains selected based on five domains and five subdomains of perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined The 142 final papers were selected based on five domains and five subdomains of Fig to ure produce 3. Dist arcomplete ibution ofsystem. studies T ov able er t4 op shows 15 count thatrie the s of distribution first authoof rs.‘Decision Making’ perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the to produce a complete system. Table 4 shows that the distribution of ‘Decision Making’ lowest 11 papers individually. 2.8. Analysis of Publication Domains section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the The 142 final papers were selected based on five domains and five subdomains of lowest 11 papers individually. perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined to produce a complete system. Table 4 shows that the distribution of ‘Decision Making’ section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the lowest 11 papers individually. Appl. Sci. 2022, 12, 6831 7 of 51 Appl. Sci. 2022, 12, x FOR PEER REVIEW 7 of 52 Table 4. Literature taxonomy of AVS using deep learning approaches. Table 4. Literature taxonomy of AVS using deep learning approaches. Domain Sub-Domain References Domain Sub-Domain References Vehicle Detection [34–45] Vehicle Detection [34–45] TTr raf a fic ffic S Sign igand n an Light d Lig Recognition ht Recognition [46 [4 –6 59 –59] ] Perception Pedestrian Detection [60–78] Perception Pedestrian Detection [60–78] Lane Detection and Tracking [44,79–101] Lane Detection and Tracking [44,79–101] Traffic Scene Analysis [55,102–120] Traffic Scene Analysis [55,102–120] Decision Making - [121–143] Decision Making - [121–143] End-to-End Controlling and Prediction - [144–163] End-to-End Controlling and Prediction - [144–163] Path and Motion Planning - [164–175] Path and Motion Planning - [164–175] AR-HUD - [176–186] AR-HUD - [176–186] To visualize the leading algorithms of each domain or subdomain, Figure 4 presents the distribution of algorithms, where the reviewed algorithm-centered approaches have a To visualize the leading algorithms of each domain or subdomain, Figure 4 presents predominant role in AVS development. Figure 5 shows the dataset clustering which was the distribution of algorithms, where the reviewed algorithm-centered approaches have a used for the reviewed approaches. Only the subdomains of perception showed depend- predominant role in AVS development. Figure 5 shows the dataset clustering which was ency on dataset, where “Traffic Sign and Light Recognition” and “Lane Detection and used for the reviewed approaches. Only the subdomains of perception showed dependency Tracking” applied to 6 datasets each, and only 3 datasets were adopted in “Traffic Scene on dataset, where “Traffic Sign and Light Recognition” and “Lane Detection and Tracking” Analysis”. applied to 6 datasets each, and only 3 datasets were adopted in “Traffic Scene Analysis”. Figure 4. Taxonomy algorithms for each domain. Figure 4. Taxonomy algorithms for each domain. Appl. Sci. 2022, 12, 6831 8 of 51 Appl. Sci. 2022, 12, x FOR PEER REVIEW 8 of 52 Figure 5. Clustering of dataset for subdomains of perception. Figure 5. Clustering of dataset for subdomains of perception. 3 3. . A Analysis nalysis of of Doma Domain in Each domain was analyzed by reviewing several approaches and methods based on Each domain was analyzed by reviewing several approaches and methods based on evaluating and discussing advantages, disadvantages, outcomes and significance. The evaluating and discussing advantages, disadvantages, outcomes and significance. The fol- following analysis of each domain was carried out with the aim of accelerating development lowing analysis of each domain was carried out with the aim of accelerating development of level 4 or 5 AVS. of level 4 or 5 AVS. 3.1. Perception 3.1. Perception 3.1.1. Vehicle Detection 3.1.1. Vehicle Detection The identification and detection of an on-road vehicle for AVS together form one of the The identification and detection of an on-road vehicle for AVS together form one of predominant and most challenging issues due to versions, combined fast multitasking and the predominant and most challenging issues due to versions, combined fast multitasking visual difficulties. For fast and more accurate vehicle detection and recognition in different and visual difficulties. For fast and more accurate vehicle detection and recognition in and uncertain driving conditions, deep learning algorithms are analyzed in this section. different and uncertain driving conditions, deep learning algorithms are analyzed in this For instance, an online network framework for detecting and tracking vehicles was section. proposed by Hu et al., who predicted full 3D vehicle bounding box mapping from a For instance, an online network framework for detecting and tracking vehicles was monocular camera using both the environment and camera coordinates by reprojecting [34]. proposed by Hu et al., who predicted full 3D vehicle bounding box mapping from a mo- Eventually, the framework tracked the movement of instances in a global coordinate system nocular camera using both the environment and camera coordinates by reprojecting [34]. and revised 3D poses with a trajectory approximation of LSTM, implementing on a KITTI Eventually, the framework tracked the movement of instances in a global coordinate sys- dataset, where the outcomes surpassed the outcomes of LiDAR in long range [187]. In a tem and revised 3D poses with a trajectory approximation of LSTM, implementing on a 30 m range, LiDAR obtained 350.50 false negative and the method scored 395.33, while KITTI dataset, where the outcomes surpassed the outcomes of LiDAR in long range [187]. vehicle detection was 11.3% higher, indicating the limitation of the framework. However, it In a 30 m range, LiDAR obtained 350.50 false negative and the method scored 395.33, performed better in 50 m and 10 m, where the false negative scores were 857.08 and 1572.33 while vehicle detection was 11.3% higher, indicating the limitation of the framework. when the LiDAR-based method obtained false negative values of 1308.25 and 2445.30, However, it performed better in 50 m and 10 m, where the false negative scores were respectively. The decreased false negative in 10 and 50 m showed that the method was 857.08 and 1572.33 when the LiDAR-based method obtained false negative values of able to overcome the performance of LiDAR using only camera and deep learning despite 1308.25 and 2445.30, respectively. The decreased false negative in 10 and 50 m showed reduced accuracy in some real-time implementations. that the method was able to overcome the performance of LiDAR using only camera and To tackle accuracy-based issues, improve slow detection and recognition speed, and deep learning despite reduced accuracy in some real-time implementations. address the lack of categorization ability, Sang et al. introduced a novel YOLOv2_Vehicle To tackle accuracy-based issues, improve slow detection and recognition speed, and architecture [35]. For multiple scales of vehicles that influenced the detection framework, a normalization ddress the lack had of ca been tego used rizatio for nthe abili impr ty, S ovement ang et al. of inthe troduc method ed a n of ove measuring l YOLOv2 losses _Vehicle for a boundary rchitecturbox e [35length ]. For m and ultiple width scales of after clustering vehicles thand at inbounding fluenced th boxes e deteapplying ction fram k-mean ework, normalization had been used for the improvement of the method of measuring losses for Appl. Sci. 2022, 12, 6831 9 of 51 ++ algorithm to the training dataset [36] along with applying multilayer feature fusion to boost the network extraction capabilities and repeatedly eliminating convolutional layers in high layers. The method implementing the BIT dataset could obtain a mean average precision (mAP) exceeding 94.78% in 0.038 s, which was found to be much faster and more accurate than compared existing methods. In another work, an AdaBoost combined with a pixel look-up features-based approach was demonstrated by Ohn-Bar et al., where the methods included mining orientation, object geometry and occlusion pattern by clustering, and 81.94%, 66.32%, 51.10% accuracy in easy, moderate and hard scenarios, respectively, was obtained for vehicle detection. However, performance decreased when a 70% overlap evaluation threshold was used instead of 50% and during heavy occlusion [37]. Nonetheless, the method was inappropriate for rough conditions as performance decreased when 70% overlap evaluation threshold was used instead of 50%, and it showed poor accuracy in heavy occlusion. Further, Chen et al. presented a new method to identify five distinct vehicle classes, that is, car, van, bus, truck and tractor, using the AdaBoost and CNN algorithms applied to CompCars and their custom dataset containing rear views of vehicles [38]. They employed CNN as a function extractor with a Support Vector Machine (SVM) for training the features separately, and further AdaBoost algorithms were applied for integration. They obtained optimum results even with faulty images and high computing costs, with average accuracy of 99.50% in 0.028 s, which was 13% higher than the other mentioned fusion methods, for instance, SIFT + SVM, HoG + SVM and SURF + SVM. However, the method was only deployed and considered in simple but low-quality images and daytime scenarios. Moreover, one of the biggest issues was the low resolution of images in the real- time traffic surveillance method due to either low-vision RGB cameras or environmental features such as low light condition or foggy weather. For this problem, vehicles in low- resolution images and videos were analyzed in terms of the efficiency of the CNN by Bautista et al. [39]. The neural network used an activation function that worked in two phases: first, the detection of high-level attributes; second, the detection of low-level attributes. It tested the comportment of the model to detect vehicles with lower input resolution at different levels, as well as the number and size of filters. Results demonstrate that CNN was remarkably successful even with low resolution in the identification and classification of vehicles with an average precision fit for real time applications. Lee et al. showed a hierarchical system for detecting and tracking vehicles in an urban area at night based on taillights [40]. The system focused primarily on effective detection and pairing of taillights, considering their innate variety and observing all aspects of the layers and interrelationships in a hierarchical framework which increases the efficiency of vehicle detection and tracking in comparison with traditional methods, with recall of 78.48% and precision of 90.78%. However, performance decreased for short-distance vehicles due to headlight illumination of host vehicles. This approach could be considered as one the most suitable methods for nighttime vehicle detection. Hu et al. demonstrated a novel CNN architecture called scale-insensitive convolu- tional neural networking (SI-Net) [41] to enhance the performance of vehicle detection for autonomous vehicles, solving the issue of limited CNN-based vehicle detection [39]. The framework improved the limitation-scale insensitive CNN, deploying context-aware region of interest (ROI) pooling to preserve the real structure of small-scale objects. The state-of- the-art method outperformed the others in terms of measuring, scoring 89.6%, 90.60% and 77.75% for accuracy in moderate, easy and complex moods, respectively, in 0.11 s execution time on the KITTI benchmark as well as a custom highway dataset with different variance of scaled objects. Thus, the method was able to maintain a good performance in multiple traffic scenarios. Targeting the runtime of previous works, Wang et al. combined anchor size, receptive field and anchor generation optimization (AGO) with Fast R-CNN to ensure that an acceptable number of vehicle features could be accessed by the network in the shortest amount of time [42]. Using the anchor shape, it efficiently detects vehicles in large, medium Appl. Sci. 2022, 12, 6831 10 of 51 and short fields of view with 87.2% average precision in 0.055 s. The anchor shape-based detection process is a very coherent technique for AVS for reducing computational cost by not taking the whole field of vision for processing. In another work, which combined Faster R-CNN training parameters with a region proposal network (RPN)-based approach, Suhao et al. implemented vehicle-type detection in a real traffic area including MIT and Caltech datasets with ZF and VGG-16 networks in multiple scenarios [43]. The research results increased the average accuracy of the detection systems and the rate of detection compared with the CNN models. The proposed architecture classified vehicles from three categories where they achieved the best accuracy: 84.4% for car, 83.8% for minibus and 78.3% for SUV using the VGG-16 model in 81 ms. The execution cost of the proposed method outperformed Fast R-CNN and Faster R-CNN applied to complex scenarios. However, a lightweight YOLO network that was built with a YOLO v3 algorithm with generalized IoU loss was combined with loss function as well as with the integration of two different focal length cameras by Liu et al. to indicate less computer complexity for AVS [44]. The method was implemented on their self-made dataset where the network obtained 90.38% precision and 82.87% recall within 44.5 ms. This could be a milestone for AVS in terms of a faster and more accurate method for different field of view and day or nighttime implementation. Leung et al. compared deep learning-based techniques for vehicle detection effi- ciency [45], and proposed solutions for data collection along with the nighttime data labelling convention to resolve different types of detection. The research also recommends a framework based on a quicker region-based CNN model, which was precisely optimized, merging with RestNet101, the VGG-16 model obtaining a mean average precision (mAP) of 84.97%. The experimental result showed a high detection accuracy in urban nighttime with version lighting conditions including extreme low light and no lighting. Thus, this method became one of the most suitable methods for AVS in challenging lighting conditions. Overall, the Deep CNN and AdaBoost-based approach achieved 99.50% accuracy in daytime (the highest) with the fastest computational time (0.028 s), but lightweight YOLO and the quicker region-based CNN model showed practical outcomes in both daytime and nighttime scenarios for vehicle detection. Multiple deep learning methods showed efficient performance by improving slow detection, recognition and categorization, enabling deployment in complex scenarios and night-time deployment with good accuracy, even surpassing accuracy outcomes of LiDAR in terms of long field of view. However, some challenges remained, for example, limited dataset of vehicle categories, performance dropping in low light and rough weather conditions for some methods, low accuracy for vehicle detection in short distance for headlight illumination at nighttime and fast execution time in real-time implementation. An overview of the methods evaluated for the detection and recognition of vehicles for AVS is provided in Table 5. 3.1.2. Traffic Sign and Light Recognition One of the most important aspects of a safe and better decision-making process for automotive driving system was traffic sign and light identification by regulating traffic, monitoring and avoiding accidents through warning the drivers. Traffic sign and light recognition systems follow a double-step process, detection and classification, where detec- tion denotes correctly spotting the geometric position in the image and classification means identification of the category in which the detected sign or light signal appears [28,188]. A bio-mechanism inspired novel architecture named Branch Convolutional Neural Network (BNN) was proposed by Hu et al. for traffic sign recognition [46]. To improve the recognition machine speed and accuracy, a branch-output mechanism which was placed between pooling and convolutional layer and added to the framework. Furthermore, instead of the initial output layer, the sign in the preceding branch was projected by the BCNN that results perfect prediction in partial visibility of road signs with 98.52% Appl. Sci. 2022, 12, 6831 11 of 51 accuracy based on German Traffic Sign Recognition Benchmark (GTSRB). For complex visual scenarios, BCNN based approach worked very well for traffic sign recognition. Table 5. Summary of multiple deep learning methods for vehicle detection. Ref. Method Outcomes Advantages Limitations 65.52% more 3D vehicle bounding Exceeded the outcomes of Unstable accuracy for certain [34] enhancement in 50 m box mapping LiDAR in long range. error properties. than LiDAR. YOLO V2 and Faster detection Trained with vehicle types [35,36] mAP 94.78% in 0.038 s. K-mean ++ and recognition. of data. Improved result with Performance decreased AdaBoost with pixel 81.94% accuracy in [37] occlusion pattern when 70% overlap and lookup features best-case scenarios. by clustering. heavy occlusion. Highest accuracy in daytime Deep CNN and 99.50% accuracy Not applicable in low-light [38] and trained with AdaBoost in 0.028 s. or complex scenarios. low-quality frames. CNN with 96.47% accuracy in Fast classification in low Higher execution time [39] Caffe framework 51.28 ms. resolution input. compared to other methods. Recall 78.48% and Decreased in performance ANN and Nighttime taillight-based [40] precision 90.78% in for headlight illumination in Kalman Filter detection in urban area. urban scenario. short distance. Improved the Not applicable in poor 90.60% accuracy in [41] SI-NET limitation-scale insensitivity lighting, nighttime or best-case scenarios. of CNN. complex weather scenarios. Enable to detect vehicle in Not applicable in urban or Faster-RCNN 87.2% average [42] large, medium and challenging with AGO precision in 0.055 s. short field. lighting conditions. Highest 84.4% accuracy Unsatisfactory accuracy Faster R-CNN Outperformed Faster R-CNN [43] for car detection in compared with the obtained with RPN in terms of execution time. 81 ms. execution time. 86.81% and 78.91% Applicable in both day and Lightweight precision and recall Was not tested in urban or [44] night scenarios in YOLO Network within 44.5 ms, crowded environment. multiple FoV. respectively. Able to detect in low-light Quicker 84.97% mAP Required huge manual [45] and almost region-based CNN nighttime scenarios. data labelling. no-light conditions. Jung et al. trained the 16 different forms of Korean traffic sign with LeNet-5 CNN architecture for a real-time traffic sign recognition system where the training set had 25,000 positive and 78,000 false samples [47]. The method obtained up to 99.95% within a fast-processing time. The applied color property-based CNN approach could be very effi- cient for lightweight traffic sign detectors for AVS as well as achieving the highest accuracy. An improved traffic sign recognition algorithm was demonstrated by Cao et al. for an intelligent driving system [48]. For accurate detection spatial threshold segmentation, the HSV color space was utilized, and traffic signs were identified accurately depending on shape features and processed with LeNet-5 CNN architecture with Gabor kernel, which was the primary convolutional kernel, and batch normalization was applied after the pooling layer. The Adam model was also implemented as the optimizer algorithm. The proposed methodology was applied to the German Traffic Sign Recognition Benchmark and obtained 99.75% accuracy with 5.4 ms per frame on average, which was higher in both sectors than [189,190], where the accuracies were 98.54% in 22 ms and 95.90 in 5.4 ms, respectively, adopting HoG + PCA and multilayer perceptions methods. Appl. Sci. 2022, 12, 6831 12 of 51 On the other hand, the parallel architecture weighted multi-convolutional neural network took 4.6 ms more to process but still achieved constant high efficiency, scoring 99.75% in GTSDB and 99.59% accuracy in the GTSRB dataset, where low and complex lighted scenarios were also considered. Despite occasional accuracy drops for blurry vision, this method could be one of the most suitable approaches for AVS [49]. To detect traffic signs, Wang et al. suggested a red bitmap extraction and SVM- based method where the detected images were color-segmented and afterwards the shape detection of ROI (ROI) was carried out on the basis of rim detail [50]. The methodology scored recall values of 97% and 99% for danger and prohibitor for the GTSDB dataset, respectively. This technique obtained good detection accuracy, but the major limitation was that this method was only applied to red circular signs. Zhang et al. demonstrated a modified YOLO V2 algorithm to develop an improved Chinese traffic sign detection system, as well as constructing the database [51]. In order to create a single convolutional network, they used several 1  1 convolutional layers for the intermediary and fewer convolutional layers for the top layers. Fine grid was also used to separate images with the goal of identifying small-sized road signs. Their technique was found to be the outcome of the CCTSDB and GTSDB databases, where AUC values for mandatory and danger signs were 96.81%, 94.02% and 96.12% in 0.017 s. Another approach applied CapsNet, which resolved the major limitation of CNN, that is, loss of max pooling layer retaining spatial relations [52]. The approach obtained 98.72% accuracy for recognizing traffic light with shape. It can be a useful method-based approach for AVS’s traffic sign-recognition methods. Furthermore, a unified deep convolutional traffic light-identification feature for au- tomated driving systems was proposed by Bach et al., based on Faster R-CNN that was suitable for detection of traffic lights, and the recognition and classification of types or states [53]. They achieved 92% average precision applying on a large-scale dataset named DriverU traffic light. When the width was greater than 8 px and smaller than these, it scored 93% for average precision. However, there were still limitations for suitable number of false positives which can be reduced by applying RNN or an integrated approach. DeepTLR was the real-time vision-dependent, in-depth and deeply convoluted traffic light-identification and classification system that did not require position details or temporal principles; these were proposed by Weber et al. [54]. On the basis of the single-frame assessment of a challenging collection of urban scenes, the authors presented noteworthy outcomes, showing that in regular images, DeepTLR achieves frame rates of up to 33 Hz. DeepTLR also ran at frame rates of up to 33 Hz. DeepTLR also ran at frame rates of 13 Hz on images with a resolution of 1280  960 pixels. The capacity for more transport lights was high in the architecture, scoring 93.5% F1 score for 1280  960 resolution and 88.1% F1 score for 640  480 in 80 ms and 28.8 ms. Li et al. developed a framework of robust traffic-light recognition with fusion detection in complex scenes [55]. To increase accuracy for each traffic light type and the creation of a fusion detective framework, a set of enhanced methods was adopted based on an optimized channel function (OCF) system after using aspect ratio, field, location and traffic lights background as prior knowledge to minimize computational redundancy and create a task model for the identification of traffic light. Furthermore, they utilized the detection knowledge of the previous system to change the original ideas, which further increased the accuracy. The framework was applied to a VIVA dataset where a combination of multi-size detectors, bulb detectors and fuzzy detectors were implemented, which improved the AUC indicator, with 7.79% for red, 9.87% for red left, 11.57% for green and 3.364% for green left, compared with general ACF on VIVA validation dataset and achieved an AUC indicator of 91.97% for red light and 89.32% for green light on the channel-modified LARA validation dataset. In addition, to reduce complexity, Lee et al. adopted the concept of upper-half clipping frame so that the model could pick only those frames that would allow it to recognize traffic lights rather than taillights [56]. The system was built based on a YOLO algorithm Appl. Sci. 2022, 12, 6831 13 of 51 and obtained 42.03% mAP and 49.1% mAP enhanced, and improved results applied to the Bosch-TL and LISA-TL datasets, but the author did not consider nighttime scenarios. Other than this issue, the method was exceptionally successful for traffic signs and light identification systems for AVS. Behrendt et al. implemented YOLO for 3D localization and tracking of traffic lights [57]. A wide field of view was considered, and the YOLO-based approach was deployed in the Bosch-TL dataset. The method showed 99% accuracy in just 0.06 ms. However, the method required huge pre-labelled data which created an obstacle to fluent performance. In this section, both traffic signs and traffic light detection and recognitions are dis- cussed and summarized in Table 6, for which most of the deep learning approaches were trained with the GTSRB dataset. Among all the deep learning methods, LetNet-5-based CNN on self-made dataset with spatial threshold segmentation with the HSV color space and Gabor filter on the GTSRB dataset performed best for traffic-sign recognition despite reduction in performance in a complex environment and detection of separated signs due to the proposed region. In addition, the YOLO method-based approached obtained highest accuracy in the fastest time for traffic-light detection with recognizing inner signs despite having pre-labelled data dependency. Table 6. Summary of multiple deep learning methods for traffic sign and light recognition. Ref. Method Outcomes Advantages Limitations 98.52% accuracy for Branch-output mechanism Implementation from [46] BCNN low visibility of enhanced recognition speed moving platform was road signs. and accuracy. not tested. Lightweight color segmentation Detected two traffic signs for [47] LetNet-5 based CNN 99.95% accuracy. before classification, improved having different location of processing speed. proposed region. Obtained highest accuracy and Accuracy of 99.75% Performance decreased in [48] CNN with Gabor filter enhanced the performance in 5.4 ms. complicated backgrounds. of LetNet-5. Classification in high-speed Struggled to classify in Highest 99.75% [49] Weighted multi CNN driving and outperformed in challenging blurry accuracy in 10 ms. low-light conditions. vision condition. Faster detection of prohibitory Recall score 97% and Only applied to red [50] SVM and danger signs in poor 99% for two cases. circular signs. illumination. Highest 96.81% AUC Faster and simple Decrease in performance for [51] YOLO V2 values in 0.017 s. processing pipeline. small traffic signs. Resolved loss of max pooling to Did not consider complex [52] CapsNet 98.72% accuracy. boost CNN’s performance. lighting condition 100% recall and 92% Detected traffic light as well as [53] Fast R-CNN Showed high false positive. average precision. indicating signs. Did not require position details [54] DeepTLR Highest 93.5% F1 score. Lower precision rate. or temporal principles. Better detection in complex 11.5% improved AUC Unsatisfied accuracy for red [55] OCF scenes with low luminous value for green light. left traffic light identification. objects combining. 42.03% and 49.16% Took upper half of frame to Did not deploy in nighttime [56] YOLO mAP higher on eliminate search area and traffic scene where reflection two datasets. vehicle taillights. creates confusion. 3D localization and tracking of [57] YOLO 99% in 0.06 s. traffic lights in large field Required more labelling. of view. Appl. Sci. 2022, 12, 6831 14 of 51 3.1.3. Pedestrian Detection Effectively detecting and localizing pedestrians on roads in various scenarios was one of the major vision-based problems for autonomous driving systems. A study shows that only in the USA has the fatality rate for road crossing increased up to 30% in seven years from 2009. In 2016, a total of 6000 pedestrians were killed, which is a record in the last three decades [58]. Moreover, based on vulnerable road users in the ASEN region, 13% of the deaths on roads are related to pedestrians [59]. In order to prevent deaths, the detection and localization of pedestrians have become a major focus of the study of autonomous vehicles. Several studies have been successfully conducted on reducing accident cases and creating a sustainable and more accurate approach to autonomous driving systems. For instance, Angelova et al. proposed a deep network named large-field-of-view (LFOV) to perform complex image processing continuously for pedestrian detection and localization [60]. The purpose of the proposed Large-Field-of-View deep network was to understand, simultaneously and effectively, as well as to make classification decisions in many places. The LFOV network processes vast regions at much higher speeds than traditional deep networks and therefore can re-use calculations implicitly. With 280 ms per image on the GPU and 35.85 on the average miss rate on the Caltech Pedestrian Detection Benchmark, the pedestrian detection system showed a promising performance for real-world deployment. A vision based pedestrian detection and pedestrian behavior classification technique was proposed by Zhan et al. [61], where YOLOv3-TINY was used for quick segmentation and multitarget tracking of detected pedestrians with the DeepSort algorithm [62]. Finally, to identify the behavior of pedestrians, an improved and customized AlexNet algorithm was adopted. The proposed model performed efficiently in real time at a rate of 20 frames per second along with a designed warning area binding each pedestrian. Convolutional neural network is one of the most popular deep learning models and has been adopted in several studies for pedestrian detection. Ghosh et al. used a novel CNN architecture model for pedestrian detection [63]. To train the model, they applied transfer learning as well as synthetic images using an uncovered region proposal of a bounding box to avoid the annotation of pedestrians’ positions. It obtained a 26% missing rate in CUHK08 and a 14% missing rate in the Caltech pedestrian dataset, where crowded scenes were considered. The biggest advantage was that it required no explicit detection while training and did not need any region proposal algorithm. A similar concept was used by Wang et al., who combined part-level fully convolu- tional networks (FCN) and CNN to generate a confidence map and pedestrian location based on the aligned bounded box concept [64]. The proposed framework was compared with CifarNet and achieved 6.83% improved outcomes. A novel single shot detector method based on late-fusion CNN architecture was introduced by Hou to analyze data of a multispectral system that performed with higher accuracy at nighttime [65]. The combined architecture was applied to a KAIST multispectral pedestrian benchmark where the late-fusion CNN architectures worked efficiently. In terms of log average miss rate, it decreased by more than 10% and developed for suitable deployment during both day and nighttime. As a result, it became one of the best practical CNN-based pedestrian detectors of all the accepted AVS methods. For identifying pedestrians in low resolution learning from low-level image features, a single image-based novel resolution-aware CNN-based framework was proposed by Yamada et al. [66]. The authors also developed a multiresolution image pyramid and obtained the original input image to identify pedestrian size. Moreover, it learnt feature extraction from a low-level image with resolution information and achieved 3.3% lower log-average miss rate than CNN which made the architecture more acceptable for AVS. In another work, Zhang et al. implemented an optimized multiclass pedestrian identification system, using a Faster RCNN-based neural network [67]. The analysis indicated that the framework for pedestrian detection in blurred fields of view were able Appl. Sci. 2022, 12, 6831 15 of 51 to increase speed with average precision of 86.6%. This approach could be suitable for distorted images for pedestrian detection tracking. Dong et al. proposed a region proposal framework for pedestrian detection imple- menting R-CNN combined with an ACF model, where the ACF model was applied to produce only pedestrian class-bounding region, which was a very useful application for autonomous vehicle systems [68]. Moreover, the proposed framework cost less execution time during training and testing. Though most of the studies showed pedestrian detection in the daytime or in clear weather, this task becomes more complex in low-light condition, haze or fog because these create vision difficulties and this kind of condition causes a higher number of accidents [69,70] and increases the possibility of traffic accidents by 13% [71]. Correspondingly, de-hazing algorithms were one of the solutions to fix vision problems which can be implemented in detection of pedestrians in haze conditions. For instance, a related approach for pedestrian detection in haze conditions was proposed by Ding et al., implementing a synthesized haze version of the INRIA dataset using dark channel prior-based linear SVM and HOG algorithm [72]. Nonetheless, the approach received poor recall value, scoring 67.88% in predicting constant depths of input images. Although the method is a good approach for pedestrian detection, the limitations could be solved taking into account the pre-trained pedestrians’ depths in multiple haze environments. Moreover, Huang et al. provided a Laplacian distribution model that featured a combined HTE (haze thickness estimation) and IVR (image visibility restoration) for solving problems of pedestrians [73]. Implementation of this algorithm could enhance the performance for detecting pedestrians in haze conditions, but the most difficult haze conditions occur in dim light conditions, where they achieved 4.1 mAP and 3.98 mAP based on expert and ordinary view, respectively. For alleviating the haze problem while detecting pedestrians, Li et al. [74] proposed three approaches named Simple-YOLO, VggPrioriBoxes-YOLO and MNPrioriBoxes for pedestrian detection based on YOLO [75]. Deep separable convolution and linear bottleneck capabilities were implemented to minimize parameters and enhanced processing speed, making the network far more usable. The average precision of their three methods was 78.0%, 80.5% and 80.5%, respectively, where precisions were 89.4%, 90.8% and 89.3%. The lowest 22.2 FPS were 22.2, 81.7 and 151.9 applied to the combined data of their HazePerson dataset and INRIA person dataset after dataset augmentation. Although this approach was one of the preferable approaches for AVS to detect and localize pedestrians in day and night haze conditions, the higher missing rate in complex scenarios was an issue which could be resolved by adopting key point detection methods. Xu et al. proposed a ground plane context aggregation network (GPCANet) for detecting pedestrians in ground plane areas of Caltech, SCUT and EuroCity datasets, where the best result was achieved for the SCUT dataset with 96.2% recall value, and obtained 25.29% and 25.10% log average miss rate for the rest of the dataset serially [76]. However, it might have slightly higher log average miss rate, but the outcomes were in crowded traffic complex scenarios, which made the approach more practicable for AVS. Moreover, CNN-based work was demonstrated with only 5.5% miss rate to localize distracting pedestrians [77]. Similarly, CNN cascaded with AdaBoost was deployed for pedestrians in night images [78]. It obtained a maximum 9% log-average miss rate, although both methods are not evaluated in complex scenarios. In summary, multiple deep learning methods were reviewed (shown in Table 7) where a CNN-based method was deployed for faster pedestrian detection and localization where the methods showed 94.5% success rate and provided an improved dataset built on the Caltech dataset. FCN achieved 6.83% improved outcomes compared with CifarNet, while in terms of estimating distance of pedestrians from the vehicle, it showed a higher missing rate. Moreover, GPCANet performed best on the SCUT dataset, scoring 96.2% recall in 320 ms and deployed in diverse scenarios in both day and night conditions. However, it scored a high missing rate and could not deal with complex scenes in terms of occluded road objects. However, when of the other methods showed stable efficient outcomes, Appl. Sci. 2022, 12, 6831 16 of 51 the main challenges remained for crowded traffic scenes and complicated visual and weather conditions. Table 7. Summary of multiple deep learning methods for pedestrian detection. Ref. Method Outcomes Advantages Limitations Processed in large field images, Higher missing rate with 35.85 on the average [60] DNN continuous detection in comparatively miss rate. complex scenes. slow detection Designed faster warning area Only considered daytime YOLOv3-TINY and 80.3% accuracy in [61] bounding by each pedestrian scenarios and DeepSort complex environment. with direction labelling. lower accuracy. 26% and 14% missing Did not apply for motion Did not require explicit [63] CNN rates on two images or detection in crowded scenario. datasets, respectively. real-time problems. 6.83% improved Estimated accurate distances Part-level FCN High missing rate for [64] outcomes compared of pedestrians generating the and CNN practical implementation. with CifarNet. confidence map using FCN. Decreased by more than SSD-based Most applicable in Slower detection and [65] 10% of log average late-fusion CNN nighttime implementation. complex parameter tuning. miss rate. Learnt feature extraction from Was not applied in complex Resolution 3.3% lower log-average [66] low-level image with traffic or aware-based CNN miss rate than CNN. resolution information. crowded environment. 86.6% average precision Outperformed on distorted Did not consider low-light or [67] Faster R-CNN in 0.104 s. and blurry frames. traffic scenarios. 14.1%, 15.3%, 45.6% R-CNN with Reduced the number of region [68] miss-rate on —- ACF model proposals, and costs less time. three datasets. Pedestrian detection and Dark 81.63% precision and Presumed constant depths in [72] position estimation from channel-based SVM 67.88% recall. input images. haze condition. 4.1 mAP and 3.98 mAP Laplacian Pedestrian detection in Was not applied in real-time [73] on expert and ordinary distribution model complex dim-light condition. driving scenes. view, respectively. Highest average Minimized number of Early dependency on Multiple [75] precision 80.5% parameters and outperformed preliminary boxes during YOLO methods in 81.7 ms. state-of-art methods. detection process. Improved outcomes in both Higher log average missing 96.2% recall in 320 ms on [76] GPCANet day and night including far rate for occluded SCUT dataset. FoV and crowded traffic. on-road objects. Localized distracting Did not test in cases for [77] CNN Showed 5.5% miss rate. pedestrian and improved crowded or complex scenes. detection annotations. CNN cascaded Generated the maximum Combined thermal images for Might fail in complex urban [78] with AdaBoost 9% log-average miss rate. nighttime detection. traffic scenarios. 3.1.4. Lane Detection and Tracking One of the core fundamentals for AVS was to identify lane and tracking curves in real time where the controlling would depend on the lane and curves. Several studies have been conducted on this field based on different camera visions implementing deep learning and computer vision approaches considering color, texture, feature extraction in different scenarios for lane detection, lane shifting, lane keeping and overtaking assisting. Appl. Sci. 2022, 12, 6831 17 of 51 A road scene sectioning framework was adopted by Alvarez et al. using a CNN-based algorithm to retrieve the 3D scene layout of the street image from noisy labels combining online and offline learning [79]. The proposed method built with color plane fusion and CNN was able to achieve 95.5% to extract a single image of a lane without manual labelling. This CNN-based approach could be considered as the most efficient method for deploying in unknown environments for AVS road-feature extraction. However, for each pixel of the image, which was a path or a lane, authors Dong et al. considered the visual road-detection challenge applying a U-Net-prior network with the DAM (Domain Adaptation Model) to reduce the disparity between the training images and the test image [80]. The proposed model was compared to other state-of-art methods such as RBNet [191], StixeNet II and MultiNet [192], where the max-F measures were 94.97%, 94.88% and 94.88%, respectively, in 0.18 s, 1.2 s and 1.7 s. Their methodology obtained 95.57% max F-measurement in 0.15 s faster and more accurately than others, which indicates that their monocular-vision-based systems achieve high precision for a lower running time. Another kind of approach for storing processes of previous stages, a method based on a combination of CNN and recurrent neural network (RNN), was proposed by Li et al., which was able to identify lane markers using geometry feedback with maximum 99% AUC value [81]. However, since no image pre-processing was conducted, this process took a lot of time in sorting unrelated image areas. In addition, these methods were either time-consuming or inefficient in a true, dynamic world, which does not fulfil the maximum efficiency restriction of a critical function. The Bayesian method for estimating multihyperbola parameters splitting frames in multiple patches was demonstrated by Fakhfakh et al. to recognize curved lanes under difficult conditions using [82]. The lane line was represented on each component by a hyperbola which was determined using the proposed Bayesian hierarchical model with an average of 91.83% true positive rate (TPR) on the ROMA dataset. To sum up the theory, it could be made more practical by adopting sampling techniques such as Hamiltonian schemes to enhance the model’s performance. Yang et al. suggested a substitution of image pre-processing to reduce the uncertainty about lane state [83]. Their approach uses profound lane detection based on deep learning as a substitute for practical lane detection with UNet encoder including high-grade GPU processing. The paper also states that the CNN-based UNet with Progressive Probabilistic Hough Transformation, UNet, Kalman filter were far more inefficient in terms of identifi- cation than the feature-based approaches, such as Hough Transformation (HOG) for lane tracking in real time [84–86]. For predicting lane line under the most challenging conditions, a spatiotemporal- based hybrid architecture after encoding–decoding SCNN and ConvLSTM [87]. This is the very first approach which improves temporal correlation with spatial relation of feature extraction with 98.19% accuracy and 91.8% F1 score. However, although this is one of the strongest approaches, the authors did not apply it to complex weather and nighttime scenarios. Furthermore, to resolve instance level and complex fork and dense line-detection issue, a novel approach was implemented, CondLaneNet, using recurrent instance module applied to a CULane dataset [88]. The approach obtained an 86.10% F1 score while detecting curve lane in complex scenarios despite the lack of proper refining of contextual features. Multiple deep learning methods were studied regarding the lane curve tracking system. For instance, Dorj et al. deployed circle equation models and parabola equations to redesign the Kalman filter for curved lane tracking with a view to calculating curving parameters in far field view [89]. Although the algorithm had an independent threshold mechanism to compensate for various light conditions, such as low light, further research was needed to identify lane reflections and shadows. The limitation of Dorj et al. was solved in [90], where the authors applied a local adaptive threshold and RANSAC feedback algorithm to prevent misdetection of the lane by estimating two-lane parameter-based issues. Nevertheless, the algorithm did not allow a close-loop lane to maintain lane control Appl. Sci. 2022, 12, 6831 18 of 51 while following the road lane, showing a slightly higher false positive rate (FPR) and slow execution for processing in the CPU only. However, it achieved 99.9% precision, 98.9% accuracy and 99.4% F-measurement in 0.677 fps complex visual and lighting conditions. Similarly, for overcoming lane detection in complex shadow and lighting conditions full of obstacles, a CNN-based method was presented by Wang et al. [91]. From an inverse perspective, the application of a fixed transformation matrix generated errors as changes occurred, allowing the predicted exhaust point to infinitely shift upward or downward. The authors trained a neural network with a custom loss function that predicted the transformable matrix parameter valued dynamically. The method was implemented on the TuSimple dataset and obtained high accuracy for insufficient light, shadow, missing lane and normal road compared to other deep learning methods, such as Spatial CNN, CNN-FCN and UNet. As an approach to preventing lighting condition problems for lane detection and tracking, a novel CNN-based model was proposed by Ye et al. [92]. In the pre-processing stage they adopted Yam angle prediction and filtering, followed by segmenting ROIs implementing waveform generation that generated on average 99.25% accuracy in the BIT dataset considering nine cases where day and nighttime accuracies were 99.34% and 98.66%, respectively, as well as a 1.78% average error rate for the Caltech dataset. However, this methodology could be the most suitable candidate only if it is performed with similar outcomes in real-life experiments of level 4 or 5 AVS. A similar CNN-based approach combined with CNN-LSTM, SegNet and UNet was applied by Zou et al. for lane detection from occlusion scenarios [93]. The method obtained 96.78% accuracy for SegNet and 96.46% for UNet within 46 ms, which was much faster than the average of the other methods. With faster processing and high accuracy, this approach could be considered as one of the most acceptable methods for AVS lane detection. Jhon et al. proposed a lane-detection algorithm, which calculated the semantic road lane by using the extra tree-based decision forest and DNN from a street scene where hue, saturation, depth (HSD) combined with a deconvolutional network were fine-tuned [94]. In the final stage, a separate extra tree regressor was trained within each lane applying the depths and the manually annotated lane marker locations on the image. The methodology was applied to the TTI and TMD datasets, where it achieved 98.80% and 97.45% accuracy, respectively, for lane detection. Further, encoder–decoder dilated convolution and finely tuned improvements were implemented by Chen et al. to create a modified CNN road lane detection system called Lane Mark Detector (LMD) which increased the accuracy of the CamVid dataset to 65.2%, obtained 79.6% class average accuracy and increased the test speed to 34.4 fps as well as improved the inference time (29.1 ms) and smaller model size of 66 mb [95]. Moreover, Ghafoorian et al. used Embedding Loss-Driven Generative Adversarial Networks (EL-GAN) for detecting road lanes [96]. This led to even more secure training with stronger discrimination and stabilized the mechanism of adverse preparation. This significantly stabilized the process of opposing training. EL-GAN was also applied to the TuSimple dataset and achieved 96.39% accuracy despite requiring the tuning of a suitable number of parameters. As the loss of embedding into classification boosted the maximum efficiency of the lane marking method, it was one of the best and most appropriate approaches for continuous lane detection and tracking. Tracking lane during nighttime was one of the most difficult tasks of AVS. He et al. solved the issue by developing a vision-based lane detection system, where they pre- processed with a Gabor filter, continuing adaptive splay ROI and Hough transformation to detect the lane marker [97]. Despite lacking an appropriate self-switching system for defining lanes in all circumstances in pre-processing, the detection rates were 97.31% and 98.15% using two clips of Guangzhou where frame numbers were 3274 and 2231. However, the method faced difficulties when tackling bright light reflection, missing lane marks and lane cracks as well. Appl. Sci. 2022, 12, 6831 19 of 51 Neven et al. formulated a solution using LaneNet and HNet for the problem of lane detection with an instance segmentation problem in which each lane constituted its own instance to be end-to-end trained [98]. In addition, to a set “bird’s-eye view”, they introduced a learning transfer to the perspective, which was contingent on the image and achieved 96.4% accuracy within 50 fps (frames per second) for the TuSimple dataset. The method was robust enough to adjust the pitch of the ground plane by adapting the transition parameters accordingly, which was the main reason for accurate visualization, and detected lane and lane curves. Moreover, Kim et al. proposed a fast-learning environment using extreme learning CNN (EL-CNN) combining extreme learning machine (ELM) calculating weights among output and hidden layers in one iteration with CNN for lane marking extraction in complex scenarios to overcome computing of large dataset [99]. It reduced training time 1/50 for the KNU dataset, and 1/200 for the Caltech dataset compared to CNN. Experimental results demonstrate that it obtained maximum weights effectively while maintaining performance of 98.9% accuracy applied to the Caltech dataset. In another work, Van Gansbeke et al. implemented ERFNet with differentiable least- squares fitting (DLSF) for end-to-end lane detection [100]. The approach used dynamic backpropagation to perform an experiment on a lane detection task that demonstrated that, despite the poor supervision signal, the end-to-end approach exceeded a two-step procedure, scoring 95.80% accuracy in 70 fps applied to the TuSimple dataset. The accuracy was not the maximum, but the weight map did not require post processing for accurate lane estimation. Hou et al. proposed a lane detection CNN by self-attention distillation (SAD) which had self-learning ability in the training phase and boosted the visual attention of multiple layers in different networks and increased the efficiency of narrow-lane detection sys- tems [101]. The method obtained 96.64% accuracy in the CULane, BDD100 K and TuSimple datasets, although the hyperparameter adjustment was complicated by an insufficient training process and loss functions. In another work, Liu et al. used a lightweight YOLO network for lane curve detection and tracking with 90.32% precision and 83.76% recall in 50.6 ms [44]. The method was applied to a custom dataset which was evaluated for day and night scenarios. However, the efficiency could be better suited to proper AVS if it solved the interruption of vehicles during lane detection. In conclusion, most of the approaches have performed well enough to be adopted for practical implementation of AVS. However, modified CNN [92] was able to detect lanes with highest accuracy for both day and nighttime, and the CNN-LSTM-based SegNet and UNet combined approach was [93] able to segment roads within the fastest runtime. The analysis presented some advantages of deep learning methods for lane and road curve detection, for instance, training without manual labelling, reducing computational complexing while in a single frame, lane detection where markers were not clear, in sharp turns and even challenging weather and shadow or low-light conditions. On the other hand, some methods showed huge dependency on dataset pre-labelling, which was inefficient in the long field of view, resource hunger and even not being evaluated in urban traffic scenarios or challenging road conditions. An overview of the deep learning methods reviewed for the detection of lane and road curves is shown in Table 8. 3.1.5. Traffic Scene Analysis Driving scene and driving behavior analysis of autonomous vehicle systems were denoted as the understanding and classifying of driving environment and traffic scene. To discuss the contribution of deep learning to understanding and analyzing complex traffic scenes, several studies were conducted. Appl. Sci. 2022, 12, 6831 20 of 51 Table 8. Summary of multiple deep learning methods for lane detection and tracking. Ref. Method Outcomes Advantages Limitations Reduced dependency of manual Deployment for testing did not 95.5% accuracy for [79] Modified CNN labelling and processing time in consider urban or single frame. single frame. crowded scenarios. 95.57% max Smooth segmentation of road surface Huge dependency on manual [80] U-Net F-measurement in 0.15 s. with multiple objects as obstacles. pre-labelling. Recognized region in complex traffic Higher computational cost and [81] Multitask CNN and RNN Max AUC value of 99%. and visualized spatially inefficient for large field of view. distributed cues. Automated curve detection in rural Lighting conditions were not [82] Bayesian Model 91.83% true positive rate. and challenging roads with lower considered and slow processing. error rate. 2.5% and 9.75% lateral Obtained less lateral error and Limited to a simple close-loop [83] UNet and Kalman Filter error generated in 10 ms. overcame slow feature extraction. circle in TORCS simulator. Solved the lane detection in fork and Contextual features need to [88] CondLaneNet 86.10% F1 score. dense scenarios. be refined. Prevent misdetection by estimating Slower execution, slightly high 99.9% precision and 99.4% [90] RANSAC parameters when FPR and did not consider urban F-measurement. illumination changes. traffic road. Created errors during shifting the Obtained highest accuracy Outperformed in shadow and roads [91] CNN ground to predict (97.85%). with obstacles. disappearing point Most suitable deployment in two Did not test in real-time [92] Modified CNN Average 99.25% accuracy. datasets with faster runtime for driving senses. nine classes Combined method of CNN and Raised performance in [93] CNN-LSTM with SegNet 96.78% accuracy in 46 ms. RNN were resource hungry occlusion scenarios. and slow. Modified Decision forest 98.80% and 97.45% High accuracy in surface and [94] High computational cost. and DNN accuracy, respectively. road detection. Modified CNN to achieve low 65.2% mIoU, 79.6% class Was not tested in crowded [95] CNN complexity and maintained average accuracy. traffic environment. similar accuracy. The loss of embedding in detector Required tuning of huge number [96] EL-GAN 96.39% accuracy. boosted the performance closest to of parameters. the mark. Did not require post processing, 96.4% accuracy within 19 Faced challenges in long field of [98] LaneNet and HNet pixel-wise segmentation or fix ms. view while detecting curves. lane number. Reduced training time 1/50 and 1/200 Required matrix inversion for [99] EL-CNN 98.9% accuracy. in KNU and Caltech datasets, better execution time in high respectively. dimensional data. Did not need post processing to Was not tested in urban or [100] ERFNet-DLSF 95.80% accuracy. estimate line coordinates using complex lighting conditions. weight map. Complex hyperparameter Has self-learning ability and increased [101] CNN with SAD 96.64% accuracy. adjustment for inadequate the efficiency of narrow-lane detection. training process. Lightweight 90.32% precision and Applicable in both day and night High interruption for [44] YOLO Network 83.76% recall in 50.6 ms. scenarios and multiple fields of view. obscured vehicles. To contribute to this field for developing traffic scene analysis for AVS, Geiger et al. proposed a novel method of generative probabilism to understand traffic scenes with the Markov Chain Monte Carlo, which was used to deal with the dynamic relationship between crossroads and feature presentation [102]. The human-inspired method took the benefit from a wide range of visual cues through the form of vehicle directions, vanishing points, semantic scene labels, scenario flow and grids rather than requiring sensor values Appl. Sci. 2022, 12, 6831 21 of 51 such as LiDAR and GPS, where most of the standard methods struggled for most of the intersections due to the lack of these attribute labels. Moreover, the method can accurately identify urban intersections with up to 90% accuracy at 113 real-world intersections. Another scene semantic segmentation approach is the High-Resolution Network (HRNet) proposed by Wang et al. [103], where the method obtained 81.1% mIoU. HRNet linked the high-to-low resolution convolution streams in parallel and transferred data across repeatedly. The advantage of the method was that the resulting representation was richer semantically and spatially. However, it required huge memory size due to high resolution-wise segmentation. Additionally, the same author improved their previous work applying contrastive loss to previous architecture (HRNet), which explored pairwise pixel-to-pixel dependencies applied to the Cityscape dataset and obtained 1.1% higher mIoU [104]. Although the proposed method demonstrated effective performance, which is applicable for top-tier AVS, it was unable to achieve success during contrastive learning in few parts of the labelled dataset. To tackle this issue, Zhao et al. [105] presented a contrastive approach following previous research [103,104] and proposing SoftMax tuning rather than applying contrastive loss and cross-entropy at once. The authors demonstrated three variants of label and pixel-wise contrastive losses by adopting DeepLabV3 with ResNet-50 with 256 channels of convolution layers and bilinear resizing for input resolution for semantic segmentation. This approach showed 79% and 74.6 mIoU, respectively, for Cityscape and PASCAL VOC 2012 datasets but using 50% less labelled dataset. Thus, powerful semantic segmentation with a fine-tuned pretrained method can be a major pathway for higher level AVS for scene analysis. Furthermore, to develop a scene recognition framework, Tang et al. demonstrated GoogleNet for multi-stage feature fusion, named G-MS2F, segmented into three layers to feature extractions and scoring scene understanding, that can be efficiently employed for autonomous driving systems [106]. The framework obtained 92.90%, 79.63% and 64.06% accuracy, respectively, when applied to the Scenel5, MIT67 and SUN397 datasets for image scene recognition. Similarly, a multiresolution convolutional neural network architecture was proposed by Wang et al. for driving scene understanding in different scales where they used two categories of resolution images in the input layer [107]. A combination of fine-resolution CNN and coarse-resolution CNN was included for recording small and comparatively large-scale visual frameworks. To obtain visual information with more accurate resolution and enhanced spatial information, on an inception layer, three convolutional layers were added. They implemented the architecture on the Place365 dataset where the lowest error rate was 13.2%. Moreover, a 2D-LSTM model was proposed to learn information from surrounding context data of scene labels as well as spatial dependencies in [108] within a single model that generated each image’s class probabilities. They obtained 78.52% accuracy when deploying on the Standford background dataset. Fu et al. introduced an integrated channel contextual framework and spatial contextual framework as a contextual deconvolution network (CDN) that used both local and global features [109]. In an attempt to optimize the visualization of the semantic data, the decoder network utilized hierarchical supervision for multilevel feature maps in the Cityscapes dataset and achieved 80.5% mean IoU. Following the work, an optimized model of a deep neural network was proposed with two distinct output directions by Oeljeklaus et al. Their method foresaw road topology along with pixel-dense categorization of images at the same time, and lower computing costs were offered in real-time autonomous applications via a proposed architecture com- bined with a novel Hadamard layer with element-wise weights using Caffe and achieved 0.65 F1, 0.67 precision and recall 0.64 after fine-tuning the architecture with 10,000 it- erations [110]. Although strong restrictions placed by the double-loss function on the DNN feature maps caused difficulties in optimizing the process, research in relation to the Appl. Sci. 2022, 12, 6831 22 of 51 Cityscapes dataset showed that a sufficient representation of traffic scene understanding was achieved relying on broad traffic components. In another work, Xue et al. presented a CNN with Overlapping Pyramid Pooling (OPP) applied to sematic segmentation of city traffic area based on a fisheye camera with wider vision [111]. The OPP was demonstrated for the exploratory study of the local, global and pyramidal local context information to resolve the complicated scenario in the fisheye image. Furthermore, they built novel zoom augmentation for augmenting fisheye images to boost performance of the method where it scored 54.5 mIoU, which is higher than the standard OPP-Net and Dilation10 method. This approach could be highly suitable for short FoV traffic scene understanding in urban areas. Pan et al. proposed Spatial CNN, a CNN-like framework for efficient spatial distribu- tion of information through slice-by-slice message passing from the top hidden layer [112]. It was tested at two roles: lane recognition and traffic scene perception. The analysis showed that the continuity of the long, small structure was appropriately preserved by SCNN, while its diffusion effects have proven positive for large objects in semantic segmen- tation. However, SCNN can master the spatial relationship for the structural production and increase operating efficiency, showing that SCNN was 8.7% and 4.6% superior to the recurrent neural network (RNN) focused on ReNet and MRF + CNN (MRFNet). It scored 68.2 mIoU for semantic segmentation and achieved 96.53% on the TuSimple Benchmark Lane Detection Challenge combined with traffic scene analysis. Mou et al. proposed a vision-based vehicle behavior prediction system by incorporat- ing vehicle behavior structural information into the learning process, obtaining a discrete numerical label from the detected vehicle [113]. The OPDNN (overfitting-preventing DNN) was constructed using the structured label as final prediction architecture, and after more than 7000 iterations, 44.18% more accuracy on-road vehicle action than CNN was achieved. In addition, the method decreased the issue of overfitting in a small-scale training set and was highly efficient for analysis of on-road vehicle behavior predicting turning angles. In another work, Jeon et al. proposed a model built on the CNN and Long Short- Term Memory (LSTM) networks to predict risk of accidents and minimize accidents and analyzing traffic scenes differing conditions of driving such as lane merging, tollgate and unsigned intersections [114]. They implemented a multi-channel occupancy Grid Map (OGM) as a bird’s-eye view that ostensibly included the features of many interaction groups to represent the traffic scene [85]. Additionally, the CNN was used to derive numerous inter-vehicle interactions from the grid and to estimate possible time-serial predictions of the derived functions. For instance, Lui et al. demonstrated a deep understanding of the vehicle-specific scene understanding state-of-art in terms of using traffic environment as object joining automatic scene segmentation and object detection, which reduced the person manipulation [55]. A SegNet network whose weight was initialized by the VGG19 network was used for semantic segmentation on the Auckland traffic dataset [115]. Afterwards, a Faster RCNN-based approach transformed feature maps in the ROI (ROI) and transferred those to the classification mode. It had an accuracy of 91% for sky detect, 90% for bus lane, 86% for road, 70% for lane and 81% building classes applying VGG19-SegNet. However, it suffered from false rate for not having a high-resolution labelled dataset and a weak vehicle detection process. Furthermore, two state-of-the-art versions of machine learning and deep learning (DNN) were used by Theofilatos et al. to estimate the incidence of a crash in real time where the dataset comprised historical accident information and combined current traffic and weather information from Attica Tollway, Greece [116]. The method achieved accuracy, precision, recall and AUC of 68.95%, 52.1%, 77% and 64.1%, respectively. The limitation was the transferability while returning the parameters and the absence of good interplay during comparison and insufficiently clarified unexpected heterogeneity. The possible solution offered by the authors was to apply a sensitivity analysis that was not used when applying a binary logistic model in their work to determine risk of crashes. Appl. Sci. 2022, 12, 6831 23 of 51 Moreover, Huegle et al. proposed a Graph-Q and DeepScene-Q off-policy reinforce- ment learning-based approach for traffic scene analysis and understanding applied to a custom dataset [117]. The proposed method used dynamic awareness-based scene under- standing for AVS, although it was not tested in a real driving environment and was unable to track lanes while moving quickly. With a view to understanding hazardous or damaged roads in a driving situation for a smooth autonomous driving experience, deep learning approaches can also provide a solution. Nguyen et al. used CNN architecture to identify damages and cracks in the road that reduced false detection and without pre-processing, which helped to decrease compu- tational time [118]. On the other hand, authors adopted a principal component analysis (PCA) method and CNN to classify and sense damaged roads with their own dataset. Another Deep CNN-based approach with discriminative features for understanding road crack identification was developed by Zhang et al., which also could be a pathway to implement in AVS [119]. The core advantage of the framework was self-learning features that did not rely on manual labelling and geometrical pavement predictions. An alternative method for autonomous road cracks alongside pothole detection was demonstrated by Anand et al. as part of an analysis of traffic scene [120]. SegNet was applied with texture that relied on features to separate roads from traffic scene in order to build a first division of mask and concatenated with the second masking, which was created with a 2-canny edge algorithm and dilation. Further, SqueezeNet was applied to the GAPs dataset along with being prepared for deployment in a self-driving vehicle. Compared with a similar approach of Zhang [119], it achieved higher precision, recall and F1 score, leaving one drawback where it failed to recognize cracked road that was misinterpreted as under-construction surface texture. For this outcome, the method of Anand et al. [120] was a more suitable approach for identifying damage road surface. In summary, deep learning approaches such as fine-resolution CNN and coarse- resolution CNN, 2D-LSTM model RNN, HRNet, Deep CNN, Contextual Deconvolution Network, DNN and CNN with pyramid pooling were analyzed which demonstrated high-accuracy traffic-scene understanding from a crowded movable platform, showed less model complexity, being applicable in different scales, avoiding confusion of ambiguous labels by increasing the contrast among pixels, in some cases developing more expressive spatial features and predicting risk of accident. However, some approaches were limited for implementation because of the requirement of re-weighting, which was inapplicable in uncertain environments, slower computational time, low accuracy and inability to focus on objects in dim light and foggy vision. The overall summary is presented in Table 9. 3.2. Decision Making As the world economy and technology have grown and developed, vehicular own- ership has increased rapidly, along with over one million traffic incidents worldwide per year. Statistics indicate that 89.8% of incidents took place because of wrong driver decision- making [193]. To solve this issue with the concept of AVS, the decision-making process was one of the key fields for studying a combined deep learning and deep reinforcement learning-based approach to take humanlike driving decisions when accelerating and de- celerating, lane shifting, overtaking and emergency braking, collision avoidance, vehicle behavior analysis and safety assessment. For instance, the automated driving coordination problem was defined as a problem of the Markov Decision Process (MDP) in the research of Yu et al., during the simulation of vehicle interactions applying multi-agent reinforcement learning (MARL) with a dynamic coordination graph to follow lead vehicles or overtaking in certain driving scenarios [121]. The advantage of the method was when most of the study focused on single vehicle policy, the proposed mechanism resolved the limitation of coordination problem in autonomous driving during overtaking and lane-shifting maneuvers, obtaining higher rewards than rule-based approaches. Appl. Sci. 2022, 12, 6831 24 of 51 Table 9. Summary of multiple deep learning methods for traffic scene analysis. Ref. Method Outcomes Advantages Limitations Efficient in specified scene Highest 91% Showed false rate for not having [55] VGG-19 SegNet understanding, reducing the classification accuracy. high-resolution labelled dataset. person manipulation. Identified intersections from Independent tractlets caused Markov Chain Identify intersections with [102] challenging and crowded unpredictable collision in Monte Carlo 90% accuracy. urban scenario. complex scenarios. Able to perform semantic [103] HRNet 81.1% mIoU. Required huge memory size. segmentation with high resolution. Did not show success of Contrastive loss with pixel-to-pixel [104] HRNet + contrastive loss 82.2% mIoU. contrastive learning in limited dependencies enhanced performance. data-labelled cases. DeepLabV3 79% mIoU with 50% less Reduce dependency on huge labelled [105] Dependency on labelled dataset. and ResNet-50 labelled dataset. data with softmax fine-tuning. Less model complexity and three times Did not demonstrate for [106] Multistage Deep CNN Highest 92.90% accuracy. less time complexity than GoogleNet. challenging scenes. Fine- and Multilabel classification from scene [107] 13.2% error rate. Applicable at different scale. coarse-resolution CNN was missing. Able to avoid the confusion of Suffered scene segmentation in [108] 2D-LSTM with RNN 78.52% accuracy. ambiguous labels by increasing foggy vision. the contrast. Fixed image semantic information and Achieved 80.5% Unable to focus on each object in [109] CDN outperformed expressive mean IoU. low-resolution images. spatial feature. Foresaw road topology with Restrictions by the double-loss DNN with 0.65 F1 score, 0.67 [110] pixel-dense categorization and less function caused difficulties in Hadamard layer precision and 0.64 recall. computing cost. optimizing the process. CNN with Developed novel image augmentation [111] Scored 54.5 mIoU. Not applicable for far field of view. pyramid pooling technique from fisheye images. Performance dropped significantly 96.53% accuracy and Re-architected CNN for long [112] Spatial CNN during low-light and 68.2% mIoU. continuous road and traffic scenarios. rainy scenarios. Required re-weighting for 91.1% accuracy after Decreased the issue of overfitting in [113] OP-DNN improved result but inapplicable 7000 iterations. small-scale training set. in uncertain environment. Slower computational time and Predict risk of accidents lane merging, [114] CNN and LSTM 90% accuracy in 3 s. tested in similar kinds of tollgate and unsigned intersections. traffic scenes. 68.95% accuracy and Determined risk of class from Sensitivity analysis was not used [116] DNN 77% recall. traffic scene. for crack detection. Developed dynamic Graph-Q and Obtained p-value Unable to see fast lane result and [117] interaction-aware-based scene DeepScene-Q of 0.0011. slow performance of agent. understanding for AVS. High accuracy for Identified damages and cracks in the Required manual labelling which [118] PCA with CNN transverse classification. road, without pre-processing. was time consuming. 92.51%, 89.65% recall and Automatic learning feature and tested Had not performed in real-time [119] CNN F1 score, respectively. in complex background. driving environment. Highest accuracy (98.93%) Identified potholes with Failed cases due to confusing with [120] SegNet and SqueezedNet in GAPs dataset. texture-reliant approach. texture of the restoration patches. In another work, the Driving Decision-Making Mechanism (DDM) was built by Zhang et al., using an SVM algorithm, optimized with the weighted hybrid kernel function and a Particle Swarm Optimization algorithm to solve decision-making issues including free driving, tracking car and lane changing [122]. The proposed decision-making mechanism obtained 92% accuracy optimizing an SVM model compared with RBF kernel and BPNN model, where the evaluated performance shows that free driving achieved 93.1% and tracking car and lane changing achieved 94.7% and 89.1% accuracy, respectively, in different Appl. Sci. 2022, 12, 6831 25 of 51 traffic environments within 4 ms for average reasoning time. The authors presented a hypothesis when analyzing the results: for driving decisions, road conditions have nearly no effect on heavy traffic density. Despite achieving good accuracy, some limitations were mentioned, such as not applying to real-world driving environments and not yet investigating critical driving scenes such as sudden presence of pedestrians or objects. This issue of [122], was solved by Fu et al., who proposed autonomous braking, analyzing a lane-changing behavior decision-making system for emergency situations, implementing an actor-critic-based DRL (AC-DRL) with deep deterministic policy gradient (DDPG) and setting up a multi-object reward function [123,124], obtaining 1.43% collision rate. The authors mentioned that using a large training dataset online can be tough and expensive, and the continuous action function decreased the convergence rate and can quickly be lowered to the maximum local. Moreover, to overcome the limitation of reinforcement learning in complex urban areas, Chen et al. used model-free deep reinforcement learning approaches named Double Deep Q-Network (DDQN), Twin Delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) to obtain low dimensional latent states with visual encoding [125]. They improved performance by implementing a CARLA simulator by altering frame dropping, exploring strategies and using a modified reward and network design. The method was evaluated in one of the most complicated tasks, a busy roundabout, and obtained improved performance compared to baseline. In the 50 min test, the three approaches were able to enter with high success rate but performance of DDQN and TD3 decreased after covering a long distance. In the best case, SAC achieved 86%, 80%, 74%, 64%, 58% success rate for first, second, third, desired exits and goal point, respectively, where DDQN and TD3 had an almost zero success rate for desired exit and goal point arriving. To avoid training complexity in a simulation environment, the DDPG algorithm with actor-critic method was applied in [124] using deep reinforcement learning (DRL), consid- ering three reward function braking scenarios: braking too early and too late, and too-quick braking deceleration. The outcomes of their proposed methodology showed that the error collision rate was 1.43% which was gained by evaluating the performance of the diverse initial positions and initial speed strategies. The ratio of obtaining maximum deceleration was 5.98% and exceeding jerk was 9.21%, which were much improved compared to DDPG with steering and DQN with discrete deceleration. A dueling deep Q-network approach was demonstrated by Liao et al. to make a strategy of highway decision making [126]. The method was built for lane-changing decisions to make a strategy for AVS on highways where the lateral and longitudinal motions of the host and surrounding vehicles were manipulated by a hierarchical control system. The outcomes showed that after 1300, 1700, 1950 episodes, the approach was able to avoid collision after 6 h of training and 26.56 s of testing. In another study, Hoel et al. introduced a tactical framework for a decision-making process of AVS combining planning with a DRL-extended Alpha Go algorithm [127]. The planning phase was carried out with a modification in the Monte Carlo Tree Search, which builds a random sampling search tree and obtained a 70% success rate in highway cases. The contrast between traditional MCTS and the variant in this search was that a neural network formed through DRL aimed towards the search tree’s most major aspects and decreased the essential sample size and helped to identify long temporal correlations with the MCTS portion. However, the proposed process considered 20 simulation parameters and 11 inputs to a neural network which were very efficient and made more suitable for practical implementation. Overtaking maneuvers for intelligent decision making while applying a mixed observ- able Markov decision process was introduced by Sezer, solving overtaking maneuvers on two-track roads [128]. In this paper, the author presented a new formulation for the issue of double-way overtaking by the resources of the mixed observability MDP (MOMDP) to identify the best strategy considering uncertainties. This was used for overcoming the problem, and was illustrated by the active solvers’ growth and in cognitive technological Appl. Sci. 2022, 12, 6831 26 of 51 advances by reducing time-to-collision (TTC) methods in different simulations. The method surpassed nine periods, relative to both MDP and conventional TTC methods. However, the limitation of proper discretion can also be considered with respect to the actual speed and distance values. A higher number of states that were specifically connected for computing and MOMDP algorithm tend to be required as the actual implementation hindrance. To overcome the issue of vehicle overtaking which needs an agent to resolve several requirements in a wide variety of ways, a multigoal reinforcement learning (MGRL)-based framework was introduced to tackle this issue by Ngai et al. [129]. A good range of cases of overtaking were simulated to demonstrate the feasibility of the suggested approach. When evaluating seven different targets, either Q-Learning or Double Action QL was being used with a fusion function to assess individual decisions depending upon the interaction of the other vehicle with the agent. The hypothesis of the work was that this proposal was very efficient at taking accurate decisions while overtaking, collision avoiding, arriving on target timely, maintaining steady speed and steering angle. Brännström et al. presented a collision-avoiding decision-making system adopting a Bayesian network-based probabilistic framework [130]. A driver model enabled the devel- oper to carry out early actions in many circumstances in which the driver finds it impossible to anticipate the potential direction of other road users. Furthermore, both calculation and prediction uncertainties were formally discussed in the theoretical framework, both when evaluating driver adoption of an action and when predicting whether the decision-making method could avoid collision. Another important decision-making task is intelligent vehicle lane-changing policy. Based on the area of acceleration and braking mechanism, a method was introduced by Zhu et al. [131]. First, velocity and relative distance acceleration area was developed based on a braking mechanism and acceleration was used as a safety assessment predictor and then, a method for lane changing with the accelerating field was built, while the driver ’s behaviors, performance and safety were taken into consideration. In compliance with the simulation findings, the use of lane-changing decision-making strategies based on the acceleration can be optimized with driver behaviors for lane-change steps, including starting line, span and speed establishing safety at the same time. Although previous approaches presented a decision-making mechanism for lane changing, most of them did not show DMS for behavior prediction while lane chang- ing [132]. A fuzzy interface system with an LSTM-based method for AVS was proposed by Wang et al. to analyze behavior of surrounding vehicles to ensure safety while lane changing with 92.40% accuracy. The novelty of their work was the adjustment of motion state dynamically in advance. Li et al. proposed a framework for the analysis of the behavior, using a gradient- boosting decision tree (GBDT), merging acceleration or deceleration behavior with the data from the trajectory of the vehicle processed in the noise method on the U.S. highway 101 [133]. The partial dependency plots demonstrated that the effect on the fusion of acceleration or deceleration in independent variables by understanding the key impacts of multiple variables, was non-linear and thus distinct from the car tracking behavior with 0.3517 MAD (Mean Absolute Deviation) value, which suggested that the adoption of typical vehicle models in combination results cannot reflect characteristic behavior. Further, DRL with Q-masking was applied by Mukadam et al. to make tactical decisions for shifting lanes [134]. They introduced a system which provided a more organized and data-efficient alternative to a comprehensive policy learning on issues where high-level policies are difficult to formulate through conventional optimization or methods based on laws. The success rate of 91% was 21% higher than human perception and the 0% collision was 24% lower than human perception. This method of DRL with Q-masking worked best in the case of avoiding collision while lane shifting. Similarly, Wang et al. adopted DRL but combined with rule-based constraints to take lane-changing decisions for AVS in a simulated environment and MDP, which was challenging for high-level policy to develop through conventional methods of optimization Appl. Sci. 2022, 12, 6831 27 of 51 or regulation [135]. The training agent could take the required action in multiple situations due to the environment of state representation, the award feature and the fusion of a high level of lateral decision making and a rule-based longitudinal regulation and trajectory adjustment. The method was able to obtain a 0.8 safety rate with superior average speed and lane-changing time. Chae et al. demonstrated an emergency braking system applying DQN [136]. The problem of brake control model was conceived in Markov’s decision-making process (MDP), where the status was provided by the relative location of the hazard and the speed of the vehicle and the operating space specified as the collection of brake actions including no braking, weak, medium and heavy braking operation, combining vehicle, pedestrian and multiple road conditions scenarios, and the obtained collision rate decreased from 61.29% to 0% for a TTC value from 0.9 s to 1.5 s. As a result, this DQN-based approach was selected as one of the most practical systems for SVM in terms of autonomous braking. Furthermore, to analyze high-accuracy braking action from a driving situation declar- ing four variables, that is, speed of host vehicle, time to collision, relative speed and distance between host and lead vehicle, Wang et al. used hidden Markov and Gaussian mixture-based (HMGM) approach [137]. The efficient technique was able to obtain high specificity and 89.41% accuracy despite not considering kinematic characteristics of lead or host vehicle for braking. However, the analysis of four variants while braking could be a pathway to develop an improved version of braking decision making for AVS. When most of the approaches had dependency on datasets, methods such as DRL that combined DL and RL were extremely efficient for driving decision making in an unknown environment. For example, Chen et al. developed a brain-inspired simulation based on deep recurrent reinforcement Q-learning (DRQL) for self-driving agents with better action and state space inputting only screen pixels [138]. Although the training process was long, it resulted in better-than-human driving ability and Stanford driving agent in terms of reward gain, which indicates that this approach was one of the most suitable for applying in AVS. Another DRL-based approach combined with automatically generated curriculum (AGC), was extremely efficient for intersection scenarios with less training cost [139]. The method obtained 98.69% and 82.1% mean average reward while intersection approaching and traverse. However, the approach might lack proper finishing or goal researching in some cases of intersection traverse, but it is still very efficient for not depending on pre-trained datasets. Similarly, continuous decision-making for intersection cases in top three accident- prone crossing paths in a Carla simulator using DDPG and CNN surpassed the limitation of single scenario with discrete behavior outputs fulfilling the criteria for safe AVS [140]. DDQG was utilized to address the MDP problem and find the best driving strategy by mapping the link between traffic photos and vehicle operations through CNN that solved the common drawback of rule-based RL methods deployed in intersection cases. The method obtained standard deviation (SD) values for left turn across path opposite direction and lateral direction, straight crossing path 0.50 m/s, 0.48 m/s and 0.63 m/s, respectively, although it only considered lateral maneuvers and two vehicles in the intersection. In contrast, approach was introduced by Deshpande et al. for dealing with behavioral decision making for environments full of pedestrians [141]. Deep recurrent Q-network (DRQN) was used for taking safe decisions to reach a goal without collision and succeeded in 70% of cases. With the comparatively lower accuracy, this approach also could be very appropriate if deep learning agents were added for better feature analysis. For AVS navigation avoiding on-road obstacles, a double deep Q-learning (DDQN) and Faster R-CNN in a stochastic environment obtained stable average reward value after only 120 epochs with maximum 94% accuracy after 180,000 training steps with hyper- parameter tuning [142]. However, this approach only considered vehicles in parallel and did not show how DDQN and Faster R-CNN are fused. Moreover, the approach was still unable to obtain stable performance in uncertain moments. Appl. Sci. 2022, 12, 6831 28 of 51 Mo et al. demonstrated reinforcement learning agent and an MCTS-based approach to reduce safe decision making and behaviors by safe policy search and risk state prediction module [143]. This research assessed the challenge of decision making for a two-lane overtaking situation using the proposed safe RL approach and comparing it with MOBIL and DRQN. The proposed model outperformed MOBIL and DRQN by scoring 24.7% and 14.3% higher overtaking rate with 100% collision-free episodes and highest speed. Therefore, the proposed Safe RL could be a pathway for current AVS for risk-free trajectory decision making. In conclusion, decision making is the most vital part of an intelligent system, and to obtain acceptable human-like driving decisions, multiple deep learning and deep reinforce- ment learning methods were analyzed (shown in Table 10). The discussed approaches where able to resolve severe limitations and outperformed in overtaking, braking, behav- ioral analysis and significant segments of decision making for full AVS. Table 10. Summary of multiple deep learning methods for decision-making process. Ref. Method Outcomes Advantages Limitations Obtained higher Resolved the limitation of Individual learning of RL [121] MARL rewards than expert coordination problem in agents involved high rule-based approach. autonomous driving. computational cost. Max 94.7% accuracy for Faster decision making in Yet to demonstrate for critical [122] Weighted hybrid SVM lane changing task. different traffic conditions. and uncertain driving scenes. Complexity with large Autonomous braking system training dataset and [131,132] AC-DRL with DDPG 1.43% collision rate. while lane shifting with decreased the high accuracy. convergence rate. In best case SAC Lack of exploration caused Decreased sample complexity [125] DDQN, TD3, SAC achieved 86% failure cases of DDQN with visual encoding. success rate. and TD3. Able to avoid collision Develop lane-changing Still needed to improve [126] Dueling DQN after lowest decision-making strategy for training process for feasible 1300 episodes. AVS on highway. decision making strategy. Combines planning stage to Required huge training Monte Carlo 70% success rate for [127] make efficient driving decision samples and did not consider Tree Search highway exist case. on highway. lead vehicles’ behavior. 91.67% les collision and Efficient overtaking decision Did not consider real-time [128] MOMD + SARSOP 25.9% enhanced without rule-based system and speed and distance value. response rate. optimum actions. Almost 100% safety Outperformed overtaking, Insignificant performance to [129] MGRL index to reach goal collision avoiding, arriving at keep lane while overtaking. without collision. seven RL goals. Higher driver Collision avoidance with path Testing in real and [130] Bayesian network acceptance while prediction and challenging driving scene avoiding pedestrians. threat assignation. was not mentioned. Better lateral and Able to perform emergency Only performed best in [131] Polynomial trajectory steering angle than braking decision at straight-road scenario. ground truth. safe distance. Urban or real-life traffic LSTM with Decision based on the behavior [132] 92.40% accuracy. conditions were Fuzzy Logic of surrounding vehicles. not considered. Appl. Sci. 2022, 12, 6831 29 of 51 Table 10. Cont. Ref. Method Outcomes Advantages Limitations Understand the key impacts of Implemented on old dataset Calibration approach [133] GBDT multiple variables on and driver features were scored 0.3517 MAD. acceleration of fusion. not analyzed. 91% success rate with Effective on high-level policies Did not analyze real-time [134] DRL using Q-masking 24% lower learning through and complex collision rate. conventional optimization. road challenges. Required explicit driving Safety rate 0.8 on Productive data alternative to path while training that [135] Rule-based DQN policy average speed and end-to-end policy learning on caused low performance in lane-changing time. challenge for high-level policy. complex scenes. Efficient and quick emergency Collision rate decreased Ambiguous training and [136] DQN braking in from 61.29% to 0%. environment setting. complex environment. Analysis of kinematic Achieved specificity Highly accurate braking action characteristics for both host [137] HMGRM 97.41% and from driving situation. and lead vehicle 89.41% accuracy. was missing. Obtained maximum Did not require a labelled [138] DRQN 64.48% more reward dataset and only took screen Training time consuming. than human. pixel as input. 98.69% higher average Efficient on intersection Showed few collisions and [139] AGC based DRL mean reward scenarios and reduced DRL unfinished cases during at intersection. training time. intersection traverse. Overcame drawback of single Lowest 0.48 SD for left Only considered [140] CNN + DDPG scenario with discrete behavior turn across path. lateral maneuvers. in intersections. Tackled high-level behavioral 70% success rate for [141] DRQN decision in pedestrian-filled Low accuracy. collision free episodes. environment. Maximum 94% Applied for four driving Limited decision making for [142] DDQL and FRC accuracy with decision for navigation vehicles in parallel sides. stable reward. avoiding obstacle. Outperformed DRQN and 100% collision Did not consider [143] RL + MCTS MOBIL method for safe free episodes. urban scenarios. lane shifting 3.3. End-to-End Controlling and Prediction End-to-end controlling is one of the major fields of study for AVS. Human mistakes were the main cause of road accidents, and fully autonomous vehicles can help reduce these accidents. To improve the control system of AVS analyzing driving scenarios for lane changing, An et al. [144] proposed a system that tried to approximate driver ’s actions based on the data obtained from an uncertain environment that were used as parameters while transferring to parameterized stochastic bird statecharts (stohChart(p)) in order to describe the interactions of agents with multiple machine learning algorithms. Following that, a mapping approach was presented to convert stohChart(p) to networks of probabilistic timed automata (NPTA) and this statistical model was built to verify quantitative properties [145]. In the learning case, weighted KNN achieved highest accuracy combined with the proposed method considering training speed and accuracy, where it achieved 85.5% accuracy in 0.223 s and in the best case, time cost for probability distribution time for aggressive, conservative and moderate driving styles was 0.094, 0.793 and 0.113 s, respectively. The authors categorized Appl. Sci. 2022, 12, 6831 30 of 51 their work into learning phase, modelling phase and quantitative analyzing phase in order to develop the driving decision-taking phase. A method was demonstrated by Pan et al. to control independently at high speeds using human-like imitation learning, involving constant steering and acceleration mo- tions [146]. The dataset’s reference policy was derived from a costly high-resolution model predictive controller, which the CNN subsequently trained to emulate using just low-cost camera sensors for observations. The approach was initially validated in ROS Gazebo sim- ulations before being applied to a real-world 30 m-long dirt track using a one-fifth-scale car. The sub-scale vehicle successfully learnt to navigate the track at speeds of up to 7.5 m/s. Chen et al. focused on a lane-keeping end-to-end learning model predicting steering angle [147]. The authors employed CNN to the current NVIDIA Autonomous Driving Architecture, where both incorporated driving image extraction and asserting steering angle values. To test the steering angle prediction while driving, they considered the difference among ground truth angle which was generated by human drivers vs. predicted angle where they acquired higher steering prediction accuracy with 2.42 mean absolute error and suggested for data augmentation for training to achieve a better performance. In another work, a technically applied system of multitask learning in order to estimate end-to-end steering angle and speed control, was proposed in [148]. It was counted as one of the major challenging issues for measuring and estimating speed only based on visual perceptions. Throughout their research, the authors projected separation of speed control functions to accelerate or decelerate, using the front-view camera, when the front view was impeded or clear. Nevertheless, it also showed some shortcomings in precision and pre-fixed speed controls. By combining previous feedback speed data as a complement for better and more stable control, they improved the speed control system. This method could be stated to solve error accumulation in fail-case scenarios of driving data. They scored 1.26 Mean Absolute Error (MAE) in estimating real-time angles along with 0.19 m/s and 0.45 MAE on both datasets for velocity prediction. Thus, the improved result made the method one of the most applicable versions of CNN and data-driven AV controlling. While driving, people identify the structures and positions of different objects including pedestrians, cars, signs and lanes with human vision. Upon recognizing several objects, people realize the relation between objects and grasp the driving role. In the spatial processing of single images by the application of three-dimensional vectors, CNN has certain shortcoming in the study of time series. However, this issue cannot be overcome using CNN alone. To solve this limitation Lee et al. demonstrated an end-to-end self-driving control framework combining a CNN and LSTM-based time-series image dataset applied in a Euro Truck simulator [149]. The system created a driving plan which takes the changes into account over time by using the feature map to formulate the next driving plan for the sequence. Moreover, NVIDIA currently has succeeded in training a ConvNet for converting raw camera images into control steering angles [150]. It resolved end-to-end control by predicting steering angle without explicating labels with approximately 90% autonomy value and 98% autonomous of the testing period. This approach was one of the most demonstrated approaches that boosted research of AVS applying deep learning methods. A similar method, deep ConvNet, was used by Chen et al. to train for directly extracting the identified accessories from the front camera [151]. A basic control system, based on affordance principles, provided steering directions and the decision to overtake proceeding vehicles. Rather than using lane-marking detection methods as well as other objects to assess indirect activity specifications of the car, a variety of driving measures allowances were specified. This method included the vehicle location, the gap to the surrounding lane markers and records of previous car driving. While this was a very trendy concept, for many reasons it may be challenging to handle traffic with complex driving maneuvers and make a human-like autonomous vehicle controlling system. To deploy a human-like autonomous vehicle speed-control decision-making system Zhang et al. proposed a double Q-network-based approach utilizing naturalistic driving Appl. Sci. 2022, 12, 6831 31 of 51 data built on the roads of Shanghai inputting low dimensional sensor data and high- dimensional image data obtained from video analysis [152]. They combined deep neural networks and double Q-learning (DDQL) [194–196] to construct the deep Q-network (DQN) model which was able to understand and make optimal control decisions in simultaneous environmental and behavioral states. Moreover, real-world data assessment reveals that DDQN can be used on a scale to effectively minimize these unreliable DQN problems, resulting in more consistent and efficient learning. DDQN had increased both in terms of interest precision and policy efficiency. The model performed 271.13% better than DQN in terms of speed-control decision making. Even so, the proposed approach could be more applicable to an unknown driving environment with combined CNN agent for feature extraction. Chi et al. formulated a ST-LSM network that incorporates spatial and temporal data from previously multiple frames from a camera’s front view [153]. Several ST-Conv layers were used in the ST-LSTM model to collect spatial information and a layer of Conv-LSTM was used to store temporarily data at the minimal resolution on the upper layer. However, the spatial and temporal connection among various feature layers was ignored by this end- to-end model. They obtained a benchmarking 0.0637 RMSE value on the Udacity dataset, creating the smallest 0.4802 MB memory and 37.107 MB model weight. The limitation of the paper was that all present end-to-end driving models were only equipped by focusing on the ground truth of the current frame steering angle, which indicated a lack of further spatiotemporal data. Furthermore, to obtain a better control system, the previous issue was tackled, and an end-to-end steering control system was implemented by Wu et al. by concatenating future spatiotemporal features [154]. They introduced the encoding for an advanced autonomous driving control system of spatiotemporal data on a different scale for steering angle approximation using the Conv-LSTM neural framework with a wide-spectrum spatiotemporal interface module. Sequential data were utilized to improve the space- time expertise of the model during development. This proposed work was compared with end-to-end driving models such as CgNet, NVIDIA’s PilotNet [155] and ST-LSTM Network [153], where the root mean square error (RMSE) was 0.1779, 0.1589 and 0.0622, respectively, and showed the lowest RMSE value of 0.0491 to predict steering angles, which was claimed to be more accurate than an expert human driver. Thus, this approach was applicable for a level 4 or 5 autonomous vehicle control system. Moreover, a deep neural network-based approach with weighted N-version Program- ming (NVP) was introduced for resilient AV steering controlling [156]. Compared to the other three networks (chauffeur, autumn, rambo), the proposed network showed 40% less RMSE retrieving steering angles in clear, rain, snow, fog and contrast lighting condi- tions. However, there was a high failure rate for the large developing cost for training an individual DNN model. Aiming to build a vehicle motion estimation system for diversity awareness while driving, Huang et al., via latent semantic sampling [157], developed a new method to generate practical and complex trajectories for vehicles. First, they expanded to include semantic sampling as merging and turning the generative adversarial network (GAN) structure with a low-dimensional semantic domain, formed the space and constructed it. It obtained 8% improvement on the Argoverse validation dataset baseline. They therefore sampled the estimated distribution from this space in a way which helped the method to monitor the representation of semantically different scenarios. A CNN and state-transitive LSTM-based approach was demonstrated with multi- auxiliary tasks for retrieving dynamic temporal information from different driving scenarios to estimated steering angles and velocity simultaneously [158]. The method applied the vehicle’s current location to determine the end-to-end driving model sub-goal angle to boost the steering angle estimation accuracy, which forecasted that the efficiency of the driving model would improve significantly. The combined method obtained 2.58 and 3.16 MAE for steering angle prediction and 0.66 m/s and 0.93 m/s speed MAE in GTA Appl. Sci. 2022, 12, 6831 32 of 51 V and Guangzhou Automotive Cooperate datasets, respectively. Nevertheless, it showed a slow response in unknown environment, so this method might not be applicable in practical implementation. In a similar manner, Toromanoff et al. presented a CNN-based model for lateral control of AVS using a fisheye camera with label augmentation technique for accurate corrections labelling under lateral control rule to tackle ceases of lateral control error in wide FoV [159]. This method compares with pure offline methods where feedback was not implemented from a prediction which resulted in 99.5% and 98.7% autonomy in urban areas and highways after training with 10,000 km and 200 h driving video. On the other hand, Smolyakov et al. reduced a huge number of parameters of CNN to avoid overfitting along with helping to find dependency on data sequence and implement in a CarND Udacity Simulator for predicting steering angles. However, the obtained unsatisfactory result was comparable to other reviewed results, where the accuracy was 78.5% [160]. Similarly, a CNN-based approach was applied for both lateral and longitudinal mo- tion controlling of AVS obtaining 100% autonomy on e-road track on TORCS simulator. Although it had performed very well, contributing to both kinds of motion controlling, it lacked training data for practical implementation and memory consumption for training two different neural networks for speed and steering angle prediction. This method could be better approached by implementing in real scenarios with a good amount of training data [161]. In another proposal, a reinforcement learning-enabled throttle and brake control system was proposed by Zhu et al. [162], focusing on a one leader and one follower formation. A neural dynamic programming algorithm evaluating with trial-and-error method was directly applied for adopting near-optimal control law. The control policy included the necessary throttle and brake control commands for the follower according to the timely modified corresponding condition. Simulation experiments were carried out using the well-known CarSim vehicle dynamic simulator to show the reliability of the approach provided. To overcome traditional sensor-based pipeline for controlling AVS where there is a tendency to learn from direct mapping, Xiao et al. demonstrated multimodal end-to-end AVS applying conditional imitation learning (CIL), taking an RGBD image as raw data in a Carla simulator environment [163]. The CNN-based CIL algorithm was evaluated in different weather modes to identify the performance for end-to-end control. The success rate of controlling in one turn and dynamic environment were 95% and 84%, respectively, which could be boosted through early fusion by changing the number of color channels from three (RGB) to four (RGBD). However, performance dropped almost 18.37% and 13.37% during controlling AVS with RGB image input for one turn and dynamic environment, respectively, in a new map of Carla simulators which could be considered as uncertain area In brief, most of the deep learning approaches for end-to-end controlling and motion predications were based on CNN, showing efficient outcomes suitable for practical level 4 or 5 AVS. Most of the methods were deployed for estimating continuous steering angle and velocity, some controlling approaches taking into account resolving blind spot, gap estimation, overcoming slow drifting, both lateral and longitudinal motion controlling with methods such as multimodal multitask-based CNN, CNN-LSTM, Deep ConvNet, ST-LSTM, neural dynamic programming-based reinforcement learning with actor-critic network and RL. These methods faced challenges, such as noise created by human factor reasoning speed changes causing lower accuracy, only equipped by focusing on the ground truth of the current frame steering angle and not applying in a practical or complex environment. The overall summary of discussed methods is presented in Table 11. Appl. Sci. 2022, 12, 6831 33 of 51 Table 11. Summary of multiple deep learning methods for end-to-end controlling and prediction. Ref. Method Outcomes Advantages Limitations Performed safe control during Unsafe driving behavior and Gained 85.5% accuracy in [144] Hybrid weighted KNN lane changing in did not consider 0.223 s in best case. uncertain environment. complex conditions. Successfully learnt to High-speed driving control Trained only for elliptical CNN + LSTM + [146] navigate the track at and robustness to racetracks with no State method speeds of up to 7.5 m/s. compound errors. other vehicles. Obtained appropriate Able to overcome slow drifting Improper dataset for [147] CNN with comma.ai steering angles with 2.42 from human driving data. practical implementation. mean absolute error. Scored 1.26 MAE for Accurate estimation Noise of human factor for Multimodal-based [148] angles and 0.45 MAE continuous steering angles speed changes caused CNN for velocity. and velocity. lower accuracy. Almost similar steering Resolved the limitation of Lack of driving data collection [149] CNN and LSTM prediction value as CNN and blind-spot problem. from vehicle. ground truth. 90% autonomy value and Required fewer training data Robustness were not successful [150] ConvNet 98% autonomous with no manual in internal processing phase. approximately. decomposition. 0.033 and 0.025 steering Considered lane gap and [151] Deep ConvNet angle MAE on GIST and records of previous Only tested on simple cases. Caltech baseline. car driving. Make optimal control 271.13% better than DQL Measured few errors on [152] DDQN increasing precision and for speed control. uneven roads. policy efficiency. Obtained 0.0637 RMSE Implemented in challenging Only focused on the ground [153] ST-LSTM Network value on Udacity with lighting conditions. truth of steering angle. model weight. Showed lowest RMSE Spatiotemporal Overcome the limitation Did not test in [154] value 0.0491 to predict Conv-LSTM of [153]. busy environment. steering angles. Showed on average 40% High developing cost for Predicted steering angles in [156] DNN less RMSE retrieving training individual multiple conditions. steering angles. DNN model. 8% improvement for Better trajectory building Had not tested in real time and [157] GAN Argoverse validation layers while motion prediction. only used simple semantics. dataset baseline. Used current position and CNN and Predicted 2.58 and 3.16 Slow prediction rate in [158] subgoal angle for steering state-transitive LSTM MAE for steering angle. unknown environment. angle prediction. Achieved 99.5% and 98.7% accuracy in urban areas Solved lateral controlling using Autonomy dropped while [159] CNN and fisheye camera. sharp turning. highways, respectively. Reduced parameters of CNN [160] CNN Achieved 78.5% accuracy. to avoid overfitting on Noticeable performance drop. data sequence. Appl. Sci. 2022, 12, 6831 34 of 51 Table 11. Cont. Ref. Method Outcomes Advantages Limitations Obtained 100% autonomy Showed both lateral and Lack of training data and [161] CNN on e-road track on TORCS. longitudinal motion control. memory consuming. Robust throttle and brake Learned controlling policy Environmental surroundings [162] RL with ACN value of the host vehicle. while following lead vehicle. were not stated. Demonstrated successful 84% success rate in Up to 18.37% performance [163] CIL multimodal approach in dynamic environment. drop in unknown map. four cases. 3.4. Path and Motion Planning Precipitation-based autonomous navigation including path and motion planning in an unknown or complex environment is one of the critical concerns for developing AVS. To tackle the current problem and analyze the contribution, multiple deep learning and deep reinforcement learning (DRL) combined methods for path and motion planning are reviewed in this section. Initially, You et al. focused on the issue of path planning of autonomous vehicles in traffic in order to repeat decision making by replicating the optimum driving technique of expert drivers’ actions for lane changing, lane and speed maintenance, acceleration and braking in MDPs on highways [164]. The optimal control policy for the proposed MDP was resolved using deep inverse reinforcement learning (DIRL) and three MaxEnt IRL algorithms by utilizing a reward function in terms of a linear combination of parameterized function to solve model-free MDP. The trajectory proposals were executed at the time of overtaking and the policy recovery was reduced to 99%, even though there was insufficient evidence for the reflection of stochastic behavior. To solve limitations of rule-based methods for safe navigation and better intersection problems for AVS, a vision-based path and motion planning formula was used by Isele et al., adopting DRL [165]. Each wait action was proceeded by another wait or go action, meaning that each pathway was a series of waiting decisions that concluded in a go decision as well as the agent not being permitted to wait after the go action had been chosen. The method secured a success rate for forward, right, left and turn and challenge of 99.96%, 99.99%, 99.78% and 98.46%, respectively, which was 28% faster than the TTC (time-to-collision) method, although performance decreased three times and average time doubled during this challenging situation. Zhang et al. proposed a risk analysis and motion planning system for autonomously operated vehicles focused on highway scenario motion prediction of surrounding vehi- cles [166]. An interactive multiple model (IMM) and constant turn rate and acceleration (CTRA) model were used for surrounding vehicle motion prediction, and model predictive control (MPC) was used for trajectory planning that scored 3.128 RMSE after 5 s dur- ing motion prediction. Although it was designed for connected AVS, it is efficient for vision-based approaches. Another approach, local and global path planning methodology, was presented in an RoS-based environment for AVS by Marin-Plaza et al., where they used the Dijkstra and time elastic bands (TEB) method [167]. The path planning model was able to reach the goal with modest error by calculating Euclidean distance for comparing local and global pan waypoints, where it scored 1.41 m, which is very efficient. However, it was applicable only if the model was not specifically calibrated for the vehicle’s kinematics or if the vehicle was out of track, and did not consider complex scenarios. In another work, Islam et al. established a vision-based autonomous driving system that relied on DNN, which handled a region with unforeseen roadway hazards and could safely maneuver the AVS in this environment [168]. In order to overcome an unsafe navigational problem, they presented object detection and structural segmentation-based deep learning architecture, where it Appl. Sci. 2022, 12, 6831 35 of 51 obtained an RMSE value of 0.52, 0.07 and 0.23 for cases 1 to 3, respectively, and 21% safety enhancement adding hazard avoiding method. Ma et al. proposed an efficient RRT algorithm that implemented a policy framework based on the traffic scenes and an intense search tree extension strategy to tackle traditional RRT problems where it faced a meandering route, an unreliable terminal state and sluggish exploration, and established more sustainable motion planning for AVS [169]. In addition, the integrated method of the proposed fast RRT algorithm and the configuration time space could be adopted in complex obstacle-laden environments to enhance the efficiency of the expected trajectory and re-planning. A significant set of experimental results showed that the system was much quicker and more successful in addressing on-road autonomous driv- ing planning queries and demonstrating its better performance over previous approaches. In another work, an optimum route planner integrated with vehicle dynamics was designed by Gu et al. implementing an artificial potential field to provide maximum workable movement that ensured the stability of the vehicle’s path [170]. The obstacles and road edges were typically used with constraints and not with any arbitrary feature in this method in the optimal control problem. Therefore, when designing the optimum route using vehicle dynamics, the path-planning method was able to treat various obstacles and road structures sharply in a CarSim simulator. The analysis showed that the method reduced computational costs by estimating convex function while path planning. A similar method was proposed by Wahid et al., where they used an artificial potential field with adaptive multispeed scheduler for a collision-avoidance motion planning strategy [171]. Cai et al. demonstrated a novel method combining CNN, LSTM and state model which was an uncertainty-aware vision-based trajectory generation network for AVS’s path-planning approach in an urban traffic scene [172]. The work was divided into two major parts: the first one was a CNN bottleneck extractor, and the second component included a self-attention module for calculating recurrent history and an LSTM module for processing spatiotemporal characteristics. Finally, they designed the probable collision-free path planning with speeds and lateral or longitudinal locations for the next 3.0 s after taking image stream and state information in the past 1.5 s considering as input. The method obtained more centralized error distribution and lower error medium. For safe navigation for AVS in road scenarios with obstacles, a model prediction control-based advanced dynamic window (ADW) method was introduced by Kiss et al. [173]. The method demonstrated differential drive that reached the destination location ignoring the desired orientation and did not require any weighted objective function. A motion planning model based on the spatiotemporal LSTM network (SLN), which had three major structural components, was proposed by Bai et al. It was able to produce real-time feedback based on the extraction of spatial knowledge [174]. First, convolutional long-term memory (Conv-LSTM) was applied in sequential image databases to retrieve hidden attributes. Secondly, to extract spatiotemporal information, a 3D CNN was used, and precise visual motion planning was displayed constructing a control model for the AV steering angle with fully connected neural networks. The outcome showed almost 98.5% accuracy and better stable performance compared with Hotz’s method [147]. Nonetheless, the method was found to minimize state after generating overfitting on antecedent data for time-series data of previous steps, causing more computational cost and time. Another motion planning avoiding-obstacle-based approach was proposed in a simu- lation environment [175]. The motion planning method had the ability to infer and replicate human-like control thinking in ambiguous circumstances, although it was difficult to estab- lish a rule base to tackle unstructured conditions. The approach was able to execute 45.6 m path planning with 50.2 s. In conclusion, very few works have adopted a perception-based path and motion planning for AVS but the existing research adopting deep inverse reinforcement learning and MaxEnt IRL, deep Q-network time-to-go method, Dijkstra and time elastic bands method, DNN, advance RRT, artificial potential field, ADW using model predictive control and fuzzy logic made a remarkable contribution, with high accuracy, collision-free path Appl. Sci. 2022, 12, 6831 36 of 51 planning, 21% safety enhancement adding hazard-avoiding method planning motion in a multilane turn-based intersection. Nevertheless, these methods were not practically implemented or theoretical, and some of the high-performing approaches were not tested in a real-life environment with heavy traffic. An overview of the deep learning methods selected for analysis to improve AVS is presented in Table 12. Table 12. Summary of multiple deep learning methods for path and motion planning. Ref. Method Outcomes Advantages Limitations Insufficient training data for 99% policy recovery Avoid cost function and [164] DIRL stochastic within less data length. manual labelling. behavior representation. Solved limitation of Performance decreased three 28% faster than [165] DRL with TTC rule-based times during TTC method. intersection problems. challenging situation. Score 3.128 RMSE after Considered motions of Lower accuracy in far [166] IMM with MPC 5 s during surrounding vehicles. predicted horizon. motion prediction. Obtained efficient Applicable only if vehicle was Dijkstra and TEB Reach goal with [167] Euclidean out of track in method modest error. distance 1.41 m. simple scenarios. 21% safety Safe navigation adding enhancement adding Tested only on simple [168] DNN hazard detection and hazard-avoiding open highway. segmentation method. method. Took 5 ms and 48 ms Novel cost function to select Limited to [169] Advance RRT for path selection and path and non-rule-based approach. path generation. obstacle-avoiding feature. Visualized potential Reduce computational cost Effects of local minimum issue [170,171] Artificial potential field field in nine by estimating that led AV to be stuck in different scenarios. convex function. a position. Lower error medium Vehicle motion planning Did not consider traffic light CNN + LSTM + [172] and more centralized predicted in multilane and weather condition for State method error distribution. turn-based intersection. performance evaluation. Reached destination Did not require any Limitation occurred with [173] ADW with MPC location ignoring the weighted objective function. constrained kinematics. desired orientation. Almost 98.5% average Able to learn time-serial Minimized state after [174] 3D-CNN accuracy and features from generating overfitting on stable outcome. traffic environment. time-series data. Difficult to establish a rule 45.6 m path planning Human-like control thinking [175] Fuzzy logic base to tackle with 50.2 s. in ambiguous circumstances. unstructured conditions. 3.5. AR-HUD Augmented reality (AR) in head-up display (HUD) or displaying in windshield for autonomous driving system as a medium of final visualizing of activities outcomes from the deep learning approach was overlayed with an autonomous driving system. The AR-based vehicular display system was essential for driving situation awareness, navigation and overall deployment as a user interface. Yoon et al. demonstrated an improved forward collision alert system detection of cars and pedestrians fused into the HUD with augmented reality through using stereo cameras and visualized early alerts where SVM classifier was applied for object recognition and obtained an F1 score of 86.75% for car identification and 84.17% for pedestrian iden- Appl. Sci. 2022, 12, 6831 37 of 51 tification [176]. The limitation of the work was noticed when the observed object moved rapidly and the car suddenly turned; it was visualized with delay. The proposed system yet needed to optimize efficiency and acceleration which in diverse vehicle conditions responds robustly to different and high speeds. An analysis showed personal navigation with AR navigation assist equipped for use with a volumetric 3D-HUD and utilizing its parameters. An interface was developed for assisting to turn faster by locating turn points quicker than during regular navigation [177]. The interface also helped to maintain user eyes and to fix them more precisely on the driving environment after analyzing traffic scenes with deep learning algorithm with proper registration of applications via spatial orientation of AR views on interface. On the basis of the results, however, the inadequate perception of the depth of a specified 2D HUD distance is obvious and the navigation system’s AR interface was ineffective without a 3D HUD. An automatic AR based on a road tracking information method registration was introduced by Yoon et al., with a SIFT matching function and homography measurement method, which defined matching between camera and HUD providing the driver ’s view was positioned to the front, which detected vehicle and pedestrians and converted them into AR contents after projective transformation [178]. This solution was good enough for daytime performance but had limitations at nighttime. Nevertheless, the procedure had the ability to automate the matching without user interference, but it is inconvenient while projecting outcomes which occurred due to misreading local correspondence. Park et al. demonstrated an AR-HUD-based driving safety instruction by identifying vehicle and pedestrians using the INRIA dataset [179]. The identification method was built using SVM and HOG with 72% and 74% in fps accuracy and detected partial obstacles, respectively, applying a billboard sweep stereo (BSS) algorithm. The detected vehicles and pedestrians were overlapped on the HUD with the AR technique. Despite detecting obstacles in sunny and rainy scenarios, it was not deployed for nighttime scenarios. In order to integrate outcomes with AR, the system was divided into two parts by Rao et al., 3D object detection and 3D surface reconstruction, to develop object-level 3D reconstruction using Gaussian Process Latent Variable Model (GPLVM) with SegNet and VPNet for in-vehicle augmented reality UI and parking system [180]. Their AR-based visualization system was built with monocular 3D shaping, which was a very cost-efficient model and needed only a single frame in the input layer. Furthermore, a new traffic sign-recognition framework based on AR was constructed by Abdi and Meddeb to overlay traffic signs with more recognizable icons overlapped in an AR-HUD to increase the visualization of a driver aiming to improve safety [181]. The Haar Cascade detector and the verification of the theory using BoVW were combined with the relative spatial data between visual words, which had proven to be a reasonable balance between resource efficiency and overall results. A classifier with an ROI and allocated 3D traffic sign was subsequently developed using a linear support vector machine that required less training and computation time. During the decision-making process, this state-of-the-art methodology influenced the distribution of visual attention and could be more consistent with the improved approach of deep learning recognition relying on the GPU. Addressing the challenge of overtaking an on-road slow vehicle, a see-through effect- based marker-less real-time driving system had been demonstrated by Rameau et al., applying AR [182]. To overcome the occlusion and produce a seamless see-through effect, a 3D map of the surroundings was created using an upper-mounted camera and implement- ing an in-vehicle pose predictor system. With up to 15 FPS, they presented a faster novel real-time 2D–3D tracking strategy for localization of rear in a 3D map. For the purpose of decreasing bandwidth usage, the ROI was switched to the rear car impacted by an occlusion conflict. This tracking method on AR-HUD showed great efficiency and easy adoption capability for vehicle displaying systems. Appl. Sci. 2022, 12, 6831 38 of 51 To reduce the accident cases, Abdi et al. proposed augmented reality-based head-up display providing more essential surrounding traffic data as well as increasing interactions between drivers and vehicles to enhance drivers’ focus on the road [183]. A custom deep CNN architecture was implemented to identify obstacles and final outputs will be projected in the AR head-up display. For AR-based projection in HUD, firstly, pose prediction of targeted ROIs were carried out and obtained 3D coordinates with points after achieving camera projection matrix to recognize AR 3D registration. This step created a 6-DOF pose of translation and rotation parameters which will be helpful for motion estimation calculation with planar homograph. Afterwards, the RANSAC method was applied to compute the homograph matrix, and OpenGL real camera was synchronized with a virtual camera that showed a projection matrix to map 2D points utilizing 3D surface points and developed a marker-less approach. Lindemann et al. demonstrated an augmented reality-based windshield display system for autonomous vehicle with a view to assisting driving situation awareness in city areas and increase automated driving level from level 4 to 5 [184]. This AR-based windshield display UI was developed based on deep learning-applied object detection to enhance situation awareness, aiming at both clear and lower-visibility conditions where they obtained very different situation awareness scores in low-visibility conditions in disabled windshield display but failed to obtain a good score when windshield UI was enabled. Nevertheless, it worked significantly better in clear weather conditions. Park et al. presented a 2D histogram of oriented gradient (HOG) tracker and an online support vector machine (SVM) re-detector based on training of the TLD (tracking-learning- detector) functional vehicle tracking system for AR-HUD using equi-height mosaicking image (EHMI) [185]. The system initially performed tracking on the pre-computed 2D HOG EHMI, when the vehicle was identified in the last frame. If the tracking failed, the system started re-detection using an online learning-based SVM classification. The tracking system conducted online learning frequently after the vehicle had been registered and minimized the further calculation necessary for tracking as the HOG descriptor for EHMI was already determined in the detection phase. The technique was perfect for deploying in various lighting and occlusion scenes since it adopted online learning. Refining the algorithm to make optimized hardware or embedded device and to identify other dangerous obstacles effectively in road scenes, this lightweight architecture-based proposed work could be a more acceptable approach for faster tracking and visualizing in HUD. To represent driving situation awareness data, Park et al. introduced a vehicle augmented-reality system that deducts drivers’ distractions with an AR-based windshield of the Genesis DH model from Hyundai motors [186]. The system presented driving conditions and warned a driver using a head-up monitor via the augmented reality. The system included a range of sub-modules, including vehicle and pedestrian recognition based on the deep learning model of [179], vehicle state data, driving data, time to collision (TTC), hazard evaluation, alert policy and display modules. During most experiments, on the basis of TTC values and driver priority, the threat levels and application of augmented EHMI was already determined in the detection phase. In this section, a combination of deep learning algorithms and their outcomes were visualized as the final task of AVS showing them in an AR-based HUD for better driving assistance. AR-HUD was adopted due to visualization in front display for early warning, navigation, object marking by overlapping, ensuring safety and better tracking. Although these studies had successful demonstrations, some major limitations were detected when analyzing the studies, such as visual delay for the case of sudden turn or rapid-moving objects, misreading of local correspondence, high computational cost while 3D shaping, visualizing challenges in extreme contrast and distraction for complex UI. Table 13 provides a summary of the section. Appl. Sci. 2022, 12, 6831 39 of 51 Table 13. Summary of multiple deep learning methods for AR-HUD. Ref. Purpose Methods Advantages Limitations Improved collision alert system Visualization delay while [176] Early warning SVM detecting cars and pedestrians observing rapid moving and fused into the HUD. sudden turning of vehicle. Helped to turn faster and more The insufficient depth Custom deep learning [177] Navigation confidently locating turn points perception of the defined 2D based scene analysis. quicker. HUD distance was apparent. Detected road objects are Automatic matching ability is SIFT and homography [178] Object marking converted into AR contents after inconvenient due to misreading measurement method. projective transformation. of local correspondence. Applicable in sunny and rainy Poor detection accuracy and not SVM, HoG and BSS [179] Safety scenarios for overlapping of applicable for nighttime algorithm. detected objects and obstacles. scenarios. 3D shaping cost-efficient model GPLVM with SegNet Computational complexity is [180] Assistance and needs only single frame in and VPNet. higher for algorithm fusion. input layer. Overlay traffic signs with more Haar cascade detector Lack of implementation in [183] Safety recognizable icons in AR-HUD with BoVW method. complex scenarios. to improve safety. Assist overtaking an on-road The detection methods for 2D–3D tracking [182] Tracking slow vehicle via marker-less deployment in the real-life field strategy with ROI. real-time driving system. had yet to become apparent. Providing more essential Was not deployed in complex [183] Safety CNN and RANSAC. surrounding traffic data to traffic scenarios or nighttime increase interactions and focus. environments. Custom deep learning Boost situation awareness Failed to achieve good [184] Safety applied object aiming at both clear and visualization in lower-visibility detection. lower-visibility conditions conditions. Applicable in various lighting Required a very wide range of [185] Tracking TLD using SVM. and occlusion scenes, since it views. adopted online learning. Enhanced a driver ’s intuitive Not computation cost efficient [186] Safety and awareness SVM and HoG. reasoning and minimized driver and complex UI. distraction calculating TTC. 4. Evaluation Methods In this section, commonly used evaluation metrics throughout the systematic review are presented in Table 14. Several evaluation techniques with equations and description are shown which will give a better understanding, as evaluation techniques are different from the reviewed methodology. Appl. Sci. 2022, 12, 6831 40 of 51 Table 14. Evaluation techniques. Ref. Technique Equations Remarks [40,44,50,53,72,76,82,90, TP is the true positive and FN is T P Sensitivity/TPR/Recall (R) R = T P+F N 110,116,119,120,137] the false negative detection. [40,44,67,72,90,110,116, T P Precision (P) FP is the false positive detection. P = T P+FP 119,120] [54,80,88,90,110,119,120, RP F1 Score - F1 = 2 R+P 176] [37–39,41,43,46– 49,52,57,60,61,64,79,91– X is the number of successes X Pred Pred 94,96,98–101,106,108,112– Accuracy Accuracy = GT and X is the ground truth. GT 114,118,120,122,137,142, 144,152,179] Region o f intersection I oU = Region o f union [95,103–105,111,112] mIoU - 1 T P m I oU = k+1 T P+FP+F N F N [34,63,77] FNR - F N R = T P+F N m AP = k denotes each episode and n is k=n [35,42,45,56,73,75,92] mAP the total episodes. ((R(k) R(k + 1)) P) k=1 Log Average Miss n a is the series positive values [65,66,68,76,78] 1 Log = ex p ln a A MR i Rate (Log ) n correlated with the missing rate. A MR i=1 TPR is the true positive rate, and [51,55,81,116] Area Under Curve (AUC) AUC = T PR d(FPR) FPR is the false positive rate. Dy is center of gravity, # is yaw [83,89,131] Lateral Error Dy = Dy + L # angle towards the road and L is ( ) the distance. t is dimension of state, d is n n min t , t > ( ) 1 2 > measured distance, d is safe sn i f l = 1 and d , d > 3 1 2 distance, l is overtaking lane and r = min(t , t ) 3 4 [121,138,139,142] Reward (r) v is the longitudinal velocity. i f l = 2 and d , d > 3 1 2 Here, n = 1,2 refers host and lead 5 else vehicle in driving lane and n = 3,4 d d n sn where, t = n refers to overtaking lane. N is the total collision number N col col [123,126,128,130,136] Collision Rate (C ) C = rate rate l a p while completing total C laps. l a p Success Counts [102,124,127,134,165] Success Rate (SR) - SR =  100% Total Test Number N is total episode without wc wc [129,135] Safety Index (SI) s = collision and N is the e p e p total episodes. x and x ˆ are the real and i i [133,147,148,151,158] MAD/MAE predicted value, respectively, and M AE/ M AD = å jx x ˆ j i i i=1 n is the total episodes. T N [137] Specificity (SPC) T N is the true negative value. SPC = FP+T N N is total interventions and T is [150,159] Autonomy value (Av) Av = 1  6  100% e elapsed time. n 2 x is the ground truth and x ˆ is the i i (x x ˆ ) å i i [153–156,166–168] RMSE i=1 R MSE = n predicted truth. 0 0 y is the sample proportion, y is 0 0 y y [117] p-Value p the assumed proportion, n is Z = 0 0 y (1y ) sample size. c Appl. Sci. 2022, 12, 6831 41 of 51 5. Case Study 5.1. AVS in Roundabout Cases A roundabout is one of the most difficult scenarios for driving with AVS due to tight time constraints on vehicle behavior because of the yield and merging of maneuvers with high-quality vision requirements for estimating the state of other vehicles, and multi-factor decision-making based on these state estimations. It is particularly tough to forecast the actions of other cars in a roundabout due to the availability of numerous exit locations as discrete options mixed with continuous vehicle dynamics. The entry risk at roundabouts grows with decreasing distance as the ego vehicle must account for cars in the circle passing. Okumura et al. proposed a neural network to map observations to actions in a roundabout that are handled as a combination of turns in order to emphasize the deep learning-based AVS for roundabout cases [197]. This method concentrated on route planning and speed estimation for the roundabout, as well as detection, tracking and generating predictions about the environment using sensor fusion, but ignored interactions between cars. This method concentrated on route planning and speed estimation for the roundabout, as well as detection, tracking and generating predictions about the surroundings using sensor fusion, but ignored interactions between cars [198]. This could be improved by a strategy for forecasting whether a vehicle will exit the roundabout based on its anticipated yaw rate. In a roundabout scenario, the projected yaw rate is a significant indication of whether a car will turn next to the ego vehicle. Although the system was proved to be capable of human-like judgments for a certain roundabout situation, only the center of mass and velocity were calculated to quantify detection of turning cars. This method may be a viable solution for the roundabout research of [197]; however, it may result in errors in roundabouts with no traffic lights or heavy traffic. One of the main reasons of vision-based AVS is to reduce the dependency in terms of safety and collision-free driving; therefore, combined multi-thread architecture of al- gorithms such as Spatial CNN (SCNN) and Deep Recurrent Q-Network (DRQN) could be a major solution for roundabout cases. The spatial features of SCNN for traffic scene understanding in dense traffic conditions as well as the ability of extreme efficient traffic scene-analysis demonstration incorporating multi-threading with self-decision making improved DRL approaches such as DRQN or DDQL could be a vast improvement in the research of AVS in roundabout cases. 5.2. AVS in Uncertain Environmental Cases Even at the current development level, it is challenging for AVS to operate au- tonomously in unknown or uncertain environments. The uncertainty may be because of variable traffic conditions, unknown terrain, unmarked or untrained settings, or even a situation including an extended obstruction. In an unexpected driving environments, even the performance of Waymo, the self-driving vehicle of Google, is at a conditional level 4 of autonomy based on NHTSA autonomous functions, and Tesla’s self-driving vehicles are only at level 2 of autonomy. In contrast, the authors of one study addressed safety issues posed by ambiguity in DL approaches: insufficient training data, locational shift, inconsistencies between training and operating parameters, and uncertainty in pre- diction [199]. The most controversial incident occurred when a Tesla Model S was involved in a deadly collision in which the driver was killed when its autopilot system failed to notice a tractor-trailer 18-wheeler that turned in front of the vehicle [200]. To reduce un- intended occurrence in unknown or uncertain situations and environments, it might be possible to develop level 4 or 5 AVS with safe perception analysis, path planning, decision making and controlling by removing dependence on labelled data and adopting deep reinforcement learning-based approaches. Moreover, several techniques, such as those in [83,128,130,144,172], which were effective in avoiding collision, lane shifting, detection, and safe decision making in unknown or dynamic situations, can be a means of reducing the constraints in uncertain environments. Appl. Sci. 2022, 12, 6831 42 of 51 6. Discussion Deep learning is fast becoming a successful alternative approach for perception-based AVS as it reduces both cost and dependency on sensor fusion. With this aim in mind, total categories of primary domains of AVS were reviewed in this paper to identify efficient methods and algorithms, their contributions and limitations. From the study, it was found that recent deep learning algorithms obtained high accuracy while detecting and identifying road vehicle types, and in some cases, the results surpassed LiDAR’s outcome in both short and long range for 3D bounding vehicles [34]. Moreover, some recent methods such as YOLO V2 [35], deep CNN [38], SINET [41] and Faster R-CNN [42] achieved high accuracy within a very short period of time from low- quality training images to challenging nighttime scenarios. However, there were several limitations, for example, in certain lighting conditions and higher execution costs. Fol- lowing that, a massive contribution to lane and curve detection along with tracking was presented by studies where 95.5% road scene extraction was demonstrated, for example, in [79], for lane edge segmentation without manual labelling using a modified CNN ar- chitecture. As discussed in previous sections, challenges such as higher computational cost [81], insufficient for far field of view [82], not testing in complex scenarios [79] and poor luminance made some proposals tough for practical implementation in present AVS. In addition, a good amount of attention was given to developing safe AVS systems for pedestrian detection. Multiple deep learning approaches such as DNN, CNN, YOLO V3- Tiny, DeepSort R-CNN, single-shot late-fusion CNN, Faster R-CNN, R-CNN combined ACF model, dark channel prior-based SVM, attention-guided encoder–decoder CNN outper- formed the baseline of applied datasets that presented a faster warning area by bounding each pedestrian in real time [61], detection in crowded environments, and dim lighting or haze scenarios [62,72] for position estimation [72], minimizing computational cost and outperforming state-of-the-art methods [120]. The approaches offer an ideal pedestrian method once their technical challenges have been overcome, for example, dependency on preliminary boxing during detection, presumption of constant depths in input image and improvement to avoid missing rate when dealing with a complex environment. Moreover, to estimate steering angles, velocity alongside controlling for lane keeping or changing, overcome slow drifting, take action on a human’s weak zone such as a blind spot and decreasing manual labelling for data training, multiple methods, such as multimodal multitask-based CNN [148], CNN with LSTM [149] and ST-LSTM [153], were studied in this literature review for AVS’s end-to-end control system. Furthermore, one of the most predominant segments of AVS, traffic scene analysis, was covered to understand scenes from a challenging and crowded movable environment [102], improve performance by making more expensive spatial-feature risk prediction [112] and on-road damage detection [120]. For this purpose, HRNet + contrastive loss [104], Multi- Stage Deep CNN [106], 2D-LSTM with RNN [108], DNN with Hadamard layer [110], Spatial CNN [112], OP-DNN [113] and the methods mentioned in Table 9 were reviewed. However, there are still some limitations, for instance, data dependency or relying on pre-labelled data, decreased accuracy in challenging traffic or at nighttime. Taking into account all taxonomies as features, the decision-making process for AVS was broadly analyzed where driving decisions such as overtaking, emergency braking, lane shifting with collision and driving safety in intersections adopting methods such as deep recurrent reinforcement learning [127], actor-critic-based DRL with DDPG [123], double DQN, TD3, SAC [124], dueling DQN [126], gradient boosting decision tree [133], deep RL using Q-masking and autonomically generated curriculum-based DRL [139]. De- spite solving most of the tasks for safe deployment in level 4 or 5 AVS, challenges remain, such as complex training cost, lack of proper surrounding vehicles’ behavior analysis and unfinished case in complex scenarios. Some problems remain to be resolved for better outcomes, such as the requirement of a larger labelled dataset [57], struggle to classify in blurry visual conditions [49] and small traffic signs from a far field of view [51], background complexity [48] and detecting two traffic signs rather than one, which occurred for different Appl. Sci. 2022, 12, 6831 43 of 51 locations of the proposed region [47]. Apart from these, one of the most complicated tasks for AVS, only vision-based path and motion planning were analyzed by reviewing approaches such as deep inverse reinforcement learning, DQN time-to-go method, MPC, Dijkstra with TEB method, DNN, discrete optimizer-based approach, artificial potential field, MPC with LSTM-RNN, advance dynamic window using, 3D-CNN, spatio-temporal LSTM and fuzzy logic, where solutions were provided by avoiding cost function and manual labelling, reducing the limitation of rule-based methods for safe navigation [164] and better path planning for intersections [165], motion planning by analyzing risks and predicting motions of surrounding vehicles [166], hazard detection-based safe naviga- tion [168], avoiding obstacles for smooth planning in multilane scenarios [169], decreasing computational cost [170] and path planning by replicating human-like control thinking in ambiguous circumstances. Nevertheless, these approaches faced challenges such as lack of live testing, low accuracy in far predicted horizon, impaired performance in complex situations or being limited to non-rule-based approaches and constrained kinematics or even difficulty in establishing a rule base to tackle unstructured conditions. Finally, to visualize overlaying outcomes generated from the previous methods su- perimposed on the front head-up display or smart windshield, augmented reality-based approaches combining deep learning methods were reviewed in the last section. AR-HUD based solutions such as 3D surface reconstruction, object marking, path overlaying, re- ducing drivers’ attention, boosting visualization in tough hazy or low-light conditions by overlapping lanes, traffic signs as well as on-road objects to reduce accidents using deep CNN, RANSAC, TTC methods and so on. However, there are still many challenges for practical execution, such as human adoption of AR-based HUD UI, limited visualization in bright daytime conditions, overlapping non-superior objects as well as visualization delay for fast moving on-road objects. In summary, the literature review established for vision-based deep learning approaches of 10 taxonomies for AVS with discussion of out- comes, challenges and limitations could be a pathway to improve and rapidly develop cost-efficient level 4 or 5 AVS without depending on expensive and complex sensor fusion. 7. Conclusions The results of the mixed method studies in the field of implementation and application of deep learning algorithms for autonomous driving systems help us to achieve a clear understanding of the future of transportation. These results prove that it has the ability to provide intelligent mobility for our constantly evolving modern world as deep learning was one of the key components to resolve the limitations and bottlenecking of traditional techniques. Despite containing a good number of studies on autonomous driving systems, only a few make an impact on recent developments in the autonomous driving industry. To overcome this challenge and build a safer and more secure sensor-independent transporta- tion system with the aim of building infrastructure of futuristic smart cities, in this paper, through a systematic review of the literature, studies of AV were selected that used deep learning and the field was reviewed in terms of decision making, path planning and navi- gation, controlling, prediction and visualizing the outcomes in augmented reality-based head-up displays. We analyzed the existing proposal of deep learning models in real-world implementation for AVS, described the methodologies, designed the flow of solutions for the limitations of other methodologies, and compared outcomes and evaluation techniques. Nevertheless, as the research field of autonomous driving systems is still growing, many of the theoretical methodologies were not applied practically, but along with the research trend of this expanding field, these are potentially excellent solutions that require further development. Thus, the large-scale distributions of the paper in the major areas of au- tonomous driving systems will be essential for further research and development of the autonomous vehicle industry into a cost-efficient, secure intelligent transport system. Funding: This research was funded by Universiti Kebangsaan Malaysia [GUP-2020-060]. Institutional Review Board Statement: Not applicable. Appl. Sci. 2022, 12, 6831 44 of 51 Informed Consent Statement: Not applicable. Data Availability Statement: Not applicable. Conflicts of Interest: The authors declare no conflict of interest. References 1. Alawadhi, M.; Almazrouie, J.; Kamil, M.; Khalil, K.A. A systematic literature review of the factors influencing the adoption of autonomous driving. Int. J. Syst. Assur. Eng. Manag. 2020, 11, 1065–1082. [CrossRef] 2. Pandey, P.; Shukla, A.; Tiwari, R. Three-dimensional path planning for unmanned aerial vehicles using glowworm swarm optimization algorithm. Int. J. Syst. Assur. Eng. Manag. 2018, 9, 836–852. [CrossRef] 3. Dirsehan, T.; Can, C. Examination of trust and sustainability concerns in autonomous vehicle adoption. Technol. Soc. 2020, 63, 101361. [CrossRef] 4. Khamis, N.K.; Deros, B.M.; Nuawi, M.Z.; Omar, R.B. Driving fatigue among long distance heavy vehicle drivers in Klang Valley, Malaysia. Appl. Mech. Mater. 2014, 663, 567–573. [CrossRef] 5. Naujoks, F.; Wiedemann, K.; Schömig, N.; Hergeth, S.; Keinath, A. Towards guidelines and verification methods for automated vehicle HMIs. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 121–136. [CrossRef] 6. Li, D.; Wagner, P. Impacts of gradual automated vehicle penetration on motorway operation: A comprehensive evaluation. Eur. Transp. Res. Rev. 2019, 11, 36. [CrossRef] 7. Mutz, F.; Veronese, L.P.; Oliveira-Santos, T.; De Aguiar, E.; Cheein, F.A.A.; De Souza, A.F. Large-scale mapping in complex field scenarios using an autonomous car. Expert Syst. Appl. 2016, 46, 439–462. [CrossRef] 8. Gandia, R.M.; Antonialli, F.; Cavazza, B.H.; Neto, A.M.; Lima, D.A.d.; Sugano, J.Y.; Nicolai, I.; Zambalde, A.L. Autonomous vehicles: Scientometric and bibliometric review. Transp. Rev. 2019, 39, 9–28. [CrossRef] 9. Maurer, M.; Christian, G.; Lenz, B.; Winner, H. Autonomous Driving: Technical, Legal and Social Aspects; Springer Nature: Berlin/Heidelberg, Germany, 2016. [CrossRef] 10. Levinson, J.; Askeland, J.; Becker, J.; Dolson, J.; Held, D.; Kammel, S.; Kolter, J.Z.; Langer, D.; Pink, O.; Pratt, V. Towards fully autonomous driving: Systems and algorithms. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 163–168. 11. Yu, H.; Yang, S.; Gu, W.; Zhang, S. Baidu driving dataset and end-to-end reactive control model. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 341–346. 12. Hashim, H.; Omar, M. Towards autonomous vehicle implementation: Issues and opportunities. J. Soc. Automot. Eng. Malays. 2017, 1, 111–123. 13. Chen, L.; Li, Q.; Li, M.; Zhang, L.; Mao, Q. Design of a multi-sensor cooperation travel environment perception system for autonomous vehicle. Sensors 2012, 12, 12386–12404. [CrossRef] 14. Li, Q.; Chen, L.; Li, M.; Shaw, S.-L.; Nüchter, A. A sensor-fusion drivable-region and lane-detection system for autonomous vehicle navigation in challenging road scenarios. IEEE Trans. Veh. Technol. 2013, 63, 540–555. [CrossRef] 15. Rahman, A.H.A.; Ariffin, K.A.Z.; Sani, N.S.; Zamzuri, H. Pedestrian Detection using Triple Laser Range Finders. Int. J. Electr. Comput. Eng. 2017, 7, 3037. [CrossRef] 16. Wang, H.; Wang, B.; Liu, B.; Meng, X.; Yang, G. Pedestrian recognition and tracking using 3D LiDAR for autonomous vehicle. Robot. Auton. Syst. 2017, 88, 71–78. [CrossRef] 17. Wang, L.; Zhang, Y.; Wang, J. Map-based localization method for autonomous vehicles using 3D-LIDAR. IFAC-Pap. 2017, 50, 276–281. 18. Kong, P.-Y. Computation and Sensor Offloading for Cloud-Based Infrastructure-Assisted Autonomous Vehicles. IEEE Syst. J. 2020, 14, 3360–3370. [CrossRef] 19. Zhao, J.; Xu, H.; Liu, H.; Wu, J.; Zheng, Y.; Wu, D. Detection and tracking of pedestrians and vehicles using roadside LiDAR sensors. Transp. Res. Part C Emerg. Technol. 2019, 100, 68–87. [CrossRef] 20. Huval, B.; Wang, T.; Tandon, S.; Kiske, J.; Song, W.; Pazhayampallil, J.; Andriluka, M.; Rajpurkar, P.; Migimatsu, T.; Cheng, Y.R. An empirical evaluation of deep learning on highway driving. arXiv 2015, arXiv:01716. 21. Leon, F.; Gavrilescu, M. A review of tracking, prediction and decision making methods for autonomous driving. arXiv 2019, arXiv:07707. 22. Ma, Y.; Wang, Z.; Yang, H.; Yang, L. Artificial intelligence applications in the development of autonomous vehicles: A survey. IEEE/CAA J. Autom. Sin. 2020, 7, 315–329. [CrossRef] 23. Paden, B.; Cáp, M.; Yong, S.Z.; Yershov, D.; Frazzoli, E. A survey of motion planning and control techniques for self-driving urban vehicles. IEEE Trans. Intell. Veh. 2016, 1, 33–55. [CrossRef] 24. Yurtsever, E.; Lambert, J.; Carballo, A.; Takeda, K. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access 2020, 8, 58443–58469. [CrossRef] 25. Schwarting, W.; Alonso-Mora, J.; Rus, D. Planning and decision-making for autonomous vehicles. Annu. Rev. Control Robot. Auton. Syst. 2018, 1, 187–210. [CrossRef] 26. Van Brummelen, J.; O’Brien, M.; Gruyer, D.; Najjaran, H. Autonomous vehicle perception: The technology of today and tomorrow. Transp. Res. Part C Emerg. Technol. 2018, 89, 384–406. [CrossRef] Appl. Sci. 2022, 12, 6831 45 of 51 27. Badue, C.; Guidolini, R.; Carneiro, R.V.; Azevedo, P.; Cardoso, V.B.; Forechi, A.; Jesus, L.; Berriel, R.; Paixao, T.M.; Mutz, F.J. Self-driving cars: A survey. Expert Syst. Appl. 2021, 165, 113816. [CrossRef] 28. Sirohi, D.; Kumar, N.; Rana, P.S. Convolutional neural networks for 5G-enabled intelligent transportation system: A systematic review. Comput. Commun. 2020, 153, 459–498. [CrossRef] 29. Zhang, Y.; Qi, Y.; Liu, J.; Wang, Y. Decade of Vision-Based Pedestrian Detection for Self-Driving: An Experimental Survey and Evaluation; 0148-7191; SAE Publishing: Thousand Oaks, CA, USA, 2018. 30. Ni, J.; Chen, Y.; Chen, Y.; Zhu, J.; Ali, D.; Cao, W. A survey on theories and applications for self-driving cars based on deep learning methods. Appl. Sci. 2020, 10, 2749. [CrossRef] 31. Fayyad, J.; Jaradat, M.A.; Gruyer, D.; Najjaran, H. Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review. Sensors 2020, 20, 4220. [CrossRef] 32. Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Al Sallab, A.A.; Yogamani, S.; Pérez, P. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4909–4926. [CrossRef] 33. Grigorescu, S.; Trasnea, B.; Cocias, T.; Macesanu, G. A survey of deep learning techniques for autonomous driving. J. Field Robot. 2020, 37, 362–386. [CrossRef] 34. Hu, H.-N.; Cai, Q.-Z.; Wang, D.; Lin, J.; Sun, M.; Krahenbuhl, P.; Darrell, T.; Yu, F. Joint monocular 3D vehicle detection and tracking. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 5390–5399. 35. Sang, J.; Wu, Z.; Guo, P.; Hu, H.; Xiang, H.; Zhang, Q.; Cai, B. An improved YOLOv2 for vehicle detection. Sensors 2018, 18, 4272. [CrossRef] 36. Shakya, V.; Makwana, R.R.S. Feature selection based intrusion detection system using the combination of DBSCAN, K-Mean++ and SMO algorithms. In Proceedings of the 2017 International Conference on Trends in Electronics and Informatics (ICEI), Tirunelveli, India, 11–12 May 2017; pp. 928–932. 37. Ohn-Bar, E.; Trivedi, M.M. Learning to detect vehicles by clustering appearance patterns. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2511–2521. [CrossRef] 38. Chen, W.; Sun, Q.; Wang, J.; Dong, J.-J.; Xu, C. A novel model based on AdaBoost and deep CNN for vehicle classification. IEEE Access 2018, 6, 60445–60455. [CrossRef] 39. Bautista, C.M.; Dy, C.A.; Mañalac, M.I.; Orbe, R.A.; Cordel, M. Convolutional neural network for vehicle detection in low resolution traffic videos. In Proceedings of the 2016 IEEE Region 10 Symposium (TENSYMP), Bali, Indonesia, 9–11 May 2016; pp. 277–281. 40. Lee, H.J.; Moon, B.; Kim, G. Hierarchical scheme of vehicle detection and tracking in nighttime urban environment. Int. J. Automot. Technol. 2018, 19, 369–377. [CrossRef] 41. Hu, X.; Xu, X.; Xiao, Y.; Chen, H.; He, S.; Qin, J.; Heng, P.-A. SINet: A scale-insensitive convolutional neural network for fast vehicle detection. IEEE Trans. Intell. Transp. Syst. 2018, 20, 1010–1019. [CrossRef] 42. Wang, Y.; Liu, Z.; Deng, W. Anchor generation optimization and region of interest assignment for vehicle detection. Sensors 2019, 19, 1089. [CrossRef] 43. Suhao, L.; Jinzhao, L.; Guoquan, L.; Tong, B.; Huiqian, W.; Yu, P. Vehicle type detection based on deep learning in traffic scene. Procedia Comput. Sci. 2018, 131, 564–572. [CrossRef] 44. Liu, J.; Zhang, R. Vehicle Detection and Ranging Using Two Different Focal Length Cameras. J. Sens. 2020, 2020, 4372847. 45. Leung, H.K.; Chen, X.-Z.; Yu, C.-W.; Liang, H.-Y.; Wu, J.-Y.; Chen, Y.-L. A Deep-Learning-Based Vehicle Detection Approach for Insufficient and Nighttime Illumination Conditions. Appl. Sci. 2019, 9, 4769. [CrossRef] 46. Hu, W.; Zhuo, Q.; Zhang, C.; Li, J. Fast branch convolutional neural network for traffic sign recognition. IEEE Intell. Transp. Syst. Mag. 2017, 9, 114–126. [CrossRef] 47. Jung, S.; Lee, U.; Jung, J.; Shim, D.H. Real-time Traffic Sign Recognition system with deep convolutional neural network. In Proceedings of the 2016 13th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Xi’an, China, 19–22 August 2016; pp. 31–34. 48. Cao, J.; Song, C.; Peng, S.; Xiao, F.; Song, S. Improved traffic sign detection and recognition algorithm for intelligent vehicles. Sensors 2019, 19, 4021. [CrossRef] 49. Natarajan, S.; Annamraju, A.K.; Baradkar, C.S. Traffic sign recognition using weighted multi-convolutional neural network. IET Intell. Transp. Syst. 2018, 12, 1396–1405. [CrossRef] 50. Wang, G.; Ren, G.; Jiang, L.; Quan, T. Hole-based traffic sign detection method for traffic signs with red rim. Vis. Comput. 2014, 30, 539–551. [CrossRef] 51. Zhang, J.; Huang, M.; Jin, X.; Li, X. A real-time chinese traffic sign detection algorithm based on modified YOLOv2. Algorithms 2017, 10, 127. [CrossRef] 52. Liu, X.; Yan, W.Q. Traffic-light sign recognition using Capsule network. Multimed. Tools Appl. 2021, 80, 15161–15171. [CrossRef] 53. Bach, M.; Stumper, D.; Dietmayer, K. Deep convolutional traffic light recognition for automated driving. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 851–858. 54. Weber, M.; Wolf, P.; Zöllner, J.M. DeepTLR: A single deep convolutional network for detection and classification of traffic lights. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 19–22 June 2016; pp. 342–348. Appl. Sci. 2022, 12, 6831 46 of 51 55. Liu, X.; Neuyen, M.; Yan, W.Q. Vehicle-related scene understanding using deep learning. In Asian Conference on Pattern Recognition; Springer: Sinapore, 2020; pp. 61–73. 56. Lee, E.; Kim, D. Accurate traffic light detection using deep neural network with focal regression loss. Image Vis. Comput. 2019, 87, 24–36. [CrossRef] 57. Behrendt, K.; Novak, L.; Botros, R. A deep learning approach to traffic lights: Detection, tracking, and classification. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 1370–1377. 58. Yanagisawa, M.; Swanson, E.; Najm, W.G. Target Crashes and Safety Benefits Estimation Methodology for Pedestrian Crash Avoid- ance/Mitigation Systems; United States National Highway Traffic Safety Administration: Sacramento, CA, USA, 2014. 59. Baharuddin, M.; Khamis, N.; Kassim, K.A.; Mansor, M.R.A. Autonomous Emergency Brake (AEB) for Pedestrian for ASEAN NCAP Safety Rating Consideration: A Review. J. Soc. Automot. Eng. Malays. 2019, 3, 63–73. 60. Angelova, A.; Krizhevsky, A.; Vanhoucke, V. Pedestrian detection with a large-field-of-view deep network. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 704–711. 61. Zhan, H.; Liu, Y.; Cui, Z.; Cheng, H. Pedestrian Detection and Behavior Recognition Based on Vision. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 771–776. 62. Qiu, Z.; Zhao, N.; Zhou, L.; Wang, M.; Yang, L.; Fang, H.; He, Y.; Liu, Y. Vision-based moving obstacle detection and tracking in paddy field using improved yolov3 and deep SORT. Sensors 2020, 20, 4082. [CrossRef] 63. Ghosh, S.; Amon, P.; Hutter, A.; Kaup, A. Reliable pedestrian detection using a deep neural network trained on pedestrian counts. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 685–689. 64. Wang, X.; Jung, C.; Hero, A.O. Part-level fully convolutional networks for pedestrian detection. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2267–2271. 65. Hou, Y.-L.; Song, Y.; Hao, X.; Shen, Y.; Qian, M.; Chen, H. Multispectral pedestrian detection based on deep convolutional neural networks. Infrared Phys. Technol. 2018, 94, 69–77. [CrossRef] 66. Yamada, K. Pedestrian detection with a resolution-aware convolutional network. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 591–596. 67. Zhang, J.; Xiao, J.; Zhou, C.; Peng, C. A multi-class pedestrian detection network for distorted pedestrians. In Proceedings of the 2018 13th IEEE Conference on Industrial Electronics and Applications (ICIEA), Wuhan, China, 31 May–June 2018; pp. 1079–1083. 68. Dong, P.; Wang, W. Better region proposals for pedestrian detection with R-CNN. In Proceedings of the 2016 Visual Communica- tions and Image Processing (VCIP), Chengdu, China, 27–30 November 2016; pp. 1–4. 69. Wood, J.M. Nighttime driving: Visual, lighting and visibility challenges. Ophthalmic Physiol. Opt. 2020, 40, 187–201. [CrossRef] 70. Baratian-Ghorghi, F.; Zhou, H.; Shaw, J. Overview of wrong-way driving fatal crashes in the United States. ITE J. 2014, 84, 41–47. 71. Mohammed, A.A.; Ambak, K.; Mosa, A.M.; Syamsunur, D. Traffic accidents in Iraq: An analytical study. J. Adv. Res. Civ. Environ. Eng. 2018, 5, 10–22. 72. Ding, B.; Liu, Z.; Sun, Y. Pedestrian Detection in Haze Environments Using Dark Channel Prior and Histogram of Oriented Gradi- ent. In Proceedings of the 2018 Eighth International Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC), Harbin, China, 19–21 July 2018; pp. 1003–1008. 73. Huang, S.-C.; Ye, J.-H.; Chen, B.-H. An advanced single-image visibility restoration algorithm for real-world hazy scenes. IEEE Trans. Ind. Electron. 2014, 62, 2962–2972. [CrossRef] 74. Li, G.; Yang, Y.; Qu, X. Deep learning approaches on pedestrian detection in hazy weather. IEEE Trans. Ind. Electron. 2019, 67, 8889–8899. [CrossRef] 75. Ju, M.; Luo, H.; Wang, Z.; Hui, B.; Chang, Z. The application of improved YOLO V3 in multi-scale target detection. Appl. Sci. 2019, 9, 3775. [CrossRef] 76. Xu, Z.; Vong, C.-M.; Wong, C.-C.; Liu, Q. Ground Plane Context Aggregation Network for Day-and-Night on Vehicular Pedestrian Detection. IEEE Trans. Intell. Transp. Syst. 2020, 22, 6395–6406. [CrossRef] 77. Zhao, Y.; Yuan, Z.; Chen, B. Accurate pedestrian detection by human pose regression. IEEE Trans. Image Process. 2019, 29, 1591–1605. [CrossRef] 78. Chen, Z.; Huang, X. Pedestrian detection for autonomous vehicle using multi-spectral cameras. IEEE Trans. Intell. Veh. 2019, 4, 211–219. [CrossRef] 79. Alvarez, J.M.; Gevers, T.; LeCun, Y.; Lopez, A.M. Road scene segmentation from a single image. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 376–389. 80. Dong, M.; Zhao, X.; Fan, X.; Shen, C.; Liu, Z. Combination of modified U-Net and domain adaptation for road detection. IET Image Process. 2019, 13, 2735–2743. [CrossRef] 81. Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep Neural Network for Structural Prediction and Lane Detection in Traffic Scene. IEEE Trans. Neural Netw. Learn Syst. 2017, 28, 690–703. [CrossRef] 82. Fakhfakh, M.; Chaari, L.; Fakhfakh, N. Bayesian curved lane estimation for autonomous driving. J. Ambient Intell. Humaniz. Comput. 2020, 11, 4133–4143. [CrossRef] 83. Yang, S.; Wu, J.; Shan, Y.; Yu, Y.; Zhang, S. A Novel Vision-Based Framework for Real-Time Lane Detection and Tracking; SAE Technical Paper: Warrendale, PA, USA, 2019. Appl. Sci. 2022, 12, 6831 47 of 51 84. Mongkonyong, P.; Nuthong, C.; Siddhichai, S.; Yamakita, M. Lane detection using randomized Hough transform. In Proceedings of the 8th TSME-International Conference on Mechanical Engineering, Bangkok, Thailand, 12–15 December 2017. 85. Javadi, M.; Hannan, M.; Samad, S.; Hussain, A. A robust vision-based lane boundaries detection approach for intelligent vehicles. Inf. Technol. J. 2012, 11, 1184–1192. [CrossRef] 86. El Hajjouji, I.; Mars, S.; Asrih, Z.; El Mourabit, A. A novel fpga implementation of hough transform for straight lane detection. Eng. Sci. Technol. Int. J. 2020, 23, 274–280. [CrossRef] 87. Dong, Y.; Patil, S.; van Arem, B.; Farah, H. A Hybrid Spatial-temporal Deep Learning Architecture for Lane Detection. arXiv 2021, arXiv:04079. [CrossRef] 88. Liu, L.; Chen, X.; Zhu, S.; Tan, P. CondLaneNet: A Top-to-down Lane Detection Framework Based on Conditional Convolution. arXiv 2021, arXiv:05003. 89. Dorj, B.; Hossain, S.; Lee, D.-J. Highly Curved Lane Detection Algorithms Based on Kalman Filter. Appl. Sci. 2020, 10, 2372. [CrossRef] 90. Son, Y.; Lee, E.S.; Kum, D. Robust multi-lane detection and tracking using adaptive threshold and lane classification. Mach. Vis. Appl. 2019, 30, 111–124. [CrossRef] 91. Wang, W.; Lin, H.; Wang, J. CNN based lane detection with instance segmentation in edge-cloud computing. J. Cloud Comput. 2020, 9, 27. [CrossRef] 92. Ye, Y.Y.; Hao, X.L.; Chen, H.J. Lane detection method based on lane structural analysis and CNNs. IET Intell. Transp. Syst. 2018, 12, 513–520. [CrossRef] 93. Zou, Q.; Jiang, H.; Dai, Q.; Yue, Y.; Chen, L.; Wang, Q. Robust lane detection from continuous driving scenes using deep neural networks. IEEE Trans. Veh. Technol. 2019, 69, 41–54. [CrossRef] 94. John, V.; Liu, Z.; Mita, S.; Guo, C.; Kidono, K. Real-time road surface and semantic lane estimation using deep features. Signal Image Video Process. 2018, 12, 1133–1140. [CrossRef] 95. Chen, P.-R.; Lo, S.-Y.; Hang, H.-M.; Chan, S.-W.; Lin, J.-J. Efficient road lane marking detection with deep learning. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. 96. Ghafoorian, M.; Nugteren, C.; Baka, N.; Booij, O.; Hofmann, M. El-gan: Embedding loss driven generative adversarial networks for lane detection. In European Conference on Computer Vision (ECCV); Springer: Cham, Switzerland, 2018. 97. He, X.; Duan, Z.; Chen, C.; You, F. Video-based lane detection and tracking during night. In Proceedings of the 19th COTA International Conference of Transportation Professionals, Nanjing, China, 6–8 July 2019; pp. 5794–5807. 98. Neven, D.; De Brabandere, B.; Georgoulis, S.; Proesmans, M.; Van Gool, L. Towards end-to-end lane detection: An instance segmentation approach. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 286–291. 99. Kim, J.; Kim, J.; Jang, G.-J.; Lee, M. Fast learning method for convolutional neural networks using extreme learning machine and its application to lane detection. Neural Netw. 2017, 87, 109–121. [CrossRef] [PubMed] 100. Van Gansbeke, W.; De Brabandere, B.; Neven, D.; Proesmans, M.; Van Gool, L. End-to-end lane detection through differentiable least-squares fitting. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019. 101. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection cnns by self attention distillation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1013–1021. 102. Geiger, A.; Lauer, M.; Wojek, C.; Stiller, C.; Urtasun, R. 3D Traffic Scene Understanding From Movable Platforms. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1012–1025. [CrossRef] [PubMed] 103. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. 2020, 43, 3349–3364. [CrossRef] [PubMed] 104. Wang, W.; Zhou, T.; Yu, F.; Dai, J.; Konukoglu, E.; Van Gool, L. Exploring cross-image pixel contrast for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7303–7313. 105. Zhao, X.; Vemulapalli, R.; Mansfield, P.A.; Gong, B.; Green, B.; Shapira, L.; Wu, Y. Contrastive Learning for Label Efficient Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10623–10633. 106. Tang, P.; Wang, H.; Kwong, S. G-MS2F: GoogLeNet based multi-stage feature fusion of deep CNN for scene recognition. Neurocomputing 2017, 225, 188–197. [CrossRef] 107. Wang, L.; Guo, S.; Huang, W.; Xiong, Y.; Qiao, Y. Knowledge guided disambiguation for large-scale scene classification with multi-resolution CNNs. IEEE Trans. Image Process. 2017, 26, 2055–2068. [CrossRef] 108. Byeon, W.; Breuel, T.M.; Raue, F.; Liwicki, M. Scene labeling with lstm recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3547–3555. 109. Fu, J.; Liu, J.; Li, Y.; Bao, Y.; Yan, W.; Fang, Z.; Lu, H. Contextual deconvolution network for semantic segmentation. Pattern Recognit. 2020, 101, 107152. [CrossRef] Appl. Sci. 2022, 12, 6831 48 of 51 110. Oeljeklaus, M.; Hoffmann, F.; Bertram, T. A combined recognition and segmentation model for urban traffic scene understanding. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. 111. Xue, J.-R.; Fang, J.-W.; Zhang, P. A survey of scene understanding by event reasoning in autonomous driving. Int. J. Autom. Comput. 2018, 15, 249–266. [CrossRef] 112. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial cnn for traffic scene understanding. arXiv 2017, arXiv:06080. 113. Mou, L.; Xie, H.; Mao, S.; Zhao, P.; Chen, Y. Vision-based vehicle behaviour analysis: A structured learning approach via convolutional neural networks. IET Intell. Transp. Syst. 2020, 14, 792–801. [CrossRef] 114. Jeon, H.-S.; Kum, D.-S.; Jeong, W.-Y. Traffic scene prediction via deep learning: Introduction of multi-channel occupancy grid map as a scene representation. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1496–1501. 115. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [CrossRef] 116. Theofilatos, A.; Chen, C.; Antoniou, C. Comparing Machine Learning and Deep Learning Methods for Real-Time Crash Prediction. Transp. Res. Rec. J. Transp. Res. Board 2019, 2673, 169–178. [CrossRef] 117. Huegle, M.; Kalweit, G.; Werling, M.; Boedecker, J. Dynamic interaction-aware scene understanding for reinforcement learning in autonomous driving. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 4329–4335. 118. Nguyen, N.T.H.; Le, T.H.; Perry, S.; Nguyen, T.T. Pavement crack detection using convolutional neural network. In Proceedings of the Ninth International Symposium on Information and Communication Technology, Sharjah, United Arab Emirates, 18–19 November 2019; pp. 251–256. 119. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road crack detection using deep convolutional neural network. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3708–3712. 120. Anand, S.; Gupta, S.; Darbari, V.; Kohli, S. Crack-pot: Autonomous road crack and pothole detection. In Proceedings of the 2018 Digital Image Computing: Techniques and Applications (DICTA), Canberra, ACT, Australia, 10–13 December 2018; pp. 1–6. 121. Yu, C.; Wang, X.; Xu, X.; Zhang, M.; Ge, H.; Ren, J.; Sun, L.; Chen, B.; Tan, G. Distributed multiagent coordinated learning for autonomous driving in highways based on dynamic coordination graphs. IEEE Trans. Intell. Transp. Syst. 2019, 21, 735–748. [CrossRef] 122. Zhang, J.; Liao, Y.; Wang, S.; Han, J.J.A.S. Study on driving decision-making mechanism of autonomous vehicle based on an optimized support vector machine regression. Apply Sci. 2018, 8, 13. [CrossRef] 123. Fu, Y.; Li, C.; Yu, F.R.; Luan, T.H.; Zhang, Y. A Decision-Making Strategy for Vehicle Autonomous Braking in Emergency via Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2020, 69, 5876–5888. [CrossRef] 124. Munk, J.; Kober, J.; Babuška, R. Learning state representation for deep actor-critic control. In Proceedings of the 2016 IEEE 55th Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 4667–4673. 125. Chen, J.; Yuan, B.; Tomizuka, M. Model-free deep reinforcement learning for urban autonomous driving. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 2765–2771. 126. Liao, J.; Liu, T.; Tang, X.; Mu, X.; Huang, B.; Cao, D. Decision-making Strategy on Highway for Autonomous Vehicles using Deep Reinforcement Learning. IEEE Access 2020, 8, 177804–177814. [CrossRef] 127. Hoel, C.-J.; Driggs-Campbell, K.; Wolff, K.; Laine, L.; Kochenderfer, M.J. Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving. IEEE Trans. Intell. Veh. 2020, 5, 294–305. [CrossRef] 128. Sezer, V. Intelligent decision making for overtaking maneuver using mixed observable Markov decision process. J. Intell. Transp. Syst. 2017, 22, 201–217. [CrossRef] 129. Ngai, D.C.K.; Yung, N.H.C. A multiple-goal reinforcement learning method for complex vehicle overtaking maneuvers. IEEE Trans. Intell. Transp. Syst. 2011, 12, 509–522. [CrossRef] 130. Brannstrom, M.; Sandblom, F.; Hammarstrand, L. A Probabilistic Framework for Decision-Making in Collision Avoidance Systems. IEEE Trans. Intell. Transp. Syst. 2013, 14, 637–648. [CrossRef] 131. Zhu, B.; Liu, S.; Zhao, J. A Lane-Changing Decision-Making Method for Intelligent Vehicle Based on Acceleration Field; SAE Technical Paper Series; SAE Publishing: Thousand Oaks, CA, USA, 2020. 132. Wang, W.; Qie, T.; Yang, C.; Liu, W.; Xiang, C.; Huang, K.J.I.T.o.I.E. An intelligent Lane-Changing Behavior Prediction and Decision-Making strategy for an Autonomous Vehicle. IEEE Trans. Ind. Electron. 2021, 69, 2927–2937. [CrossRef] 133. Gómez-Huélamo, C.; Egido, J.D.; Bergasa, L.M.; Barea, R.; López-Guillén, E.; Arango, F.; Araluce, J.; López, J. Train here, drive there: Simulating real-world use cases with fully-autonomous driving architecture in carla simulator. In Workshop of Physical Agents; Springer: Cham, Switzerland, 2020; pp. 44–59. 134. Mukadam, M.; Cosgun, A.; Nakhaei, A.; Fujimura, K. Tactical decision making for lane changing with deep reinforcement learning. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. 135. Wang, J.; Zhang, Q.; Zhao, D.; Chen, Y. Lane change decision-making through deep reinforcement learning with rule-based constraints. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–6. Appl. Sci. 2022, 12, 6831 49 of 51 136. Chae, H.; Kang, C.M.; Kim, B.; Kim, J.; Chung, C.C.; Choi, J.W. Autonomous braking system via deep reinforcement learning. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. 137. Wang, W.; Xi, J.; Zhao, D. Learning and inferring a driver ’s braking action in car-following scenarios. IEEE Trans. Veh. Technol. 2018, 67, 3887–3899. [CrossRef] 138. Chen, J.; Chen, J.; Zhang, R.; Hu, X. Towards Brain-inspired System: Deep Recurrent Reinforcement Learning for Simulated Self-driving Agent. Front. Neurorobot. 2019, 13, 40. [CrossRef] [PubMed] 139. Qiao, Z.; Muelling, K.; Dolan, J.M.; Palanisamy, P.; Mudalige, P. Automatically generated curriculum based reinforcement learning for autonomous vehicles in urban environment. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1233–1238. 140. Li, G.; Li, S.; Li, S.; Qu, X. Continuous decision-making for autonomous driving at intersections using deep deterministic policy gradient. IET Intell. Transp. Syst. 2021. [CrossRef] 141. Deshpande, N.; Vaufreydaz, D.; Spalanzani, A. Behavioral decision-making for urban autonomous driving in the presence of pedestrians using Deep Recurrent Q-Network. In Proceedings of the 16th International Conference on Control, Automation, Robotics and Vision (ICARCV), Shenzhen, China, 13–15 December 2020. 142. Bin Issa, R.; Das, M.; Rahman, M.S.; Barua, M.; Rhaman, M.K.; Ripon, K.S.N.; Alam, M.G.R. Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment. Sensors 2021, 21, 1468. [CrossRef] 143. Mo, S.; Pei, X.; Wu, C. Safe reinforcement learning for autonomous vehicle using monte carlo tree search. IEEE Trans. Intell. Transp. Syst. 2021. [CrossRef] 144. An, D.; Liu, J.; Zhang, M.; Chen, X.; Chen, M.; Sun, H. Uncertainty Modeling and Runtime Verification for Autonomous Vehicles Driving Control: A Machine Learning-based Approach. J. Syst. Softw. 2020, 167, 110617. [CrossRef] 145. Jovanovic, ´ A.; Kwiatkowska, M. Parameter synthesis for probabilistic timed automata using stochastic game abstractions. Theor. Comput. Sci. 2018, 735, 64–81. [CrossRef] 146. Pan, Y.; Cheng, C.-A.; Saigol, K.; Lee, K.; Yan, X.; Theodorou, E.A.; Boots, B. Imitation learning for agile autonomous driving. Int. J. Robot. Res. 2020, 39, 286–302. [CrossRef] 147. Chen, Z.; Huang, X. End-to-end learning for lane keeping of self-driving cars. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 1856–1860. 148. Yang, Z.; Zhang, Y.; Yu, J.; Cai, J.; Luo, J. End-to-end multi-modal multi-task vehicle control for self-driving cars with visual perceptions. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2289–2294. 149. Lee, M.-J.; Ha, Y.-G. Autonomous Driving Control Using End-to-End Deep Learning. In Proceedings of the 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), Busan, Korea, 19–22 February 2020; pp. 470–473. 150. Bojarski, M.; Del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J. End to end learning for self-driving cars. arXiv 2016, arXiv:07316. 151. Chen, C.; Seff, A.; Kornhauser, A.; Xiao, J. Deepdriving: Learning affordance for direct perception in autonomous driving. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2722–2730. 152. Zhang, Y.; Sun, P.; Yin, Y.; Lin, L.; Wang, X. Human-like autonomous vehicle speed control by deep reinforcement learning with double Q-learning. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1251–1256. 153. Chi, L.; Mu, Y. Deep steering: Learning end-to-end driving model from spatial and temporal visual cues. arXiv 2017, arXiv:03798. 154. Wu, T.; Luo, A.; Huang, R.; Cheng, H.; Zhao, Y. End-to-End Driving Model for Steering Control of Autonomous Vehicles with Future Spatiotemporal Features. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 950–955. 155. Bojarski, M.; Yeres, P.; Choromanska, A.; Choromanski, K.; Firner, B.; Jackel, L.; Muller, U. Explaining how a deep neural network trained with end-to-end learning steers a car. arXiv 2017, arXiv:07911. 156. Wu, A.; Rubaiyat, A.H.M.; Anton, C.; Alemzadeh, H. Model Fusion: Weighted N-version programming for resilient autonomous vehicle steering control. In Proceedings of the 2018 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW), Memphis, TN, USA, 15–18 October 2018; pp. 144–145. 157. Huang, X.; McGill, S.G.; DeCastro, J.A.; Fletcher, L.; Leonard, J.J.; Williams, B.C.; Rosman, G. Diversitygan: Diversity-aware vehicle motion prediction via latent semantic sampling. IEEE Robot. Autom. Lett. 2020, 5, 5089–5096. [CrossRef] 158. Wang, D.; Wen, J.; Wang, Y.; Huang, X.; Pei, F. End-to-End Self-Driving Using Deep Neural Networks with Multi-auxiliary Tasks. Automot. Innov. 2019, 2, 127–136. [CrossRef] 159. Toromanoff, M.; Wirbel, E.; Wilhelm, F.; Vejarano, C.; Perrotton, X.; Moutarde, F. End to end vehicle lateral control using a single fisheye camera. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 3613–3619. 160. Smolyakov, M.; Frolov, A.; Volkov, V.; Stelmashchuk, I. Self-Driving Car Steering Angle Prediction Based On Deep Neural Network An Example Of CarND Udacity Simulator. In Proceedings of the 2018 IEEE 12th International Conference on Application of Information and Communication Technologies (AICT), Almaty, Kazakhstan, 17–19 October 2018; pp. 1–5. Appl. Sci. 2022, 12, 6831 50 of 51 161. Sharma, S.; Tewolde, G.; Kwon, J. Lateral and Longitudinal Motion Control of Autonomous Vehicles using Deep Learning. In Proceedings of the 2019 IEEE International Conference on Electro Information Technology (EIT), Brookings, SD, USA, 20–22 May 2019; pp. 1–5. 162. Zhu, Q.; Huang, Z.; Sun, Z.; Liu, D.; Dai, B. Reinforcement learning based throttle and brake control for autonomous vehicle following. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 6657–6662. 163. Xiao, Y.; Codevilla, F.; Gurram, A.; Urfalioglu, O.; López, A.M. Multimodal end-to-end autonomous driving. IEEE Trans. Intell. Transp. Syst. 2020, 23, 537–547. [CrossRef] 164. You, C.; Lu, J.; Filev, D.; Tsiotras, P. Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning. Robot. Auton. Syst. 2019, 114, 1–18. [CrossRef] 165. Isele, D.; Rahimi, R.; Cosgun, A.; Subramanian, K.; Fujimura, K. Navigating occluded intersections with autonomous vehicles using deep reinforcement learning. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 2034–2039. 166. Zhang, L.; Xiao, W.; Zhang, Z.; Meng, D. Surrounding Vehicles Motion Prediction for Risk Assessment and Motion Planning of Autonomous Vehicle in Highway Scenarios. IEEE Access 2020, 8, 209356–209376. [CrossRef] 167. Marin-Plaza, P.; Hussein, A.; Martin, D.; de la Escalera, A. Global and Local Path Planning Study in a ROS-Based Research Platform for Autonomous Vehicles. J. Adv. Transp. 2018, 2018, 6392697. [CrossRef] 168. Islam, M.; Chowdhury, M.; Li, H.; Hu, H. Vision-Based Navigation of Autonomous Vehicles in Roadway Environments with Unexpected Hazards. Transp. Res. Rec. J. Transp. Res. Board 2019, 2673, 494–507. [CrossRef] 169. Ma, L.; Xue, J.; Kawabata, K.; Zhu, J.; Ma, C.; Zheng, N. Efficient sampling-based motion planning for on-road autonomous driving. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1961–1976. [CrossRef] 170. Gu, T.; Dolan, J.M. On-road motion planning for autonomous vehicles. In Proceedings of the International Conference on Intelligent Robotics and Applications, Montreal, QC, Canada, 3–5 October 2012; pp. 588–597. 171. Wahid, N.; Zamzuri, H.; Amer, N.H.; Dwijotomo, A.; Saruchi, S.A.; Mazlan, S.A. Vehicle collision avoidance motion planning strategy using artificial potential field with adaptive multi-speed scheduler. IET Intell. Transp. Syst. 2020, 14, 1200–1209. [CrossRef] 172. Cai, P.; Sun, Y.; Chen, Y.; Liu, M. Vision-based trajectory planning via imitation learning for autonomous vehicles. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 2736–2742. 173. Kiss, D.; Tevesz, G. Advanced dynamic window based navigation approach using model predictive control. In Proceedings of the 2012 17th International Conference on Methods & Models in Automation & Robotics (MMAR), Miedzyzdroje, Poland, 27–30 August 2012; pp. 148–153. 174. Bai, Z.; Cai, B.; ShangGuan, W.; Chai, L. Deep learning based motion planning for autonomous vehicle using spatiotemporal LSTM network. In Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; pp. 1610–1614. 175. Li, X.; Choi, B.-J. Design of obstacle avoidance system for mobile robot using fuzzy logic systems. Int. J. Smart Home 2013, 7, 321–328. 176. Yoon, C.; Kim, K.; Park, H.S.; Park, M.W.; Jung, S.K. Development of augmented forward collision warning system for Head-Up Display. In Proceedings of the 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), Qingdao, China, 8–11 October 2014; pp. 2277–2279. 177. Bark, K.; Tran, C.; Fujimura, K.; Ng-Thow-Hing, V. Personal Navi: Benefits of an augmented reality navigational aid using a see-thru 3D volumetric HUD. In Proceedings of the 6th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Seattle, WA, USA, 17–19 September 2014; pp. 1–8. 178. Yoon, C.; Kim, K.-H. Augmented reality information registration for head-up display. In Proceedings of the 2015 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Korea, 28–30 October 2015; pp. 1135–1137. 179. Park, H.S.; Park, M.W.; Won, K.H.; Kim, K.H.; Jung, S.K. In-vehicle AR-HUD system to provide driving-safety information. ETRI J. 2013, 35, 1038–1047. [CrossRef] 180. Rao, Q.; Chakraborty, S.J. In-Vehicle Object-Level 3D Reconstruction of Traffic Scenes. IEEE Trans. Intell. Transp. Syst. 2020, 22, 7747–7759. [CrossRef] 181. Abdi, L.; Meddeb, A. Driver information system: A combination of augmented reality, deep learning and vehicular Ad-hoc networks. Multimed. Tools Appl. 2018, 77, 14673–14703. [CrossRef] 182. Rameau, F.; Ha, H.; Joo, K.; Choi, J.; Park, K.; Kweon, I.S. A real-time augmented reality system to see-through cars. IEEE Trans. Vis. Comput. Graph. 2016, 22, 2395–2404. [CrossRef] 183. Abdi, L.; Meddeb, A. In-vehicle augmented reality system to provide driving safety information. J. Vis. 2018, 21, 163–184. [CrossRef] 184. Lindemann, P.; Lee, T.-Y.; Rigoll, G. Supporting Driver Situation Awareness for Autonomous Urban Driving with an Augmented- Reality Windshield Display. In Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Munich, Germany, 16–20 October 2018; pp. 358–363. 185. Park, M.W.; Jung, S.K. TLD based vehicle tracking system for AR-HUD using HOG and online SVM in EHMI. In Proceedings of the 2015 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 9–12 January 2015; pp. 289–290. Appl. Sci. 2022, 12, 6831 51 of 51 186. Park, B.-J.; Yoon, C.; Lee, J.-W.; Kim, K.-H. Augmented reality based on driving situation awareness in vehicle. In Proceedings of the 2015 17th International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 1–3 July 2015; pp. 593–595. 187. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [CrossRef] 188. Gudigar, A.; Chokkadi, S.; Raghavendra, U.J.M.T. A review on automatic detection and recognition of traffic sign. Multimed. Tools Appl. 2016, 75, 333–364. [CrossRef] 189. Vashisth, S.; Saurav, S. Histogram of Oriented Gradients based reduced feature for traffic sign recognition. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; pp. 2206–2212. 190. Gonzalez-Reyna, S.E.; Avina-Cervantes, J.G.; Ledesma-Orozco, S.E.; Cruz-Aceves, I. Eigen-gradients for traffic sign recognition. Math. Probl. Eng. 2013, 2013, 364305. [CrossRef] 191. Yu, D.; Hu, X.; Liang, K. A two-scaled fully convolutional learning network for road detection. IET Image Process. 2021, 16, 948–957. [CrossRef] 192. Teichmann, M.; Weber, M.; Zoellner, M.; Cipolla, R.; Urtasun, R. Multinet: Real-time joint semantic reasoning for autonomous driving. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1013–1020. 193. Lv, W.; Song, W.-G.; Liu, X.-D.; Ma, J. A microscopic lane changing process model for multilane traffic. Phys. A Stat. Mech. Its Appl. 2013, 392, 1142–1152. [CrossRef] 194. Van Hasselt, H.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. arXiv 2015, arXiv:06461. 195. Pan, J.; Wang, X.; Cheng, Y.; Yu, Q. Multisource transfer double DQN based on actor learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2227–2238. [CrossRef] [PubMed] 196. Shuai, B.; Zhou, Q.; Li, J.; He, Y.; Li, Z.; Williams, H.; Xu, H.; Shuai, S. Heuristic action execution for energy efficient charge- sustaining control of connected hybrid vehicles with model-free double Q-learning. Appl. Energy 2020, 267, 114900. [CrossRef] 197. Okumura, B.; James, M.R.; Kanzawa, Y.; Derry, M.; Sakai, K.; Nishi, T.; Prokhorov, D. Challenges in perception and decision making for intelligent automotive vehicles: A case study. IEEE Trans. Intell. Veh. 2016, 1, 20–32. [CrossRef] 198. Muffert, M.; Pfeiffer, D.; Franke, U. A stereo-vision based object tracking approach at roundabouts. IEEE Intell. Transp. Syst. Mag. 2013, 5, 22–32. [CrossRef] 199. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamzadeh, M.; Fieguth, P.; Cao, X.; Khosravi, A.; Acharya, U.R. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf. Fusion 2021, 76, 243–297. [CrossRef] 200. Endsley, M.R. Autonomous driving systems: A preliminary naturalistic study of the Tesla Model S. J. Cogn. Eng. Decis. Mak. 2017, 11, 225–238. [CrossRef] http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Applied Sciences Multidisciplinary Digital Publishing Institute

Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review

Loading next page...
 
/lp/multidisciplinary-digital-publishing-institute/vision-based-autonomous-vehicle-systems-based-on-deep-learning-a-Kt9iZRIa42

References

References for this paper are not available at this time. We will be adding them shortly, thank you for your patience.

Publisher
Multidisciplinary Digital Publishing Institute
Copyright
© 1996-2022 MDPI (Basel, Switzerland) unless otherwise stated Disclaimer The statements, opinions and data contained in the journals are solely those of the individual authors and contributors and not of the publisher and the editor(s). MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. Terms and Conditions Privacy Policy
ISSN
2076-3417
DOI
10.3390/app12146831
Publisher site
See Article on Publisher Site

Abstract

applied sciences Review Vision-Based Autonomous Vehicle Systems Based on Deep Learning: A Systematic Literature Review Monirul Islam Pavel , Siok Yee Tan * and Azizi Abdullah Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia; pavel@ieee.org (M.I.P.); azizia@ukm.edu.my (A.A.) * Correspondence: esther@ukm.edu.my Abstract: In the past decade, autonomous vehicle systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. However, the AVS is still far away from mass production because of the high cost of sensor fusion and a lack of combination of top-tier solutions to tackle uncertainty on roads. To reduce sensor dependency and to increase manufacturing along with enhancing research, deep learning-based approaches could be the best alternative for developing practical AVS. With this vision, in this systematic review paper, we broadly discussed the literature of deep learning for AVS from the past decade for real-life implementation in core fields. The systematic review on AVS implementing deep learning is categorized into several modules that cover activities including perception analysis (vehicle detection, traffic signs and light identification, pedestrian detection, lane and curve detection, road object localization, traffic scene analysis), decision making, end-to-end controlling and prediction, path and motion planning and augmented reality-based HUD, analyzing research works from 2011 to 2021 that focus on RGB camera vision. The literature is also analyzed for final representative outcomes as visualization in augmented reality-based head-up display (AR-HUD) with categories such as early warning, road markings for improved navigation and enhanced safety with overlapping on vehicles and pedestrians Citation: Pavel, M.I.; Tan, S.Y.; in extreme visual conditions to reduce collisions. The contribution of the literature review includes Abdullah, A. Vision-Based detailed analysis of current state-of-the-art deep learning methods that only rely on RGB camera Autonomous Vehicle Systems Based on Deep Learning: A Systematic vision rather than complex sensor fusion. It is expected to offer a pathway for the rapid development Literature Review. Appl. Sci. 2022, 12, of cost-efficient and more secure practical autonomous vehicle systems. 6831. https://doi.org/10.3390/ app12146831 Keywords: autonomous controlling; deep learning; decision making; intelligent vehicle; perception; self-driving Academic Editor: Seong-Ik Han Received: 11 May 2022 Accepted: 20 June 2022 Published: 6 July 2022 1. Introduction Publisher’s Note: MDPI stays neutral Recently, the autonomous vehicle system (AVS) has become one of the most trend- with regard to jurisdictional claims in ing research domains that focus on driverless intelligent transport for better safety and published maps and institutional affil- reliability on roads [1]. One of the main motives for enhancing AVS developments is iations. its ability to overcome human driving mistakes, including distraction, discomfort and lack of experience, that cause nearly 94% of accidents, according to a statistical survey by the National Highway Traffic Safety Administration (NHTSA) [2]. In addition, almost 50 million people are severely injured by road collisions, and over 1.25 million people Copyright: © 2022 by the authors. worldwide are killed annually in highway accidents. The possible reasons for these injuries Licensee MDPI, Basel, Switzerland. may derive from less emphasis on educating drivers with behavior guidance and poorly This article is an open access article developed drivers’ training procedures, fatigue while driving, visual complexities, that distributed under the terms and is, human error, which can be potentially solved by adopting highly efficient self-driving conditions of the Creative Commons vehicles [3,4]. The NHTSA and the U.S. Department of Transportation formed the SAE Attribution (CC BY) license (https:// International levels of driving automation, identifying autonomous vehicles (AV) from creativecommons.org/licenses/by/ 0 0 ‘level 0 to the ‘level 5 [5], where levels 3 to 5 are considered to be fully AV. However, as of 4.0/). Appl. Sci. 2022, 12, 6831. https://doi.org/10.3390/app12146831 https://www.mdpi.com/journal/applsci Appl. Sci. 2022, 12, 6831 2 of 51 2019, the manufacturing of level 1 to 3 vehicle systems has been achieved but level 4 vehicle systems are in the testing phase [6]. Moreover, it is highly anticipated that autonomous vehicles will be employed to support people in need of mobility as well as reduce the costs and times of transport systems and provide assistance to people who cannot drive [7,8]. In the past couple of years, not only the autonomous driving academic institutions but also giant tech companies like Google, Baidu, Uber and Nvidia have shown great interest [9–11] and vehicle manufacturing companies such as Toyota, BMW and Tesla are already working on launching AVSes within the first half of this decade [12]. Although different sensors such as radar, lidar, geodimetric, computer views, Kinect and GPS are used by conventional AVS to perceive the environment [13–17], it is indeed expensive to equip vehicles with these sensors and the high costs of these sensors are often limited to on-road vehicles [18]. Table 1 shows a comparison of three major vision sensors based on a total of nine factors. While the concept of driverless vehicles has existed for decades, the exorbitant costs have inhibited development for large-scale deployment [19]. To resolve this issue and build a system that is cost efficient with high accuracy, deep learning applied vision-based systems are becoming more popular where RGB vision is used as the only camera sensor. The recent developments in this field of deep learning have accelerated the potential of profound learning applications for the solution of complex real-world challenges [20]. Table 1. Comparison of vision sensors. VS VR FoV Cost PT DA AAD FE LLP AWP Camera High High Low Medium Medium High High Medium Medium Lidar High Medium High Medium High Medium Medium High Medium Radar Medium Low Medium High High Low Low High Low VS = Vision Sensor, VR = Visibility Range, FoV = Field of View, PT = Processing Time, DA = Distance Accuracy, AAD = AI Algorithm Deployment, FE = Feature Engineering, LLP = Low-Light Performance, AWP = All-Weather Performance. In this systematic review paper, a broad discussion and survey of the implementation of deep learning are applied to aspects of AVS such as vehicle detection (VD), traffic signs and light identification (TSL), pedestrian detection (PD), lane detection and tracking (LDT), traffic scene analysis (TSA), decision making (DM), end-to-end controlling and prediction (E2EP), path and motion planning (PMP) and augmented reality-based HUD (ARH) ana- lyzing research articles from 2011 to 2021 research articles on deep learning-applied AVS to reduce the dependency on sensor fusion and the high cost of manufacturing and to enhance the focus on developing a level 5 autonomous driving vehicle. We represent and thoroughly discuss the best deep learning algorithms for each domain, provide solutions to their limitations and analyze their performance for increasing practical implementation concepts. Moreover, this systematic review explored the most complete and predominate domains compared to other surveys from [21–33] (shown in Table 2) that indicates its impact on AVS implementing deep learning where the review article covered all aspects of the human–machine interface (HMI). The overall contribution of the research is set out below: Analyzed recent solution of state-of-the-art deep learning algorithms for cost-efficient AVS using RGB camera. Detailed literature review covering major domains and most subcategories to decrease vision sensor complexities. Discussed the key advantages and disadvantages of deep learning methods applied to AVS. Appl. Sci. 2022, 12, 6831 3 of 51 Table 2. Comparison of existing studies. Survey Coverage Ref. Year VD LRCD PD TSL E2EC TSA PMP DM ARH HMI [21] 2019 4 X 4 X X X 4 4 X 4 [22] 2020 X X X 4 X 4 4 4 X X [23] 2016 X X X X 4 X 4 X X X [24] 2020 X X X X 4 X 4 X X 4 [25] 2018 X X X X 4 X 4 X X X [26] 2018 X X X X 4 X 4 X X X [27] 2021 X X X 4 4 X 4 4 X 4 [28] 2020 4 4 4 4 4 4 X X X 4 [29] 2018 X X 4 X X X X X X X [30] 2020 4 4 4 X X X 4 X X X [31] 2020 X 4 4 4 4 4 4 X X X [32] 2021 X 4 X X X X 4 4 X X [33] 2020 4 X X X 4 4 4 X X 4 Ours 2022 4 4 4 4 4 4 4 4 4 X 2. Methodology 2.1. Review Planning The study is based on a systematic review methodology, an approach for analyzing and evaluating accessible studies related to a particular issue of current research where the core three phases are preparing the review, conducting the review, and creating a report that summarizes the review. In this systematic review paper, the researchers have included 142 papers containing deep learning and belonging to a different domain of AVS. To finalize the papers, we initially focused on the entire domain of autonomous driving, then we restricted our search to the usage of deep learning in AVS. Only papers with full text in English from renowned journals, conferences and book chapters that were published between 2011 and 2021 were selected. Due to an increase in the scope of advanced autonomous transportation, we finally limited our search to the vision-based application of deep learning in AVS, and the rest were rejected. We also took the most complete edition to avoid dealing with duplication. The key plan and protocol of the review includes source of data, searching criteria and procedure, research questions, data selection and data extraction. 2.2. Sources of Data Research papers were gathered from various famous research databases to incorporate specific field and research questions. Irrelevant research papers that could not address or endorse our research questions were dismissed. To achieve a broad coverage for the literature review, we used the following databases as our key resources: Web of Science, Scopus, IEEE Xplorer, ScienceDirect, MDPI, Springer, Wiley Library and ACM. 2.3. Research Questions Research questions were formed to refine the survey and maintain the aim of the topic. The following research questions are answered throughout the discussion in the different sections of the paper. How does deep learning reduce sensor dependency? How are on-road objects detected and localized? What decision-making processes are solved for AVS? How does deep learning contribute to end-to-end controlling and path planning? How should final outcomes be represented in AR-HUD? Appl. Sci. 2022, 12, 6831 4 of 51 2.4. Searching Criteria To find research papers according to the methodology, a pattern was followed to gather suitable papers which were mostly necessary for our study. We adopted a Boolean searching method with multiple AND, OR in the advance search options of each data source. During the search for the relevant papers, we selected “Autonomous Driving” and “Autonomous Vehicle” or “Intelligent Vehicle” or “Self-Driving” and “Deep Learning” as the main phrases. For a further refined search, various keywords were included to obtain the desired research papers according to our aim in this review. The following queries were developed based on Boolean operations: ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Object) AND ([Vehicle] OR [Pedestrians] [Traffic Sign] AND [Traffic Light])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self-Driving) AND (Deep Learning) AND ([Traffic Scene] OR [Localization] OR [Segmentation])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Lane) AND ([Track] OR [Shift] OR [Segmentation])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND (Control) AND ([Steering] OR [Motion])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND ([Deep Learning] OR [Deep Reinforcement Learning]) AND (Decision Making) AND ([Uncertainty] OR [Lane Keeping] OR [Overtaking] OR [Braking] OR [Acceleration])) ((Autonomous Driving) OR (Autonomous Vehicle) OR (Intelligent Vehicle) OR (Self- Driving) AND (Deep Learning) AND ([Augmented Reality] AND [Head Up Display] OR [HUD])) 2.5. Searching and Extraction Procedure The selection procedure for choosing papers includes four core iteration filtering processes. As the aim of the study is to discuss implementation of deep learning and comprehensive literature searches to analyze the frameworks and system designs, first, a total of 760 papers were selected from eight data sources based on the queries mentioned in the searching criteria (Section 2.4). Web of Science had the highest 151 and ACM had the lowest 40 papers. Then, the selected papers had to be processed through an eligibility stage where 209 duplicated papers were eliminated at first. Furthermore, 121 papers were screened out during abstract scanning and 276 papers were chosen after full text reading. In the next iteration, studies containing domains of deep learning in relation to AVS were selected where all the papers were published between 2011 and 2021. The final dataset contains a total of 142 papers that covers the literature on the implementation of deep learning methods for AVS. The structure of the whole selection process is presented in Figure 1. Table 3 presents the final calculation for the selection of 142 papers according to these steps and based on the most relatable topics and in-depth analysis. 2.6. Analysis of Publication by Year Out of 142 final papers for review, the studies published between 2011 and 2021 were selected. The year 2019 had the highest number of selected research papers, with 31, which is 17.9% of the total, and 2011 had the lowest number of papers (2). The distribution of publication is visualized in Figure 2. Appl. Sci. 2022, 12, x FOR PEER REVIEW 5 of 52 Springer 96 53 21 Wily Library 71 44 8 Appl. Sci. 2022, 12, 6831 5 of 51 MDPI 112 62 21 Total 760 430 142 Figure 1. Framework for searching the literature and making the selection. Figure 1. Framework for searching the literature and making the selection. 2.6. Analysis of Publication by Year Table 3. Paper selection from multiple sources. Out of 142 final papers for review, the studies published between 2011 and 2021 were Scheme 151 Primary Candidate Selected selected. The year 2019 had the highest number of selected research papers, with 31, which Web of Science 151 72 24 is 17.9% of the total, and 2011 had the lowest number of papers (2). The distribution of IEEE Xplore 124 69 22 publication is visualized in Figure 2. Scopus 72 47 20 ScienceDirect 94 51 19 ACM 40 32 7 Springer 96 53 21 Wily Library 71 44 8 MDPI 112 62 21 Total 760 430 142 Appl. Sci. 2022, 12, x FOR PEER REVIEW 6 of 52 Appl. Sci. 2022, 12, x FOR PEER REVIEW 6 of 52 Appl. Sci. 2022, 12, 6831 6 of 51 Figure 2. Distribution of studies in terms of year of publication (2011–2021). 2.Figure 7. Analy 2. Distribution sis of Publicat of studies ion by in Country terms of year of publication (2011–2021). Figure 2. Distribution of studies in terms of year of publication (2011–2021). Among the 142 selected papers for the literature review, 56 countries contributed to 2.7. Analysis of Publication by Country autonomous vehicle system development. Figure 3 shows the top 10 countries and the 2.7. Analy Among sis the of P 142 ublicat selected ion bpapers y Country for the literature review, 56 countries contributed to number of papers they contributed before the final selection. The graphical representation autonomous vehicle system development. Figure 3 shows the top 10 countries and the Among the 142 selected papers for the literature review, 56 countries contributed to shows that China made the largest contribution, with 34 papers, and the USA contributed number of papers they contributed before the final selection. The graphical representation autonomous vehicle system development. Figure 3 shows the top 10 countries and the 21 shows papers that , wh China ich wa made s th the e secon largest d la contribution, rgest. with 34 papers, and the USA contributed number of papers they contributed before the final selection. The graphical representation 21 papers, which was the second largest. shows that China made the largest contribution, with 34 papers, and the USA contributed 21 papers, which was the second largest. Figure 3. Distribution of studies over top 15 countries of first authors. Figure 3. Distribution of studies over top 15 countries of first authors. 2.8. Analysis of Publication Domains 2.8. Analy Thes142 is of final Publicat papers ion D wer om e ains selected based on five domains and five subdomains of perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined The 142 final papers were selected based on five domains and five subdomains of Fig to ure produce 3. Dist arcomplete ibution ofsystem. studies T ov able er t4 op shows 15 count thatrie the s of distribution first authoof rs.‘Decision Making’ perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the to produce a complete system. Table 4 shows that the distribution of ‘Decision Making’ lowest 11 papers individually. 2.8. Analysis of Publication Domains section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the The 142 final papers were selected based on five domains and five subdomains of lowest 11 papers individually. perceptions, shown in the literature taxonomy of AVS in Table 4, which were combined to produce a complete system. Table 4 shows that the distribution of ‘Decision Making’ section has highest 20 papers, and ‘Path and Motion Planning’ and ‘AR-HUD’ have the lowest 11 papers individually. Appl. Sci. 2022, 12, 6831 7 of 51 Appl. Sci. 2022, 12, x FOR PEER REVIEW 7 of 52 Table 4. Literature taxonomy of AVS using deep learning approaches. Table 4. Literature taxonomy of AVS using deep learning approaches. Domain Sub-Domain References Domain Sub-Domain References Vehicle Detection [34–45] Vehicle Detection [34–45] TTr raf a fic ffic S Sign igand n an Light d Lig Recognition ht Recognition [46 [4 –6 59 –59] ] Perception Pedestrian Detection [60–78] Perception Pedestrian Detection [60–78] Lane Detection and Tracking [44,79–101] Lane Detection and Tracking [44,79–101] Traffic Scene Analysis [55,102–120] Traffic Scene Analysis [55,102–120] Decision Making - [121–143] Decision Making - [121–143] End-to-End Controlling and Prediction - [144–163] End-to-End Controlling and Prediction - [144–163] Path and Motion Planning - [164–175] Path and Motion Planning - [164–175] AR-HUD - [176–186] AR-HUD - [176–186] To visualize the leading algorithms of each domain or subdomain, Figure 4 presents the distribution of algorithms, where the reviewed algorithm-centered approaches have a To visualize the leading algorithms of each domain or subdomain, Figure 4 presents predominant role in AVS development. Figure 5 shows the dataset clustering which was the distribution of algorithms, where the reviewed algorithm-centered approaches have a used for the reviewed approaches. Only the subdomains of perception showed depend- predominant role in AVS development. Figure 5 shows the dataset clustering which was ency on dataset, where “Traffic Sign and Light Recognition” and “Lane Detection and used for the reviewed approaches. Only the subdomains of perception showed dependency Tracking” applied to 6 datasets each, and only 3 datasets were adopted in “Traffic Scene on dataset, where “Traffic Sign and Light Recognition” and “Lane Detection and Tracking” Analysis”. applied to 6 datasets each, and only 3 datasets were adopted in “Traffic Scene Analysis”. Figure 4. Taxonomy algorithms for each domain. Figure 4. Taxonomy algorithms for each domain. Appl. Sci. 2022, 12, 6831 8 of 51 Appl. Sci. 2022, 12, x FOR PEER REVIEW 8 of 52 Figure 5. Clustering of dataset for subdomains of perception. Figure 5. Clustering of dataset for subdomains of perception. 3 3. . A Analysis nalysis of of Doma Domain in Each domain was analyzed by reviewing several approaches and methods based on Each domain was analyzed by reviewing several approaches and methods based on evaluating and discussing advantages, disadvantages, outcomes and significance. The evaluating and discussing advantages, disadvantages, outcomes and significance. The fol- following analysis of each domain was carried out with the aim of accelerating development lowing analysis of each domain was carried out with the aim of accelerating development of level 4 or 5 AVS. of level 4 or 5 AVS. 3.1. Perception 3.1. Perception 3.1.1. Vehicle Detection 3.1.1. Vehicle Detection The identification and detection of an on-road vehicle for AVS together form one of the The identification and detection of an on-road vehicle for AVS together form one of predominant and most challenging issues due to versions, combined fast multitasking and the predominant and most challenging issues due to versions, combined fast multitasking visual difficulties. For fast and more accurate vehicle detection and recognition in different and visual difficulties. For fast and more accurate vehicle detection and recognition in and uncertain driving conditions, deep learning algorithms are analyzed in this section. different and uncertain driving conditions, deep learning algorithms are analyzed in this For instance, an online network framework for detecting and tracking vehicles was section. proposed by Hu et al., who predicted full 3D vehicle bounding box mapping from a For instance, an online network framework for detecting and tracking vehicles was monocular camera using both the environment and camera coordinates by reprojecting [34]. proposed by Hu et al., who predicted full 3D vehicle bounding box mapping from a mo- Eventually, the framework tracked the movement of instances in a global coordinate system nocular camera using both the environment and camera coordinates by reprojecting [34]. and revised 3D poses with a trajectory approximation of LSTM, implementing on a KITTI Eventually, the framework tracked the movement of instances in a global coordinate sys- dataset, where the outcomes surpassed the outcomes of LiDAR in long range [187]. In a tem and revised 3D poses with a trajectory approximation of LSTM, implementing on a 30 m range, LiDAR obtained 350.50 false negative and the method scored 395.33, while KITTI dataset, where the outcomes surpassed the outcomes of LiDAR in long range [187]. vehicle detection was 11.3% higher, indicating the limitation of the framework. However, it In a 30 m range, LiDAR obtained 350.50 false negative and the method scored 395.33, performed better in 50 m and 10 m, where the false negative scores were 857.08 and 1572.33 while vehicle detection was 11.3% higher, indicating the limitation of the framework. when the LiDAR-based method obtained false negative values of 1308.25 and 2445.30, However, it performed better in 50 m and 10 m, where the false negative scores were respectively. The decreased false negative in 10 and 50 m showed that the method was 857.08 and 1572.33 when the LiDAR-based method obtained false negative values of able to overcome the performance of LiDAR using only camera and deep learning despite 1308.25 and 2445.30, respectively. The decreased false negative in 10 and 50 m showed reduced accuracy in some real-time implementations. that the method was able to overcome the performance of LiDAR using only camera and To tackle accuracy-based issues, improve slow detection and recognition speed, and deep learning despite reduced accuracy in some real-time implementations. address the lack of categorization ability, Sang et al. introduced a novel YOLOv2_Vehicle To tackle accuracy-based issues, improve slow detection and recognition speed, and architecture [35]. For multiple scales of vehicles that influenced the detection framework, a normalization ddress the lack had of ca been tego used rizatio for nthe abili impr ty, S ovement ang et al. of inthe troduc method ed a n of ove measuring l YOLOv2 losses _Vehicle for a boundary rchitecturbox e [35length ]. For m and ultiple width scales of after clustering vehicles thand at inbounding fluenced th boxes e deteapplying ction fram k-mean ework, normalization had been used for the improvement of the method of measuring losses for Appl. Sci. 2022, 12, 6831 9 of 51 ++ algorithm to the training dataset [36] along with applying multilayer feature fusion to boost the network extraction capabilities and repeatedly eliminating convolutional layers in high layers. The method implementing the BIT dataset could obtain a mean average precision (mAP) exceeding 94.78% in 0.038 s, which was found to be much faster and more accurate than compared existing methods. In another work, an AdaBoost combined with a pixel look-up features-based approach was demonstrated by Ohn-Bar et al., where the methods included mining orientation, object geometry and occlusion pattern by clustering, and 81.94%, 66.32%, 51.10% accuracy in easy, moderate and hard scenarios, respectively, was obtained for vehicle detection. However, performance decreased when a 70% overlap evaluation threshold was used instead of 50% and during heavy occlusion [37]. Nonetheless, the method was inappropriate for rough conditions as performance decreased when 70% overlap evaluation threshold was used instead of 50%, and it showed poor accuracy in heavy occlusion. Further, Chen et al. presented a new method to identify five distinct vehicle classes, that is, car, van, bus, truck and tractor, using the AdaBoost and CNN algorithms applied to CompCars and their custom dataset containing rear views of vehicles [38]. They employed CNN as a function extractor with a Support Vector Machine (SVM) for training the features separately, and further AdaBoost algorithms were applied for integration. They obtained optimum results even with faulty images and high computing costs, with average accuracy of 99.50% in 0.028 s, which was 13% higher than the other mentioned fusion methods, for instance, SIFT + SVM, HoG + SVM and SURF + SVM. However, the method was only deployed and considered in simple but low-quality images and daytime scenarios. Moreover, one of the biggest issues was the low resolution of images in the real- time traffic surveillance method due to either low-vision RGB cameras or environmental features such as low light condition or foggy weather. For this problem, vehicles in low- resolution images and videos were analyzed in terms of the efficiency of the CNN by Bautista et al. [39]. The neural network used an activation function that worked in two phases: first, the detection of high-level attributes; second, the detection of low-level attributes. It tested the comportment of the model to detect vehicles with lower input resolution at different levels, as well as the number and size of filters. Results demonstrate that CNN was remarkably successful even with low resolution in the identification and classification of vehicles with an average precision fit for real time applications. Lee et al. showed a hierarchical system for detecting and tracking vehicles in an urban area at night based on taillights [40]. The system focused primarily on effective detection and pairing of taillights, considering their innate variety and observing all aspects of the layers and interrelationships in a hierarchical framework which increases the efficiency of vehicle detection and tracking in comparison with traditional methods, with recall of 78.48% and precision of 90.78%. However, performance decreased for short-distance vehicles due to headlight illumination of host vehicles. This approach could be considered as one the most suitable methods for nighttime vehicle detection. Hu et al. demonstrated a novel CNN architecture called scale-insensitive convolu- tional neural networking (SI-Net) [41] to enhance the performance of vehicle detection for autonomous vehicles, solving the issue of limited CNN-based vehicle detection [39]. The framework improved the limitation-scale insensitive CNN, deploying context-aware region of interest (ROI) pooling to preserve the real structure of small-scale objects. The state-of- the-art method outperformed the others in terms of measuring, scoring 89.6%, 90.60% and 77.75% for accuracy in moderate, easy and complex moods, respectively, in 0.11 s execution time on the KITTI benchmark as well as a custom highway dataset with different variance of scaled objects. Thus, the method was able to maintain a good performance in multiple traffic scenarios. Targeting the runtime of previous works, Wang et al. combined anchor size, receptive field and anchor generation optimization (AGO) with Fast R-CNN to ensure that an acceptable number of vehicle features could be accessed by the network in the shortest amount of time [42]. Using the anchor shape, it efficiently detects vehicles in large, medium Appl. Sci. 2022, 12, 6831 10 of 51 and short fields of view with 87.2% average precision in 0.055 s. The anchor shape-based detection process is a very coherent technique for AVS for reducing computational cost by not taking the whole field of vision for processing. In another work, which combined Faster R-CNN training parameters with a region proposal network (RPN)-based approach, Suhao et al. implemented vehicle-type detection in a real traffic area including MIT and Caltech datasets with ZF and VGG-16 networks in multiple scenarios [43]. The research results increased the average accuracy of the detection systems and the rate of detection compared with the CNN models. The proposed architecture classified vehicles from three categories where they achieved the best accuracy: 84.4% for car, 83.8% for minibus and 78.3% for SUV using the VGG-16 model in 81 ms. The execution cost of the proposed method outperformed Fast R-CNN and Faster R-CNN applied to complex scenarios. However, a lightweight YOLO network that was built with a YOLO v3 algorithm with generalized IoU loss was combined with loss function as well as with the integration of two different focal length cameras by Liu et al. to indicate less computer complexity for AVS [44]. The method was implemented on their self-made dataset where the network obtained 90.38% precision and 82.87% recall within 44.5 ms. This could be a milestone for AVS in terms of a faster and more accurate method for different field of view and day or nighttime implementation. Leung et al. compared deep learning-based techniques for vehicle detection effi- ciency [45], and proposed solutions for data collection along with the nighttime data labelling convention to resolve different types of detection. The research also recommends a framework based on a quicker region-based CNN model, which was precisely optimized, merging with RestNet101, the VGG-16 model obtaining a mean average precision (mAP) of 84.97%. The experimental result showed a high detection accuracy in urban nighttime with version lighting conditions including extreme low light and no lighting. Thus, this method became one of the most suitable methods for AVS in challenging lighting conditions. Overall, the Deep CNN and AdaBoost-based approach achieved 99.50% accuracy in daytime (the highest) with the fastest computational time (0.028 s), but lightweight YOLO and the quicker region-based CNN model showed practical outcomes in both daytime and nighttime scenarios for vehicle detection. Multiple deep learning methods showed efficient performance by improving slow detection, recognition and categorization, enabling deployment in complex scenarios and night-time deployment with good accuracy, even surpassing accuracy outcomes of LiDAR in terms of long field of view. However, some challenges remained, for example, limited dataset of vehicle categories, performance dropping in low light and rough weather conditions for some methods, low accuracy for vehicle detection in short distance for headlight illumination at nighttime and fast execution time in real-time implementation. An overview of the methods evaluated for the detection and recognition of vehicles for AVS is provided in Table 5. 3.1.2. Traffic Sign and Light Recognition One of the most important aspects of a safe and better decision-making process for automotive driving system was traffic sign and light identification by regulating traffic, monitoring and avoiding accidents through warning the drivers. Traffic sign and light recognition systems follow a double-step process, detection and classification, where detec- tion denotes correctly spotting the geometric position in the image and classification means identification of the category in which the detected sign or light signal appears [28,188]. A bio-mechanism inspired novel architecture named Branch Convolutional Neural Network (BNN) was proposed by Hu et al. for traffic sign recognition [46]. To improve the recognition machine speed and accuracy, a branch-output mechanism which was placed between pooling and convolutional layer and added to the framework. Furthermore, instead of the initial output layer, the sign in the preceding branch was projected by the BCNN that results perfect prediction in partial visibility of road signs with 98.52% Appl. Sci. 2022, 12, 6831 11 of 51 accuracy based on German Traffic Sign Recognition Benchmark (GTSRB). For complex visual scenarios, BCNN based approach worked very well for traffic sign recognition. Table 5. Summary of multiple deep learning methods for vehicle detection. Ref. Method Outcomes Advantages Limitations 65.52% more 3D vehicle bounding Exceeded the outcomes of Unstable accuracy for certain [34] enhancement in 50 m box mapping LiDAR in long range. error properties. than LiDAR. YOLO V2 and Faster detection Trained with vehicle types [35,36] mAP 94.78% in 0.038 s. K-mean ++ and recognition. of data. Improved result with Performance decreased AdaBoost with pixel 81.94% accuracy in [37] occlusion pattern when 70% overlap and lookup features best-case scenarios. by clustering. heavy occlusion. Highest accuracy in daytime Deep CNN and 99.50% accuracy Not applicable in low-light [38] and trained with AdaBoost in 0.028 s. or complex scenarios. low-quality frames. CNN with 96.47% accuracy in Fast classification in low Higher execution time [39] Caffe framework 51.28 ms. resolution input. compared to other methods. Recall 78.48% and Decreased in performance ANN and Nighttime taillight-based [40] precision 90.78% in for headlight illumination in Kalman Filter detection in urban area. urban scenario. short distance. Improved the Not applicable in poor 90.60% accuracy in [41] SI-NET limitation-scale insensitivity lighting, nighttime or best-case scenarios. of CNN. complex weather scenarios. Enable to detect vehicle in Not applicable in urban or Faster-RCNN 87.2% average [42] large, medium and challenging with AGO precision in 0.055 s. short field. lighting conditions. Highest 84.4% accuracy Unsatisfactory accuracy Faster R-CNN Outperformed Faster R-CNN [43] for car detection in compared with the obtained with RPN in terms of execution time. 81 ms. execution time. 86.81% and 78.91% Applicable in both day and Lightweight precision and recall Was not tested in urban or [44] night scenarios in YOLO Network within 44.5 ms, crowded environment. multiple FoV. respectively. Able to detect in low-light Quicker 84.97% mAP Required huge manual [45] and almost region-based CNN nighttime scenarios. data labelling. no-light conditions. Jung et al. trained the 16 different forms of Korean traffic sign with LeNet-5 CNN architecture for a real-time traffic sign recognition system where the training set had 25,000 positive and 78,000 false samples [47]. The method obtained up to 99.95% within a fast-processing time. The applied color property-based CNN approach could be very effi- cient for lightweight traffic sign detectors for AVS as well as achieving the highest accuracy. An improved traffic sign recognition algorithm was demonstrated by Cao et al. for an intelligent driving system [48]. For accurate detection spatial threshold segmentation, the HSV color space was utilized, and traffic signs were identified accurately depending on shape features and processed with LeNet-5 CNN architecture with Gabor kernel, which was the primary convolutional kernel, and batch normalization was applied after the pooling layer. The Adam model was also implemented as the optimizer algorithm. The proposed methodology was applied to the German Traffic Sign Recognition Benchmark and obtained 99.75% accuracy with 5.4 ms per frame on average, which was higher in both sectors than [189,190], where the accuracies were 98.54% in 22 ms and 95.90 in 5.4 ms, respectively, adopting HoG + PCA and multilayer perceptions methods. Appl. Sci. 2022, 12, 6831 12 of 51 On the other hand, the parallel architecture weighted multi-convolutional neural network took 4.6 ms more to process but still achieved constant high efficiency, scoring 99.75% in GTSDB and 99.59% accuracy in the GTSRB dataset, where low and complex lighted scenarios were also considered. Despite occasional accuracy drops for blurry vision, this method could be one of the most suitable approaches for AVS [49]. To detect traffic signs, Wang et al. suggested a red bitmap extraction and SVM- based method where the detected images were color-segmented and afterwards the shape detection of ROI (ROI) was carried out on the basis of rim detail [50]. The methodology scored recall values of 97% and 99% for danger and prohibitor for the GTSDB dataset, respectively. This technique obtained good detection accuracy, but the major limitation was that this method was only applied to red circular signs. Zhang et al. demonstrated a modified YOLO V2 algorithm to develop an improved Chinese traffic sign detection system, as well as constructing the database [51]. In order to create a single convolutional network, they used several 1  1 convolutional layers for the intermediary and fewer convolutional layers for the top layers. Fine grid was also used to separate images with the goal of identifying small-sized road signs. Their technique was found to be the outcome of the CCTSDB and GTSDB databases, where AUC values for mandatory and danger signs were 96.81%, 94.02% and 96.12% in 0.017 s. Another approach applied CapsNet, which resolved the major limitation of CNN, that is, loss of max pooling layer retaining spatial relations [52]. The approach obtained 98.72% accuracy for recognizing traffic light with shape. It can be a useful method-based approach for AVS’s traffic sign-recognition methods. Furthermore, a unified deep convolutional traffic light-identification feature for au- tomated driving systems was proposed by Bach et al., based on Faster R-CNN that was suitable for detection of traffic lights, and the recognition and classification of types or states [53]. They achieved 92% average precision applying on a large-scale dataset named DriverU traffic light. When the width was greater than 8 px and smaller than these, it scored 93% for average precision. However, there were still limitations for suitable number of false positives which can be reduced by applying RNN or an integrated approach. DeepTLR was the real-time vision-dependent, in-depth and deeply convoluted traffic light-identification and classification system that did not require position details or temporal principles; these were proposed by Weber et al. [54]. On the basis of the single-frame assessment of a challenging collection of urban scenes, the authors presented noteworthy outcomes, showing that in regular images, DeepTLR achieves frame rates of up to 33 Hz. DeepTLR also ran at frame rates of up to 33 Hz. DeepTLR also ran at frame rates of 13 Hz on images with a resolution of 1280  960 pixels. The capacity for more transport lights was high in the architecture, scoring 93.5% F1 score for 1280  960 resolution and 88.1% F1 score for 640  480 in 80 ms and 28.8 ms. Li et al. developed a framework of robust traffic-light recognition with fusion detection in complex scenes [55]. To increase accuracy for each traffic light type and the creation of a fusion detective framework, a set of enhanced methods was adopted based on an optimized channel function (OCF) system after using aspect ratio, field, location and traffic lights background as prior knowledge to minimize computational redundancy and create a task model for the identification of traffic light. Furthermore, they utilized the detection knowledge of the previous system to change the original ideas, which further increased the accuracy. The framework was applied to a VIVA dataset where a combination of multi-size detectors, bulb detectors and fuzzy detectors were implemented, which improved the AUC indicator, with 7.79% for red, 9.87% for red left, 11.57% for green and 3.364% for green left, compared with general ACF on VIVA validation dataset and achieved an AUC indicator of 91.97% for red light and 89.32% for green light on the channel-modified LARA validation dataset. In addition, to reduce complexity, Lee et al. adopted the concept of upper-half clipping frame so that the model could pick only those frames that would allow it to recognize traffic lights rather than taillights [56]. The system was built based on a YOLO algorithm Appl. Sci. 2022, 12, 6831 13 of 51 and obtained 42.03% mAP and 49.1% mAP enhanced, and improved results applied to the Bosch-TL and LISA-TL datasets, but the author did not consider nighttime scenarios. Other than this issue, the method was exceptionally successful for traffic signs and light identification systems for AVS. Behrendt et al. implemented YOLO for 3D localization and tracking of traffic lights [57]. A wide field of view was considered, and the YOLO-based approach was deployed in the Bosch-TL dataset. The method showed 99% accuracy in just 0.06 ms. However, the method required huge pre-labelled data which created an obstacle to fluent performance. In this section, both traffic signs and traffic light detection and recognitions are dis- cussed and summarized in Table 6, for which most of the deep learning approaches were trained with the GTSRB dataset. Among all the deep learning methods, LetNet-5-based CNN on self-made dataset with spatial threshold segmentation with the HSV color space and Gabor filter on the GTSRB dataset performed best for traffic-sign recognition despite reduction in performance in a complex environment and detection of separated signs due to the proposed region. In addition, the YOLO method-based approached obtained highest accuracy in the fastest time for traffic-light detection with recognizing inner signs despite having pre-labelled data dependency. Table 6. Summary of multiple deep learning methods for traffic sign and light recognition. Ref. Method Outcomes Advantages Limitations 98.52% accuracy for Branch-output mechanism Implementation from [46] BCNN low visibility of enhanced recognition speed moving platform was road signs. and accuracy. not tested. Lightweight color segmentation Detected two traffic signs for [47] LetNet-5 based CNN 99.95% accuracy. before classification, improved having different location of processing speed. proposed region. Obtained highest accuracy and Accuracy of 99.75% Performance decreased in [48] CNN with Gabor filter enhanced the performance in 5.4 ms. complicated backgrounds. of LetNet-5. Classification in high-speed Struggled to classify in Highest 99.75% [49] Weighted multi CNN driving and outperformed in challenging blurry accuracy in 10 ms. low-light conditions. vision condition. Faster detection of prohibitory Recall score 97% and Only applied to red [50] SVM and danger signs in poor 99% for two cases. circular signs. illumination. Highest 96.81% AUC Faster and simple Decrease in performance for [51] YOLO V2 values in 0.017 s. processing pipeline. small traffic signs. Resolved loss of max pooling to Did not consider complex [52] CapsNet 98.72% accuracy. boost CNN’s performance. lighting condition 100% recall and 92% Detected traffic light as well as [53] Fast R-CNN Showed high false positive. average precision. indicating signs. Did not require position details [54] DeepTLR Highest 93.5% F1 score. Lower precision rate. or temporal principles. Better detection in complex 11.5% improved AUC Unsatisfied accuracy for red [55] OCF scenes with low luminous value for green light. left traffic light identification. objects combining. 42.03% and 49.16% Took upper half of frame to Did not deploy in nighttime [56] YOLO mAP higher on eliminate search area and traffic scene where reflection two datasets. vehicle taillights. creates confusion. 3D localization and tracking of [57] YOLO 99% in 0.06 s. traffic lights in large field Required more labelling. of view. Appl. Sci. 2022, 12, 6831 14 of 51 3.1.3. Pedestrian Detection Effectively detecting and localizing pedestrians on roads in various scenarios was one of the major vision-based problems for autonomous driving systems. A study shows that only in the USA has the fatality rate for road crossing increased up to 30% in seven years from 2009. In 2016, a total of 6000 pedestrians were killed, which is a record in the last three decades [58]. Moreover, based on vulnerable road users in the ASEN region, 13% of the deaths on roads are related to pedestrians [59]. In order to prevent deaths, the detection and localization of pedestrians have become a major focus of the study of autonomous vehicles. Several studies have been successfully conducted on reducing accident cases and creating a sustainable and more accurate approach to autonomous driving systems. For instance, Angelova et al. proposed a deep network named large-field-of-view (LFOV) to perform complex image processing continuously for pedestrian detection and localization [60]. The purpose of the proposed Large-Field-of-View deep network was to understand, simultaneously and effectively, as well as to make classification decisions in many places. The LFOV network processes vast regions at much higher speeds than traditional deep networks and therefore can re-use calculations implicitly. With 280 ms per image on the GPU and 35.85 on the average miss rate on the Caltech Pedestrian Detection Benchmark, the pedestrian detection system showed a promising performance for real-world deployment. A vision based pedestrian detection and pedestrian behavior classification technique was proposed by Zhan et al. [61], where YOLOv3-TINY was used for quick segmentation and multitarget tracking of detected pedestrians with the DeepSort algorithm [62]. Finally, to identify the behavior of pedestrians, an improved and customized AlexNet algorithm was adopted. The proposed model performed efficiently in real time at a rate of 20 frames per second along with a designed warning area binding each pedestrian. Convolutional neural network is one of the most popular deep learning models and has been adopted in several studies for pedestrian detection. Ghosh et al. used a novel CNN architecture model for pedestrian detection [63]. To train the model, they applied transfer learning as well as synthetic images using an uncovered region proposal of a bounding box to avoid the annotation of pedestrians’ positions. It obtained a 26% missing rate in CUHK08 and a 14% missing rate in the Caltech pedestrian dataset, where crowded scenes were considered. The biggest advantage was that it required no explicit detection while training and did not need any region proposal algorithm. A similar concept was used by Wang et al., who combined part-level fully convolu- tional networks (FCN) and CNN to generate a confidence map and pedestrian location based on the aligned bounded box concept [64]. The proposed framework was compared with CifarNet and achieved 6.83% improved outcomes. A novel single shot detector method based on late-fusion CNN architecture was introduced by Hou to analyze data of a multispectral system that performed with higher accuracy at nighttime [65]. The combined architecture was applied to a KAIST multispectral pedestrian benchmark where the late-fusion CNN architectures worked efficiently. In terms of log average miss rate, it decreased by more than 10% and developed for suitable deployment during both day and nighttime. As a result, it became one of the best practical CNN-based pedestrian detectors of all the accepted AVS methods. For identifying pedestrians in low resolution learning from low-level image features, a single image-based novel resolution-aware CNN-based framework was proposed by Yamada et al. [66]. The authors also developed a multiresolution image pyramid and obtained the original input image to identify pedestrian size. Moreover, it learnt feature extraction from a low-level image with resolution information and achieved 3.3% lower log-average miss rate than CNN which made the architecture more acceptable for AVS. In another work, Zhang et al. implemented an optimized multiclass pedestrian identification system, using a Faster RCNN-based neural network [67]. The analysis indicated that the framework for pedestrian detection in blurred fields of view were able Appl. Sci. 2022, 12, 6831 15 of 51 to increase speed with average precision of 86.6%. This approach could be suitable for distorted images for pedestrian detection tracking. Dong et al. proposed a region proposal framework for pedestrian detection imple- menting R-CNN combined with an ACF model, where the ACF model was applied to produce only pedestrian class-bounding region, which was a very useful application for autonomous vehicle systems [68]. Moreover, the proposed framework cost less execution time during training and testing. Though most of the studies showed pedestrian detection in the daytime or in clear weather, this task becomes more complex in low-light condition, haze or fog because these create vision difficulties and this kind of condition causes a higher number of accidents [69,70] and increases the possibility of traffic accidents by 13% [71]. Correspondingly, de-hazing algorithms were one of the solutions to fix vision problems which can be implemented in detection of pedestrians in haze conditions. For instance, a related approach for pedestrian detection in haze conditions was proposed by Ding et al., implementing a synthesized haze version of the INRIA dataset using dark channel prior-based linear SVM and HOG algorithm [72]. Nonetheless, the approach received poor recall value, scoring 67.88% in predicting constant depths of input images. Although the method is a good approach for pedestrian detection, the limitations could be solved taking into account the pre-trained pedestrians’ depths in multiple haze environments. Moreover, Huang et al. provided a Laplacian distribution model that featured a combined HTE (haze thickness estimation) and IVR (image visibility restoration) for solving problems of pedestrians [73]. Implementation of this algorithm could enhance the performance for detecting pedestrians in haze conditions, but the most difficult haze conditions occur in dim light conditions, where they achieved 4.1 mAP and 3.98 mAP based on expert and ordinary view, respectively. For alleviating the haze problem while detecting pedestrians, Li et al. [74] proposed three approaches named Simple-YOLO, VggPrioriBoxes-YOLO and MNPrioriBoxes for pedestrian detection based on YOLO [75]. Deep separable convolution and linear bottleneck capabilities were implemented to minimize parameters and enhanced processing speed, making the network far more usable. The average precision of their three methods was 78.0%, 80.5% and 80.5%, respectively, where precisions were 89.4%, 90.8% and 89.3%. The lowest 22.2 FPS were 22.2, 81.7 and 151.9 applied to the combined data of their HazePerson dataset and INRIA person dataset after dataset augmentation. Although this approach was one of the preferable approaches for AVS to detect and localize pedestrians in day and night haze conditions, the higher missing rate in complex scenarios was an issue which could be resolved by adopting key point detection methods. Xu et al. proposed a ground plane context aggregation network (GPCANet) for detecting pedestrians in ground plane areas of Caltech, SCUT and EuroCity datasets, where the best result was achieved for the SCUT dataset with 96.2% recall value, and obtained 25.29% and 25.10% log average miss rate for the rest of the dataset serially [76]. However, it might have slightly higher log average miss rate, but the outcomes were in crowded traffic complex scenarios, which made the approach more practicable for AVS. Moreover, CNN-based work was demonstrated with only 5.5% miss rate to localize distracting pedestrians [77]. Similarly, CNN cascaded with AdaBoost was deployed for pedestrians in night images [78]. It obtained a maximum 9% log-average miss rate, although both methods are not evaluated in complex scenarios. In summary, multiple deep learning methods were reviewed (shown in Table 7) where a CNN-based method was deployed for faster pedestrian detection and localization where the methods showed 94.5% success rate and provided an improved dataset built on the Caltech dataset. FCN achieved 6.83% improved outcomes compared with CifarNet, while in terms of estimating distance of pedestrians from the vehicle, it showed a higher missing rate. Moreover, GPCANet performed best on the SCUT dataset, scoring 96.2% recall in 320 ms and deployed in diverse scenarios in both day and night conditions. However, it scored a high missing rate and could not deal with complex scenes in terms of occluded road objects. However, when of the other methods showed stable efficient outcomes, Appl. Sci. 2022, 12, 6831 16 of 51 the main challenges remained for crowded traffic scenes and complicated visual and weather conditions. Table 7. Summary of multiple deep learning methods for pedestrian detection. Ref. Method Outcomes Advantages Limitations Processed in large field images, Higher missing rate with 35.85 on the average [60] DNN continuous detection in comparatively miss rate. complex scenes. slow detection Designed faster warning area Only considered daytime YOLOv3-TINY and 80.3% accuracy in [61] bounding by each pedestrian scenarios and DeepSort complex environment. with direction labelling. lower accuracy. 26% and 14% missing Did not apply for motion Did not require explicit [63] CNN rates on two images or detection in crowded scenario. datasets, respectively. real-time problems. 6.83% improved Estimated accurate distances Part-level FCN High missing rate for [64] outcomes compared of pedestrians generating the and CNN practical implementation. with CifarNet. confidence map using FCN. Decreased by more than SSD-based Most applicable in Slower detection and [65] 10% of log average late-fusion CNN nighttime implementation. complex parameter tuning. miss rate. Learnt feature extraction from Was not applied in complex Resolution 3.3% lower log-average [66] low-level image with traffic or aware-based CNN miss rate than CNN. resolution information. crowded environment. 86.6% average precision Outperformed on distorted Did not consider low-light or [67] Faster R-CNN in 0.104 s. and blurry frames. traffic scenarios. 14.1%, 15.3%, 45.6% R-CNN with Reduced the number of region [68] miss-rate on —- ACF model proposals, and costs less time. three datasets. Pedestrian detection and Dark 81.63% precision and Presumed constant depths in [72] position estimation from channel-based SVM 67.88% recall. input images. haze condition. 4.1 mAP and 3.98 mAP Laplacian Pedestrian detection in Was not applied in real-time [73] on expert and ordinary distribution model complex dim-light condition. driving scenes. view, respectively. Highest average Minimized number of Early dependency on Multiple [75] precision 80.5% parameters and outperformed preliminary boxes during YOLO methods in 81.7 ms. state-of-art methods. detection process. Improved outcomes in both Higher log average missing 96.2% recall in 320 ms on [76] GPCANet day and night including far rate for occluded SCUT dataset. FoV and crowded traffic. on-road objects. Localized distracting Did not test in cases for [77] CNN Showed 5.5% miss rate. pedestrian and improved crowded or complex scenes. detection annotations. CNN cascaded Generated the maximum Combined thermal images for Might fail in complex urban [78] with AdaBoost 9% log-average miss rate. nighttime detection. traffic scenarios. 3.1.4. Lane Detection and Tracking One of the core fundamentals for AVS was to identify lane and tracking curves in real time where the controlling would depend on the lane and curves. Several studies have been conducted on this field based on different camera visions implementing deep learning and computer vision approaches considering color, texture, feature extraction in different scenarios for lane detection, lane shifting, lane keeping and overtaking assisting. Appl. Sci. 2022, 12, 6831 17 of 51 A road scene sectioning framework was adopted by Alvarez et al. using a CNN-based algorithm to retrieve the 3D scene layout of the street image from noisy labels combining online and offline learning [79]. The proposed method built with color plane fusion and CNN was able to achieve 95.5% to extract a single image of a lane without manual labelling. This CNN-based approach could be considered as the most efficient method for deploying in unknown environments for AVS road-feature extraction. However, for each pixel of the image, which was a path or a lane, authors Dong et al. considered the visual road-detection challenge applying a U-Net-prior network with the DAM (Domain Adaptation Model) to reduce the disparity between the training images and the test image [80]. The proposed model was compared to other state-of-art methods such as RBNet [191], StixeNet II and MultiNet [192], where the max-F measures were 94.97%, 94.88% and 94.88%, respectively, in 0.18 s, 1.2 s and 1.7 s. Their methodology obtained 95.57% max F-measurement in 0.15 s faster and more accurately than others, which indicates that their monocular-vision-based systems achieve high precision for a lower running time. Another kind of approach for storing processes of previous stages, a method based on a combination of CNN and recurrent neural network (RNN), was proposed by Li et al., which was able to identify lane markers using geometry feedback with maximum 99% AUC value [81]. However, since no image pre-processing was conducted, this process took a lot of time in sorting unrelated image areas. In addition, these methods were either time-consuming or inefficient in a true, dynamic world, which does not fulfil the maximum efficiency restriction of a critical function. The Bayesian method for estimating multihyperbola parameters splitting frames in multiple patches was demonstrated by Fakhfakh et al. to recognize curved lanes under difficult conditions using [82]. The lane line was represented on each component by a hyperbola which was determined using the proposed Bayesian hierarchical model with an average of 91.83% true positive rate (TPR) on the ROMA dataset. To sum up the theory, it could be made more practical by adopting sampling techniques such as Hamiltonian schemes to enhance the model’s performance. Yang et al. suggested a substitution of image pre-processing to reduce the uncertainty about lane state [83]. Their approach uses profound lane detection based on deep learning as a substitute for practical lane detection with UNet encoder including high-grade GPU processing. The paper also states that the CNN-based UNet with Progressive Probabilistic Hough Transformation, UNet, Kalman filter were far more inefficient in terms of identifi- cation than the feature-based approaches, such as Hough Transformation (HOG) for lane tracking in real time [84–86]. For predicting lane line under the most challenging conditions, a spatiotemporal- based hybrid architecture after encoding–decoding SCNN and ConvLSTM [87]. This is the very first approach which improves temporal correlation with spatial relation of feature extraction with 98.19% accuracy and 91.8% F1 score. However, although this is one of the strongest approaches, the authors did not apply it to complex weather and nighttime scenarios. Furthermore, to resolve instance level and complex fork and dense line-detection issue, a novel approach was implemented, CondLaneNet, using recurrent instance module applied to a CULane dataset [88]. The approach obtained an 86.10% F1 score while detecting curve lane in complex scenarios despite the lack of proper refining of contextual features. Multiple deep learning methods were studied regarding the lane curve tracking system. For instance, Dorj et al. deployed circle equation models and parabola equations to redesign the Kalman filter for curved lane tracking with a view to calculating curving parameters in far field view [89]. Although the algorithm had an independent threshold mechanism to compensate for various light conditions, such as low light, further research was needed to identify lane reflections and shadows. The limitation of Dorj et al. was solved in [90], where the authors applied a local adaptive threshold and RANSAC feedback algorithm to prevent misdetection of the lane by estimating two-lane parameter-based issues. Nevertheless, the algorithm did not allow a close-loop lane to maintain lane control Appl. Sci. 2022, 12, 6831 18 of 51 while following the road lane, showing a slightly higher false positive rate (FPR) and slow execution for processing in the CPU only. However, it achieved 99.9% precision, 98.9% accuracy and 99.4% F-measurement in 0.677 fps complex visual and lighting conditions. Similarly, for overcoming lane detection in complex shadow and lighting conditions full of obstacles, a CNN-based method was presented by Wang et al. [91]. From an inverse perspective, the application of a fixed transformation matrix generated errors as changes occurred, allowing the predicted exhaust point to infinitely shift upward or downward. The authors trained a neural network with a custom loss function that predicted the transformable matrix parameter valued dynamically. The method was implemented on the TuSimple dataset and obtained high accuracy for insufficient light, shadow, missing lane and normal road compared to other deep learning methods, such as Spatial CNN, CNN-FCN and UNet. As an approach to preventing lighting condition problems for lane detection and tracking, a novel CNN-based model was proposed by Ye et al. [92]. In the pre-processing stage they adopted Yam angle prediction and filtering, followed by segmenting ROIs implementing waveform generation that generated on average 99.25% accuracy in the BIT dataset considering nine cases where day and nighttime accuracies were 99.34% and 98.66%, respectively, as well as a 1.78% average error rate for the Caltech dataset. However, this methodology could be the most suitable candidate only if it is performed with similar outcomes in real-life experiments of level 4 or 5 AVS. A similar CNN-based approach combined with CNN-LSTM, SegNet and UNet was applied by Zou et al. for lane detection from occlusion scenarios [93]. The method obtained 96.78% accuracy for SegNet and 96.46% for UNet within 46 ms, which was much faster than the average of the other methods. With faster processing and high accuracy, this approach could be considered as one of the most acceptable methods for AVS lane detection. Jhon et al. proposed a lane-detection algorithm, which calculated the semantic road lane by using the extra tree-based decision forest and DNN from a street scene where hue, saturation, depth (HSD) combined with a deconvolutional network were fine-tuned [94]. In the final stage, a separate extra tree regressor was trained within each lane applying the depths and the manually annotated lane marker locations on the image. The methodology was applied to the TTI and TMD datasets, where it achieved 98.80% and 97.45% accuracy, respectively, for lane detection. Further, encoder–decoder dilated convolution and finely tuned improvements were implemented by Chen et al. to create a modified CNN road lane detection system called Lane Mark Detector (LMD) which increased the accuracy of the CamVid dataset to 65.2%, obtained 79.6% class average accuracy and increased the test speed to 34.4 fps as well as improved the inference time (29.1 ms) and smaller model size of 66 mb [95]. Moreover, Ghafoorian et al. used Embedding Loss-Driven Generative Adversarial Networks (EL-GAN) for detecting road lanes [96]. This led to even more secure training with stronger discrimination and stabilized the mechanism of adverse preparation. This significantly stabilized the process of opposing training. EL-GAN was also applied to the TuSimple dataset and achieved 96.39% accuracy despite requiring the tuning of a suitable number of parameters. As the loss of embedding into classification boosted the maximum efficiency of the lane marking method, it was one of the best and most appropriate approaches for continuous lane detection and tracking. Tracking lane during nighttime was one of the most difficult tasks of AVS. He et al. solved the issue by developing a vision-based lane detection system, where they pre- processed with a Gabor filter, continuing adaptive splay ROI and Hough transformation to detect the lane marker [97]. Despite lacking an appropriate self-switching system for defining lanes in all circumstances in pre-processing, the detection rates were 97.31% and 98.15% using two clips of Guangzhou where frame numbers were 3274 and 2231. However, the method faced difficulties when tackling bright light reflection, missing lane marks and lane cracks as well. Appl. Sci. 2022, 12, 6831 19 of 51 Neven et al. formulated a solution using LaneNet and HNet for the problem of lane detection with an instance segmentation problem in which each lane constituted its own instance to be end-to-end trained [98]. In addition, to a set “bird’s-eye view”, they introduced a learning transfer to the perspective, which was contingent on the image and achieved 96.4% accuracy within 50 fps (frames per second) for the TuSimple dataset. The method was robust enough to adjust the pitch of the ground plane by adapting the transition parameters accordingly, which was the main reason for accurate visualization, and detected lane and lane curves. Moreover, Kim et al. proposed a fast-learning environment using extreme learning CNN (EL-CNN) combining extreme learning machine (ELM) calculating weights among output and hidden layers in one iteration with CNN for lane marking extraction in complex scenarios to overcome computing of large dataset [99]. It reduced training time 1/50 for the KNU dataset, and 1/200 for the Caltech dataset compared to CNN. Experimental results demonstrate that it obtained maximum weights effectively while maintaining performance of 98.9% accuracy applied to the Caltech dataset. In another work, Van Gansbeke et al. implemented ERFNet with differentiable least- squares fitting (DLSF) for end-to-end lane detection [100]. The approach used dynamic backpropagation to perform an experiment on a lane detection task that demonstrated that, despite the poor supervision signal, the end-to-end approach exceeded a two-step procedure, scoring 95.80% accuracy in 70 fps applied to the TuSimple dataset. The accuracy was not the maximum, but the weight map did not require post processing for accurate lane estimation. Hou et al. proposed a lane detection CNN by self-attention distillation (SAD) which had self-learning ability in the training phase and boosted the visual attention of multiple layers in different networks and increased the efficiency of narrow-lane detection sys- tems [101]. The method obtained 96.64% accuracy in the CULane, BDD100 K and TuSimple datasets, although the hyperparameter adjustment was complicated by an insufficient training process and loss functions. In another work, Liu et al. used a lightweight YOLO network for lane curve detection and tracking with 90.32% precision and 83.76% recall in 50.6 ms [44]. The method was applied to a custom dataset which was evaluated for day and night scenarios. However, the efficiency could be better suited to proper AVS if it solved the interruption of vehicles during lane detection. In conclusion, most of the approaches have performed well enough to be adopted for practical implementation of AVS. However, modified CNN [92] was able to detect lanes with highest accuracy for both day and nighttime, and the CNN-LSTM-based SegNet and UNet combined approach was [93] able to segment roads within the fastest runtime. The analysis presented some advantages of deep learning methods for lane and road curve detection, for instance, training without manual labelling, reducing computational complexing while in a single frame, lane detection where markers were not clear, in sharp turns and even challenging weather and shadow or low-light conditions. On the other hand, some methods showed huge dependency on dataset pre-labelling, which was inefficient in the long field of view, resource hunger and even not being evaluated in urban traffic scenarios or challenging road conditions. An overview of the deep learning methods reviewed for the detection of lane and road curves is shown in Table 8. 3.1.5. Traffic Scene Analysis Driving scene and driving behavior analysis of autonomous vehicle systems were denoted as the understanding and classifying of driving environment and traffic scene. To discuss the contribution of deep learning to understanding and analyzing complex traffic scenes, several studies were conducted. Appl. Sci. 2022, 12, 6831 20 of 51 Table 8. Summary of multiple deep learning methods for lane detection and tracking. Ref. Method Outcomes Advantages Limitations Reduced dependency of manual Deployment for testing did not 95.5% accuracy for [79] Modified CNN labelling and processing time in consider urban or single frame. single frame. crowded scenarios. 95.57% max Smooth segmentation of road surface Huge dependency on manual [80] U-Net F-measurement in 0.15 s. with multiple objects as obstacles. pre-labelling. Recognized region in complex traffic Higher computational cost and [81] Multitask CNN and RNN Max AUC value of 99%. and visualized spatially inefficient for large field of view. distributed cues. Automated curve detection in rural Lighting conditions were not [82] Bayesian Model 91.83% true positive rate. and challenging roads with lower considered and slow processing. error rate. 2.5% and 9.75% lateral Obtained less lateral error and Limited to a simple close-loop [83] UNet and Kalman Filter error generated in 10 ms. overcame slow feature extraction. circle in TORCS simulator. Solved the lane detection in fork and Contextual features need to [88] CondLaneNet 86.10% F1 score. dense scenarios. be refined. Prevent misdetection by estimating Slower execution, slightly high 99.9% precision and 99.4% [90] RANSAC parameters when FPR and did not consider urban F-measurement. illumination changes. traffic road. Created errors during shifting the Obtained highest accuracy Outperformed in shadow and roads [91] CNN ground to predict (97.85%). with obstacles. disappearing point Most suitable deployment in two Did not test in real-time [92] Modified CNN Average 99.25% accuracy. datasets with faster runtime for driving senses. nine classes Combined method of CNN and Raised performance in [93] CNN-LSTM with SegNet 96.78% accuracy in 46 ms. RNN were resource hungry occlusion scenarios. and slow. Modified Decision forest 98.80% and 97.45% High accuracy in surface and [94] High computational cost. and DNN accuracy, respectively. road detection. Modified CNN to achieve low 65.2% mIoU, 79.6% class Was not tested in crowded [95] CNN complexity and maintained average accuracy. traffic environment. similar accuracy. The loss of embedding in detector Required tuning of huge number [96] EL-GAN 96.39% accuracy. boosted the performance closest to of parameters. the mark. Did not require post processing, 96.4% accuracy within 19 Faced challenges in long field of [98] LaneNet and HNet pixel-wise segmentation or fix ms. view while detecting curves. lane number. Reduced training time 1/50 and 1/200 Required matrix inversion for [99] EL-CNN 98.9% accuracy. in KNU and Caltech datasets, better execution time in high respectively. dimensional data. Did not need post processing to Was not tested in urban or [100] ERFNet-DLSF 95.80% accuracy. estimate line coordinates using complex lighting conditions. weight map. Complex hyperparameter Has self-learning ability and increased [101] CNN with SAD 96.64% accuracy. adjustment for inadequate the efficiency of narrow-lane detection. training process. Lightweight 90.32% precision and Applicable in both day and night High interruption for [44] YOLO Network 83.76% recall in 50.6 ms. scenarios and multiple fields of view. obscured vehicles. To contribute to this field for developing traffic scene analysis for AVS, Geiger et al. proposed a novel method of generative probabilism to understand traffic scenes with the Markov Chain Monte Carlo, which was used to deal with the dynamic relationship between crossroads and feature presentation [102]. The human-inspired method took the benefit from a wide range of visual cues through the form of vehicle directions, vanishing points, semantic scene labels, scenario flow and grids rather than requiring sensor values Appl. Sci. 2022, 12, 6831 21 of 51 such as LiDAR and GPS, where most of the standard methods struggled for most of the intersections due to the lack of these attribute labels. Moreover, the method can accurately identify urban intersections with up to 90% accuracy at 113 real-world intersections. Another scene semantic segmentation approach is the High-Resolution Network (HRNet) proposed by Wang et al. [103], where the method obtained 81.1% mIoU. HRNet linked the high-to-low resolution convolution streams in parallel and transferred data across repeatedly. The advantage of the method was that the resulting representation was richer semantically and spatially. However, it required huge memory size due to high resolution-wise segmentation. Additionally, the same author improved their previous work applying contrastive loss to previous architecture (HRNet), which explored pairwise pixel-to-pixel dependencies applied to the Cityscape dataset and obtained 1.1% higher mIoU [104]. Although the proposed method demonstrated effective performance, which is applicable for top-tier AVS, it was unable to achieve success during contrastive learning in few parts of the labelled dataset. To tackle this issue, Zhao et al. [105] presented a contrastive approach following previous research [103,104] and proposing SoftMax tuning rather than applying contrastive loss and cross-entropy at once. The authors demonstrated three variants of label and pixel-wise contrastive losses by adopting DeepLabV3 with ResNet-50 with 256 channels of convolution layers and bilinear resizing for input resolution for semantic segmentation. This approach showed 79% and 74.6 mIoU, respectively, for Cityscape and PASCAL VOC 2012 datasets but using 50% less labelled dataset. Thus, powerful semantic segmentation with a fine-tuned pretrained method can be a major pathway for higher level AVS for scene analysis. Furthermore, to develop a scene recognition framework, Tang et al. demonstrated GoogleNet for multi-stage feature fusion, named G-MS2F, segmented into three layers to feature extractions and scoring scene understanding, that can be efficiently employed for autonomous driving systems [106]. The framework obtained 92.90%, 79.63% and 64.06% accuracy, respectively, when applied to the Scenel5, MIT67 and SUN397 datasets for image scene recognition. Similarly, a multiresolution convolutional neural network architecture was proposed by Wang et al. for driving scene understanding in different scales where they used two categories of resolution images in the input layer [107]. A combination of fine-resolution CNN and coarse-resolution CNN was included for recording small and comparatively large-scale visual frameworks. To obtain visual information with more accurate resolution and enhanced spatial information, on an inception layer, three convolutional layers were added. They implemented the architecture on the Place365 dataset where the lowest error rate was 13.2%. Moreover, a 2D-LSTM model was proposed to learn information from surrounding context data of scene labels as well as spatial dependencies in [108] within a single model that generated each image’s class probabilities. They obtained 78.52% accuracy when deploying on the Standford background dataset. Fu et al. introduced an integrated channel contextual framework and spatial contextual framework as a contextual deconvolution network (CDN) that used both local and global features [109]. In an attempt to optimize the visualization of the semantic data, the decoder network utilized hierarchical supervision for multilevel feature maps in the Cityscapes dataset and achieved 80.5% mean IoU. Following the work, an optimized model of a deep neural network was proposed with two distinct output directions by Oeljeklaus et al. Their method foresaw road topology along with pixel-dense categorization of images at the same time, and lower computing costs were offered in real-time autonomous applications via a proposed architecture com- bined with a novel Hadamard layer with element-wise weights using Caffe and achieved 0.65 F1, 0.67 precision and recall 0.64 after fine-tuning the architecture with 10,000 it- erations [110]. Although strong restrictions placed by the double-loss function on the DNN feature maps caused difficulties in optimizing the process, research in relation to the Appl. Sci. 2022, 12, 6831 22 of 51 Cityscapes dataset showed that a sufficient representation of traffic scene understanding was achieved relying on broad traffic components. In another work, Xue et al. presented a CNN with Overlapping Pyramid Pooling (OPP) applied to sematic segmentation of city traffic area based on a fisheye camera with wider vision [111]. The OPP was demonstrated for the exploratory study of the local, global and pyramidal local context information to resolve the complicated scenario in the fisheye image. Furthermore, they built novel zoom augmentation for augmenting fisheye images to boost performance of the method where it scored 54.5 mIoU, which is higher than the standard OPP-Net and Dilation10 method. This approach could be highly suitable for short FoV traffic scene understanding in urban areas. Pan et al. proposed Spatial CNN, a CNN-like framework for efficient spatial distribu- tion of information through slice-by-slice message passing from the top hidden layer [112]. It was tested at two roles: lane recognition and traffic scene perception. The analysis showed that the continuity of the long, small structure was appropriately preserved by SCNN, while its diffusion effects have proven positive for large objects in semantic segmen- tation. However, SCNN can master the spatial relationship for the structural production and increase operating efficiency, showing that SCNN was 8.7% and 4.6% superior to the recurrent neural network (RNN) focused on ReNet and MRF + CNN (MRFNet). It scored 68.2 mIoU for semantic segmentation and achieved 96.53% on the TuSimple Benchmark Lane Detection Challenge combined with traffic scene analysis. Mou et al. proposed a vision-based vehicle behavior prediction system by incorporat- ing vehicle behavior structural information into the learning process, obtaining a discrete numerical label from the detected vehicle [113]. The OPDNN (overfitting-preventing DNN) was constructed using the structured label as final prediction architecture, and after more than 7000 iterations, 44.18% more accuracy on-road vehicle action than CNN was achieved. In addition, the method decreased the issue of overfitting in a small-scale training set and was highly efficient for analysis of on-road vehicle behavior predicting turning angles. In another work, Jeon et al. proposed a model built on the CNN and Long Short- Term Memory (LSTM) networks to predict risk of accidents and minimize accidents and analyzing traffic scenes differing conditions of driving such as lane merging, tollgate and unsigned intersections [114]. They implemented a multi-channel occupancy Grid Map (OGM) as a bird’s-eye view that ostensibly included the features of many interaction groups to represent the traffic scene [85]. Additionally, the CNN was used to derive numerous inter-vehicle interactions from the grid and to estimate possible time-serial predictions of the derived functions. For instance, Lui et al. demonstrated a deep understanding of the vehicle-specific scene understanding state-of-art in terms of using traffic environment as object joining automatic scene segmentation and object detection, which reduced the person manipulation [55]. A SegNet network whose weight was initialized by the VGG19 network was used for semantic segmentation on the Auckland traffic dataset [115]. Afterwards, a Faster RCNN-based approach transformed feature maps in the ROI (ROI) and transferred those to the classification mode. It had an accuracy of 91% for sky detect, 90% for bus lane, 86% for road, 70% for lane and 81% building classes applying VGG19-SegNet. However, it suffered from false rate for not having a high-resolution labelled dataset and a weak vehicle detection process. Furthermore, two state-of-the-art versions of machine learning and deep learning (DNN) were used by Theofilatos et al. to estimate the incidence of a crash in real time where the dataset comprised historical accident information and combined current traffic and weather information from Attica Tollway, Greece [116]. The method achieved accuracy, precision, recall and AUC of 68.95%, 52.1%, 77% and 64.1%, respectively. The limitation was the transferability while returning the parameters and the absence of good interplay during comparison and insufficiently clarified unexpected heterogeneity. The possible solution offered by the authors was to apply a sensitivity analysis that was not used when applying a binary logistic model in their work to determine risk of crashes. Appl. Sci. 2022, 12, 6831 23 of 51 Moreover, Huegle et al. proposed a Graph-Q and DeepScene-Q off-policy reinforce- ment learning-based approach for traffic scene analysis and understanding applied to a custom dataset [117]. The proposed method used dynamic awareness-based scene under- standing for AVS, although it was not tested in a real driving environment and was unable to track lanes while moving quickly. With a view to understanding hazardous or damaged roads in a driving situation for a smooth autonomous driving experience, deep learning approaches can also provide a solution. Nguyen et al. used CNN architecture to identify damages and cracks in the road that reduced false detection and without pre-processing, which helped to decrease compu- tational time [118]. On the other hand, authors adopted a principal component analysis (PCA) method and CNN to classify and sense damaged roads with their own dataset. Another Deep CNN-based approach with discriminative features for understanding road crack identification was developed by Zhang et al., which also could be a pathway to implement in AVS [119]. The core advantage of the framework was self-learning features that did not rely on manual labelling and geometrical pavement predictions. An alternative method for autonomous road cracks alongside pothole detection was demonstrated by Anand et al. as part of an analysis of traffic scene [120]. SegNet was applied with texture that relied on features to separate roads from traffic scene in order to build a first division of mask and concatenated with the second masking, which was created with a 2-canny edge algorithm and dilation. Further, SqueezeNet was applied to the GAPs dataset along with being prepared for deployment in a self-driving vehicle. Compared with a similar approach of Zhang [119], it achieved higher precision, recall and F1 score, leaving one drawback where it failed to recognize cracked road that was misinterpreted as under-construction surface texture. For this outcome, the method of Anand et al. [120] was a more suitable approach for identifying damage road surface. In summary, deep learning approaches such as fine-resolution CNN and coarse- resolution CNN, 2D-LSTM model RNN, HRNet, Deep CNN, Contextual Deconvolution Network, DNN and CNN with pyramid pooling were analyzed which demonstrated high-accuracy traffic-scene understanding from a crowded movable platform, showed less model complexity, being applicable in different scales, avoiding confusion of ambiguous labels by increasing the contrast among pixels, in some cases developing more expressive spatial features and predicting risk of accident. However, some approaches were limited for implementation because of the requirement of re-weighting, which was inapplicable in uncertain environments, slower computational time, low accuracy and inability to focus on objects in dim light and foggy vision. The overall summary is presented in Table 9. 3.2. Decision Making As the world economy and technology have grown and developed, vehicular own- ership has increased rapidly, along with over one million traffic incidents worldwide per year. Statistics indicate that 89.8% of incidents took place because of wrong driver decision- making [193]. To solve this issue with the concept of AVS, the decision-making process was one of the key fields for studying a combined deep learning and deep reinforcement learning-based approach to take humanlike driving decisions when accelerating and de- celerating, lane shifting, overtaking and emergency braking, collision avoidance, vehicle behavior analysis and safety assessment. For instance, the automated driving coordination problem was defined as a problem of the Markov Decision Process (MDP) in the research of Yu et al., during the simulation of vehicle interactions applying multi-agent reinforcement learning (MARL) with a dynamic coordination graph to follow lead vehicles or overtaking in certain driving scenarios [121]. The advantage of the method was when most of the study focused on single vehicle policy, the proposed mechanism resolved the limitation of coordination problem in autonomous driving during overtaking and lane-shifting maneuvers, obtaining higher rewards than rule-based approaches. Appl. Sci. 2022, 12, 6831 24 of 51 Table 9. Summary of multiple deep learning methods for traffic scene analysis. Ref. Method Outcomes Advantages Limitations Efficient in specified scene Highest 91% Showed false rate for not having [55] VGG-19 SegNet understanding, reducing the classification accuracy. high-resolution labelled dataset. person manipulation. Identified intersections from Independent tractlets caused Markov Chain Identify intersections with [102] challenging and crowded unpredictable collision in Monte Carlo 90% accuracy. urban scenario. complex scenarios. Able to perform semantic [103] HRNet 81.1% mIoU. Required huge memory size. segmentation with high resolution. Did not show success of Contrastive loss with pixel-to-pixel [104] HRNet + contrastive loss 82.2% mIoU. contrastive learning in limited dependencies enhanced performance. data-labelled cases. DeepLabV3 79% mIoU with 50% less Reduce dependency on huge labelled [105] Dependency on labelled dataset. and ResNet-50 labelled dataset. data with softmax fine-tuning. Less model complexity and three times Did not demonstrate for [106] Multistage Deep CNN Highest 92.90% accuracy. less time complexity than GoogleNet. challenging scenes. Fine- and Multilabel classification from scene [107] 13.2% error rate. Applicable at different scale. coarse-resolution CNN was missing. Able to avoid the confusion of Suffered scene segmentation in [108] 2D-LSTM with RNN 78.52% accuracy. ambiguous labels by increasing foggy vision. the contrast. Fixed image semantic information and Achieved 80.5% Unable to focus on each object in [109] CDN outperformed expressive mean IoU. low-resolution images. spatial feature. Foresaw road topology with Restrictions by the double-loss DNN with 0.65 F1 score, 0.67 [110] pixel-dense categorization and less function caused difficulties in Hadamard layer precision and 0.64 recall. computing cost. optimizing the process. CNN with Developed novel image augmentation [111] Scored 54.5 mIoU. Not applicable for far field of view. pyramid pooling technique from fisheye images. Performance dropped significantly 96.53% accuracy and Re-architected CNN for long [112] Spatial CNN during low-light and 68.2% mIoU. continuous road and traffic scenarios. rainy scenarios. Required re-weighting for 91.1% accuracy after Decreased the issue of overfitting in [113] OP-DNN improved result but inapplicable 7000 iterations. small-scale training set. in uncertain environment. Slower computational time and Predict risk of accidents lane merging, [114] CNN and LSTM 90% accuracy in 3 s. tested in similar kinds of tollgate and unsigned intersections. traffic scenes. 68.95% accuracy and Determined risk of class from Sensitivity analysis was not used [116] DNN 77% recall. traffic scene. for crack detection. Developed dynamic Graph-Q and Obtained p-value Unable to see fast lane result and [117] interaction-aware-based scene DeepScene-Q of 0.0011. slow performance of agent. understanding for AVS. High accuracy for Identified damages and cracks in the Required manual labelling which [118] PCA with CNN transverse classification. road, without pre-processing. was time consuming. 92.51%, 89.65% recall and Automatic learning feature and tested Had not performed in real-time [119] CNN F1 score, respectively. in complex background. driving environment. Highest accuracy (98.93%) Identified potholes with Failed cases due to confusing with [120] SegNet and SqueezedNet in GAPs dataset. texture-reliant approach. texture of the restoration patches. In another work, the Driving Decision-Making Mechanism (DDM) was built by Zhang et al., using an SVM algorithm, optimized with the weighted hybrid kernel function and a Particle Swarm Optimization algorithm to solve decision-making issues including free driving, tracking car and lane changing [122]. The proposed decision-making mechanism obtained 92% accuracy optimizing an SVM model compared with RBF kernel and BPNN model, where the evaluated performance shows that free driving achieved 93.1% and tracking car and lane changing achieved 94.7% and 89.1% accuracy, respectively, in different Appl. Sci. 2022, 12, 6831 25 of 51 traffic environments within 4 ms for average reasoning time. The authors presented a hypothesis when analyzing the results: for driving decisions, road conditions have nearly no effect on heavy traffic density. Despite achieving good accuracy, some limitations were mentioned, such as not applying to real-world driving environments and not yet investigating critical driving scenes such as sudden presence of pedestrians or objects. This issue of [122], was solved by Fu et al., who proposed autonomous braking, analyzing a lane-changing behavior decision-making system for emergency situations, implementing an actor-critic-based DRL (AC-DRL) with deep deterministic policy gradient (DDPG) and setting up a multi-object reward function [123,124], obtaining 1.43% collision rate. The authors mentioned that using a large training dataset online can be tough and expensive, and the continuous action function decreased the convergence rate and can quickly be lowered to the maximum local. Moreover, to overcome the limitation of reinforcement learning in complex urban areas, Chen et al. used model-free deep reinforcement learning approaches named Double Deep Q-Network (DDQN), Twin Delayed Deep Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) to obtain low dimensional latent states with visual encoding [125]. They improved performance by implementing a CARLA simulator by altering frame dropping, exploring strategies and using a modified reward and network design. The method was evaluated in one of the most complicated tasks, a busy roundabout, and obtained improved performance compared to baseline. In the 50 min test, the three approaches were able to enter with high success rate but performance of DDQN and TD3 decreased after covering a long distance. In the best case, SAC achieved 86%, 80%, 74%, 64%, 58% success rate for first, second, third, desired exits and goal point, respectively, where DDQN and TD3 had an almost zero success rate for desired exit and goal point arriving. To avoid training complexity in a simulation environment, the DDPG algorithm with actor-critic method was applied in [124] using deep reinforcement learning (DRL), consid- ering three reward function braking scenarios: braking too early and too late, and too-quick braking deceleration. The outcomes of their proposed methodology showed that the error collision rate was 1.43% which was gained by evaluating the performance of the diverse initial positions and initial speed strategies. The ratio of obtaining maximum deceleration was 5.98% and exceeding jerk was 9.21%, which were much improved compared to DDPG with steering and DQN with discrete deceleration. A dueling deep Q-network approach was demonstrated by Liao et al. to make a strategy of highway decision making [126]. The method was built for lane-changing decisions to make a strategy for AVS on highways where the lateral and longitudinal motions of the host and surrounding vehicles were manipulated by a hierarchical control system. The outcomes showed that after 1300, 1700, 1950 episodes, the approach was able to avoid collision after 6 h of training and 26.56 s of testing. In another study, Hoel et al. introduced a tactical framework for a decision-making process of AVS combining planning with a DRL-extended Alpha Go algorithm [127]. The planning phase was carried out with a modification in the Monte Carlo Tree Search, which builds a random sampling search tree and obtained a 70% success rate in highway cases. The contrast between traditional MCTS and the variant in this search was that a neural network formed through DRL aimed towards the search tree’s most major aspects and decreased the essential sample size and helped to identify long temporal correlations with the MCTS portion. However, the proposed process considered 20 simulation parameters and 11 inputs to a neural network which were very efficient and made more suitable for practical implementation. Overtaking maneuvers for intelligent decision making while applying a mixed observ- able Markov decision process was introduced by Sezer, solving overtaking maneuvers on two-track roads [128]. In this paper, the author presented a new formulation for the issue of double-way overtaking by the resources of the mixed observability MDP (MOMDP) to identify the best strategy considering uncertainties. This was used for overcoming the problem, and was illustrated by the active solvers’ growth and in cognitive technological Appl. Sci. 2022, 12, 6831 26 of 51 advances by reducing time-to-collision (TTC) methods in different simulations. The method surpassed nine periods, relative to both MDP and conventional TTC methods. However, the limitation of proper discretion can also be considered with respect to the actual speed and distance values. A higher number of states that were specifically connected for computing and MOMDP algorithm tend to be required as the actual implementation hindrance. To overcome the issue of vehicle overtaking which needs an agent to resolve several requirements in a wide variety of ways, a multigoal reinforcement learning (MGRL)-based framework was introduced to tackle this issue by Ngai et al. [129]. A good range of cases of overtaking were simulated to demonstrate the feasibility of the suggested approach. When evaluating seven different targets, either Q-Learning or Double Action QL was being used with a fusion function to assess individual decisions depending upon the interaction of the other vehicle with the agent. The hypothesis of the work was that this proposal was very efficient at taking accurate decisions while overtaking, collision avoiding, arriving on target timely, maintaining steady speed and steering angle. Brännström et al. presented a collision-avoiding decision-making system adopting a Bayesian network-based probabilistic framework [130]. A driver model enabled the devel- oper to carry out early actions in many circumstances in which the driver finds it impossible to anticipate the potential direction of other road users. Furthermore, both calculation and prediction uncertainties were formally discussed in the theoretical framework, both when evaluating driver adoption of an action and when predicting whether the decision-making method could avoid collision. Another important decision-making task is intelligent vehicle lane-changing policy. Based on the area of acceleration and braking mechanism, a method was introduced by Zhu et al. [131]. First, velocity and relative distance acceleration area was developed based on a braking mechanism and acceleration was used as a safety assessment predictor and then, a method for lane changing with the accelerating field was built, while the driver ’s behaviors, performance and safety were taken into consideration. In compliance with the simulation findings, the use of lane-changing decision-making strategies based on the acceleration can be optimized with driver behaviors for lane-change steps, including starting line, span and speed establishing safety at the same time. Although previous approaches presented a decision-making mechanism for lane changing, most of them did not show DMS for behavior prediction while lane chang- ing [132]. A fuzzy interface system with an LSTM-based method for AVS was proposed by Wang et al. to analyze behavior of surrounding vehicles to ensure safety while lane changing with 92.40% accuracy. The novelty of their work was the adjustment of motion state dynamically in advance. Li et al. proposed a framework for the analysis of the behavior, using a gradient- boosting decision tree (GBDT), merging acceleration or deceleration behavior with the data from the trajectory of the vehicle processed in the noise method on the U.S. highway 101 [133]. The partial dependency plots demonstrated that the effect on the fusion of acceleration or deceleration in independent variables by understanding the key impacts of multiple variables, was non-linear and thus distinct from the car tracking behavior with 0.3517 MAD (Mean Absolute Deviation) value, which suggested that the adoption of typical vehicle models in combination results cannot reflect characteristic behavior. Further, DRL with Q-masking was applied by Mukadam et al. to make tactical decisions for shifting lanes [134]. They introduced a system which provided a more organized and data-efficient alternative to a comprehensive policy learning on issues where high-level policies are difficult to formulate through conventional optimization or methods based on laws. The success rate of 91% was 21% higher than human perception and the 0% collision was 24% lower than human perception. This method of DRL with Q-masking worked best in the case of avoiding collision while lane shifting. Similarly, Wang et al. adopted DRL but combined with rule-based constraints to take lane-changing decisions for AVS in a simulated environment and MDP, which was challenging for high-level policy to develop through conventional methods of optimization Appl. Sci. 2022, 12, 6831 27 of 51 or regulation [135]. The training agent could take the required action in multiple situations due to the environment of state representation, the award feature and the fusion of a high level of lateral decision making and a rule-based longitudinal regulation and trajectory adjustment. The method was able to obtain a 0.8 safety rate with superior average speed and lane-changing time. Chae et al. demonstrated an emergency braking system applying DQN [136]. The problem of brake control model was conceived in Markov’s decision-making process (MDP), where the status was provided by the relative location of the hazard and the speed of the vehicle and the operating space specified as the collection of brake actions including no braking, weak, medium and heavy braking operation, combining vehicle, pedestrian and multiple road conditions scenarios, and the obtained collision rate decreased from 61.29% to 0% for a TTC value from 0.9 s to 1.5 s. As a result, this DQN-based approach was selected as one of the most practical systems for SVM in terms of autonomous braking. Furthermore, to analyze high-accuracy braking action from a driving situation declar- ing four variables, that is, speed of host vehicle, time to collision, relative speed and distance between host and lead vehicle, Wang et al. used hidden Markov and Gaussian mixture-based (HMGM) approach [137]. The efficient technique was able to obtain high specificity and 89.41% accuracy despite not considering kinematic characteristics of lead or host vehicle for braking. However, the analysis of four variants while braking could be a pathway to develop an improved version of braking decision making for AVS. When most of the approaches had dependency on datasets, methods such as DRL that combined DL and RL were extremely efficient for driving decision making in an unknown environment. For example, Chen et al. developed a brain-inspired simulation based on deep recurrent reinforcement Q-learning (DRQL) for self-driving agents with better action and state space inputting only screen pixels [138]. Although the training process was long, it resulted in better-than-human driving ability and Stanford driving agent in terms of reward gain, which indicates that this approach was one of the most suitable for applying in AVS. Another DRL-based approach combined with automatically generated curriculum (AGC), was extremely efficient for intersection scenarios with less training cost [139]. The method obtained 98.69% and 82.1% mean average reward while intersection approaching and traverse. However, the approach might lack proper finishing or goal researching in some cases of intersection traverse, but it is still very efficient for not depending on pre-trained datasets. Similarly, continuous decision-making for intersection cases in top three accident- prone crossing paths in a Carla simulator using DDPG and CNN surpassed the limitation of single scenario with discrete behavior outputs fulfilling the criteria for safe AVS [140]. DDQG was utilized to address the MDP problem and find the best driving strategy by mapping the link between traffic photos and vehicle operations through CNN that solved the common drawback of rule-based RL methods deployed in intersection cases. The method obtained standard deviation (SD) values for left turn across path opposite direction and lateral direction, straight crossing path 0.50 m/s, 0.48 m/s and 0.63 m/s, respectively, although it only considered lateral maneuvers and two vehicles in the intersection. In contrast, approach was introduced by Deshpande et al. for dealing with behavioral decision making for environments full of pedestrians [141]. Deep recurrent Q-network (DRQN) was used for taking safe decisions to reach a goal without collision and succeeded in 70% of cases. With the comparatively lower accuracy, this approach also could be very appropriate if deep learning agents were added for better feature analysis. For AVS navigation avoiding on-road obstacles, a double deep Q-learning (DDQN) and Faster R-CNN in a stochastic environment obtained stable average reward value after only 120 epochs with maximum 94% accuracy after 180,000 training steps with hyper- parameter tuning [142]. However, this approach only considered vehicles in parallel and did not show how DDQN and Faster R-CNN are fused. Moreover, the approach was still unable to obtain stable performance in uncertain moments. Appl. Sci. 2022, 12, 6831 28 of 51 Mo et al. demonstrated reinforcement learning agent and an MCTS-based approach to reduce safe decision making and behaviors by safe policy search and risk state prediction module [143]. This research assessed the challenge of decision making for a two-lane overtaking situation using the proposed safe RL approach and comparing it with MOBIL and DRQN. The proposed model outperformed MOBIL and DRQN by scoring 24.7% and 14.3% higher overtaking rate with 100% collision-free episodes and highest speed. Therefore, the proposed Safe RL could be a pathway for current AVS for risk-free trajectory decision making. In conclusion, decision making is the most vital part of an intelligent system, and to obtain acceptable human-like driving decisions, multiple deep learning and deep reinforce- ment learning methods were analyzed (shown in Table 10). The discussed approaches where able to resolve severe limitations and outperformed in overtaking, braking, behav- ioral analysis and significant segments of decision making for full AVS. Table 10. Summary of multiple deep learning methods for decision-making process. Ref. Method Outcomes Advantages Limitations Obtained higher Resolved the limitation of Individual learning of RL [121] MARL rewards than expert coordination problem in agents involved high rule-based approach. autonomous driving. computational cost. Max 94.7% accuracy for Faster decision making in Yet to demonstrate for critical [122] Weighted hybrid SVM lane changing task. different traffic conditions. and uncertain driving scenes. Complexity with large Autonomous braking system training dataset and [131,132] AC-DRL with DDPG 1.43% collision rate. while lane shifting with decreased the high accuracy. convergence rate. In best case SAC Lack of exploration caused Decreased sample complexity [125] DDQN, TD3, SAC achieved 86% failure cases of DDQN with visual encoding. success rate. and TD3. Able to avoid collision Develop lane-changing Still needed to improve [126] Dueling DQN after lowest decision-making strategy for training process for feasible 1300 episodes. AVS on highway. decision making strategy. Combines planning stage to Required huge training Monte Carlo 70% success rate for [127] make efficient driving decision samples and did not consider Tree Search highway exist case. on highway. lead vehicles’ behavior. 91.67% les collision and Efficient overtaking decision Did not consider real-time [128] MOMD + SARSOP 25.9% enhanced without rule-based system and speed and distance value. response rate. optimum actions. Almost 100% safety Outperformed overtaking, Insignificant performance to [129] MGRL index to reach goal collision avoiding, arriving at keep lane while overtaking. without collision. seven RL goals. Higher driver Collision avoidance with path Testing in real and [130] Bayesian network acceptance while prediction and challenging driving scene avoiding pedestrians. threat assignation. was not mentioned. Better lateral and Able to perform emergency Only performed best in [131] Polynomial trajectory steering angle than braking decision at straight-road scenario. ground truth. safe distance. Urban or real-life traffic LSTM with Decision based on the behavior [132] 92.40% accuracy. conditions were Fuzzy Logic of surrounding vehicles. not considered. Appl. Sci. 2022, 12, 6831 29 of 51 Table 10. Cont. Ref. Method Outcomes Advantages Limitations Understand the key impacts of Implemented on old dataset Calibration approach [133] GBDT multiple variables on and driver features were scored 0.3517 MAD. acceleration of fusion. not analyzed. 91% success rate with Effective on high-level policies Did not analyze real-time [134] DRL using Q-masking 24% lower learning through and complex collision rate. conventional optimization. road challenges. Required explicit driving Safety rate 0.8 on Productive data alternative to path while training that [135] Rule-based DQN policy average speed and end-to-end policy learning on caused low performance in lane-changing time. challenge for high-level policy. complex scenes. Efficient and quick emergency Collision rate decreased Ambiguous training and [136] DQN braking in from 61.29% to 0%. environment setting. complex environment. Analysis of kinematic Achieved specificity Highly accurate braking action characteristics for both host [137] HMGRM 97.41% and from driving situation. and lead vehicle 89.41% accuracy. was missing. Obtained maximum Did not require a labelled [138] DRQN 64.48% more reward dataset and only took screen Training time consuming. than human. pixel as input. 98.69% higher average Efficient on intersection Showed few collisions and [139] AGC based DRL mean reward scenarios and reduced DRL unfinished cases during at intersection. training time. intersection traverse. Overcame drawback of single Lowest 0.48 SD for left Only considered [140] CNN + DDPG scenario with discrete behavior turn across path. lateral maneuvers. in intersections. Tackled high-level behavioral 70% success rate for [141] DRQN decision in pedestrian-filled Low accuracy. collision free episodes. environment. Maximum 94% Applied for four driving Limited decision making for [142] DDQL and FRC accuracy with decision for navigation vehicles in parallel sides. stable reward. avoiding obstacle. Outperformed DRQN and 100% collision Did not consider [143] RL + MCTS MOBIL method for safe free episodes. urban scenarios. lane shifting 3.3. End-to-End Controlling and Prediction End-to-end controlling is one of the major fields of study for AVS. Human mistakes were the main cause of road accidents, and fully autonomous vehicles can help reduce these accidents. To improve the control system of AVS analyzing driving scenarios for lane changing, An et al. [144] proposed a system that tried to approximate driver ’s actions based on the data obtained from an uncertain environment that were used as parameters while transferring to parameterized stochastic bird statecharts (stohChart(p)) in order to describe the interactions of agents with multiple machine learning algorithms. Following that, a mapping approach was presented to convert stohChart(p) to networks of probabilistic timed automata (NPTA) and this statistical model was built to verify quantitative properties [145]. In the learning case, weighted KNN achieved highest accuracy combined with the proposed method considering training speed and accuracy, where it achieved 85.5% accuracy in 0.223 s and in the best case, time cost for probability distribution time for aggressive, conservative and moderate driving styles was 0.094, 0.793 and 0.113 s, respectively. The authors categorized Appl. Sci. 2022, 12, 6831 30 of 51 their work into learning phase, modelling phase and quantitative analyzing phase in order to develop the driving decision-taking phase. A method was demonstrated by Pan et al. to control independently at high speeds using human-like imitation learning, involving constant steering and acceleration mo- tions [146]. The dataset’s reference policy was derived from a costly high-resolution model predictive controller, which the CNN subsequently trained to emulate using just low-cost camera sensors for observations. The approach was initially validated in ROS Gazebo sim- ulations before being applied to a real-world 30 m-long dirt track using a one-fifth-scale car. The sub-scale vehicle successfully learnt to navigate the track at speeds of up to 7.5 m/s. Chen et al. focused on a lane-keeping end-to-end learning model predicting steering angle [147]. The authors employed CNN to the current NVIDIA Autonomous Driving Architecture, where both incorporated driving image extraction and asserting steering angle values. To test the steering angle prediction while driving, they considered the difference among ground truth angle which was generated by human drivers vs. predicted angle where they acquired higher steering prediction accuracy with 2.42 mean absolute error and suggested for data augmentation for training to achieve a better performance. In another work, a technically applied system of multitask learning in order to estimate end-to-end steering angle and speed control, was proposed in [148]. It was counted as one of the major challenging issues for measuring and estimating speed only based on visual perceptions. Throughout their research, the authors projected separation of speed control functions to accelerate or decelerate, using the front-view camera, when the front view was impeded or clear. Nevertheless, it also showed some shortcomings in precision and pre-fixed speed controls. By combining previous feedback speed data as a complement for better and more stable control, they improved the speed control system. This method could be stated to solve error accumulation in fail-case scenarios of driving data. They scored 1.26 Mean Absolute Error (MAE) in estimating real-time angles along with 0.19 m/s and 0.45 MAE on both datasets for velocity prediction. Thus, the improved result made the method one of the most applicable versions of CNN and data-driven AV controlling. While driving, people identify the structures and positions of different objects including pedestrians, cars, signs and lanes with human vision. Upon recognizing several objects, people realize the relation between objects and grasp the driving role. In the spatial processing of single images by the application of three-dimensional vectors, CNN has certain shortcoming in the study of time series. However, this issue cannot be overcome using CNN alone. To solve this limitation Lee et al. demonstrated an end-to-end self-driving control framework combining a CNN and LSTM-based time-series image dataset applied in a Euro Truck simulator [149]. The system created a driving plan which takes the changes into account over time by using the feature map to formulate the next driving plan for the sequence. Moreover, NVIDIA currently has succeeded in training a ConvNet for converting raw camera images into control steering angles [150]. It resolved end-to-end control by predicting steering angle without explicating labels with approximately 90% autonomy value and 98% autonomous of the testing period. This approach was one of the most demonstrated approaches that boosted research of AVS applying deep learning methods. A similar method, deep ConvNet, was used by Chen et al. to train for directly extracting the identified accessories from the front camera [151]. A basic control system, based on affordance principles, provided steering directions and the decision to overtake proceeding vehicles. Rather than using lane-marking detection methods as well as other objects to assess indirect activity specifications of the car, a variety of driving measures allowances were specified. This method included the vehicle location, the gap to the surrounding lane markers and records of previous car driving. While this was a very trendy concept, for many reasons it may be challenging to handle traffic with complex driving maneuvers and make a human-like autonomous vehicle controlling system. To deploy a human-like autonomous vehicle speed-control decision-making system Zhang et al. proposed a double Q-network-based approach utilizing naturalistic driving Appl. Sci. 2022, 12, 6831 31 of 51 data built on the roads of Shanghai inputting low dimensional sensor data and high- dimensional image data obtained from video analysis [152]. They combined deep neural networks and double Q-learning (DDQL) [194–196] to construct the deep Q-network (DQN) model which was able to understand and make optimal control decisions in simultaneous environmental and behavioral states. Moreover, real-world data assessment reveals that DDQN can be used on a scale to effectively minimize these unreliable DQN problems, resulting in more consistent and efficient learning. DDQN had increased both in terms of interest precision and policy efficiency. The model performed 271.13% better than DQN in terms of speed-control decision making. Even so, the proposed approach could be more applicable to an unknown driving environment with combined CNN agent for feature extraction. Chi et al. formulated a ST-LSM network that incorporates spatial and temporal data from previously multiple frames from a camera’s front view [153]. Several ST-Conv layers were used in the ST-LSTM model to collect spatial information and a layer of Conv-LSTM was used to store temporarily data at the minimal resolution on the upper layer. However, the spatial and temporal connection among various feature layers was ignored by this end- to-end model. They obtained a benchmarking 0.0637 RMSE value on the Udacity dataset, creating the smallest 0.4802 MB memory and 37.107 MB model weight. The limitation of the paper was that all present end-to-end driving models were only equipped by focusing on the ground truth of the current frame steering angle, which indicated a lack of further spatiotemporal data. Furthermore, to obtain a better control system, the previous issue was tackled, and an end-to-end steering control system was implemented by Wu et al. by concatenating future spatiotemporal features [154]. They introduced the encoding for an advanced autonomous driving control system of spatiotemporal data on a different scale for steering angle approximation using the Conv-LSTM neural framework with a wide-spectrum spatiotemporal interface module. Sequential data were utilized to improve the space- time expertise of the model during development. This proposed work was compared with end-to-end driving models such as CgNet, NVIDIA’s PilotNet [155] and ST-LSTM Network [153], where the root mean square error (RMSE) was 0.1779, 0.1589 and 0.0622, respectively, and showed the lowest RMSE value of 0.0491 to predict steering angles, which was claimed to be more accurate than an expert human driver. Thus, this approach was applicable for a level 4 or 5 autonomous vehicle control system. Moreover, a deep neural network-based approach with weighted N-version Program- ming (NVP) was introduced for resilient AV steering controlling [156]. Compared to the other three networks (chauffeur, autumn, rambo), the proposed network showed 40% less RMSE retrieving steering angles in clear, rain, snow, fog and contrast lighting condi- tions. However, there was a high failure rate for the large developing cost for training an individual DNN model. Aiming to build a vehicle motion estimation system for diversity awareness while driving, Huang et al., via latent semantic sampling [157], developed a new method to generate practical and complex trajectories for vehicles. First, they expanded to include semantic sampling as merging and turning the generative adversarial network (GAN) structure with a low-dimensional semantic domain, formed the space and constructed it. It obtained 8% improvement on the Argoverse validation dataset baseline. They therefore sampled the estimated distribution from this space in a way which helped the method to monitor the representation of semantically different scenarios. A CNN and state-transitive LSTM-based approach was demonstrated with multi- auxiliary tasks for retrieving dynamic temporal information from different driving scenarios to estimated steering angles and velocity simultaneously [158]. The method applied the vehicle’s current location to determine the end-to-end driving model sub-goal angle to boost the steering angle estimation accuracy, which forecasted that the efficiency of the driving model would improve significantly. The combined method obtained 2.58 and 3.16 MAE for steering angle prediction and 0.66 m/s and 0.93 m/s speed MAE in GTA Appl. Sci. 2022, 12, 6831 32 of 51 V and Guangzhou Automotive Cooperate datasets, respectively. Nevertheless, it showed a slow response in unknown environment, so this method might not be applicable in practical implementation. In a similar manner, Toromanoff et al. presented a CNN-based model for lateral control of AVS using a fisheye camera with label augmentation technique for accurate corrections labelling under lateral control rule to tackle ceases of lateral control error in wide FoV [159]. This method compares with pure offline methods where feedback was not implemented from a prediction which resulted in 99.5% and 98.7% autonomy in urban areas and highways after training with 10,000 km and 200 h driving video. On the other hand, Smolyakov et al. reduced a huge number of parameters of CNN to avoid overfitting along with helping to find dependency on data sequence and implement in a CarND Udacity Simulator for predicting steering angles. However, the obtained unsatisfactory result was comparable to other reviewed results, where the accuracy was 78.5% [160]. Similarly, a CNN-based approach was applied for both lateral and longitudinal mo- tion controlling of AVS obtaining 100% autonomy on e-road track on TORCS simulator. Although it had performed very well, contributing to both kinds of motion controlling, it lacked training data for practical implementation and memory consumption for training two different neural networks for speed and steering angle prediction. This method could be better approached by implementing in real scenarios with a good amount of training data [161]. In another proposal, a reinforcement learning-enabled throttle and brake control system was proposed by Zhu et al. [162], focusing on a one leader and one follower formation. A neural dynamic programming algorithm evaluating with trial-and-error method was directly applied for adopting near-optimal control law. The control policy included the necessary throttle and brake control commands for the follower according to the timely modified corresponding condition. Simulation experiments were carried out using the well-known CarSim vehicle dynamic simulator to show the reliability of the approach provided. To overcome traditional sensor-based pipeline for controlling AVS where there is a tendency to learn from direct mapping, Xiao et al. demonstrated multimodal end-to-end AVS applying conditional imitation learning (CIL), taking an RGBD image as raw data in a Carla simulator environment [163]. The CNN-based CIL algorithm was evaluated in different weather modes to identify the performance for end-to-end control. The success rate of controlling in one turn and dynamic environment were 95% and 84%, respectively, which could be boosted through early fusion by changing the number of color channels from three (RGB) to four (RGBD). However, performance dropped almost 18.37% and 13.37% during controlling AVS with RGB image input for one turn and dynamic environment, respectively, in a new map of Carla simulators which could be considered as uncertain area In brief, most of the deep learning approaches for end-to-end controlling and motion predications were based on CNN, showing efficient outcomes suitable for practical level 4 or 5 AVS. Most of the methods were deployed for estimating continuous steering angle and velocity, some controlling approaches taking into account resolving blind spot, gap estimation, overcoming slow drifting, both lateral and longitudinal motion controlling with methods such as multimodal multitask-based CNN, CNN-LSTM, Deep ConvNet, ST-LSTM, neural dynamic programming-based reinforcement learning with actor-critic network and RL. These methods faced challenges, such as noise created by human factor reasoning speed changes causing lower accuracy, only equipped by focusing on the ground truth of the current frame steering angle and not applying in a practical or complex environment. The overall summary of discussed methods is presented in Table 11. Appl. Sci. 2022, 12, 6831 33 of 51 Table 11. Summary of multiple deep learning methods for end-to-end controlling and prediction. Ref. Method Outcomes Advantages Limitations Performed safe control during Unsafe driving behavior and Gained 85.5% accuracy in [144] Hybrid weighted KNN lane changing in did not consider 0.223 s in best case. uncertain environment. complex conditions. Successfully learnt to High-speed driving control Trained only for elliptical CNN + LSTM + [146] navigate the track at and robustness to racetracks with no State method speeds of up to 7.5 m/s. compound errors. other vehicles. Obtained appropriate Able to overcome slow drifting Improper dataset for [147] CNN with comma.ai steering angles with 2.42 from human driving data. practical implementation. mean absolute error. Scored 1.26 MAE for Accurate estimation Noise of human factor for Multimodal-based [148] angles and 0.45 MAE continuous steering angles speed changes caused CNN for velocity. and velocity. lower accuracy. Almost similar steering Resolved the limitation of Lack of driving data collection [149] CNN and LSTM prediction value as CNN and blind-spot problem. from vehicle. ground truth. 90% autonomy value and Required fewer training data Robustness were not successful [150] ConvNet 98% autonomous with no manual in internal processing phase. approximately. decomposition. 0.033 and 0.025 steering Considered lane gap and [151] Deep ConvNet angle MAE on GIST and records of previous Only tested on simple cases. Caltech baseline. car driving. Make optimal control 271.13% better than DQL Measured few errors on [152] DDQN increasing precision and for speed control. uneven roads. policy efficiency. Obtained 0.0637 RMSE Implemented in challenging Only focused on the ground [153] ST-LSTM Network value on Udacity with lighting conditions. truth of steering angle. model weight. Showed lowest RMSE Spatiotemporal Overcome the limitation Did not test in [154] value 0.0491 to predict Conv-LSTM of [153]. busy environment. steering angles. Showed on average 40% High developing cost for Predicted steering angles in [156] DNN less RMSE retrieving training individual multiple conditions. steering angles. DNN model. 8% improvement for Better trajectory building Had not tested in real time and [157] GAN Argoverse validation layers while motion prediction. only used simple semantics. dataset baseline. Used current position and CNN and Predicted 2.58 and 3.16 Slow prediction rate in [158] subgoal angle for steering state-transitive LSTM MAE for steering angle. unknown environment. angle prediction. Achieved 99.5% and 98.7% accuracy in urban areas Solved lateral controlling using Autonomy dropped while [159] CNN and fisheye camera. sharp turning. highways, respectively. Reduced parameters of CNN [160] CNN Achieved 78.5% accuracy. to avoid overfitting on Noticeable performance drop. data sequence. Appl. Sci. 2022, 12, 6831 34 of 51 Table 11. Cont. Ref. Method Outcomes Advantages Limitations Obtained 100% autonomy Showed both lateral and Lack of training data and [161] CNN on e-road track on TORCS. longitudinal motion control. memory consuming. Robust throttle and brake Learned controlling policy Environmental surroundings [162] RL with ACN value of the host vehicle. while following lead vehicle. were not stated. Demonstrated successful 84% success rate in Up to 18.37% performance [163] CIL multimodal approach in dynamic environment. drop in unknown map. four cases. 3.4. Path and Motion Planning Precipitation-based autonomous navigation including path and motion planning in an unknown or complex environment is one of the critical concerns for developing AVS. To tackle the current problem and analyze the contribution, multiple deep learning and deep reinforcement learning (DRL) combined methods for path and motion planning are reviewed in this section. Initially, You et al. focused on the issue of path planning of autonomous vehicles in traffic in order to repeat decision making by replicating the optimum driving technique of expert drivers’ actions for lane changing, lane and speed maintenance, acceleration and braking in MDPs on highways [164]. The optimal control policy for the proposed MDP was resolved using deep inverse reinforcement learning (DIRL) and three MaxEnt IRL algorithms by utilizing a reward function in terms of a linear combination of parameterized function to solve model-free MDP. The trajectory proposals were executed at the time of overtaking and the policy recovery was reduced to 99%, even though there was insufficient evidence for the reflection of stochastic behavior. To solve limitations of rule-based methods for safe navigation and better intersection problems for AVS, a vision-based path and motion planning formula was used by Isele et al., adopting DRL [165]. Each wait action was proceeded by another wait or go action, meaning that each pathway was a series of waiting decisions that concluded in a go decision as well as the agent not being permitted to wait after the go action had been chosen. The method secured a success rate for forward, right, left and turn and challenge of 99.96%, 99.99%, 99.78% and 98.46%, respectively, which was 28% faster than the TTC (time-to-collision) method, although performance decreased three times and average time doubled during this challenging situation. Zhang et al. proposed a risk analysis and motion planning system for autonomously operated vehicles focused on highway scenario motion prediction of surrounding vehi- cles [166]. An interactive multiple model (IMM) and constant turn rate and acceleration (CTRA) model were used for surrounding vehicle motion prediction, and model predictive control (MPC) was used for trajectory planning that scored 3.128 RMSE after 5 s dur- ing motion prediction. Although it was designed for connected AVS, it is efficient for vision-based approaches. Another approach, local and global path planning methodology, was presented in an RoS-based environment for AVS by Marin-Plaza et al., where they used the Dijkstra and time elastic bands (TEB) method [167]. The path planning model was able to reach the goal with modest error by calculating Euclidean distance for comparing local and global pan waypoints, where it scored 1.41 m, which is very efficient. However, it was applicable only if the model was not specifically calibrated for the vehicle’s kinematics or if the vehicle was out of track, and did not consider complex scenarios. In another work, Islam et al. established a vision-based autonomous driving system that relied on DNN, which handled a region with unforeseen roadway hazards and could safely maneuver the AVS in this environment [168]. In order to overcome an unsafe navigational problem, they presented object detection and structural segmentation-based deep learning architecture, where it Appl. Sci. 2022, 12, 6831 35 of 51 obtained an RMSE value of 0.52, 0.07 and 0.23 for cases 1 to 3, respectively, and 21% safety enhancement adding hazard avoiding method. Ma et al. proposed an efficient RRT algorithm that implemented a policy framework based on the traffic scenes and an intense search tree extension strategy to tackle traditional RRT problems where it faced a meandering route, an unreliable terminal state and sluggish exploration, and established more sustainable motion planning for AVS [169]. In addition, the integrated method of the proposed fast RRT algorithm and the configuration time space could be adopted in complex obstacle-laden environments to enhance the efficiency of the expected trajectory and re-planning. A significant set of experimental results showed that the system was much quicker and more successful in addressing on-road autonomous driv- ing planning queries and demonstrating its better performance over previous approaches. In another work, an optimum route planner integrated with vehicle dynamics was designed by Gu et al. implementing an artificial potential field to provide maximum workable movement that ensured the stability of the vehicle’s path [170]. The obstacles and road edges were typically used with constraints and not with any arbitrary feature in this method in the optimal control problem. Therefore, when designing the optimum route using vehicle dynamics, the path-planning method was able to treat various obstacles and road structures sharply in a CarSim simulator. The analysis showed that the method reduced computational costs by estimating convex function while path planning. A similar method was proposed by Wahid et al., where they used an artificial potential field with adaptive multispeed scheduler for a collision-avoidance motion planning strategy [171]. Cai et al. demonstrated a novel method combining CNN, LSTM and state model which was an uncertainty-aware vision-based trajectory generation network for AVS’s path-planning approach in an urban traffic scene [172]. The work was divided into two major parts: the first one was a CNN bottleneck extractor, and the second component included a self-attention module for calculating recurrent history and an LSTM module for processing spatiotemporal characteristics. Finally, they designed the probable collision-free path planning with speeds and lateral or longitudinal locations for the next 3.0 s after taking image stream and state information in the past 1.5 s considering as input. The method obtained more centralized error distribution and lower error medium. For safe navigation for AVS in road scenarios with obstacles, a model prediction control-based advanced dynamic window (ADW) method was introduced by Kiss et al. [173]. The method demonstrated differential drive that reached the destination location ignoring the desired orientation and did not require any weighted objective function. A motion planning model based on the spatiotemporal LSTM network (SLN), which had three major structural components, was proposed by Bai et al. It was able to produce real-time feedback based on the extraction of spatial knowledge [174]. First, convolutional long-term memory (Conv-LSTM) was applied in sequential image databases to retrieve hidden attributes. Secondly, to extract spatiotemporal information, a 3D CNN was used, and precise visual motion planning was displayed constructing a control model for the AV steering angle with fully connected neural networks. The outcome showed almost 98.5% accuracy and better stable performance compared with Hotz’s method [147]. Nonetheless, the method was found to minimize state after generating overfitting on antecedent data for time-series data of previous steps, causing more computational cost and time. Another motion planning avoiding-obstacle-based approach was proposed in a simu- lation environment [175]. The motion planning method had the ability to infer and replicate human-like control thinking in ambiguous circumstances, although it was difficult to estab- lish a rule base to tackle unstructured conditions. The approach was able to execute 45.6 m path planning with 50.2 s. In conclusion, very few works have adopted a perception-based path and motion planning for AVS but the existing research adopting deep inverse reinforcement learning and MaxEnt IRL, deep Q-network time-to-go method, Dijkstra and time elastic bands method, DNN, advance RRT, artificial potential field, ADW using model predictive control and fuzzy logic made a remarkable contribution, with high accuracy, collision-free path Appl. Sci. 2022, 12, 6831 36 of 51 planning, 21% safety enhancement adding hazard-avoiding method planning motion in a multilane turn-based intersection. Nevertheless, these methods were not practically implemented or theoretical, and some of the high-performing approaches were not tested in a real-life environment with heavy traffic. An overview of the deep learning methods selected for analysis to improve AVS is presented in Table 12. Table 12. Summary of multiple deep learning methods for path and motion planning. Ref. Method Outcomes Advantages Limitations Insufficient training data for 99% policy recovery Avoid cost function and [164] DIRL stochastic within less data length. manual labelling. behavior representation. Solved limitation of Performance decreased three 28% faster than [165] DRL with TTC rule-based times during TTC method. intersection problems. challenging situation. Score 3.128 RMSE after Considered motions of Lower accuracy in far [166] IMM with MPC 5 s during surrounding vehicles. predicted horizon. motion prediction. Obtained efficient Applicable only if vehicle was Dijkstra and TEB Reach goal with [167] Euclidean out of track in method modest error. distance 1.41 m. simple scenarios. 21% safety Safe navigation adding enhancement adding Tested only on simple [168] DNN hazard detection and hazard-avoiding open highway. segmentation method. method. Took 5 ms and 48 ms Novel cost function to select Limited to [169] Advance RRT for path selection and path and non-rule-based approach. path generation. obstacle-avoiding feature. Visualized potential Reduce computational cost Effects of local minimum issue [170,171] Artificial potential field field in nine by estimating that led AV to be stuck in different scenarios. convex function. a position. Lower error medium Vehicle motion planning Did not consider traffic light CNN + LSTM + [172] and more centralized predicted in multilane and weather condition for State method error distribution. turn-based intersection. performance evaluation. Reached destination Did not require any Limitation occurred with [173] ADW with MPC location ignoring the weighted objective function. constrained kinematics. desired orientation. Almost 98.5% average Able to learn time-serial Minimized state after [174] 3D-CNN accuracy and features from generating overfitting on stable outcome. traffic environment. time-series data. Difficult to establish a rule 45.6 m path planning Human-like control thinking [175] Fuzzy logic base to tackle with 50.2 s. in ambiguous circumstances. unstructured conditions. 3.5. AR-HUD Augmented reality (AR) in head-up display (HUD) or displaying in windshield for autonomous driving system as a medium of final visualizing of activities outcomes from the deep learning approach was overlayed with an autonomous driving system. The AR-based vehicular display system was essential for driving situation awareness, navigation and overall deployment as a user interface. Yoon et al. demonstrated an improved forward collision alert system detection of cars and pedestrians fused into the HUD with augmented reality through using stereo cameras and visualized early alerts where SVM classifier was applied for object recognition and obtained an F1 score of 86.75% for car identification and 84.17% for pedestrian iden- Appl. Sci. 2022, 12, 6831 37 of 51 tification [176]. The limitation of the work was noticed when the observed object moved rapidly and the car suddenly turned; it was visualized with delay. The proposed system yet needed to optimize efficiency and acceleration which in diverse vehicle conditions responds robustly to different and high speeds. An analysis showed personal navigation with AR navigation assist equipped for use with a volumetric 3D-HUD and utilizing its parameters. An interface was developed for assisting to turn faster by locating turn points quicker than during regular navigation [177]. The interface also helped to maintain user eyes and to fix them more precisely on the driving environment after analyzing traffic scenes with deep learning algorithm with proper registration of applications via spatial orientation of AR views on interface. On the basis of the results, however, the inadequate perception of the depth of a specified 2D HUD distance is obvious and the navigation system’s AR interface was ineffective without a 3D HUD. An automatic AR based on a road tracking information method registration was introduced by Yoon et al., with a SIFT matching function and homography measurement method, which defined matching between camera and HUD providing the driver ’s view was positioned to the front, which detected vehicle and pedestrians and converted them into AR contents after projective transformation [178]. This solution was good enough for daytime performance but had limitations at nighttime. Nevertheless, the procedure had the ability to automate the matching without user interference, but it is inconvenient while projecting outcomes which occurred due to misreading local correspondence. Park et al. demonstrated an AR-HUD-based driving safety instruction by identifying vehicle and pedestrians using the INRIA dataset [179]. The identification method was built using SVM and HOG with 72% and 74% in fps accuracy and detected partial obstacles, respectively, applying a billboard sweep stereo (BSS) algorithm. The detected vehicles and pedestrians were overlapped on the HUD with the AR technique. Despite detecting obstacles in sunny and rainy scenarios, it was not deployed for nighttime scenarios. In order to integrate outcomes with AR, the system was divided into two parts by Rao et al., 3D object detection and 3D surface reconstruction, to develop object-level 3D reconstruction using Gaussian Process Latent Variable Model (GPLVM) with SegNet and VPNet for in-vehicle augmented reality UI and parking system [180]. Their AR-based visualization system was built with monocular 3D shaping, which was a very cost-efficient model and needed only a single frame in the input layer. Furthermore, a new traffic sign-recognition framework based on AR was constructed by Abdi and Meddeb to overlay traffic signs with more recognizable icons overlapped in an AR-HUD to increase the visualization of a driver aiming to improve safety [181]. The Haar Cascade detector and the verification of the theory using BoVW were combined with the relative spatial data between visual words, which had proven to be a reasonable balance between resource efficiency and overall results. A classifier with an ROI and allocated 3D traffic sign was subsequently developed using a linear support vector machine that required less training and computation time. During the decision-making process, this state-of-the-art methodology influenced the distribution of visual attention and could be more consistent with the improved approach of deep learning recognition relying on the GPU. Addressing the challenge of overtaking an on-road slow vehicle, a see-through effect- based marker-less real-time driving system had been demonstrated by Rameau et al., applying AR [182]. To overcome the occlusion and produce a seamless see-through effect, a 3D map of the surroundings was created using an upper-mounted camera and implement- ing an in-vehicle pose predictor system. With up to 15 FPS, they presented a faster novel real-time 2D–3D tracking strategy for localization of rear in a 3D map. For the purpose of decreasing bandwidth usage, the ROI was switched to the rear car impacted by an occlusion conflict. This tracking method on AR-HUD showed great efficiency and easy adoption capability for vehicle displaying systems. Appl. Sci. 2022, 12, 6831 38 of 51 To reduce the accident cases, Abdi et al. proposed augmented reality-based head-up display providing more essential surrounding traffic data as well as increasing interactions between drivers and vehicles to enhance drivers’ focus on the road [183]. A custom deep CNN architecture was implemented to identify obstacles and final outputs will be projected in the AR head-up display. For AR-based projection in HUD, firstly, pose prediction of targeted ROIs were carried out and obtained 3D coordinates with points after achieving camera projection matrix to recognize AR 3D registration. This step created a 6-DOF pose of translation and rotation parameters which will be helpful for motion estimation calculation with planar homograph. Afterwards, the RANSAC method was applied to compute the homograph matrix, and OpenGL real camera was synchronized with a virtual camera that showed a projection matrix to map 2D points utilizing 3D surface points and developed a marker-less approach. Lindemann et al. demonstrated an augmented reality-based windshield display system for autonomous vehicle with a view to assisting driving situation awareness in city areas and increase automated driving level from level 4 to 5 [184]. This AR-based windshield display UI was developed based on deep learning-applied object detection to enhance situation awareness, aiming at both clear and lower-visibility conditions where they obtained very different situation awareness scores in low-visibility conditions in disabled windshield display but failed to obtain a good score when windshield UI was enabled. Nevertheless, it worked significantly better in clear weather conditions. Park et al. presented a 2D histogram of oriented gradient (HOG) tracker and an online support vector machine (SVM) re-detector based on training of the TLD (tracking-learning- detector) functional vehicle tracking system for AR-HUD using equi-height mosaicking image (EHMI) [185]. The system initially performed tracking on the pre-computed 2D HOG EHMI, when the vehicle was identified in the last frame. If the tracking failed, the system started re-detection using an online learning-based SVM classification. The tracking system conducted online learning frequently after the vehicle had been registered and minimized the further calculation necessary for tracking as the HOG descriptor for EHMI was already determined in the detection phase. The technique was perfect for deploying in various lighting and occlusion scenes since it adopted online learning. Refining the algorithm to make optimized hardware or embedded device and to identify other dangerous obstacles effectively in road scenes, this lightweight architecture-based proposed work could be a more acceptable approach for faster tracking and visualizing in HUD. To represent driving situation awareness data, Park et al. introduced a vehicle augmented-reality system that deducts drivers’ distractions with an AR-based windshield of the Genesis DH model from Hyundai motors [186]. The system presented driving conditions and warned a driver using a head-up monitor via the augmented reality. The system included a range of sub-modules, including vehicle and pedestrian recognition based on the deep learning model of [179], vehicle state data, driving data, time to collision (TTC), hazard evaluation, alert policy and display modules. During most experiments, on the basis of TTC values and driver priority, the threat levels and application of augmented EHMI was already determined in the detection phase. In this section, a combination of deep learning algorithms and their outcomes were visualized as the final task of AVS showing them in an AR-based HUD for better driving assistance. AR-HUD was adopted due to visualization in front display for early warning, navigation, object marking by overlapping, ensuring safety and better tracking. Although these studies had successful demonstrations, some major limitations were detected when analyzing the studies, such as visual delay for the case of sudden turn or rapid-moving objects, misreading of local correspondence, high computational cost while 3D shaping, visualizing challenges in extreme contrast and distraction for complex UI. Table 13 provides a summary of the section. Appl. Sci. 2022, 12, 6831 39 of 51 Table 13. Summary of multiple deep learning methods for AR-HUD. Ref. Purpose Methods Advantages Limitations Improved collision alert system Visualization delay while [176] Early warning SVM detecting cars and pedestrians observing rapid moving and fused into the HUD. sudden turning of vehicle. Helped to turn faster and more The insufficient depth Custom deep learning [177] Navigation confidently locating turn points perception of the defined 2D based scene analysis. quicker. HUD distance was apparent. Detected road objects are Automatic matching ability is SIFT and homography [178] Object marking converted into AR contents after inconvenient due to misreading measurement method. projective transformation. of local correspondence. Applicable in sunny and rainy Poor detection accuracy and not SVM, HoG and BSS [179] Safety scenarios for overlapping of applicable for nighttime algorithm. detected objects and obstacles. scenarios. 3D shaping cost-efficient model GPLVM with SegNet Computational complexity is [180] Assistance and needs only single frame in and VPNet. higher for algorithm fusion. input layer. Overlay traffic signs with more Haar cascade detector Lack of implementation in [183] Safety recognizable icons in AR-HUD with BoVW method. complex scenarios. to improve safety. Assist overtaking an on-road The detection methods for 2D–3D tracking [182] Tracking slow vehicle via marker-less deployment in the real-life field strategy with ROI. real-time driving system. had yet to become apparent. Providing more essential Was not deployed in complex [183] Safety CNN and RANSAC. surrounding traffic data to traffic scenarios or nighttime increase interactions and focus. environments. Custom deep learning Boost situation awareness Failed to achieve good [184] Safety applied object aiming at both clear and visualization in lower-visibility detection. lower-visibility conditions conditions. Applicable in various lighting Required a very wide range of [185] Tracking TLD using SVM. and occlusion scenes, since it views. adopted online learning. Enhanced a driver ’s intuitive Not computation cost efficient [186] Safety and awareness SVM and HoG. reasoning and minimized driver and complex UI. distraction calculating TTC. 4. Evaluation Methods In this section, commonly used evaluation metrics throughout the systematic review are presented in Table 14. Several evaluation techniques with equations and description are shown which will give a better understanding, as evaluation techniques are different from the reviewed methodology. Appl. Sci. 2022, 12, 6831 40 of 51 Table 14. Evaluation techniques. Ref. Technique Equations Remarks [40,44,50,53,72,76,82,90, TP is the true positive and FN is T P Sensitivity/TPR/Recall (R) R = T P+F N 110,116,119,120,137] the false negative detection. [40,44,67,72,90,110,116, T P Precision (P) FP is the false positive detection. P = T P+FP 119,120] [54,80,88,90,110,119,120, RP F1 Score - F1 = 2 R+P 176] [37–39,41,43,46– 49,52,57,60,61,64,79,91– X is the number of successes X Pred Pred 94,96,98–101,106,108,112– Accuracy Accuracy = GT and X is the ground truth. GT 114,118,120,122,137,142, 144,152,179] Region o f intersection I oU = Region o f union [95,103–105,111,112] mIoU - 1 T P m I oU = k+1 T P+FP+F N F N [34,63,77] FNR - F N R = T P+F N m AP = k denotes each episode and n is k=n [35,42,45,56,73,75,92] mAP the total episodes. ((R(k) R(k + 1)) P) k=1 Log Average Miss n a is the series positive values [65,66,68,76,78] 1 Log = ex p ln a A MR i Rate (Log ) n correlated with the missing rate. A MR i=1 TPR is the true positive rate, and [51,55,81,116] Area Under Curve (AUC) AUC = T PR d(FPR) FPR is the false positive rate. Dy is center of gravity, # is yaw [83,89,131] Lateral Error Dy = Dy + L # angle towards the road and L is ( ) the distance. t is dimension of state, d is n n min t , t > ( ) 1 2 > measured distance, d is safe sn i f l = 1 and d , d > 3 1 2 distance, l is overtaking lane and r = min(t , t ) 3 4 [121,138,139,142] Reward (r) v is the longitudinal velocity. i f l = 2 and d , d > 3 1 2 Here, n = 1,2 refers host and lead 5 else vehicle in driving lane and n = 3,4 d d n sn where, t = n refers to overtaking lane. N is the total collision number N col col [123,126,128,130,136] Collision Rate (C ) C = rate rate l a p while completing total C laps. l a p Success Counts [102,124,127,134,165] Success Rate (SR) - SR =  100% Total Test Number N is total episode without wc wc [129,135] Safety Index (SI) s = collision and N is the e p e p total episodes. x and x ˆ are the real and i i [133,147,148,151,158] MAD/MAE predicted value, respectively, and M AE/ M AD = å jx x ˆ j i i i=1 n is the total episodes. T N [137] Specificity (SPC) T N is the true negative value. SPC = FP+T N N is total interventions and T is [150,159] Autonomy value (Av) Av = 1  6  100% e elapsed time. n 2 x is the ground truth and x ˆ is the i i (x x ˆ ) å i i [153–156,166–168] RMSE i=1 R MSE = n predicted truth. 0 0 y is the sample proportion, y is 0 0 y y [117] p-Value p the assumed proportion, n is Z = 0 0 y (1y ) sample size. c Appl. Sci. 2022, 12, 6831 41 of 51 5. Case Study 5.1. AVS in Roundabout Cases A roundabout is one of the most difficult scenarios for driving with AVS due to tight time constraints on vehicle behavior because of the yield and merging of maneuvers with high-quality vision requirements for estimating the state of other vehicles, and multi-factor decision-making based on these state estimations. It is particularly tough to forecast the actions of other cars in a roundabout due to the availability of numerous exit locations as discrete options mixed with continuous vehicle dynamics. The entry risk at roundabouts grows with decreasing distance as the ego vehicle must account for cars in the circle passing. Okumura et al. proposed a neural network to map observations to actions in a roundabout that are handled as a combination of turns in order to emphasize the deep learning-based AVS for roundabout cases [197]. This method concentrated on route planning and speed estimation for the roundabout, as well as detection, tracking and generating predictions about the environment using sensor fusion, but ignored interactions between cars. This method concentrated on route planning and speed estimation for the roundabout, as well as detection, tracking and generating predictions about the surroundings using sensor fusion, but ignored interactions between cars [198]. This could be improved by a strategy for forecasting whether a vehicle will exit the roundabout based on its anticipated yaw rate. In a roundabout scenario, the projected yaw rate is a significant indication of whether a car will turn next to the ego vehicle. Although the system was proved to be capable of human-like judgments for a certain roundabout situation, only the center of mass and velocity were calculated to quantify detection of turning cars. This method may be a viable solution for the roundabout research of [197]; however, it may result in errors in roundabouts with no traffic lights or heavy traffic. One of the main reasons of vision-based AVS is to reduce the dependency in terms of safety and collision-free driving; therefore, combined multi-thread architecture of al- gorithms such as Spatial CNN (SCNN) and Deep Recurrent Q-Network (DRQN) could be a major solution for roundabout cases. The spatial features of SCNN for traffic scene understanding in dense traffic conditions as well as the ability of extreme efficient traffic scene-analysis demonstration incorporating multi-threading with self-decision making improved DRL approaches such as DRQN or DDQL could be a vast improvement in the research of AVS in roundabout cases. 5.2. AVS in Uncertain Environmental Cases Even at the current development level, it is challenging for AVS to operate au- tonomously in unknown or uncertain environments. The uncertainty may be because of variable traffic conditions, unknown terrain, unmarked or untrained settings, or even a situation including an extended obstruction. In an unexpected driving environments, even the performance of Waymo, the self-driving vehicle of Google, is at a conditional level 4 of autonomy based on NHTSA autonomous functions, and Tesla’s self-driving vehicles are only at level 2 of autonomy. In contrast, the authors of one study addressed safety issues posed by ambiguity in DL approaches: insufficient training data, locational shift, inconsistencies between training and operating parameters, and uncertainty in pre- diction [199]. The most controversial incident occurred when a Tesla Model S was involved in a deadly collision in which the driver was killed when its autopilot system failed to notice a tractor-trailer 18-wheeler that turned in front of the vehicle [200]. To reduce un- intended occurrence in unknown or uncertain situations and environments, it might be possible to develop level 4 or 5 AVS with safe perception analysis, path planning, decision making and controlling by removing dependence on labelled data and adopting deep reinforcement learning-based approaches. Moreover, several techniques, such as those in [83,128,130,144,172], which were effective in avoiding collision, lane shifting, detection, and safe decision making in unknown or dynamic situations, can be a means of reducing the constraints in uncertain environments. Appl. Sci. 2022, 12, 6831 42 of 51 6. Discussion Deep learning is fast becoming a successful alternative approach for perception-based AVS as it reduces both cost and dependency on sensor fusion. With this aim in mind, total categories of primary domains of AVS were reviewed in this paper to identify efficient methods and algorithms, their contributions and limitations. From the study, it was found that recent deep learning algorithms obtained high accuracy while detecting and identifying road vehicle types, and in some cases, the results surpassed LiDAR’s outcome in both short and long range for 3D bounding vehicles [34]. Moreover, some recent methods such as YOLO V2 [35], deep CNN [38], SINET [41] and Faster R-CNN [42] achieved high accuracy within a very short period of time from low- quality training images to challenging nighttime scenarios. However, there were several limitations, for example, in certain lighting conditions and higher execution costs. Fol- lowing that, a massive contribution to lane and curve detection along with tracking was presented by studies where 95.5% road scene extraction was demonstrated, for example, in [79], for lane edge segmentation without manual labelling using a modified CNN ar- chitecture. As discussed in previous sections, challenges such as higher computational cost [81], insufficient for far field of view [82], not testing in complex scenarios [79] and poor luminance made some proposals tough for practical implementation in present AVS. In addition, a good amount of attention was given to developing safe AVS systems for pedestrian detection. Multiple deep learning approaches such as DNN, CNN, YOLO V3- Tiny, DeepSort R-CNN, single-shot late-fusion CNN, Faster R-CNN, R-CNN combined ACF model, dark channel prior-based SVM, attention-guided encoder–decoder CNN outper- formed the baseline of applied datasets that presented a faster warning area by bounding each pedestrian in real time [61], detection in crowded environments, and dim lighting or haze scenarios [62,72] for position estimation [72], minimizing computational cost and outperforming state-of-the-art methods [120]. The approaches offer an ideal pedestrian method once their technical challenges have been overcome, for example, dependency on preliminary boxing during detection, presumption of constant depths in input image and improvement to avoid missing rate when dealing with a complex environment. Moreover, to estimate steering angles, velocity alongside controlling for lane keeping or changing, overcome slow drifting, take action on a human’s weak zone such as a blind spot and decreasing manual labelling for data training, multiple methods, such as multimodal multitask-based CNN [148], CNN with LSTM [149] and ST-LSTM [153], were studied in this literature review for AVS’s end-to-end control system. Furthermore, one of the most predominant segments of AVS, traffic scene analysis, was covered to understand scenes from a challenging and crowded movable environment [102], improve performance by making more expensive spatial-feature risk prediction [112] and on-road damage detection [120]. For this purpose, HRNet + contrastive loss [104], Multi- Stage Deep CNN [106], 2D-LSTM with RNN [108], DNN with Hadamard layer [110], Spatial CNN [112], OP-DNN [113] and the methods mentioned in Table 9 were reviewed. However, there are still some limitations, for instance, data dependency or relying on pre-labelled data, decreased accuracy in challenging traffic or at nighttime. Taking into account all taxonomies as features, the decision-making process for AVS was broadly analyzed where driving decisions such as overtaking, emergency braking, lane shifting with collision and driving safety in intersections adopting methods such as deep recurrent reinforcement learning [127], actor-critic-based DRL with DDPG [123], double DQN, TD3, SAC [124], dueling DQN [126], gradient boosting decision tree [133], deep RL using Q-masking and autonomically generated curriculum-based DRL [139]. De- spite solving most of the tasks for safe deployment in level 4 or 5 AVS, challenges remain, such as complex training cost, lack of proper surrounding vehicles’ behavior analysis and unfinished case in complex scenarios. Some problems remain to be resolved for better outcomes, such as the requirement of a larger labelled dataset [57], struggle to classify in blurry visual conditions [49] and small traffic signs from a far field of view [51], background complexity [48] and detecting two traffic signs rather than one, which occurred for different Appl. Sci. 2022, 12, 6831 43 of 51 locations of the proposed region [47]. Apart from these, one of the most complicated tasks for AVS, only vision-based path and motion planning were analyzed by reviewing approaches such as deep inverse reinforcement learning, DQN time-to-go method, MPC, Dijkstra with TEB method, DNN, discrete optimizer-based approach, artificial potential field, MPC with LSTM-RNN, advance dynamic window using, 3D-CNN, spatio-temporal LSTM and fuzzy logic, where solutions were provided by avoiding cost function and manual labelling, reducing the limitation of rule-based methods for safe navigation [164] and better path planning for intersections [165], motion planning by analyzing risks and predicting motions of surrounding vehicles [166], hazard detection-based safe naviga- tion [168], avoiding obstacles for smooth planning in multilane scenarios [169], decreasing computational cost [170] and path planning by replicating human-like control thinking in ambiguous circumstances. Nevertheless, these approaches faced challenges such as lack of live testing, low accuracy in far predicted horizon, impaired performance in complex situations or being limited to non-rule-based approaches and constrained kinematics or even difficulty in establishing a rule base to tackle unstructured conditions. Finally, to visualize overlaying outcomes generated from the previous methods su- perimposed on the front head-up display or smart windshield, augmented reality-based approaches combining deep learning methods were reviewed in the last section. AR-HUD based solutions such as 3D surface reconstruction, object marking, path overlaying, re- ducing drivers’ attention, boosting visualization in tough hazy or low-light conditions by overlapping lanes, traffic signs as well as on-road objects to reduce accidents using deep CNN, RANSAC, TTC methods and so on. However, there are still many challenges for practical execution, such as human adoption of AR-based HUD UI, limited visualization in bright daytime conditions, overlapping non-superior objects as well as visualization delay for fast moving on-road objects. In summary, the literature review established for vision-based deep learning approaches of 10 taxonomies for AVS with discussion of out- comes, challenges and limitations could be a pathway to improve and rapidly develop cost-efficient level 4 or 5 AVS without depending on expensive and complex sensor fusion. 7. Conclusions The results of the mixed method studies in the field of implementation and application of deep learning algorithms for autonomous driving systems help us to achieve a clear understanding of the future of transportation. These results prove that it has the ability to provide intelligent mobility for our constantly evolving modern world as deep learning was one of the key components to resolve the limitations and bottlenecking of traditional techniques. Despite containing a good number of studies on autonomous driving systems, only a few make an impact on recent developments in the autonomous driving industry. To overcome this challenge and build a safer and more secure sensor-independent transporta- tion system with the aim of building infrastructure of futuristic smart cities, in this paper, through a systematic review of the literature, studies of AV were selected that used deep learning and the field was reviewed in terms of decision making, path planning and navi- gation, controlling, prediction and visualizing the outcomes in augmented reality-based head-up displays. We analyzed the existing proposal of deep learning models in real-world implementation for AVS, described the methodologies, designed the flow of solutions for the limitations of other methodologies, and compared outcomes and evaluation techniques. Nevertheless, as the research field of autonomous driving systems is still growing, many of the theoretical methodologies were not applied practically, but along with the research trend of this expanding field, these are potentially excellent solutions that require further development. Thus, the large-scale distributions of the paper in the major areas of au- tonomous driving systems will be essential for further research and development of the autonomous vehicle industry into a cost-efficient, secure intelligent transport system. Funding: This research was funded by Universiti Kebangsaan Malaysia [GUP-2020-060]. Institutional Review Board Statement: Not applicable. Appl. Sci. 2022, 12, 6831 44 of 51 Informed Consent Statement: Not applicable. Data Availability Statement: Not applicable. Conflicts of Interest: The authors declare no conflict of interest. References 1. Alawadhi, M.; Almazrouie, J.; Kamil, M.; Khalil, K.A. A systematic literature review of the factors influencing the adoption of autonomous driving. Int. J. Syst. Assur. Eng. Manag. 2020, 11, 1065–1082. [CrossRef] 2. Pandey, P.; Shukla, A.; Tiwari, R. Three-dimensional path planning for unmanned aerial vehicles using glowworm swarm optimization algorithm. Int. J. Syst. Assur. Eng. Manag. 2018, 9, 836–852. [CrossRef] 3. Dirsehan, T.; Can, C. Examination of trust and sustainability concerns in autonomous vehicle adoption. Technol. Soc. 2020, 63, 101361. [CrossRef] 4. Khamis, N.K.; Deros, B.M.; Nuawi, M.Z.; Omar, R.B. Driving fatigue among long distance heavy vehicle drivers in Klang Valley, Malaysia. Appl. Mech. Mater. 2014, 663, 567–573. [CrossRef] 5. Naujoks, F.; Wiedemann, K.; Schömig, N.; Hergeth, S.; Keinath, A. Towards guidelines and verification methods for automated vehicle HMIs. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 121–136. [CrossRef] 6. Li, D.; Wagner, P. Impacts of gradual automated vehicle penetration on motorway operation: A comprehensive evaluation. Eur. Transp. Res. Rev. 2019, 11, 36. [CrossRef] 7. Mutz, F.; Veronese, L.P.; Oliveira-Santos, T.; De Aguiar, E.; Cheein, F.A.A.; De Souza, A.F. Large-scale mapping in complex field scenarios using an autonomous car. Expert Syst. Appl. 2016, 46, 439–462. [CrossRef] 8. Gandia, R.M.; Antonialli, F.; Cavazza, B.H.; Neto, A.M.; Lima, D.A.d.; Sugano, J.Y.; Nicolai, I.; Zambalde, A.L. Autonomous vehicles: Scientometric and bibliometric review. Transp. Rev. 2019, 39, 9–28. [CrossRef] 9. Maurer, M.; Christian, G.; Lenz, B.; Winner, H. Autonomous Driving: Technical, Legal and Social Aspects; Springer Nature: Berlin/Heidelberg, Germany, 2016. [CrossRef] 10. Levinson, J.; Askeland, J.; Becker, J.; Dolson, J.; Held, D.; Kammel, S.; Kolter, J.Z.; Langer, D.; Pink, O.; Pratt, V. Towards fully autonomous driving: Systems and algorithms. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 163–168. 11. Yu, H.; Yang, S.; Gu, W.; Zhang, S. Baidu driving dataset and end-to-end reactive control model. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 341–346. 12. Hashim, H.; Omar, M. Towards autonomous vehicle implementation: Issues and opportunities. J. Soc. Automot. Eng. Malays. 2017, 1, 111–123. 13. Chen, L.; Li, Q.; Li, M.; Zhang, L.; Mao, Q. Design of a multi-sensor cooperation travel environment perception system for autonomous vehicle. Sensors 2012, 12, 12386–12404. [CrossRef] 14. Li, Q.; Chen, L.; Li, M.; Shaw, S.-L.; Nüchter, A. A sensor-fusion drivable-region and lane-detection system for autonomous vehicle navigation in challenging road scenarios. IEEE Trans. Veh. Technol. 2013, 63, 540–555. [CrossRef] 15. Rahman, A.H.A.; Ariffin, K.A.Z.; Sani, N.S.; Zamzuri, H. Pedestrian Detection using Triple Laser Range Finders. Int. J. Electr. Comput. Eng. 2017, 7, 3037. [CrossRef] 16. Wang, H.; Wang, B.; Liu, B.; Meng, X.; Yang, G. Pedestrian recognition and tracking using 3D LiDAR for autonomous vehicle. Robot. Auton. Syst. 2017, 88, 71–78. [CrossRef] 17. Wang, L.; Zhang, Y.; Wang, J. Map-based localization method for autonomous vehicles using 3D-LIDAR. IFAC-Pap. 2017, 50, 276–281. 18. Kong, P.-Y. Computation and Sensor Offloading for Cloud-Based Infrastructure-Assisted Autonomous Vehicles. IEEE Syst. J. 2020, 14, 3360–3370. [CrossRef] 19. Zhao, J.; Xu, H.; Liu, H.; Wu, J.; Zheng, Y.; Wu, D. Detection and tracking of pedestrians and vehicles using roadside LiDAR sensors. Transp. Res. Part C Emerg. Technol. 2019, 100, 68–87. [CrossRef] 20. Huval, B.; Wang, T.; Tandon, S.; Kiske, J.; Song, W.; Pazhayampallil, J.; Andriluka, M.; Rajpurkar, P.; Migimatsu, T.; Cheng, Y.R. An empirical evaluation of deep learning on highway driving. arXiv 2015, arXiv:01716. 21. Leon, F.; Gavrilescu, M. A review of tracking, prediction and decision making methods for autonomous driving. arXiv 2019, arXiv:07707. 22. Ma, Y.; Wang, Z.; Yang, H.; Yang, L. Artificial intelligence applications in the development of autonomous vehicles: A survey. IEEE/CAA J. Autom. Sin. 2020, 7, 315–329. [CrossRef] 23. Paden, B.; Cáp, M.; Yong, S.Z.; Yershov, D.; Frazzoli, E. A survey of motion planning and control techniques for self-driving urban vehicles. IEEE Trans. Intell. Veh. 2016, 1, 33–55. [CrossRef] 24. Yurtsever, E.; Lambert, J.; Carballo, A.; Takeda, K. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access 2020, 8, 58443–58469. [CrossRef] 25. Schwarting, W.; Alonso-Mora, J.; Rus, D. Planning and decision-making for autonomous vehicles. Annu. Rev. Control Robot. Auton. Syst. 2018, 1, 187–210. [CrossRef] 26. Van Brummelen, J.; O’Brien, M.; Gruyer, D.; Najjaran, H. Autonomous vehicle perception: The technology of today and tomorrow. Transp. Res. Part C Emerg. Technol. 2018, 89, 384–406. [CrossRef] Appl. Sci. 2022, 12, 6831 45 of 51 27. Badue, C.; Guidolini, R.; Carneiro, R.V.; Azevedo, P.; Cardoso, V.B.; Forechi, A.; Jesus, L.; Berriel, R.; Paixao, T.M.; Mutz, F.J. Self-driving cars: A survey. Expert Syst. Appl. 2021, 165, 113816. [CrossRef] 28. Sirohi, D.; Kumar, N.; Rana, P.S. Convolutional neural networks for 5G-enabled intelligent transportation system: A systematic review. Comput. Commun. 2020, 153, 459–498. [CrossRef] 29. Zhang, Y.; Qi, Y.; Liu, J.; Wang, Y. Decade of Vision-Based Pedestrian Detection for Self-Driving: An Experimental Survey and Evaluation; 0148-7191; SAE Publishing: Thousand Oaks, CA, USA, 2018. 30. Ni, J.; Chen, Y.; Chen, Y.; Zhu, J.; Ali, D.; Cao, W. A survey on theories and applications for self-driving cars based on deep learning methods. Appl. Sci. 2020, 10, 2749. [CrossRef] 31. Fayyad, J.; Jaradat, M.A.; Gruyer, D.; Najjaran, H. Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review. Sensors 2020, 20, 4220. [CrossRef] 32. Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Al Sallab, A.A.; Yogamani, S.; Pérez, P. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4909–4926. [CrossRef] 33. Grigorescu, S.; Trasnea, B.; Cocias, T.; Macesanu, G. A survey of deep learning techniques for autonomous driving. J. Field Robot. 2020, 37, 362–386. [CrossRef] 34. Hu, H.-N.; Cai, Q.-Z.; Wang, D.; Lin, J.; Sun, M.; Krahenbuhl, P.; Darrell, T.; Yu, F. Joint monocular 3D vehicle detection and tracking. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 5390–5399. 35. Sang, J.; Wu, Z.; Guo, P.; Hu, H.; Xiang, H.; Zhang, Q.; Cai, B. An improved YOLOv2 for vehicle detection. Sensors 2018, 18, 4272. [CrossRef] 36. Shakya, V.; Makwana, R.R.S. Feature selection based intrusion detection system using the combination of DBSCAN, K-Mean++ and SMO algorithms. In Proceedings of the 2017 International Conference on Trends in Electronics and Informatics (ICEI), Tirunelveli, India, 11–12 May 2017; pp. 928–932. 37. Ohn-Bar, E.; Trivedi, M.M. Learning to detect vehicles by clustering appearance patterns. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2511–2521. [CrossRef] 38. Chen, W.; Sun, Q.; Wang, J.; Dong, J.-J.; Xu, C. A novel model based on AdaBoost and deep CNN for vehicle classification. IEEE Access 2018, 6, 60445–60455. [CrossRef] 39. Bautista, C.M.; Dy, C.A.; Mañalac, M.I.; Orbe, R.A.; Cordel, M. Convolutional neural network for vehicle detection in low resolution traffic videos. In Proceedings of the 2016 IEEE Region 10 Symposium (TENSYMP), Bali, Indonesia, 9–11 May 2016; pp. 277–281. 40. Lee, H.J.; Moon, B.; Kim, G. Hierarchical scheme of vehicle detection and tracking in nighttime urban environment. Int. J. Automot. Technol. 2018, 19, 369–377. [CrossRef] 41. Hu, X.; Xu, X.; Xiao, Y.; Chen, H.; He, S.; Qin, J.; Heng, P.-A. SINet: A scale-insensitive convolutional neural network for fast vehicle detection. IEEE Trans. Intell. Transp. Syst. 2018, 20, 1010–1019. [CrossRef] 42. Wang, Y.; Liu, Z.; Deng, W. Anchor generation optimization and region of interest assignment for vehicle detection. Sensors 2019, 19, 1089. [CrossRef] 43. Suhao, L.; Jinzhao, L.; Guoquan, L.; Tong, B.; Huiqian, W.; Yu, P. Vehicle type detection based on deep learning in traffic scene. Procedia Comput. Sci. 2018, 131, 564–572. [CrossRef] 44. Liu, J.; Zhang, R. Vehicle Detection and Ranging Using Two Different Focal Length Cameras. J. Sens. 2020, 2020, 4372847. 45. Leung, H.K.; Chen, X.-Z.; Yu, C.-W.; Liang, H.-Y.; Wu, J.-Y.; Chen, Y.-L. A Deep-Learning-Based Vehicle Detection Approach for Insufficient and Nighttime Illumination Conditions. Appl. Sci. 2019, 9, 4769. [CrossRef] 46. Hu, W.; Zhuo, Q.; Zhang, C.; Li, J. Fast branch convolutional neural network for traffic sign recognition. IEEE Intell. Transp. Syst. Mag. 2017, 9, 114–126. [CrossRef] 47. Jung, S.; Lee, U.; Jung, J.; Shim, D.H. Real-time Traffic Sign Recognition system with deep convolutional neural network. In Proceedings of the 2016 13th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Xi’an, China, 19–22 August 2016; pp. 31–34. 48. Cao, J.; Song, C.; Peng, S.; Xiao, F.; Song, S. Improved traffic sign detection and recognition algorithm for intelligent vehicles. Sensors 2019, 19, 4021. [CrossRef] 49. Natarajan, S.; Annamraju, A.K.; Baradkar, C.S. Traffic sign recognition using weighted multi-convolutional neural network. IET Intell. Transp. Syst. 2018, 12, 1396–1405. [CrossRef] 50. Wang, G.; Ren, G.; Jiang, L.; Quan, T. Hole-based traffic sign detection method for traffic signs with red rim. Vis. Comput. 2014, 30, 539–551. [CrossRef] 51. Zhang, J.; Huang, M.; Jin, X.; Li, X. A real-time chinese traffic sign detection algorithm based on modified YOLOv2. Algorithms 2017, 10, 127. [CrossRef] 52. Liu, X.; Yan, W.Q. Traffic-light sign recognition using Capsule network. Multimed. Tools Appl. 2021, 80, 15161–15171. [CrossRef] 53. Bach, M.; Stumper, D.; Dietmayer, K. Deep convolutional traffic light recognition for automated driving. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 851–858. 54. Weber, M.; Wolf, P.; Zöllner, J.M. DeepTLR: A single deep convolutional network for detection and classification of traffic lights. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 19–22 June 2016; pp. 342–348. Appl. Sci. 2022, 12, 6831 46 of 51 55. Liu, X.; Neuyen, M.; Yan, W.Q. Vehicle-related scene understanding using deep learning. In Asian Conference on Pattern Recognition; Springer: Sinapore, 2020; pp. 61–73. 56. Lee, E.; Kim, D. Accurate traffic light detection using deep neural network with focal regression loss. Image Vis. Comput. 2019, 87, 24–36. [CrossRef] 57. Behrendt, K.; Novak, L.; Botros, R. A deep learning approach to traffic lights: Detection, tracking, and classification. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 1370–1377. 58. Yanagisawa, M.; Swanson, E.; Najm, W.G. Target Crashes and Safety Benefits Estimation Methodology for Pedestrian Crash Avoid- ance/Mitigation Systems; United States National Highway Traffic Safety Administration: Sacramento, CA, USA, 2014. 59. Baharuddin, M.; Khamis, N.; Kassim, K.A.; Mansor, M.R.A. Autonomous Emergency Brake (AEB) for Pedestrian for ASEAN NCAP Safety Rating Consideration: A Review. J. Soc. Automot. Eng. Malays. 2019, 3, 63–73. 60. Angelova, A.; Krizhevsky, A.; Vanhoucke, V. Pedestrian detection with a large-field-of-view deep network. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 704–711. 61. Zhan, H.; Liu, Y.; Cui, Z.; Cheng, H. Pedestrian Detection and Behavior Recognition Based on Vision. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 771–776. 62. Qiu, Z.; Zhao, N.; Zhou, L.; Wang, M.; Yang, L.; Fang, H.; He, Y.; Liu, Y. Vision-based moving obstacle detection and tracking in paddy field using improved yolov3 and deep SORT. Sensors 2020, 20, 4082. [CrossRef] 63. Ghosh, S.; Amon, P.; Hutter, A.; Kaup, A. Reliable pedestrian detection using a deep neural network trained on pedestrian counts. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 685–689. 64. Wang, X.; Jung, C.; Hero, A.O. Part-level fully convolutional networks for pedestrian detection. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2267–2271. 65. Hou, Y.-L.; Song, Y.; Hao, X.; Shen, Y.; Qian, M.; Chen, H. Multispectral pedestrian detection based on deep convolutional neural networks. Infrared Phys. Technol. 2018, 94, 69–77. [CrossRef] 66. Yamada, K. Pedestrian detection with a resolution-aware convolutional network. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 591–596. 67. Zhang, J.; Xiao, J.; Zhou, C.; Peng, C. A multi-class pedestrian detection network for distorted pedestrians. In Proceedings of the 2018 13th IEEE Conference on Industrial Electronics and Applications (ICIEA), Wuhan, China, 31 May–June 2018; pp. 1079–1083. 68. Dong, P.; Wang, W. Better region proposals for pedestrian detection with R-CNN. In Proceedings of the 2016 Visual Communica- tions and Image Processing (VCIP), Chengdu, China, 27–30 November 2016; pp. 1–4. 69. Wood, J.M. Nighttime driving: Visual, lighting and visibility challenges. Ophthalmic Physiol. Opt. 2020, 40, 187–201. [CrossRef] 70. Baratian-Ghorghi, F.; Zhou, H.; Shaw, J. Overview of wrong-way driving fatal crashes in the United States. ITE J. 2014, 84, 41–47. 71. Mohammed, A.A.; Ambak, K.; Mosa, A.M.; Syamsunur, D. Traffic accidents in Iraq: An analytical study. J. Adv. Res. Civ. Environ. Eng. 2018, 5, 10–22. 72. Ding, B.; Liu, Z.; Sun, Y. Pedestrian Detection in Haze Environments Using Dark Channel Prior and Histogram of Oriented Gradi- ent. In Proceedings of the 2018 Eighth International Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC), Harbin, China, 19–21 July 2018; pp. 1003–1008. 73. Huang, S.-C.; Ye, J.-H.; Chen, B.-H. An advanced single-image visibility restoration algorithm for real-world hazy scenes. IEEE Trans. Ind. Electron. 2014, 62, 2962–2972. [CrossRef] 74. Li, G.; Yang, Y.; Qu, X. Deep learning approaches on pedestrian detection in hazy weather. IEEE Trans. Ind. Electron. 2019, 67, 8889–8899. [CrossRef] 75. Ju, M.; Luo, H.; Wang, Z.; Hui, B.; Chang, Z. The application of improved YOLO V3 in multi-scale target detection. Appl. Sci. 2019, 9, 3775. [CrossRef] 76. Xu, Z.; Vong, C.-M.; Wong, C.-C.; Liu, Q. Ground Plane Context Aggregation Network for Day-and-Night on Vehicular Pedestrian Detection. IEEE Trans. Intell. Transp. Syst. 2020, 22, 6395–6406. [CrossRef] 77. Zhao, Y.; Yuan, Z.; Chen, B. Accurate pedestrian detection by human pose regression. IEEE Trans. Image Process. 2019, 29, 1591–1605. [CrossRef] 78. Chen, Z.; Huang, X. Pedestrian detection for autonomous vehicle using multi-spectral cameras. IEEE Trans. Intell. Veh. 2019, 4, 211–219. [CrossRef] 79. Alvarez, J.M.; Gevers, T.; LeCun, Y.; Lopez, A.M. Road scene segmentation from a single image. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 376–389. 80. Dong, M.; Zhao, X.; Fan, X.; Shen, C.; Liu, Z. Combination of modified U-Net and domain adaptation for road detection. IET Image Process. 2019, 13, 2735–2743. [CrossRef] 81. Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep Neural Network for Structural Prediction and Lane Detection in Traffic Scene. IEEE Trans. Neural Netw. Learn Syst. 2017, 28, 690–703. [CrossRef] 82. Fakhfakh, M.; Chaari, L.; Fakhfakh, N. Bayesian curved lane estimation for autonomous driving. J. Ambient Intell. Humaniz. Comput. 2020, 11, 4133–4143. [CrossRef] 83. Yang, S.; Wu, J.; Shan, Y.; Yu, Y.; Zhang, S. A Novel Vision-Based Framework for Real-Time Lane Detection and Tracking; SAE Technical Paper: Warrendale, PA, USA, 2019. Appl. Sci. 2022, 12, 6831 47 of 51 84. Mongkonyong, P.; Nuthong, C.; Siddhichai, S.; Yamakita, M. Lane detection using randomized Hough transform. In Proceedings of the 8th TSME-International Conference on Mechanical Engineering, Bangkok, Thailand, 12–15 December 2017. 85. Javadi, M.; Hannan, M.; Samad, S.; Hussain, A. A robust vision-based lane boundaries detection approach for intelligent vehicles. Inf. Technol. J. 2012, 11, 1184–1192. [CrossRef] 86. El Hajjouji, I.; Mars, S.; Asrih, Z.; El Mourabit, A. A novel fpga implementation of hough transform for straight lane detection. Eng. Sci. Technol. Int. J. 2020, 23, 274–280. [CrossRef] 87. Dong, Y.; Patil, S.; van Arem, B.; Farah, H. A Hybrid Spatial-temporal Deep Learning Architecture for Lane Detection. arXiv 2021, arXiv:04079. [CrossRef] 88. Liu, L.; Chen, X.; Zhu, S.; Tan, P. CondLaneNet: A Top-to-down Lane Detection Framework Based on Conditional Convolution. arXiv 2021, arXiv:05003. 89. Dorj, B.; Hossain, S.; Lee, D.-J. Highly Curved Lane Detection Algorithms Based on Kalman Filter. Appl. Sci. 2020, 10, 2372. [CrossRef] 90. Son, Y.; Lee, E.S.; Kum, D. Robust multi-lane detection and tracking using adaptive threshold and lane classification. Mach. Vis. Appl. 2019, 30, 111–124. [CrossRef] 91. Wang, W.; Lin, H.; Wang, J. CNN based lane detection with instance segmentation in edge-cloud computing. J. Cloud Comput. 2020, 9, 27. [CrossRef] 92. Ye, Y.Y.; Hao, X.L.; Chen, H.J. Lane detection method based on lane structural analysis and CNNs. IET Intell. Transp. Syst. 2018, 12, 513–520. [CrossRef] 93. Zou, Q.; Jiang, H.; Dai, Q.; Yue, Y.; Chen, L.; Wang, Q. Robust lane detection from continuous driving scenes using deep neural networks. IEEE Trans. Veh. Technol. 2019, 69, 41–54. [CrossRef] 94. John, V.; Liu, Z.; Mita, S.; Guo, C.; Kidono, K. Real-time road surface and semantic lane estimation using deep features. Signal Image Video Process. 2018, 12, 1133–1140. [CrossRef] 95. Chen, P.-R.; Lo, S.-Y.; Hang, H.-M.; Chan, S.-W.; Lin, J.-J. Efficient road lane marking detection with deep learning. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. 96. Ghafoorian, M.; Nugteren, C.; Baka, N.; Booij, O.; Hofmann, M. El-gan: Embedding loss driven generative adversarial networks for lane detection. In European Conference on Computer Vision (ECCV); Springer: Cham, Switzerland, 2018. 97. He, X.; Duan, Z.; Chen, C.; You, F. Video-based lane detection and tracking during night. In Proceedings of the 19th COTA International Conference of Transportation Professionals, Nanjing, China, 6–8 July 2019; pp. 5794–5807. 98. Neven, D.; De Brabandere, B.; Georgoulis, S.; Proesmans, M.; Van Gool, L. Towards end-to-end lane detection: An instance segmentation approach. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 286–291. 99. Kim, J.; Kim, J.; Jang, G.-J.; Lee, M. Fast learning method for convolutional neural networks using extreme learning machine and its application to lane detection. Neural Netw. 2017, 87, 109–121. [CrossRef] [PubMed] 100. Van Gansbeke, W.; De Brabandere, B.; Neven, D.; Proesmans, M.; Van Gool, L. End-to-end lane detection through differentiable least-squares fitting. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019. 101. Hou, Y.; Ma, Z.; Liu, C.; Loy, C.C. Learning lightweight lane detection cnns by self attention distillation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1013–1021. 102. Geiger, A.; Lauer, M.; Wojek, C.; Stiller, C.; Urtasun, R. 3D Traffic Scene Understanding From Movable Platforms. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1012–1025. [CrossRef] [PubMed] 103. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. 2020, 43, 3349–3364. [CrossRef] [PubMed] 104. Wang, W.; Zhou, T.; Yu, F.; Dai, J.; Konukoglu, E.; Van Gool, L. Exploring cross-image pixel contrast for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7303–7313. 105. Zhao, X.; Vemulapalli, R.; Mansfield, P.A.; Gong, B.; Green, B.; Shapira, L.; Wu, Y. Contrastive Learning for Label Efficient Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10623–10633. 106. Tang, P.; Wang, H.; Kwong, S. G-MS2F: GoogLeNet based multi-stage feature fusion of deep CNN for scene recognition. Neurocomputing 2017, 225, 188–197. [CrossRef] 107. Wang, L.; Guo, S.; Huang, W.; Xiong, Y.; Qiao, Y. Knowledge guided disambiguation for large-scale scene classification with multi-resolution CNNs. IEEE Trans. Image Process. 2017, 26, 2055–2068. [CrossRef] 108. Byeon, W.; Breuel, T.M.; Raue, F.; Liwicki, M. Scene labeling with lstm recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3547–3555. 109. Fu, J.; Liu, J.; Li, Y.; Bao, Y.; Yan, W.; Fang, Z.; Lu, H. Contextual deconvolution network for semantic segmentation. Pattern Recognit. 2020, 101, 107152. [CrossRef] Appl. Sci. 2022, 12, 6831 48 of 51 110. Oeljeklaus, M.; Hoffmann, F.; Bertram, T. A combined recognition and segmentation model for urban traffic scene understanding. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. 111. Xue, J.-R.; Fang, J.-W.; Zhang, P. A survey of scene understanding by event reasoning in autonomous driving. Int. J. Autom. Comput. 2018, 15, 249–266. [CrossRef] 112. Pan, X.; Shi, J.; Luo, P.; Wang, X.; Tang, X. Spatial as deep: Spatial cnn for traffic scene understanding. arXiv 2017, arXiv:06080. 113. Mou, L.; Xie, H.; Mao, S.; Zhao, P.; Chen, Y. Vision-based vehicle behaviour analysis: A structured learning approach via convolutional neural networks. IET Intell. Transp. Syst. 2020, 14, 792–801. [CrossRef] 114. Jeon, H.-S.; Kum, D.-S.; Jeong, W.-Y. Traffic scene prediction via deep learning: Introduction of multi-channel occupancy grid map as a scene representation. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1496–1501. 115. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [CrossRef] 116. Theofilatos, A.; Chen, C.; Antoniou, C. Comparing Machine Learning and Deep Learning Methods for Real-Time Crash Prediction. Transp. Res. Rec. J. Transp. Res. Board 2019, 2673, 169–178. [CrossRef] 117. Huegle, M.; Kalweit, G.; Werling, M.; Boedecker, J. Dynamic interaction-aware scene understanding for reinforcement learning in autonomous driving. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 4329–4335. 118. Nguyen, N.T.H.; Le, T.H.; Perry, S.; Nguyen, T.T. Pavement crack detection using convolutional neural network. In Proceedings of the Ninth International Symposium on Information and Communication Technology, Sharjah, United Arab Emirates, 18–19 November 2019; pp. 251–256. 119. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road crack detection using deep convolutional neural network. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3708–3712. 120. Anand, S.; Gupta, S.; Darbari, V.; Kohli, S. Crack-pot: Autonomous road crack and pothole detection. In Proceedings of the 2018 Digital Image Computing: Techniques and Applications (DICTA), Canberra, ACT, Australia, 10–13 December 2018; pp. 1–6. 121. Yu, C.; Wang, X.; Xu, X.; Zhang, M.; Ge, H.; Ren, J.; Sun, L.; Chen, B.; Tan, G. Distributed multiagent coordinated learning for autonomous driving in highways based on dynamic coordination graphs. IEEE Trans. Intell. Transp. Syst. 2019, 21, 735–748. [CrossRef] 122. Zhang, J.; Liao, Y.; Wang, S.; Han, J.J.A.S. Study on driving decision-making mechanism of autonomous vehicle based on an optimized support vector machine regression. Apply Sci. 2018, 8, 13. [CrossRef] 123. Fu, Y.; Li, C.; Yu, F.R.; Luan, T.H.; Zhang, Y. A Decision-Making Strategy for Vehicle Autonomous Braking in Emergency via Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2020, 69, 5876–5888. [CrossRef] 124. Munk, J.; Kober, J.; Babuška, R. Learning state representation for deep actor-critic control. In Proceedings of the 2016 IEEE 55th Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 4667–4673. 125. Chen, J.; Yuan, B.; Tomizuka, M. Model-free deep reinforcement learning for urban autonomous driving. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 2765–2771. 126. Liao, J.; Liu, T.; Tang, X.; Mu, X.; Huang, B.; Cao, D. Decision-making Strategy on Highway for Autonomous Vehicles using Deep Reinforcement Learning. IEEE Access 2020, 8, 177804–177814. [CrossRef] 127. Hoel, C.-J.; Driggs-Campbell, K.; Wolff, K.; Laine, L.; Kochenderfer, M.J. Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving. IEEE Trans. Intell. Veh. 2020, 5, 294–305. [CrossRef] 128. Sezer, V. Intelligent decision making for overtaking maneuver using mixed observable Markov decision process. J. Intell. Transp. Syst. 2017, 22, 201–217. [CrossRef] 129. Ngai, D.C.K.; Yung, N.H.C. A multiple-goal reinforcement learning method for complex vehicle overtaking maneuvers. IEEE Trans. Intell. Transp. Syst. 2011, 12, 509–522. [CrossRef] 130. Brannstrom, M.; Sandblom, F.; Hammarstrand, L. A Probabilistic Framework for Decision-Making in Collision Avoidance Systems. IEEE Trans. Intell. Transp. Syst. 2013, 14, 637–648. [CrossRef] 131. Zhu, B.; Liu, S.; Zhao, J. A Lane-Changing Decision-Making Method for Intelligent Vehicle Based on Acceleration Field; SAE Technical Paper Series; SAE Publishing: Thousand Oaks, CA, USA, 2020. 132. Wang, W.; Qie, T.; Yang, C.; Liu, W.; Xiang, C.; Huang, K.J.I.T.o.I.E. An intelligent Lane-Changing Behavior Prediction and Decision-Making strategy for an Autonomous Vehicle. IEEE Trans. Ind. Electron. 2021, 69, 2927–2937. [CrossRef] 133. Gómez-Huélamo, C.; Egido, J.D.; Bergasa, L.M.; Barea, R.; López-Guillén, E.; Arango, F.; Araluce, J.; López, J. Train here, drive there: Simulating real-world use cases with fully-autonomous driving architecture in carla simulator. In Workshop of Physical Agents; Springer: Cham, Switzerland, 2020; pp. 44–59. 134. Mukadam, M.; Cosgun, A.; Nakhaei, A.; Fujimura, K. Tactical decision making for lane changing with deep reinforcement learning. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. 135. Wang, J.; Zhang, Q.; Zhao, D.; Chen, Y. Lane change decision-making through deep reinforcement learning with rule-based constraints. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–6. Appl. Sci. 2022, 12, 6831 49 of 51 136. Chae, H.; Kang, C.M.; Kim, B.; Kim, J.; Chung, C.C.; Choi, J.W. Autonomous braking system via deep reinforcement learning. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. 137. Wang, W.; Xi, J.; Zhao, D. Learning and inferring a driver ’s braking action in car-following scenarios. IEEE Trans. Veh. Technol. 2018, 67, 3887–3899. [CrossRef] 138. Chen, J.; Chen, J.; Zhang, R.; Hu, X. Towards Brain-inspired System: Deep Recurrent Reinforcement Learning for Simulated Self-driving Agent. Front. Neurorobot. 2019, 13, 40. [CrossRef] [PubMed] 139. Qiao, Z.; Muelling, K.; Dolan, J.M.; Palanisamy, P.; Mudalige, P. Automatically generated curriculum based reinforcement learning for autonomous vehicles in urban environment. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1233–1238. 140. Li, G.; Li, S.; Li, S.; Qu, X. Continuous decision-making for autonomous driving at intersections using deep deterministic policy gradient. IET Intell. Transp. Syst. 2021. [CrossRef] 141. Deshpande, N.; Vaufreydaz, D.; Spalanzani, A. Behavioral decision-making for urban autonomous driving in the presence of pedestrians using Deep Recurrent Q-Network. In Proceedings of the 16th International Conference on Control, Automation, Robotics and Vision (ICARCV), Shenzhen, China, 13–15 December 2020. 142. Bin Issa, R.; Das, M.; Rahman, M.S.; Barua, M.; Rhaman, M.K.; Ripon, K.S.N.; Alam, M.G.R. Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment. Sensors 2021, 21, 1468. [CrossRef] 143. Mo, S.; Pei, X.; Wu, C. Safe reinforcement learning for autonomous vehicle using monte carlo tree search. IEEE Trans. Intell. Transp. Syst. 2021. [CrossRef] 144. An, D.; Liu, J.; Zhang, M.; Chen, X.; Chen, M.; Sun, H. Uncertainty Modeling and Runtime Verification for Autonomous Vehicles Driving Control: A Machine Learning-based Approach. J. Syst. Softw. 2020, 167, 110617. [CrossRef] 145. Jovanovic, ´ A.; Kwiatkowska, M. Parameter synthesis for probabilistic timed automata using stochastic game abstractions. Theor. Comput. Sci. 2018, 735, 64–81. [CrossRef] 146. Pan, Y.; Cheng, C.-A.; Saigol, K.; Lee, K.; Yan, X.; Theodorou, E.A.; Boots, B. Imitation learning for agile autonomous driving. Int. J. Robot. Res. 2020, 39, 286–302. [CrossRef] 147. Chen, Z.; Huang, X. End-to-end learning for lane keeping of self-driving cars. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 1856–1860. 148. Yang, Z.; Zhang, Y.; Yu, J.; Cai, J.; Luo, J. End-to-end multi-modal multi-task vehicle control for self-driving cars with visual perceptions. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2289–2294. 149. Lee, M.-J.; Ha, Y.-G. Autonomous Driving Control Using End-to-End Deep Learning. In Proceedings of the 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), Busan, Korea, 19–22 February 2020; pp. 470–473. 150. Bojarski, M.; Del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J. End to end learning for self-driving cars. arXiv 2016, arXiv:07316. 151. Chen, C.; Seff, A.; Kornhauser, A.; Xiao, J. Deepdriving: Learning affordance for direct perception in autonomous driving. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2722–2730. 152. Zhang, Y.; Sun, P.; Yin, Y.; Lin, L.; Wang, X. Human-like autonomous vehicle speed control by deep reinforcement learning with double Q-learning. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1251–1256. 153. Chi, L.; Mu, Y. Deep steering: Learning end-to-end driving model from spatial and temporal visual cues. arXiv 2017, arXiv:03798. 154. Wu, T.; Luo, A.; Huang, R.; Cheng, H.; Zhao, Y. End-to-End Driving Model for Steering Control of Autonomous Vehicles with Future Spatiotemporal Features. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 950–955. 155. Bojarski, M.; Yeres, P.; Choromanska, A.; Choromanski, K.; Firner, B.; Jackel, L.; Muller, U. Explaining how a deep neural network trained with end-to-end learning steers a car. arXiv 2017, arXiv:07911. 156. Wu, A.; Rubaiyat, A.H.M.; Anton, C.; Alemzadeh, H. Model Fusion: Weighted N-version programming for resilient autonomous vehicle steering control. In Proceedings of the 2018 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW), Memphis, TN, USA, 15–18 October 2018; pp. 144–145. 157. Huang, X.; McGill, S.G.; DeCastro, J.A.; Fletcher, L.; Leonard, J.J.; Williams, B.C.; Rosman, G. Diversitygan: Diversity-aware vehicle motion prediction via latent semantic sampling. IEEE Robot. Autom. Lett. 2020, 5, 5089–5096. [CrossRef] 158. Wang, D.; Wen, J.; Wang, Y.; Huang, X.; Pei, F. End-to-End Self-Driving Using Deep Neural Networks with Multi-auxiliary Tasks. Automot. Innov. 2019, 2, 127–136. [CrossRef] 159. Toromanoff, M.; Wirbel, E.; Wilhelm, F.; Vejarano, C.; Perrotton, X.; Moutarde, F. End to end vehicle lateral control using a single fisheye camera. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 3613–3619. 160. Smolyakov, M.; Frolov, A.; Volkov, V.; Stelmashchuk, I. Self-Driving Car Steering Angle Prediction Based On Deep Neural Network An Example Of CarND Udacity Simulator. In Proceedings of the 2018 IEEE 12th International Conference on Application of Information and Communication Technologies (AICT), Almaty, Kazakhstan, 17–19 October 2018; pp. 1–5. Appl. Sci. 2022, 12, 6831 50 of 51 161. Sharma, S.; Tewolde, G.; Kwon, J. Lateral and Longitudinal Motion Control of Autonomous Vehicles using Deep Learning. In Proceedings of the 2019 IEEE International Conference on Electro Information Technology (EIT), Brookings, SD, USA, 20–22 May 2019; pp. 1–5. 162. Zhu, Q.; Huang, Z.; Sun, Z.; Liu, D.; Dai, B. Reinforcement learning based throttle and brake control for autonomous vehicle following. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 6657–6662. 163. Xiao, Y.; Codevilla, F.; Gurram, A.; Urfalioglu, O.; López, A.M. Multimodal end-to-end autonomous driving. IEEE Trans. Intell. Transp. Syst. 2020, 23, 537–547. [CrossRef] 164. You, C.; Lu, J.; Filev, D.; Tsiotras, P. Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning. Robot. Auton. Syst. 2019, 114, 1–18. [CrossRef] 165. Isele, D.; Rahimi, R.; Cosgun, A.; Subramanian, K.; Fujimura, K. Navigating occluded intersections with autonomous vehicles using deep reinforcement learning. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 2034–2039. 166. Zhang, L.; Xiao, W.; Zhang, Z.; Meng, D. Surrounding Vehicles Motion Prediction for Risk Assessment and Motion Planning of Autonomous Vehicle in Highway Scenarios. IEEE Access 2020, 8, 209356–209376. [CrossRef] 167. Marin-Plaza, P.; Hussein, A.; Martin, D.; de la Escalera, A. Global and Local Path Planning Study in a ROS-Based Research Platform for Autonomous Vehicles. J. Adv. Transp. 2018, 2018, 6392697. [CrossRef] 168. Islam, M.; Chowdhury, M.; Li, H.; Hu, H. Vision-Based Navigation of Autonomous Vehicles in Roadway Environments with Unexpected Hazards. Transp. Res. Rec. J. Transp. Res. Board 2019, 2673, 494–507. [CrossRef] 169. Ma, L.; Xue, J.; Kawabata, K.; Zhu, J.; Ma, C.; Zheng, N. Efficient sampling-based motion planning for on-road autonomous driving. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1961–1976. [CrossRef] 170. Gu, T.; Dolan, J.M. On-road motion planning for autonomous vehicles. In Proceedings of the International Conference on Intelligent Robotics and Applications, Montreal, QC, Canada, 3–5 October 2012; pp. 588–597. 171. Wahid, N.; Zamzuri, H.; Amer, N.H.; Dwijotomo, A.; Saruchi, S.A.; Mazlan, S.A. Vehicle collision avoidance motion planning strategy using artificial potential field with adaptive multi-speed scheduler. IET Intell. Transp. Syst. 2020, 14, 1200–1209. [CrossRef] 172. Cai, P.; Sun, Y.; Chen, Y.; Liu, M. Vision-based trajectory planning via imitation learning for autonomous vehicles. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 2736–2742. 173. Kiss, D.; Tevesz, G. Advanced dynamic window based navigation approach using model predictive control. In Proceedings of the 2012 17th International Conference on Methods & Models in Automation & Robotics (MMAR), Miedzyzdroje, Poland, 27–30 August 2012; pp. 148–153. 174. Bai, Z.; Cai, B.; ShangGuan, W.; Chai, L. Deep learning based motion planning for autonomous vehicle using spatiotemporal LSTM network. In Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; pp. 1610–1614. 175. Li, X.; Choi, B.-J. Design of obstacle avoidance system for mobile robot using fuzzy logic systems. Int. J. Smart Home 2013, 7, 321–328. 176. Yoon, C.; Kim, K.; Park, H.S.; Park, M.W.; Jung, S.K. Development of augmented forward collision warning system for Head-Up Display. In Proceedings of the 17th International IEEE Conference on Intelligent Transportation Systems (ITSC), Qingdao, China, 8–11 October 2014; pp. 2277–2279. 177. Bark, K.; Tran, C.; Fujimura, K.; Ng-Thow-Hing, V. Personal Navi: Benefits of an augmented reality navigational aid using a see-thru 3D volumetric HUD. In Proceedings of the 6th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Seattle, WA, USA, 17–19 September 2014; pp. 1–8. 178. Yoon, C.; Kim, K.-H. Augmented reality information registration for head-up display. In Proceedings of the 2015 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Korea, 28–30 October 2015; pp. 1135–1137. 179. Park, H.S.; Park, M.W.; Won, K.H.; Kim, K.H.; Jung, S.K. In-vehicle AR-HUD system to provide driving-safety information. ETRI J. 2013, 35, 1038–1047. [CrossRef] 180. Rao, Q.; Chakraborty, S.J. In-Vehicle Object-Level 3D Reconstruction of Traffic Scenes. IEEE Trans. Intell. Transp. Syst. 2020, 22, 7747–7759. [CrossRef] 181. Abdi, L.; Meddeb, A. Driver information system: A combination of augmented reality, deep learning and vehicular Ad-hoc networks. Multimed. Tools Appl. 2018, 77, 14673–14703. [CrossRef] 182. Rameau, F.; Ha, H.; Joo, K.; Choi, J.; Park, K.; Kweon, I.S. A real-time augmented reality system to see-through cars. IEEE Trans. Vis. Comput. Graph. 2016, 22, 2395–2404. [CrossRef] 183. Abdi, L.; Meddeb, A. In-vehicle augmented reality system to provide driving safety information. J. Vis. 2018, 21, 163–184. [CrossRef] 184. Lindemann, P.; Lee, T.-Y.; Rigoll, G. Supporting Driver Situation Awareness for Autonomous Urban Driving with an Augmented- Reality Windshield Display. In Proceedings of the 2018 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Munich, Germany, 16–20 October 2018; pp. 358–363. 185. Park, M.W.; Jung, S.K. TLD based vehicle tracking system for AR-HUD using HOG and online SVM in EHMI. In Proceedings of the 2015 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 9–12 January 2015; pp. 289–290. Appl. Sci. 2022, 12, 6831 51 of 51 186. Park, B.-J.; Yoon, C.; Lee, J.-W.; Kim, K.-H. Augmented reality based on driving situation awareness in vehicle. In Proceedings of the 2015 17th International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 1–3 July 2015; pp. 593–595. 187. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [CrossRef] 188. Gudigar, A.; Chokkadi, S.; Raghavendra, U.J.M.T. A review on automatic detection and recognition of traffic sign. Multimed. Tools Appl. 2016, 75, 333–364. [CrossRef] 189. Vashisth, S.; Saurav, S. Histogram of Oriented Gradients based reduced feature for traffic sign recognition. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; pp. 2206–2212. 190. Gonzalez-Reyna, S.E.; Avina-Cervantes, J.G.; Ledesma-Orozco, S.E.; Cruz-Aceves, I. Eigen-gradients for traffic sign recognition. Math. Probl. Eng. 2013, 2013, 364305. [CrossRef] 191. Yu, D.; Hu, X.; Liang, K. A two-scaled fully convolutional learning network for road detection. IET Image Process. 2021, 16, 948–957. [CrossRef] 192. Teichmann, M.; Weber, M.; Zoellner, M.; Cipolla, R.; Urtasun, R. Multinet: Real-time joint semantic reasoning for autonomous driving. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1013–1020. 193. Lv, W.; Song, W.-G.; Liu, X.-D.; Ma, J. A microscopic lane changing process model for multilane traffic. Phys. A Stat. Mech. Its Appl. 2013, 392, 1142–1152. [CrossRef] 194. Van Hasselt, H.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. arXiv 2015, arXiv:06461. 195. Pan, J.; Wang, X.; Cheng, Y.; Yu, Q. Multisource transfer double DQN based on actor learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2227–2238. [CrossRef] [PubMed] 196. Shuai, B.; Zhou, Q.; Li, J.; He, Y.; Li, Z.; Williams, H.; Xu, H.; Shuai, S. Heuristic action execution for energy efficient charge- sustaining control of connected hybrid vehicles with model-free double Q-learning. Appl. Energy 2020, 267, 114900. [CrossRef] 197. Okumura, B.; James, M.R.; Kanzawa, Y.; Derry, M.; Sakai, K.; Nishi, T.; Prokhorov, D. Challenges in perception and decision making for intelligent automotive vehicles: A case study. IEEE Trans. Intell. Veh. 2016, 1, 20–32. [CrossRef] 198. Muffert, M.; Pfeiffer, D.; Franke, U. A stereo-vision based object tracking approach at roundabouts. IEEE Intell. Transp. Syst. Mag. 2013, 5, 22–32. [CrossRef] 199. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamzadeh, M.; Fieguth, P.; Cao, X.; Khosravi, A.; Acharya, U.R. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf. Fusion 2021, 76, 243–297. [CrossRef] 200. Endsley, M.R. Autonomous driving systems: A preliminary naturalistic study of the Tesla Model S. J. Cogn. Eng. Decis. Mak. 2017, 11, 225–238. [CrossRef]

Journal

Applied SciencesMultidisciplinary Digital Publishing Institute

Published: Jul 6, 2022

Keywords: autonomous controlling; deep learning; decision making; intelligent vehicle; perception; self-driving

There are no references for this article.