Submitted Successfully!
To reward your contribution, here is a gift for you: A free trial for our video production service.
Thank you for your contribution! You can also upload a video entry or images related to this topic.
Version Summary Created by Modification Content Size Created at Operation
1 -- 3523 2023-03-20 15:01:45 |
2 format change Meta information modification 3523 2023-03-21 02:29:28 |

Video Upload Options

Do you have a full video?


Are you sure to Delete?
If you have any further questions, please contact Encyclopedia Editorial Office.
Dilek, E.; Dener, M. Computer Vision Applications in Intelligent Transportation Systems. Encyclopedia. Available online: (accessed on 25 June 2024).
Dilek E, Dener M. Computer Vision Applications in Intelligent Transportation Systems. Encyclopedia. Available at: Accessed June 25, 2024.
Dilek, Esma, Murat Dener. "Computer Vision Applications in Intelligent Transportation Systems" Encyclopedia, (accessed June 25, 2024).
Dilek, E., & Dener, M. (2023, March 20). Computer Vision Applications in Intelligent Transportation Systems. In Encyclopedia.
Dilek, Esma and Murat Dener. "Computer Vision Applications in Intelligent Transportation Systems." Encyclopedia. Web. 20 March, 2023.
Computer Vision Applications in Intelligent Transportation Systems

As technology continues to develop, computer vision (CV) applications are becoming increasingly widespread in the intelligent transportation systems (ITS) context. These applications are developed to improve the efficiency of transportation systems, increase their level of intelligence, and enhance traffic safety. Advances in CV play an important role in solving problems in the fields of traffic monitoring and control, incident detection and management, road usage pricing, and road condition monitoring, among many others, by providing more effective methods.

intelligent transportation systems computer vision automatic number plate recognition (ANPR) traffic sign detection vehicle detection pedestrian detection lane line detection obstacle detection anomaly detection structural damage detection autonomous vehi

1. Introduction

Smart city technologies are an important element of effectively managing the rapid industrialization of the world today, as they can help to address the economic and environmental problems resulting from the increase in urban populations. Smart cities, which integrate traditional infrastructure and public services with technology to create a more efficient, sustainable, and accessible system while meeting the needs of city residents, also transform the traditional understanding of city management. Intelligent transportation systems (ITS), which are among the key components of smart cities, are developed to improve transportation safety and mobility, reduce environmental impact, promote sustainable transportation development, and enhance productivity [1].
ITSs offer modern solutions to transportation-related problems, such as traffic jams and accidents, and help to ensure the safety of road users by utilizing data collected from surrounding vehicles, infrastructure, and other networks. ITS applications exist in a variety of forms, including collaborative highway maneuvers, sharing road safety information, optimization of traffic signals, and autonomous driving [2]. ITS, which can be defined as integrated transportation management systems consisting of advanced data communication, information processing, and traffic management technologies, can instantly process real-time data collected from heterogeneous sources and analyze it to support better decision making [3].
Decisions that were formerly made based on human experience can now be made using computers by digitizing information. Moreover, predictions and forecasting can also be improved through the use of new-generation artificial intelligence (AI) algorithms. Thanks to AI technologies, it is possible to develop systems that can make decisions based on data. These technologies have also led to radical changes in many areas, including public transportation and transportation systems, and have helped to make different modes of transportation safer, greener, smarter, and more efficient [4]. Yuan et al. [5] divide AI applications in the field of ITS into three main categories, namely (i) detection/recognition, (ii) prediction, and (iii) management. Machine learning (ML) methods, a sub-branch of AI, act as the brain function of ITS and determine the accuracy, reliability, and smartness of the systems. In particular, in recent years, it has been observed that deep learning (DL) methods, which are a subset of ML methods, are being effectively utilized in classification and prediction works in different areas of ITS [3].
Computer vision (CV) is an AI field that enables machines to derive meaningful information from digital images, videos, and other visual inputs, as well as to act based on this information [6]. CV, in which both ML and DL methods are used, addresses image and video processing problems and offers solutions that can be used in the process of automating transportation systems and making them safer. CV techniques are actively used in various ITS applications, such as automatic license plate detection and recognition, traffic sign detection and recognition, vehicle detection and classification, pedestrian detection, obstacle and lane line detection, anomaly detection in video surveillance cameras, vehicle and passenger tracking, structural damage detection, and autonomous vehicle applications. CV methods are appealing in these applications largely due to their cost-effectiveness, as well as the wide range of applications that CV can support [7].
CV methods used in ITS are categorized and examined under 10 headings, as shown in Figure 1.
Figure 1. Computer Vision Applications in Intelligent Transportation Systems.
  • CV applications in the field of ITS, along with the methods used, datasets, performance evaluation criteria, and success rates, are examined in a holistic and comprehensive way.
  • The problems and application areas addressed by CV applications in ITS are investigated.
  • The potential effects of CV studies on the transportation sector are evaluated.
  • The applicability, contributions, shortcomings, challenges, future research areas, and trends of CV applications in ITS are summarized.
  • Suggestions are made that will aid in improving the efficiency and effectiveness of transportation systems, increasing their safety levels, and making them smarter through CV studies in the future.
  • This research surveys over 300 studies that shed light on the development of CV techniques in the field of ITS. These studies have been published in journals listed in top electronic libraries and presented at leading conferences. The survey further presents recent academic papers and review articles that can be consulted by researchers aiming to conduct detailed analysis of the categories of CV applications.
  • It is believed that this survey can provide useful insights for researchers working on the potential effects of CV techniques, the automation of transportation systems, and the improvement of the efficiency and safety of ITS.

2. Computer Vision Studies in the Field of ITS

2.1. Evolution of Computer Vision Studies

While there are many methods used in CV studies in the literature, the methods most commonly used in the field of ITS are summarized in the following sections.

2.1.1. Handcrafted Techniques

Early CV researchers focused primarily on the use of different handcrafted spatiotemporal features and traditional image-processing methods [8]. Handcrafted features are those obtained with the help of several algorithms using the information that exists in the image itself. These features have been widely used in previous works using traditional ML approaches for object recognition.
Deformable part-based models, integral channel features (ICF), aggregated channel features (ACF), histograms of oriented gradients (HOG), local binary patterns (LBPs), scale-invariant feature transform (SIFT), Gabor filters, local ternary patterns (LTPs), local phase quantization (LPQ), rotation-invariant co-occurrence local binary patterns, completed local binary patterns, rotated local binary pattern images, and globally rotation-invariant multi-scale co-occurrence local binary patterns are among the handcrafted techniques that were used to extract features from images in previous studies [9]. Newer approaches, such as convolutional neural networks (CNNs), do not require such handcrafted features, as they can learn features from the image data.

2.1.2. Machine Learning and Deep Learning Methods

Machine learning, one of the most prominent subfields of AI, deals with the design and creation of algorithms for the recognition of complex patterns and decision making based on experimental data [10]. Problems handled with ML methods can be broadly categorized into (i) supervised, (ii) unsupervised, and (iii) reinforcement learning methods. In supervised learning, the goal is to estimate an output by taking feature vectors as inputs. Here, the ML algorithms establish a temporary model between the input and output values. The model attempts to estimate the output of the test data, which it has never seen before. If the ML model divides the input data into certain categories, then it is considered to be a classification; if the model tries to find continuous values using input values, then it is considered to be a regression. For both problems, the data must be labeled beforehand. The most frequently used algorithms for classification are support vector machine (SVM), collective/ensemble learning, k-nearest neighbors, and random forest (RF). Support vector regression and Gaussian process regression models are used in the literature for regression. Supervised learning models have been used for the classification of vehicles [11][12], classification of traffic lights [13], recognition and classification of license plate characters [14][15][16][17], detection of traffic signs [18], detection of pedestrians [19][20], etc.
Since assigning labels to millions of data points is a laborious and inefficient process, unlabeled data can be grouped through the use of unsupervised learning algorithms. Using different mathematical infrastructures, these algorithms classify data according to their own criteria. Among the unsupervised learning algorithms, methods such as k-means, density-based spatial clustering of applications with noise (DBSCAN), and the Gaussian Mixture Model (GMM) are used to identify groups and clusters. Unsupervised learning models have been used for the recognition of license plates [21], detection of obstacles [22], detection of road cracks [23], etc.
Based on the idea that there may be no available training data in some cases, reinforcement learning models have been developed, inspired by the knowledge acquisition processes of infants. These algorithms utilize a type of learning that tries to find the steps that a subject (a robot, an autonomous vehicle, etc.) must perform in order to receive the highest reward in the environment. Subjects working according to the reward–punishment mechanism perform actions in an attempt to understand the environment. After a range of these actions have been performed, the steps that lead to the highest reward score are saved, and these turn into behaviors. There are studies in the literature in which reinforcement learning methods were used in traffic signal control systems [24], traffic timing applications [25], and for the detection of lane lines [26].
Although traditional ML methods such as SVM [16][19][27], Bayesian networks [28], and the Kalman filter (KF) [29][30] were used in early ITS research [3], the problem-solving capabilities of algorithms have improved over time due to the development of hardware resources and the increasing amount of training data. As can be seen in Figure 2, while the concept of ML was dominant in the years between 1980 and 2010 [31], it was observed that these classical ML algorithms had difficulties processing large amounts of data; in response, artificial neural network (ANN)-based models began to emerge. However, since classical ANN models are insufficient for processing big data, modern ANN structures have been developed, which led to the development of DL models. While models that extract features from images and those that perform classification using these features are separate in ML algorithms, DL models can perform both processes in a single artificial neural stack.
Figure 2. Evolution of Artificial Intelligence, Machine Learning, and Deep Learning.

2.1.3. Deep Neural Networks (DNNs)

A DNN consists of multiple layers of ANN architectures and DNN models. It contains an input layer, one or more hidden layers, and an output layer. As a groundbreaking innovation, DNNs have produced satisfactory results on basic tasks such as the classification, detection, and segmentation of objects. Thus, AI technologies have become important in the field of ITS thanks to DNNs.
There are many types of DNN models which are used for different purposes. For example, deep belief networks (DBN) have been used for facial recognition [32] and crack detection [33]; stacked auto-encoder (SAE) networks have been used for object detection [34], image compression [35], and video retrieval [36]; restricted Boltzmann machines (RBM) have been used for face recognition [37], and YOLO (You Only Look Once)-based DL methods have been utilized in object-detection [38] tasks.

2.1.4. Convolutional Neural Networks (CNNs)

In the field of CV, the DNN most widely used to extract features from images is the CNN. In essence, CNNs try to imitate the working principles of the human brain and visual cortex, making use of multiple layers to recognize objects. One of the outstanding strengths of CNNs is their ability to classify objects into thousands of classes. Other advantages of CNNs include their relative robustness to image noise, along with their robustness to rotation and changes in the position of objects in an image. Their biggest disadvantages are their long training time and the need for a large training dataset [39]. The use of graphics cards and parallel processors during training contributes positively to the training and classification time of CNN models.
Variants of CNN networks are widely used in CV studies in the field of ITS. There are a number of CNN-based studies in the literature, such as those focused on automatic license plate recognition [40][41], traffic sign detection and recognition [25][42][43][44][45][46][47][48][49][50][51], vehicle detection [52][53][54][55], pedestrian detection [56][57][58][59][60], lane line detection [61][62][63], obstacle detection [64], video anomaly detection [65][66][67][68], structural damage detection [69][70][71][72][73][74][75][76][77][78], and steering angle detection [79][80][81][82] in autonomous vehicles. The most popular and advanced CNN-based architectures in the literature [83][84] are presented in Figure 3.
Figure 3. CNN Architectures.

2.1.5. Recurrent Neural Networks (RNNs)

RNNs are specially designed for modeling sequence data. The RNN is a powerful DL method, as it can directly learn the mapping between input and output sequences. However, traditional RNNs are impacted by the gradient vanishing problem. Long short-term memory (LSTM) networks were developed to solve this problem. An LSTM network is a type of RNN that can learn order dependence in sequence prediction tasks. In LSTM networks, memory cells are designed to maintain their state over time and learn long-term dependencies. RNNs have been used for license plate recognition [85], lane line detection [63], and crack classification [76] tasks, as well as in autonomous vehicle applications [86].
The gated recurrent unit (GRU) is a simplified variant of LSTM that does not contain discrete memory cells. The GRU is faster to train, while retaining its resilience to the vanishing gradient problem.
Convolutional LSTM networks have been used for the detection of anomalies in videos [87][88][89][90][91][92], as well as in autonomous vehicle applications [86][93], while a convolutional GRU network was used for video anomaly detection [94].

2.1.6. Generative Adversarial Networks (GANs)

The GAN is an approach based on generative modeling that uses DL methods to produce high-quality images. In recent years, GANs have been widely studied by DL communities in the context of video anomaly detection studies.
Generative modeling is an unsupervised learning method that involves automatically discovering and learning regularities or patterns in the input data, which the model can use to generate or create new examples that may be reasonably drawn from the original dataset. GANs are based on a learning approach that utilizes two sub-models, called the discriminator and generator, to train generative models. GAN is based on the idea of training implicitly through the discriminator, which is an ANN that dynamically updates itself and can gauge how realistic the input appears. Rather than minimizing the difference from a particular image, the generator learns in an unsupervised manner to fool the discriminator. GANs have been widely used in recent video anomaly detection studies [95][96][97][98][99].

2.1.7. Other Methods

Hybrid methods include a combination of multiple ML or DL methods used in CV techniques. There are many intelligent transportation applications for this approach, such as license plate recognition [85][100][101], video anomaly detection [68][89][92][102], automatic license plate recognition [25][103], vehicle detection [11][12][53][55], pedestrian detection [58][104], lane line detection [63][105], obstacle detection [106][107][108][109][110], structural damage detection [111][112][113], and autonomous vehicle applications [13][114][115].
Vaswani et al. [116] introduced an encoder–decoder architecture based on attention layers, named the transformer. A transformer neural network takes an input sentence in the form of a sequence of vectors, converts it into a vector called an encoding, and then decodes it back into another sequence. An essential part of the transformer is the attention mechanism, which represents how important other tokens in an input are for the encoding of a given token. Transformers are used for image classification, object detection, and image compression in CV applications. In the field of ITS, they have been used in license plate recognition [85], pedestrian detection [117], and driver distraction detection [118] studies.

2.2. Computer Vision Functions

Among the data emerging in the field of ITS, visual data are among the most voluminous kind. CV studies enable the analysis of both images and videos and provide detailed information about the traffic situation. Figure 4 presents some of the basic functions performed by CV techniques in the field of ITS. As can be seen from the figure, CV methods play a significant role in performing basic functions such as (i) classification, (ii) object detection, (iii) semantic segmentation, and (iv) instance segmentation [119].
Figure 4. Basic Functions Performed by Computer Vision Techniques in the Field of ITS.
Object classification can be performed by using CV techniques to process the image or video data obtained by the cameras. A label can be assigned automatically to each sub-object in the image. To achieve this, the objects are divided into parts and given to the model.
Another function performed using CV techniques is object detection. The detection of traffic objects such as vehicles and pedestrians in an image plays a vital role in the development of many applications. Important functions, such as detecting traffic density, detecting pedestrians that suddenly appear on the road, or detecting the locations of other vehicles for autonomous driving vehicles, can be performed with DL-based object detection models. The main feature that distinguishes object detection from classification is that the former can determine the coordinates of the area in which it is located, in addition to classifying each relevant object in the image. AI models of this kind can perform both classification and regression. The object with corner coordinate points becomes positionable by the machine in the image.
In the semantic segmentation context, all pixels belonging to objects are classified. As can be seen in Figure 4, cars are automatically marked in blue and pedestrians in red by CV techniques. Grouping all pixels of the object and assigning the appropriate class to each is a challenging problem. Semantic segmentation models assign the same groups of objects to a single class. However, vehicles and pedestrians in traffic sometimes need to be grouped individually. Under these circumstances, instance segmentation methods are used. The purpose of instance segmentation, like semantic segmentation, is to assign classes to pixels. With instance segmentation, objects belonging to the same class can be grouped separately, even if they overlap.
A framework outlining which problems in the field of ITS can be solved with CV techniques adapted from [120] is presented in Table 1.
Table 1. Problems Addressed by CV Techniques in the Field of ITS.

3. Computer Vision Applications in Intelligent Transportation Systems

ITSs have made many contributions to transport systems, including improving transport safety, increasing transport system efficiency, aiding law enforcement, and boosting energy conservation and emissions reduction. CV applications play an important role in this context and are thus of interest to researchers. In the Web of Science (WoS) database, there are more than a thousand studies that have been published in the field of CV in ITS since 2000. Since the field of ITS is multi-disciplinary, it has been observed that these publications extend across multiple scientific publication categories, such as electrical/electronic engineering, computer science, transportation science technology, civil engineering, telecommunications, and automation control systems.
Research into the use of CV methods in road transport systems was presented in [7] and [121], while a comprehensive review of traditional CV techniques for traffic analysis systems with a particular focus on urban environments was presented in [122]. However, those studies lack the state-of-the-art CV methods developed within the last decade. ML techniques have been used effectively to make transportation systems more efficient, especially in recent years, in. In current research, it has been noted that traditional ML models are now being replaced by new learning techniques and that DL techniques are widely used in ITS. A comprehensive study focusing on the use of DL models to increase the intelligence of transportation systems was presented by Wang et al. in [3]. Authors explored the use of DL models in various transportation applications including (i) traffic sign recognition, (ii) traffic flow prediction, (iii) traffic speed prediction, and (iv) travel time prediction. Applicability and shortcomings of DL models in the context of ITS and evolving future trends were also argued.


  1. Lin, Y.; Wang, P.; Ma, M. Intelligent Transportation System (ITS): Concept, Challenge and Opportunity. In Proceedings of the 2017 IEEE 3rd International Conference on Big Data Security On cloud (Bigdatasecurity), IEEE International Conference on High Performance and Smart Computing (HPSC), and IEEE International Conference on Intelligent Data and Security (IDS), Beijing, China, 26–28 May 2017; IEEE: Beijing, China, 2017; pp. 167–172.
  2. Porter, M. Towards Safe and Equitable Intelligent Transportation Systems: Leveraging Stochastic Control Theory in Attack Detection; The University of Michigan: Michigan, MI, USA, 2021.
  3. Wang, Y.; Zhang, D.; Liu, Y.; Dai, B.; Lee, L.H. Enhancing Transportation Systems via Deep Learning: A Survey. Transp. Res. Part C Emerg. Technol. 2019, 99, 144–163.
  4. Parveen, S.; Chadha, R.S.; Noida, C.; Kumar, I.P.; Singh, J. Artificial Intelligence in Transportation Industry. Int. J. Innov. Sci. Res. Technol. 2022, 7, 1274–1283.
  5. Yuan, Y.; Xiong, Z.; Wang, Q. An Incremental Framework for Video-Based Traffic Sign Detection, Tracking, and Recognition. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1918–1929.
  6. Sharma, V.; Gupta, M.; Kumar, A.; Mishra, D. Video Processing Using Deep Learning Techniques: A Systematic Literature Review. IEEE Access 2021, 9, 139489–139507.
  7. Loce, R.P.; Bernal, E.A.; Wu, W.; Bernal, E.A.; Bala, R. Computer Vision in Roadway Transportation Systems: A Survey Process Mining and Data Automation View Project Gait Segmentation View Project Computer Vision in Roadway Transportation Systems: A Survey Computer Vision in Roadway Transportation Systems: A Survey. Artic. J. Electron. Imaging 2013, 22, 041121.
  8. Patrikar, D.R.; Parate, M.R. Anomaly Detection Using Edge Computing in Video Surveillance System: Review. Int. J. Multimed. Inf. Retr. 2022, 11, 85–110.
  9. Nanni, L.; Ghidoni, S.; Brahnam, S. Handcrafted vs. Non-Handcrafted Features for Computer Vision Classification. Pattern Recognit. 2017, 71, 158–172.
  10. Varshney, H.; Khan, R.A.; Khan, U.; Verma, R. Approaches of Artificial Intelligence and Machine Learning in Smart Cities: Critical Review. IOP Conf. Ser. Mater Sci. Eng. 2021, 1022, 012019.
  11. Mittal, D.; Reddy, A.; Ramadurai, G.; Mitra, K.; Ravindran, B. Training a Deep Learning Architecture for Vehicle Detection Using Limited Heterogeneous Traffic Data. In Proceedings of the 2018 10th International Conference on Communication Systems & Networks (COMSNETS), Bengaluru, India, 3–7 January 2018; pp. 294–589.
  12. Alam, A.; Jaffery, Z.A.; Sharma, H. A Cost-Effective Computer Vision-Based Vehicle Detection System. Concurr. Eng. 2022, 30, 148–158.
  13. Vishal, K.; Arvind, C.S.; Mishra, R.; Gundimeda, V. Traffic Light Recognition for Autonomous Vehicles by Admixing the Traditional ML and DL. In Proceedings of the Eleventh International Conference on Machine Vision (ICMV 2018), Munich, Germany, 1–3 November 2018; SPIE: Bellingham, WA, USA, 2019; Volume 11041, pp. 126–133.
  14. Al-Shemarry, M.S.; Li, Y. Developing Learning-Based Preprocessing Methods for Detecting Complicated Vehicle Licence Plates. IEEE Access 2020, 8, 170951–170966.
  15. Greenhalgh, J.; Mirmehdi, M. Real-Time Detection and Recognition of Road Traffic Signs. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1498–1506.
  16. Maldonado-Bascón, S.; Lafuente-Arroyo, S.; Gil-Jimenez, P.; Gómez-Moreno, H.; López-Ferreras, F. Road-Sign Detection and Recognition Based on Support Vector Machines. IEEE Trans. Intell. Transp. Syst. 2007, 8, 264–278.
  17. Lafuente-Arroyo, S.; Gil-Jimenez, P.; Maldonado-Bascon, R.; López-Ferreras, F.; Maldonado-Bascon, S. Traffic Sign Shape Classification Evaluation I: SVM Using Distance to Borders. In Proceedings of the IEEE Proceedings. Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; IEEE: New York, NY, USA, 2005; pp. 557–562.
  18. Li, C.; Yang, C. The Research on Traffic Sign Recognition Based on Deep Learning. In Proceedings of the 2016 16th International Symposium on Communications and Information Technologies (ISCIT), Qingdao, China, 26–28 September 2016; IEEE: New York, NY, USA, 2016; pp. 156–161.
  19. Oren, M.; Papageorgiou, C.; Sinha, P.; Osuna, E.; Poggio, T. Pedestrian Detection Using Wavelet Templates. In Proceedings of the Proceedings of the IEEE COMPUTER society Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 1997; IEEE: New York, NY, USA, 1997; pp. 193–199.
  20. Papageorgiou, C.; Evgeniou, T.; Poggio, T. A Trainable Pedestrian Detection System. In Proceedings of the Proc. of Intelligent Vehicles, Seville, Spain, 23–24 March 1998; pp. 241–246.
  21. Pustokhina, I.V.; Pustokhin, D.A.; Rodrigues, J.J.P.C.; Gupta, D.; Khanna, A.; Shankar, K.; Seo, C.; Joshi, G.P. Automatic Vehicle License Plate Recognition Using Optimal K-Means with Convolutional Neural Network for Intelligent Transportation Systems. Ieee Access 2020, 8, 92907–92917.
  22. Hu, F.; Tian, Z.; Li, Y.; Huang, S.; Feng, M. A Combined Clustering and Image Mapping Based Point Cloud Segmentation for 3D Object Detection. In Proceedings of the 2018 Chinese Control And Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; IEEE: New York, NY, USA, 2018; pp. 1664–1669.
  23. Shan, B.; Zheng, S.; Ou, J. A Stereovision-Based Crack Width Detection Approach for Concrete Surface Assessment. KSCE J. Civ. Eng. 2016, 20, 803–812.
  24. Hurtado-Gómez, J.; Romo, J.D.; Salazar-Cabrera, R.; Pachon de la Cruz, A.; Madrid Molina, J.M. Traffic Signal Control System Based on Intelligent Transportation System and Reinforcement Learning. Electronics 2021, 10, 2363.
  25. Li, L.; Lv, Y.; Wang, F.-Y. Traffic Signal Timing via Deep Reinforcement Learning. IEEE/CAA J. Autom. Sin. 2016, 3, 247–254.
  26. Liu, Y.; Xu, P.; Zhu, L.; Yan, M.; Xue, L. Reinforced Attention Method for Real-Time Traffic Line Detection. J. Real Time Image Process. 2022, 19, 957–968.
  27. Le, T.T.; Tran, S.T.; Mita, S.; Nguyen, T.D. Real Time Traffic Sign Detection Using Color and Shape-Based Features. In Proceedings of the ACIIDS, Hue City, Vietnam, 24–26 March 2010; pp. 268–278.
  28. Song, X.; Nevatia, R. Detection and Tracking of Moving Vehicles in Crowded Scenes. In Proceedings of the 2007 IEEE Workshop on Motion and Video Computing (WMVC’07), Austin, TX, USA, 23–24 February 2007; p. 4.
  29. Messelodi, S.; Modena, C.M.; Segata, N.; Zanin, M. A Kalman Filter Based Background Updating Algorithm Robust to Sharp Illumination Changes. In Proceedings of the ICIAP, Cagliari, Italy, 6–8 September 2005; Volume 3617, pp. 163–170.
  30. Okutani, I.; Stephanedes, Y.J. Dynamic Prediction of Traffic Volume through Kalman Filtering Theory. Transp. Res. Part B Methodol. 1984, 18, 1–11.
  31. Ramrath, B.; Ari, L.; Doug, M. Insurance 2030—The Impact of AI on the Future of Insurance. 2021. Available online: (accessed on 22 January 2023).
  32. Babu, K.; Kumar, C.; Kannaiyaraju, C. Face Recognition System Using Deep Belief Network and Particle Swarm Optimization. Intell. Autom. Soft Comput. 2022, 33, 317–329.
  33. Wang, X.; Zhang, Y. The Detection and Recognition of Bridges’ Cracks Based on Deep Belief Network. In Proceedings of the 2017 IEEE International Conference on Computational Science and Engineering (CSE) and IEEE International Conference on Embedded and Ubiquitous Computing (EUC), Guangzhou, China, 21–24 July 2017; IEEE: New York, NY, USA, 2017; Volume 1, pp. 768–771.
  34. Maria, J.; Amaro, J.; Falcao, G.; Alexandre, L.A. Stacked Autoencoders Using Low-Power Accelerated Architectures for Object Recognition in Autonomous Systems. Neural Process. Lett. 2016, 43, 445–458.
  35. Theis, L.; Shi, W.; Cunningham, A.; Huszár, F. Lossy Image Compression with Compressive Autoencoders. arXiv 2017, arXiv:1703.00395.
  36. Song, J.; Zhang, H.; Li, X.; Gao, L.; Wang, M.; Hong, R. Self-Supervised Video Hashing with Hierarchical Binary Auto-Encoder. IEEE Trans. Image Process. 2018, 27, 3210–3221.
  37. Teh, Y.W.; Hinton, G.E. Rate-Coded Restricted Boltzmann Machines for Face Recognition. Adv. Neural Inf. Process. Syst. 2000, 13, 872–878.
  38. Ghahremannezhad, H.; Shi, H.; Liu, C. Real-Time Accident Detection in Traffic Surveillance Using Deep Learning. In Proceedings of the 2022 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 21–23 June 2022; IEEE: New York, NY, USA, 2022; pp. 1–6.
  39. Lange, S.; Ulbrich, F.; Goehring, D. Online Vehicle Detection Using Deep Neural Networks and Lidar Based Preselected Image Patches. In Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 22 January 2016; IEEE: New York, NY, USA, 2016; pp. 954–959.
  40. Laroca, R.; Severo, E.; Zanlorensi, L.A.; Oliveira, L.S.; Gonçalves, G.R.; Schwartz, W.R.; Menotti, D. A Robust Real-Time Automatic License Plate Recognition Based on the YOLO Detector. In Proceedings of the 2018 International joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: New York, NY, USA, 2018; pp. 1–10.
  41. Hashmi, S.N.; Kumar, K.; Khandelwal, S.; Lochan, D.; Mittal, S. Real Time License Plate Recognition from Video Streams Using Deep Learning. Int. J. Inf. Retr. Res. 2019, 9, 65–87.
  42. Cirean, D.; Meier, U.; Masci, J.; Schmidhuber, J. A Committee of Neural Networks for Traffic Sign Classification. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 1918–1921.
  43. Sermanet, P.; LeCun, Y. Traffic Sign Recognition with Multi-Scale Convolutional Networks. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; IEEE: New York, NY, USA, 2011; pp. 2809–2813.
  44. Ciresan, D.; Meier, U.; Masci, J.; Schmidhuber, J. Multi-Column Deep Neural Network for Traffic Sign Classification. Neural Netw. 2012, 32, 333–338.
  45. Jin, J.; Fu, K.; Zhang, C. Traffic Sign Recognition with Hinge Loss Trained Convolutional Neural Networks. IEEE Trans. Intell. Transp. Syst. 2014, 15, 1991–2000.
  46. Haloi, M. Traffic Sign Classification Using Deep Inception Based Convolutional Networks. arXiv 2015, arXiv:1511.02992.
  47. Qian, R.; Zhang, B.; Yue, Y.; Wang, Z.; Coenen, F. Robust Chinese Traffic Sign Detection and Recognition with Deep Convolutional Neural Network. In Proceedings of the 2015 11th International Conference on Natural Computation (ICNC), Zhangjiajie, China, 15–17 August 2015; IEEE: New York, NY, USA, 2015; pp. 791–796.
  48. Changzhen, X.; Cong, W.; Weixin, M.; Yanmei, S. A Traffic Sign Detection Algorithm Based on Deep Convolutional Neural Network. In Proceedings of the 2016 IEEE International Conference on Signal and Image Processing (ICSIP), Beijing, China, 13–15August 2016; IEEE: New York, NY, USA, 2016; pp. 676–679.
  49. Jung, S.; Lee, U.; Jung, J.; Shim, D.H. Real-Time Traffic Sign Recognition System with Deep Convolutional Neural Network. In Proceedings of the 2016 13th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Xi’an, China, 19–22 August 2016; IEEE: New York, NY, USA, 2016; pp. 31–34.
  50. Zeng, Y.; Xu, X.; Shen, D.; Fang, Y.; Xiao, Z. Traffic Sign Recognition Using Kernel Extreme Learning Machines with Deep Perceptual Features. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1647–1653.
  51. Zhang, J.; Huang, Q.; Wu, H.; Liu, Y. A Shallow Network with Combined Pooling for Fast Traffic Sign Recognition. Information 2017, 8, 45.
  52. Du, X.; Ang, M.H.; Rus, D. Car Detection for Autonomous Vehicle: LIDAR and Vision Fusion Approach through Deep Learning Framework. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24 September 2017; IEEE: New York, NY, USA, 2017; pp. 749–754.
  53. Wu, L.-T.; Lin, H.-Y. Overtaking Vehicle Detection Techniques Based on Optical Flow and Convolutional Neural Network. In Proceedings of the VEHITS, Madeira, Portugal, 16–18 March 2018; pp. 133–140.
  54. Pillai, U.K.K.; Valles, D. An Initial Deep CNN Design Approach for Identification of Vehicle Color and Type for Amber and Silver Alerts. In Proceedings of the 2021 IEEE 11th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 27–30 January 2021; IEEE: New York, NY, USA, 2021; pp. 903–908.
  55. Shvai, N.; Hasnat, A.; Meicler, A.; Nakib, A. Accurate Classification for Automatic Vehicle-Type Recognition Based on Ensemble Classifiers. IEEE Trans. Intell. Transp. Syst. 2019, 21, 1288–1297.
  56. Yi, S. Pedestrian Behavior Modeling and Understanding in Crowds. Doctoral Dissertation, The Chinese University of Hong Kong, Hong Kong, 2016.
  57. Ouyang, W.; Wang, X. Joint Deep Learning for Pedestrian Detection. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2056–2063.
  58. Fukui, H.; Yamashita, T.; Yamauchi, Y.; Fujiyoshi, H.; Murase, H. Pedestrian Detection Based on Deep Convolutional Neural Network with Ensemble Inference Network. In Proceedings of the 2015 IEEE Intelligent Vehicles Symposium (IV), Seoul, Republic of Korea, 28 June–1 July 2015; IEEE: New York, NY, USA, 2015; pp. 223–228.
  59. John, V.; Mita, S.; Liu, Z.; Qi, B. Pedestrian Detection in Thermal Images Using Adaptive Fuzzy C-Means Clustering and Convolutional Neural Networks. In Proceedings of the 2015 14th IAPR International Conference on Machine Vision Applications (MVA), Tokyo, Japan, 18–22 May 2015; IEEE: New York, NY, USA, 2015; pp. 246–249.
  60. Schlosser, J.; Chow, C.K.; Kira, Z. Fusing Lidar and Images for Pedestrian Detection Using Convolutional Neural Networks. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; IEEE: New York, NY, USA, 2016; pp. 2198–2205.
  61. Kim, J.; Lee, M. Robust Lane Detection Based on Convolutional Neural Network and Random Sample Consensus. In Proceedings of the Neural Information Processing: 21st International Conference, ICONIP 2014, Kuching, Malaysia, 3–6 November 2014; Part I 21. Springer: Berlin/Heidelberg, Germany, 2014; pp. 454–461.
  62. Huval, B.; Wang, T.; Tandon, S.; Kiske, J.; Song, W.; Pazhayampallil, J.; Andriluka, M.; Rajpurkar, P.; Migimatsu, T.; Cheng-Yue, R. An Empirical Evaluation of Deep Learning on Highway Driving. arXiv 2015, arXiv:1504.01716.
  63. Li, J.; Mei, X.; Prokhorov, D.; Tao, D. Deep Neural Network for Structural Prediction and Lane Detection in Traffic Scene. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 690–703.
  64. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969.
  65. Shao, J.; Loy, C.-C.; Kang, K.; Wang, X. Slicing Convolutional Neural Network for Crowd Video Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 5620–5628.
  66. Sabokrou, M.; Fayyaz, M.; Fathy, M.; Moayed, Z.; Klette, R. Deep-Anomaly: Fully Convolutional Neural Network for Fast Anomaly Detection in Crowded Scenes. Comput. Vis. Image Underst. 2018, 172, 88–97.
  67. Sun, J.; Shao, J.; He, C. Abnormal Event Detection for Video Surveillance Using Deep One-Class Learning. Multimed Tools Appl. 2019, 78, 3633–3647.
  68. Sabih, M.; Vishwakarma, D.K. Crowd Anomaly Detection with LSTMs Using Optical Features and Domain Knowledge for Improved Inferring. Vis. Comput. 2022, 38, 1719–1730.
  69. Zhang, A.; Wang, K.C.P.; Li, B.; Yang, E.; Dai, X.; Peng, Y.; Fei, Y.; Liu, Y.; Li, J.Q.; Chen, C. Automated Pixel-level Pavement Crack Detection on 3D Asphalt Surfaces Using a Deep-learning Network. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 805–819.
  70. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road Crack Detection Using Deep Convolutional Neural Network. In Proceedings of the 2016 IEEE international conference on image processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; IEEE: New York, NY, USA, 2016; pp. 3708–3712.
  71. Gulgec, N.S.; Takáč, M.; Pakzad, S.N. Structural Damage Detection Using Convolutional Neural Networks. In Model Validation and Uncertainty Quantification, Proceedings of the 35th IMAC, A Conference and Exposition on Structural Dynamics; Springer: Berlin/Heidelberg, Germany, 2017; Volume 3, pp. 331–337.
  72. Protopapadakis, E.; Voulodimos, A.; Doulamis, A.; Doulamis, N.; Stathaki, T. Automatic Crack Detection for Tunnel Inspection Using Deep Learning and Heuristic Image Post-Processing. Appl. Intell. 2019, 49, 2793–2806.
  73. Mandal, V.; Uong, L.; Adu-Gyamfi, Y. Automated Road Crack Detection Using Deep Convolutional Neural Networks. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; IEEE: New York, NY, USA, 2018; pp. 5212–5215.
  74. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9.
  75. Li, S.; Zhao, X. Convolutional Neural Networks-Based Crack Detection for Real Concrete Surface. In Proceedings of the Sensors and Smart Structures Technologies for Civil, Mechanical, and Aerospace Systems 2018, Denver, CO, USA, 5–8 March 2018; SPIE: Bellingham, WA, USA, 2018; Volume 10598, pp. 955–961.
  76. Ahmed, T.U.; Hossain, M.S.; Alam, M.J.; Andersson, K. An Integrated CNN-RNN Framework to Assess Road Crack. In Proceedings of the 2019 22nd International Conference on Computer and Information Technology (ICCIT), Dhaka, Bangladesh, 18–20 December 2019; IEEE: New York, NY, USA, 2019; pp. 1–6.
  77. Nguyen, N.H.T.; Perry, S.; Bone, D.; Le, H.T.; Nguyen, T.T. Two-Stage Convolutional Neural Network for Road Crack Detection and Segmentation. Expert Syst. Appl. 2021, 186, 115718.
  78. Chen, J.; Liu, Z.; Wang, H.; Liu, K. High-Speed Railway Catenary Components Detection Using the Cascaded Convolutional Neural Networks. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; IEEE: New York, NY, USA, 2017; pp. 1–6.
  79. Bojarski, M.; del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J. End to End Learning for Self-Driving Cars. arXiv 2016, arXiv:1604.07316.
  80. Farkh, R.; Alhuwaimel, S.; Alzahrani, S.; al Jaloud, K.; Tabrez Quasim, M. Deep Learning Control for Autonomous Robot. Comput. Mater. Contin. 2022, 72, 2811–2824.
  81. Nose, Y.; Kojima, A.; Kawabata, H.; Hironaka, T. A Study on a Lane Keeping System Using CNN for Online Learning of Steering Control from Real Time Images. In Proceedings of the 2019 34th International Technical Conference on Circuits/Systems, Computers and Communications (ITC-CSCC), Jeju, Republic of Korea, 23–26 June 2019; IEEE: New York, NY, USA, 2019; pp. 1–4.
  82. Chen, Z.; Huang, X. End-to-End Learning for Lane Keeping of Self-Driving Cars. In Proceedings of the 2017 IEEE intelligent vehicles symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; IEEE: New York, NY, USA, 2017; pp. 1856–1860.
  83. Rateke, T.; von Wangenheim, A. Passive Vision Road Obstacle Detection: A Literature Mapping. Int. J. Comput. Appl. 2022, 44, 376–395.
  84. Ali, R.; Chuah, J.H.; Talip, M.S.A.; Mokhtar, N.; Shoaib, M.A. Structural Crack Detection Using Deep Convolutional Neural Networks. Autom. Constr. 2022, 133, 103989.
  85. GONG, W.; SHI, Z.; Qiang, J.I. Non-Segmented Chinese License Plate Recognition Algorithm Based on Deep Neural Networks. In Proceedings of the 2020 Chinese Control and Decision Conference (CCDC), Hefei, China, 22–24 August 2020; IEEE: New York, NY, USA, 2020; pp. 66–71.
  86. Chen, S.; Zhang, S.; Shang, J.; Chen, B.; Zheng, N. Brain-Inspired Cognitive Model with Attention for Self-Driving Cars. IEEE Trans. Cogn. Dev. Syst. 2017, 11, 13–25.
  87. Medel, J.R.; Savakis, A. Anomaly Detection in Video Using Predictive Convolutional Long Short-Term Memory Networks. arXiv 2016, arXiv:1612.00390.
  88. Medel, J.R. Anomaly Detection Using Predictive Convolutional Long Short-Term Memory Units; Rochester Institute of Technology: Rochester, NY, USA, 2016; ISBN 1369443943.
  89. Luo, W.; Liu, W.; Gao, S. Remembering History with Convolutional Lstm for Anomaly Detection. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, 10–14 July 2017; IEEE: New York, NY, USA, 2017; pp. 439–444.
  90. Patraucean, V.; Handa, A.; Cipolla, R. Spatio-Temporal Video Autoencoder with Differentiable Memory. arXiv 2015, arXiv:1511.06309.
  91. Li, Y.; Cai, Y.; Liu, J.; Lang, S.; Zhang, X. Spatio-Temporal Unity Networking for Video Anomaly Detection. IEEE Access 2019, 7, 172425–172432.
  92. Wang, L.; Tan, H.; Zhou, F.; Zuo, W.; Sun, P. Unsupervised Anomaly Video Detection via a Double-Flow Convlstm Variational Autoencoder. IEEE Access 2022, 10, 44278–44289.
  93. Kim, J.; Canny, J. Interpretable Learning for Self-Driving Cars by Visualizing Causal Attention. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2942–2950.
  94. Wang, X.; Che, Z.; Jiang, B.; Xiao, N.; Yang, K.; Tang, J.; Ye, J.; Wang, J.; Qi, Q. Robust Unsupervised Video Anomaly Detection by Multipath Frame Prediction. IEEE Trans. Neural. Netw. Learn. Syst. 2021, 33, 2301–2312.
  95. Jackson, S.D.; Cuzzolin, F. SVD-GAN for Real-Time Unsupervised Video Anomaly Detection. In Proceedings of the British Machine Vision Conference (BMVC), Virtual, 22–25 November 2021.
  96. Song, H.; Sun, C.; Wu, X.; Chen, M.; Jia, Y. Learning Normal Patterns via Adversarial Attention-Based Autoencoder for Abnormal Event Detection in Videos. IEEE Trans. Multimed. 2019, 22, 2138–2148.
  97. Ganokratanaa, T.; Aramvith, S.; Sebe, N. Unsupervised Anomaly Detection and Localization Based on Deep Spatiotemporal Translation Network. IEEE Access 2020, 8, 50312–50329.
  98. Chen, D.; Yue, L.; Chang, X.; Xu, M.; Jia, T. NM-GAN: Noise-Modulated Generative Adversarial Network for Video Anomaly Detection. Pattern Recognit. 2021, 116, 107969.
  99. Huang, C.; Wen, J.; Xu, Y.; Jiang, Q.; Yang, J.; Wang, Y.; Zhang, D. Self-Supervised Attentive Generative Adversarial Networks for Video Anomaly Detection. In IEEE Transactions on Neural Networks and Learning Systems; IEEE: New York, NY, USA, 2022.
  100. Darapaneni, N.; Mogeraya, K.; Mandal, S.; Narayanan, A.; Siva, P.; Paduri, A.R.; Khan, F.; Agadi, P.M. Computer Vision Based License Plate Detection for Automated Vehicle Parking Management System. In Proceedings of the 2020 11th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), New York, NY, USA, 28–31 October 2020; IEEE: New York, NY, USA, 2020; pp. 800–805.
  101. Vetriselvi, T.; Lydia, E.L.; Mohanty, S.N.; Alabdulkreem, E.; Al-Otaibi, S.; Al-Rasheed, A.; Mansour, R.F. Deep Learning Based License Plate Number Recognition for Smart Cities. CMC Comput. Mater Contin. 2022, 70, 2049–2064.
  102. Duman, E.; Erdem, O.A. Anomaly Detection in Videos Using Optical Flow and Convolutional Autoencoder. IEEE Access 2019, 7, 183914–183923.
  103. Xing, J.; Nguyen, M.; Qi Yan, W. The Improved Framework for Traffic Sign Recognition Using Guided Image Filtering. SN Comput. Sci. 2022, 3, 461.
  104. Liu, J.; Zhang, S.; Wang, S.; Metaxas, D.N. Multispectral Deep Neural Networks for Pedestrian Detection. arXiv 2016, arXiv:1611.02644.
  105. Dewangan, D.K.; Sahu, S.P. Road Detection Using Semantic Segmentation-Based Convolutional Neural Network for Intelligent Vehicle System. In Proceedings of the Data Engineering and Communication Technology: Proceedings of ICDECT 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 629–637.
  106. Walk, S.; Schindler, K.; Schiele, B. Disparity Statistics for Pedestrian Detection: Combining Appearance, Motion and Stereo. In Proceedings of the Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, September 5–11, 2010, Proceedings, Part VI 11; Springer: Berlin/Heidelberg, Germany, 2010; pp. 182–195.
  107. Liu, Z.; Yu, C.; Zheng, B. Any Type of Obstacle Detection in Complex Environments Based on Monocular Vision. In Proceedings of the 32nd Chinese Control Conference, Xi’an, China, 26–28 July 2013; IEEE: New York, NY, USA, 2013; pp. 7692–7697.
  108. Pantilie, C.D.; Nedevschi, S. Real-Time Obstacle Detection in Complex Scenarios Using Dense Stereo Vision and Optical Flow. In Proceedings of the 13th International IEEE Conference on Intelligent Transportation Systems, Funchal, Madeira Island, Portugal, 19–22 September 2010; IEEE: New York, NY, USA, 2010; pp. 439–444.
  109. Dairi, A.; Harrou, F.; Sun, Y.; Senouci, M. Obstacle Detection for Intelligent Transportation Systems Using Deep Stacked Autoencoder and $ k $-Nearest Neighbor Scheme. IEEE Sens. J. 2018, 18, 5122–5132.
  110. Ci, W.; Xu, T.; Lin, R.; Lu, S. A Novel Method for Unexpected Obstacle Detection in the Traffic Environment Based on Computer Vision. Appl. Sci. 2022, 12, 8937.
  111. Cha, Y.; Choi, W.; Büyüköztürk, O. Deep Learning-based Crack Damage Detection Using Convolutional Neural Networks. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 361–378.
  112. Kortmann, F.; Fassmeyer, P.; Funk, B.; Drews, P. Watch out, Pothole! Featuring Road Damage Detection in an End-to-End System for Autonomous Driving. Data Knowl Eng 2022, 142, 102091.
  113. Liu, J.; Yang, X.; Lau, S.; Wang, X.; Luo, S.; Lee, V.C.; Ding, L. Automated Pavement Crack Detection and Segmentation Based on Two-step Convolutional Neural Network. Comput.-Aided Civ. Infrastruct. Eng. 2020, 35, 1291–1305.
  114. Muhammad, K.; Hussain, T.; Ullah, H.; del Ser, J.; Rezaei, M.; Kumar, N.; Hijji, M.; Bellavista, P.; de Albuquerque, V.H.C. Vision-Based Semantic Segmentation in Scene Understanding for Autonomous Driving: Recent Achievements, Challenges, and Outlooks. In IEEE Transactions on Intelligent Transportation Systems; IEEE: New York, NY, USA, 2022.
  115. Benamer, I.; Yahiouche, A.; Ghenai, A. Deep Learning Environment Perception and Self-Tracking for Autonomous and Connected Vehicles. In Proceedings of the Machine Learning for Networking: Third International Conference, MLN 2020, Paris, France, November 24–26, 2020, Revised Selected Papers 3; Springer: Berlin/Heidelberg, Germany, 2021; pp. 305–319.
  116. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neura. Inf. Process. Syst. 2017, 30, 6000–6010.
  117. Zang, X.; Li, G.; Gao, W. Multidirection and Multiscale Pyramid in Transformer for Video-Based Pedestrian Retrieval. IEEE Trans. Ind. Inf. 2022, 18, 8776–8785.
  118. Wang, H.; Chen, J.; Huang, Z.; Li, B.; Lv, J.; Xi, J.; Wu, B.; Zhang, J.; Wu, Z. FPT: Fine-Grained Detection of Driver Distraction Based on the Feature Pyramid Vision Transformer. In IEEE Transactions on Intelligent Transportation Systems; IEEE: New York, NY, USA, 2022.
  119. Kirillov, A.; He, K.; Girshick, R.; Rother, C.; Dollár, P. Panoptic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9404–9413.
  120. Rafie, M.; Zhang, Y.; Liu, S. Evaluation Framework for Video Coding for Machines. ISO/IEC JTC 2021, 1, 3–4.
  121. Manikoth, N.; Loce, R.; Bernal, E.; Wu, W. Survey of Computer Vision in Roadway Transportation Systems. In Proceedings of the Visual Information Processing and Communication III; SPIE: Bellingham, WA, USA, 2012; Volume 8305, pp. 258–276.
  122. Buch, N.; Velastin, S.A.; Orwell, J. A Review of Computer Vision Techniques for the Analysis of Urban Traffic. IEEE Trans. Intell. Transp. Syst. 2011, 12, 920–939.
Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to : ,
View Times: 830
Revisions: 2 times (View History)
Update Date: 21 Mar 2023
Video Production Service