Processes 09 01786 v2
Processes 09 01786 v2
Processes 09 01786 v2
Article
Efficient Video-based Vehicle Queue Length Estimation using
Computer Vision and Deep Learning for an Urban Traffic
Scenario
Muhammad Umair 1 , Muhammad Umar Farooq 1, * , Rana Hammad Raza 1 , Qian Chen 2 and Baher Abdulhai 3
1 Department of Electronics and Power Engineering, Pakistan Navy Engineering College (PNEC),
National University of Sciences and Technology (NUST), Karachi 75350, Pakistan;
umair.arif@pnec.nust.edu.pk (M.U.); hammad@pnec.nust.edu.pk (R.H.R.)
2 Division of Engineering Science, University of Toronto, Toronto, ON M5S 2E4, Canada;
philipq.chen@utoronto.ca
3 Department of Civil Engineering, University of Toronto, Toronto, ON M5S 1A4, Canada;
baher.abdulhai@utoronto.ca
* Correspondence: umar.farooq@pnec.nust.edu.pk
Abstract: In the Intelligent Transportation System (ITS) realm, queue length estimation is one of
an essential yet a challenging task. Queue lengths are important for determining traffic density
in traffic lanes so that possible congestion in any lane can be minimized. Smart roadside sensors
such as loop detectors, radars and pneumatic road tubes etc. are promising for such tasks though
they have a very high installation and maintenance cost. Large scale deployment of surveillance
cameras have shown a great potential in the collection of vehicular data in a flexible way and are
also cost effective. Similarly, vision-based sensors can be used independently or if required can also
Citation: Umair, M.; Farooq, M.U.;
Raza, R.H.; Chen, Q.; Abdulhai, B.
augment the functionality of other roadside sensors to effectively process queue length at prescribed
Efficient Video-based Vehicle Queue traffic lanes. In this research, a CNN-based approach for estimation of vehicle queue length in an
Length Estimation using Computer urban traffic scenario using low-resolution traffic videos is proposed. The queue length is estimated
Vision and Deep Learning for an based on count of total vehicles waiting on a signal. The proposed approach calculates queue length
Urban Traffic Scenario. Processes 2021, without the knowledge of any onsite camera calibration information. Average vehicle length is
9, 1786. https://doi.org/10.3390/ approximated to be 5 m. This caters for the vehicles at the far end of the traffic lane that appear
pr9101786 smaller in the camera view. Identification of stopped vehicles is done using Deep SORT based object
tracking. Due to robust and accurate CNN-based detection and tracking, the queue length estimated
Academic Editor: Konstantinos
by using only the cameras has been very effective. This mostly eliminates the need for fusion with
Demertzis
any roadside or in-vehicle sensors. A detailed comparative analysis of vehicle detection models
including YOLOv3, YOLOv4, YOLOv5, SSD, ResNet101, and InceptionV3 was performed. Based
Received: 2 September 2021
on this analysis, YOLOv4 was selected as a baseline model for queue length estimation. Using the
Accepted: 19 September 2021
Published: 8 October 2021
pre-trained 80-classes YOLOv4 model, an overall accuracy of 73% and 88% was achieved for vehicle
count and vehicle count-based queue length estimation, respectively. After fine-tuning of model and
Publisher’s Note: MDPI stays neutral narrowing the output classes to vehicle class only, an average accuracy of 83% and 93% was achieved,
with regard to jurisdictional claims in respectively. This shows the efficiency and robustness of the proposed approach.
published maps and institutional affil-
iations. Keywords: computer vision; deep learning; convolutional neural networks; intelligent transportation
system; vehicle queue length; YOLO
traffic require smart solutions like estimation of vehicle queue length in a traffic lane and
prediction of traffic etc [4–9].
Technological advancement in wireless communication and computing power has
transformed ITS. For example, development towards connected vehicles is contributing to
the implementation of next generation ITS. The connectivity in vehicles enable collection
of necessary data for better traffic management and control [10]. Recent advances in com-
puter vision and deep learning have also enabled improved vehicle annotation, detection
and classification in traffic surveillance application. Moreover, large-scale deployment of
surveillance cameras all over the world has shown a great potential in the collection of
vehicular data. Similarly, ITS technologies such as smart roadside sensors and surveillance
cameras with enhanced analytics functionalities are widely adapted around the world [3].
These traffic cameras can be used independently or if required can also augment the func-
tionality of other smart roadside sensors to effectively estimate queue length at prescribed
traffic lanes.
Vehicle queue length on the signalized intersection is defined as the distance from
the stop line to the tail of the last vehicle stopped in any traffic lane while the signal is
red during one signal cycle. Vehicle queue length is an important parameter to determine
traffic density in traffic lanes so that possible congestion in any lane can be minimized.
It is also important for traffic optimization on a signal in terms of prediction of traffic,
intersection delay and travel time etc [11]. Figure 1 shows a random vehicle queue at
a traffic intersection. Vehicle queue length information is generally extracted by using
fixed roadside sensors like loop detectors, pneumatic road tubes, radars and in-vehicle
sensors like GPS etc. The roadside sensors such as loop detectors are promising, however,
their installation and maintenance costs are comparatively higher. Their installation also
requires a lot of traffic disruption. Similarly, magnetic sensors can easily get affected by
temperature and humidity. Radar sensors are very accurate but their main disadvantage
is high exposure to electromagnetic interferences. Compared to these sensors, traffic
surveillance cameras can be used effectively to determine the presence and location of
vehicles, queue length, flow rate and occupancy across several traffic lanes etc.
In this research work, a state-of-the-art CNN-based vehicle detection model is used to
estimate the queue lengths of vehicles (particularly cars) in signalized traffic lanes. Using
the proposed model, a fairly accurate estimation of vehicle queues for dense urban traffic
can be made without any on-site camera calibration. The major contributions of the paper
are highlighted as follows:
• A detailed comparison of deep learning architectures including YOLOv3, YOLOv4,
YOLOv5, SSD, ResNet101, and InceptionV3 with inexpensive training dynamics for
vehicle detection.
• Proposed method for estimation of vehicle count-based queue length using images
obtained from low-resolution traffic surveillance cameras.
• Selection of an application specific dataset with a fixed far-field camera view for better
estimation of queue lengths while also considering the depth of a queue and camera
angles.
The remainder of the paper is organized as follows: Section 2 provides literature
review regarding queue length estimation. Section 3 describes the methodology adopted to
estimate the queue length. Section 4 provides the implementation of the proposed model
in detail. Section 5 discusses and demonstrates the effectiveness of the proposed method.
Finally, Section 6 concludes the paper and presents a future research direction related to
the proposed research.
Processes 2021, 9, 1786 3 of 19
2. Literature Review
Several works in the literature are available which present efficient models for queue
length estimation. A common approach is to develop a mathematical model using data
collected from roadside or in-vehicle sensors. Based on the methodology and type of
sensors used, the literature on queue length estimation and related tasks can be categorized
broadly into three groups.
• Estimation using Fixed Location Sensors. These include fixed sensors such as loop
detectors etc.
• Estimation using Mobile Sensors. These include sensors installed in vehicles such as
Global Positioning System (GPS) etc.
• Estimation using Vision based Sensors. These include processing of traffic camera feeds
using Computer vision tools.
Lee et al. determined the number of queued vehicles in each lane with the help of loop
detectors [12]. The simulations used Kalman filter to find traffic volume in each lane. Wu
et al. presented an approach for queue length detection using roadside LiDAR data [13].
The proposed approach achieved an average of 98% accuracy at various sites. Ki An et al.
devised a numerical model for queue length estimation at a metering roundabout [14]. The
framework used two roadside detectors and traffic signal for estimation of queue lengths.
The ground-truth queue length information was obtained from two different drones. The
ground truth was then used to calibrate and validate the model. The model performance
was analyzed using the R2 test and achieved an 83% value. Skabardonis et al. proposed an
intersection queue length estimation method using loop detector and traffic signal data
with 30 s intervals. Using this approach accurate results were reported for travel time
estimation [15].
Li et al. proposed a cycle-by-cycle queue length estimation approach using probe data
obtained from vehicles participating in traffic flow [16]. The results were benchmarked
compared to paper from Ramezani et al. [17] and reported improved accuracy under low
penetration rate. Blokpoel et al. compared different queue length estimation approaches
based on Cooperative Awareness Messages (CAM) received from vehicles [18]. The GPS
information from these vehicles is easily obtained through CAM instead of relying on
expensive equipment. The overall approach showed methods to improve queue length
detections with this obtained data. Wu et al. proposed queue length estimation based
on data obtained from RFID detectors [19]. Instead of relying on incoming traffic flow,
the authors exploited the queue delay of individual vehicles to measure queue length at
an intersection. Authors reported satisfactory results using this approach. Rompis et al.
performed lane identification and queue length estimation using probe vehicle data [20].
Processes 2021, 9, 1786 4 of 19
Using the optimal Bayes Rule model, promising results were obtained for lane identification
which was then used for queue length estimation.
Okaishi et al. proposed a real-time vehicular queue length measurement system for
intersections based on camera feeds [21]. The proposed system used frame differencing to
analyze motion in focused areas. When no motion is observed, vehicles in that area are
detected using CNN based SSD networks. The system works with video feeds obtained
from stationary cameras. Similarly, Zanin et al. proposed a system for real-time mea-
surement of vehicle queue parameters [22]. It is based on vehicle presence detection and
movement analysis in a set of videos acquired from stationary traffic cameras. The system
can then detect queues in lanes of interest. Albiol et al. presented an approach to estimate
traffic queue lengths in real time by collecting and evaluating low-level features such as
corners to detect the presence of vehicles [23]. After that, features are classified as either
moving or static. The algorithm only requires locations of lane masks and starting point
of queues etc. Shirazi et al. has also presented a tracking method for estimation of queue
lengths and waiting time of vehicles at junctions [11]. Background subtraction technique
was used to detect all the moving vehicles in the prescribed motion area. The vehicles
were continuously tracked until stopped. The total number of vehicles that cross the
motion area helps with estimation of queue length. Though the proposed system is unable
to detect queue lengths beyond this motion area. Li et al. collected traffic images from
custom-installed cameras, which constitute the training and testing sets for YOLO-based
models [24]. The training was performed using transfer learning with initial parameters
loaded from a pretrained model followed by fine-tuning on the obtained traffic images.
The authors reported network improvements by converting an 80-classes classifier into
a binary classifier. The final improved model achieved an average accuracy of 90% in
detecting vehicles from traffic flows of different densities.
In summary, there is limited research on deep learning and vision-based queue length
estimation at signalized intersections. The reviewed literature presents satisfactory results
on a single intersection and is not applicable on large network of intersections. Similarly,
vehicle detection and tracking in different conditions such as low-resolution traffic feeds,
varying illumination conditions, or loss of tracks for long waiting vehicles still need
enhancement.
3. Research Methodology
The overall proposed queue length estimation pipeline is shown in Figure 2. Details
of the methodology are provided in ensuing paragraphs.
• Vehicle Count-based Queue Length. It is based on calculating the number of vehicles. The
queue length is calculated by multiplying the detected vehicle count with an average
vehicle length of approximately 5 m. This estimation is done while keeping in view
the different type of vehicles such as cars, trucks, and buses etc. This method doesn’t
require camera calibration information.
Due to no requirement of camera calibration, vehicle count-based method is selected
for queue length estimation. The ground truth of queue length in each frame was es-
tablished by manually adding length of all the visible vehicles stopped at a signalized
intersection in each traffic lane. Since the location of vehicles is defined by bounding
boxes, pixel distance between the front and last stopped vehicle in the queue is therefore
used to detect the queue in a traffic lane under consideration. For easy representation,
detection of queue is shown as red line in Figure 3. As explained earlier, the queue length
is then estimated by multiplying the number of vehicles (in the detected queue) by 5 m.
Subsequently, queue length is measured for each traffic lane. Traffic lanes and the stop
line are labeled manually and are shown as green lines in Figure 3. Stop line is used as
the start of the queue in all traffic lanes. The vehicle that stops after crossing the stop
line is not considered in queue length estimation. Identification of stop vehicles on stop
lines is performed through Deep Simple Online and Realtime Tracking (Deep SORT) [25].
Deep SORT uses the appearance features to track objects in longer periods of occlusion. It
performs tracking by assigning unique tracking ID to each identified vehicle (as shown in
Figure 4a). The stopped vehicle in a traffic lane and at a stop line is identified by finding
the movement of that unique vehicle across consecutive frames. When any vehicle does
not show any motion in two consecutive frames, it is considered as a stopped vehicle.
Subsequently, queue is detected, and queue length is calculated (as shown in Figure 4b).
(a) (b)
Figure 4. (a) Unique tracking ID assignment and identification of stopped vehicles, (b) Estimation of
queue lengths after identification of stopped vehicles.
Video Frame
The inference was then performed on the images from Burlington traffic videos using
the same 80-classes pretrained YOLOv3, YOLOv4, and YOLOv5 models. While using
different YOLO models, vehicle detection performed well under varying camera angle.
The same is shown in Appendix A, Tables A3–A5. However, the pre-trained models failed
to recognize vehicles in the far end of traffic lanes. It also failed to recognize vehicles that
are stacked between queues of vehicles. Therefore, transfer learning was considered to
improve the vehicle detection accuracy for the YOLO models. Both YOLOv4 and YOLOv5
have similar performance (as shown in the tables mentioned above). Similarly, Deep SORT
implementation was also compatible with YOLOv4 implementation. Therefore, finally,
YOLOv4 was used to perform the experimentation.
Since the object detection predictions for only vehicle classes are being used in queue
length calculation, therefore, all vehicle classes were considered as one class during labeling
images for transfer learning. Instead of predicting 80-classes in the model, having a single
class object detection model simplifies the inference process as there is no need to hide other
classes in final output. It is also possible to perform transfer learning on five vehicle classes.
However, given the unbalanced image data (i.e., cars appear more often than trucks, buses,
or motorcycles in traffic images), training a single-class object-detection model is more
reasonable. As shown in Figure 7, significant improvement in vehicle detecting accuracy
can be achieved after applying transfer learning for a single-vehicle class to the original
object-detection model.
(a) (b)
Figure 7. Inference with (a) original pre-trained COCO model (cars and trucks marked as blue and
red colors respectively) and (b) improved model with transfer learning (vehicles marked as red).
• Maximum Queue Length Estimated at Each Second. All frames within a second are
inference, and we choose the longest predicted queue length among all frames
4. Implementation
4.1. Hardware and Software Configuration
The experimentation is performed on a laptop with core i5-6300hq CPU, 8 GB of RAM
and GTX960M GPU with 2 GB VRAM. The versions of Python and the software framework
used are Python 3.5.2, TensorFlow 2.2.0, and OpenCV 3.4.2. Training of YOLO model is
done using Google Colab computing platform with Intel Xeon 2 Cores CPU, 12 GB of RAM
and Nvidia Tesla K80 GPU with 12 GB of VRAM. The versions of Python and the software
framework used are Python 3.6.7 and TensorFlow 2.2.0 respectively.
Figure 9. Queue length performance evaluation results (with original pretrained 80-classes model).
Tables 2 and 3 show the improvement of queue length accuracies in terms of vehicle
count and vehicle count-based queue length estimation. Using a pre-trained 80-classes
YOLOv4 model, an overall accuracy of 73% and 88% was achieved for vehicle count and
vehicle count-based queue length estimation, respectively. After fine-tuning of model
and narrowing the output classes to vehicle class only, an average accuracy of 83% and
93% was achieved, respectively. The results for each video have been averaged across all
lanes and all queue length measuring methods. The greatest improvements are seen at
30% after applying fine-tuning implementation on sampled videos. Most of the videos
are significantly improved, except for the vehicle count accuracies in B20M video. The
improvements are result of enhanced vehicle detection in the traffic videos. Fine-tuning
Processes 2021, 9, 1786 12 of 19
enabled the YOLOv4 network to recognize vehicles which were small or had an appearance
which could easily blend into part of the road, as shown in Figure 10. Similarly, the
improved model also detects cars stacked in queues that were not picked up by the original
network, particularly in B34M and B30S videos. Confidence scores assigned to detected
vehicles are also observed to be much higher than before. Nevertheless, there is still room
for improvement as the vehicle count accuracies for the B20M video are much lower than
the other two videos. If video frames that have the same stacked vehicle density are used
in fine-tuning labels, the recognition result can improve.
(a) (b)
Figure 10. Difference of vehicle detection in Video B34M (a) before fine-tuning (cars and buses
marked as purple and green colors respectively) (b) after fine-tuning (vehicles marked as yellow)
Author Contributions: Conceptualization, B.A.; Data curation, Q.C.; Formal analysis, M.U.F.; Investi-
gation, Q.C.; Methodology, M.U.; Project administration, M.U.; Supervision, R.H.R.; Writing—original
draft, M.U.; Writing—review and editing, M.U.F. All authors have read and agreed to the published
version of the manuscript.
Funding: We acknowledge partial support from National Center of Big Data and Cloud Computing
(NCBC) and Higher Education Commission (HEC) of Pakistan for conducting this research.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: All the relevant data is publicly available to download and can be ac-
cessed at the following link: https://drive.google.com/drive/folders/1DI7-81g92__bdaJrSolHrCVUmBB8
6wKo?usp=sharing (accessed on 1 September 2021).
Conflicts of Interest: The authors declare no conflict of interest.
Appendix A
Table A1. Output of an original pre-trained YOLO model on selected traffic camera images obtained from Toronto Open
Data. Cars, buses and trucks marked as purple, green and red colors respectively.
Video Sample Original Image Model Output Terminal Output Incorrect Detections
Table A2. Comparison of vehicle detection models on selected traffic camera images obtained from Toronto Open Data.
Cars, buses and trucks marked as purple/cyan, green/beige and red colors respectively.
Video Sample YOLO Results SSD Results ResNet Results Inception Results
Loc8004w
Loc8011n
Loc8018e
Loc8037n
Loc8043w
Processes 2021, 9, 1786 15 of 19
Table A3. Vehicle detection accuracies on selected images obtained from Burlington video using YOLOv3. Cars and trucks marked as purple and red colors respectively.
Video Sample Original Image Model Output Ground Truth Vehicles Detected Accuracy
Burlington-WL-HV-SYNC-
9 5 55.60%
11090E0000000AB4-2015-01-08
Burlington-HV-SSR-SYNC-
5 3 60.0%
0A090E0000000A8D-2015-01-08
Burlington-WL-HV-
12 12 100.0%
0A090E0000000A87-2015-01-08
Burlington-HV-SSR-SYNC-
5 5 100.0%
0A090E0000000A8C-2015-01-08
Burlington-Queue-Mask3- new
26 20 76.90%
zones-11090E0000000AB6-2015-01-08
57 45 78.90% *
* Accuracy = (Total Ground Truth/Total Detected)×100%
Processes 2021, 9, 1786 16 of 19
Table A4. Vehicle detection accuracies on selected images obtained from Burlington video using YOLOv4. Cars and buses marked as red/green/blue/purple and light green colors
respectively.
Video Sample Original Image Model Output Ground Truth Vehicles Detected Accuracy
Burlington-WL-HV-SYNC-
9 7 77.78%
11090E0000000AB4-2015-01-08
Burlington-HV-SSR-SYNC-
5 4 80.0%
0A090E0000000A8D-2015-01-08
Burlington-WL-HV-
12 10 83.33%
0A090E0000000A87-2015-01-08
Burlington-HV-SSR-SYNC-
5 5 100.0%
0A090E0000000A8C-2015-01-08
Burlington-Queue-Mask3- new
26 18 69.23%
zones-11090E0000000AB6-2015-01-08
57 44 77.19% *
* Accuracy = (Total Ground Truth/Total Detected)×100%.
Processes 2021, 9, 1786 17 of 19
Table A5. Vehicle detection accuracies on selected images obtained from Burlington video using YOLOv5. Cars, buses and trucks marked as orange, neon green and light green colors
respectively.
Video Sample Original Image Model Output Ground Truth Vehicles Detected Accuracy
Burlington-WL-HV-SYNC-
9 7 77.78%
11090E0000000AB4-2015-01-08
Burlington-HV-SSR-SYNC-
5 4 80.0%
0A090E0000000A8D-2015-01-08
Burlington-WL-HV-
12 10 83.33%
0A090E0000000A87-2015-01-08
Burlington-HV-SSR-SYNC-
5 5 100.0%
0A090E0000000A8C-2015-01-08
Burlington-Queue-Mask3- new
26 22 84.61%
zones-11090E0000000AB6-2015-01-08
57 48 85.14% *
* Accuracy = (Total Ground Truth/Total Detected)×100%.
Processes 2021, 9, 1786 18 of 19
References
1. Urbanization. Available online: https://www.unfpa.org/urbanization (accessed on 1 September 2021).
2. Directive 2010/40/EU of the European Parliament and of the Council of 7 July 2010. Available online: https://eur-lex.europa.eu/
LexUriServ/LexUriServ.do?uri=OJ:L:2010:207:0001:0013:EN:PDF (accessed on 1 September 2021).
3. Makino, H.; Tamada, K.; Sakai, K.; Kamijo, S. Solutions for Urban Traffic Issues by ITS Technologies. IATSS Res. 2018, 42, 49–60.
[CrossRef]
4. Shafiq, M.; Tian, Z.; Bashir, A.K.; Jolfaei, A.; Yu, X. Data Mining and Machine Learning Methods for Sustainable Smart Cities
Traffic Classification: A Survey. Sustain. Cities Soc. 2020, 60, 102177. [CrossRef]
5. Lan, C.-L.; Chang, G.-L. Empirical Observations and Formulations of Tri-Class Traffic Flow Properties for Design of Traffic Signals.
IEEE Trans. Intell. Transp. Syst. 2019, 20, 830–842. [CrossRef]
6. Chen, X.; Li, Z.; Yang, Y.; Qi, L.; Ke, R. High-Resolution Vehicle Trajectory Extraction and Denoising from Aerial Videos. IEEE
Trans. Intell. Transp. Syst. 2021, 22, 3190–3202. [CrossRef]
7. Huang, Y.-Q.; Zheng, J.-C.; Sun, S.-D.; Yang, C.-F.; Liu, J. Optimized YOLOv3 Algorithm and Its Application in Traffic Flow
Detections. Appl. Sci. 2020, 10, 3079. [CrossRef]
8. Li, J.-Q.; Zhou, K.; Shladover, S.E.; Skabardonis, A. Estimating Queue Length under Connected Vehicle Technology. Transp. Res.
Rec. J. Transp. Res. Board 2013, 2356, 17–22. [CrossRef]
9. Kalamaras, I.; Zamichos, A.; Salamanis, A.; Drosou, A.; Kehagias, D.D.; Margaritis, G.; Papadopoulos, S.; Tzovaras, D. An
Interactive Visual Analytics Platform for Smart Intelligent Transportation Systems Management. IEEE Trans. Intell. Transp. Syst.
2018, 19, 487–496. [CrossRef]
10. Lu, N.; Cheng, N.; Zhang, N.; Shen, X.; Mark, J.W. Connected Vehicles: Solutions and Challenges. IEEE Internet Things J. 2014, 1,
289–299. [CrossRef]
11. Shirazi, M.S.; Morris, B. Vision-Based Vehicle Queue Analysis at Junctions. In Proceedings of the 2015 12th IEEE International
Conference on Advanced Video and Signal Based Surveillance (AVSS), Karlsruhe, Germany, 25–28 August 2015. [CrossRef]
12. Lee, S.; Wong, S.C.; Li, Y.C. Real-Time Estimation of Lane-Based Queue Lengths at Isolated Signalized Junctions. Transp. Res. Part
C Emerg. Technol. 2015, 56, 1–17. [CrossRef]
13. Wu, J.; Xu, H.; Zhang, Y.; Tian, Y.; Song, X. Real-Time Queue Length Detection with Roadside LiDAR Data. Sensors 2020, 20, 2342.
[CrossRef]
14. An, H.K.; Yue, W.L.; Stazic, B. Estimation of Vehicle Queuing Lengths at Metering Roundabouts. J. Traffic Transp. Eng. 2017, 4,
545–554. [CrossRef]
15. Skabardonis, A.; Geroliminis, N. Real-Time Monitoring and Control on Signalized Arterials. J. Intell. Transp. Syst. 2008, 12(2),
64–74. [CrossRef]
16. Li, F.; Tang, K.; Yao, J.; Li, K. Real-Time Queue Length Estimation for Signalized Intersections Using Vehicle Trajectory Data.
Transp. Res. Rec. J. Transp. Res. Board 2017, 2623, 49–59. [CrossRef]
17. Ramezani, M. , and Geroliminis, N. . Queue Profile Estimation in Congested Urban Networks with Probe Data. Comput.-Aided
Civ. Infrastruct. Eng. 2015, 30, 6, 414–432. [CrossRef]
18. Blokpoel, R.; Vreeswijk, J. Uses of Probe Vehicle Data in Traffic Light Control. Transp. Res. Procedia 2016, 14, 4572–4581. [CrossRef]
19. Wu, A.; Yang, X. Real-Time Queue Length Estimation of Signalized Intersections Based on RFID Data. Procedia Soc. Behav. Sci.
2013, 96, 1477–1484. [CrossRef]
20. Rompis, S.Y.R.; Cetin, M.; Habtemichael, F. Probe Vehicle Lane Identification for Queue Length Estimation at Intersections. J.
Intell. Transp. Syst. 2017, 22, 10–25.
21. Okaishi, W.A.; Zaarane, A.; Slimani, I.; Atouf, I.; Benrabh, M. A Vehicular Queue Length Measurement System in Real-Time
Based on SSD Network. Transp. Telecommun. J. 2021, 22, 29–38. [CrossRef]
22. Zanin, M.; Messelodi, S.; Modena, C.M. An Efficient Vehicle Queue Detection System Based on Image Processing. Available online:
https://ieeexplore.ieee.org/document/1234055 (accessed on 26 July 2020).
23. Albiol, A.; Albiol, A.; Mossi, J.M. Video-Based Traffic Queue Length Estimation. In Proceedings of the 2011 IEEE International
Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011. [CrossRef]
24. Li, X.; Liu, Y.; Zhao, Z.; Zhang, Y.; He, L. A Deep Learning Approach of Vehicle Multitarget Detection from Traffic Video. J. Adv.
Transp. 2018, 2018, 7075814. [CrossRef]
25. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017
IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; doi: 10.1109/ICIP.2017.8296962.
[CrossRef]
26. Open Data Dataset. Available online: https://open.toronto.ca/dataset/traffic-cameras/ (accessed on 1 September 2021).
27. LabelImg. Available online: https://github.com/tzutalin/labelImg (accessed on 1 September 2021).
28. Huang, L.; Yang, Y.; Deng, Y.; Yu, Y. DenseBox: Unifying Landmark Localization with End to End Object Detection. arXiv 2015,
arXiv:1509.04874.
Processes 2021, 9, 1786 19 of 19
29. Huang, J.; Rathod, V.; Sun, C.; Zhu, M.; Korattikara, A.; Fathi, A.; Fischer, I.; Wojna, Z.; Song, Y.; Guadarrama, S.; et al.
Speed/Accuracy Trade-Offs for Modern Convolutional Object Detectors. In Proceedings of the 2017 IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [CrossRef]
30. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016.
31. Fachrie, M. A Simple Vehicle Counting System Using Deep Learning with YOLOv3 Model. J. Resti (Rekayasa Sist. Dan Teknol. Inf.)
2020, 4, 462–468. [CrossRef]
32. Song, H.; Liang, H.; Li, H.; Dai, Z.; Yun, X. Vision-Based Vehicle Detection and Counting System Using Deep Learning in Highway
Scenes. Eur. Transp. Res. Rev. 2019, 11, 51. [CrossRef]
33. Alghyaline, S.; El-Omari, N.; Al-Khatib, R.M.; Al-Kharbshh, H. RT-VC: An Efficient Real-Time Vehicle Counting Approach.
J. Theor. Appl. Inf. Technol. 2019, 97, 2062–2075.
34. Ding, X.; Yang, R. Vehicle and Parking Space Detection Based on Improved YOLO Network Model. In Journal of Physics: Conference
Series 2019; IOP Publishing: Bristol, UK, 2019; Volume 1325, p. 012084. [CrossRef]