Drowsiness Detection 2
Drowsiness Detection 2
Drowsiness Detection 2
TRANSPORTATION SYSTEMS
Received November 26, 2019, accepted December 5, 2019, date of publication December 10, 2019,
date of current version December 23, 2019.
Digital Object Identifier 10.1109/ACCESS.2019.2958667
ABSTRACT The research work about driving drowsiness detection algorithm has great significance to
improve traffic safety. Presently, there are many fruits and literature about driving drowsiness detection
method. However, most of them are devoted to find a universal drowsiness detection method, while ignore
the individual driver differences. This paper proposes a real-time driving drowsiness detection algorithm that
considers the individual differences of driver. A deep cascaded convolutional neural network was constructed
to detect the face region, which avoids the problem of poor accuracy caused by artificial feature extraction.
Based on the Dlib toolkit, the landmarks of frontal driver facial in a frame are found. According to the
eyes landmarks, a new parameter, called Eyes Aspect Ratio, is introduced to evaluate the drowsiness of
driver in the current frame. Taking into account differences in size of driver’s eyes, the proposed algorithm
consists of two modules: offline training and online monitoring. In the first module, a unique fatigue state
classifier, based on Support Vector Machines, was trained which taking the Eyes Aspect Ratio as input. Then,
in the second module, the trained classifier is application to monitor the state of driver online. Because the
fatigue driving state is gradually produced, a variable which calculated by number of drowsy frames per unit
time is introduced to assess the drowsiness of driver. Through comparative experiments, we demonstrate
this algorithm outperforms current driving drowsiness detection approaches in both accuracy and speed.
In simulated driving applications, the proposed algorithm detects the drowsy state of driver quickly from
640∗480 resolution images at over 20fps and 94.80% accuracy. The research result can serve intelligent
transportation system, ensure driver safety and reduce the losses caused by drowsy driving.
INDEX TERMS Traffic safety, driving drowsiness detection, CNN, individual differences, SVM.
179396 This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see http://creativecommons.org/licenses/by/4.0/ VOLUME 7, 2019
F. You et al.: Real-time Driving Drowsiness Detection Algorithm With Individual Differences Consideration
Safety Law of China stipulates that the driver must not run a algorithm based on deep learning technology while ignore the
vehicle continuously for more than 4 hours [5]. By measur- specific application sometimes.
ing driving time to judge the state of drivers, it can reduce Percentage of Eyelid Closure Over the Pupil Over
the traffic accidents caused by driving drowsiness to some Time(PERCLOS) [22] is one of the most popular parame-
extent. However, this method cannot detect whether a driver is ter, which applies to driving drowsiness detection based on
fatigued or not in real-time. With the development of informa- computer vision. It firstly assesses whether the driver’s eyes
tion technology, detection system for driving drowsiness has are open or closed in the current frame according to the
become an alternative means to solve the problem. Therefore, proportion of eyelid covering the pupil. Then, PERCLOS
research on the intelligent identification of drowsy driving is calculated by the number of eyes closed frames over a
has important realistic meanings. pried of time. As we all know, the characteristics of human
In recent years, thanks to the non-invasive and low cost, facial, especially the size of eyes, have certain differences,
methods of driving drowsiness detection based on driver which introduces a new variable when it comes to PERCLOS.
behavior have become a research hotspot. However, the per- In the past literatures [23], PERCLOS-80(P80) was used to
formance of algorithms might be limited by the technolo- judge whether the eyes of driver are open or closed. That
gies of face detection and fatigue assessing in complex and is, when the proportion of eyelid covering the pupil is over
changing surroundings [6], [7]. Reference [8] uses a mask 80 percent, the eyes are identified as closed in the current
to obtain the face of driver and evaluate the driving state frame. However, this method does not take into account the
with the application of PERCLOS. Experiments show that the individual differences of driver, especially the differences of
performance of the method is well good in ideal conditions. the size of eyes, which may cause misjudgment in practical
But the generalization performance of the method is affected applications.
by the fabrication of mask. The cascade face detector pro- In this paper, we propose a new algorithm to detect driv-
posed by Viola and Jones [9] utilizes Haar-like feature and ing drowsiness, which considers the individual differences
Adaboost to train cascaded classifier, which achieve good of the drivers. It consists of off-line training module and
performance in face detection. [10] extracts the multi-scale online monitoring module. Firstly, we design a deep cas-
feature of face using Gabor wavelet transform, and trains an caded convolutional neural network (DCCNN) to detect the
Adaboost cascaded classifier to select a most recognizable face region from live video. Based on this, the eyes land-
features for driver drowsy state discrimination. However, marks are obtained by the application of Dlib toolkit. A new
quite a few works [11]–[13] indicate that the method based parameter, Eyes Aspect Ratio (EAR), which calculated by
on Adaboost may degrade significantly in read-world appli- the coordinates of landmarks, is introduced to identify the
cation with larger visual variations. Reference [14] uses an state of eyes (open or closed). In the offline training module,
active near-infrared light source to obtain the stable image for a specific driver, there are two sets of the EAR, which
of eyes and tracks it using finite state machine. The fuzzy represent eyes-open and eyes-closed, are obtained from the
system is applied to evaluate the state of driver. However, face detected by DCCNN. And a unique support vector
the method requires the installation of more sophisticated machine (SVM) classifier is constructed which takes the two
sensors, which increases the cost. With the development sets of data into the input. In the online monitoring module,
of artificial intelligence technology, the capable of com- the driver’s face is detected by DCCNN from live video
puting has been greatly improved. The driving drowsiness firstly. Then the landmarks of driver’s face are obtained with
detection algorithms based on deep learning are increasingly the application of the Dlib toolkit as well. The Eyes Aspect
attracting attention. Reference [15] proposed a fatigue detec- Ratio can be calculated by the coordinates of the eyes. Finally,
tion algorithm based on the Convolutional Neural Network the state of eyes in the current frame can be classified by
(CNN). The algorithm classified human eyes and non-human the unique SVM. Thanks to this offline training and online
eyes by training the first network, detected the position of monitoring module, the performance of the algorithm can be
the eye feature points with the second network, calculated notably improved.
the eye-opening degree according to the feature point posi- The major contributions of this paper are summarized as
tion, and judged the fatigue state of driver by PERCLOS. follows:
Reference [16] introduced a detection method based on facial 1. We design a new deep cascaded convolutional neural
behavior analysis. It combined with Adaboost and kernel cor- network to detect the face of a driver, which effectively
relation filter (KCF) for face detection and tracking in facial improve the performance.
image captured from infrared acquisition device. The cascad- 2. We introduce a new parameter based on the Dlib
ing regression is exploited to locate the feature points and toolkit to assess whether the eyes of a driver are open or
extracted the eye and mouth regions. Also, the convolutional closed.
neural network (CNN) is employed to recognize the status of 3. Extensive experiments are conducted, to show the EAR
eyes and mouth. However, the deep learning methods are still is work in different drivers and the significant performance
in the exploration stage for the driving drowsiness detection. improved of the proposed approach in both accuracy and
A lot of works [17]–[21] were focus on exploring a universal speed.
4. Comparative experiments result show that the proposed Finally, take the two types of the EAR as positive and negative
algorithm, which with consideration of individual differ- samples, we training a unique SVM classifier for the specific
ences, is reasonable and more accurate. driver to judge whether the eyes are open or closed.
The structure of this paper is organized as follows. Online Monitoring: It is a real-time module to detect
In Section II, we describe the proposed methods in detail, driving drowsiness form live video. All frames are fed to
including the deep learning model for face detection and DCCNN during driving to find the face of a driver. If the
facial landmarks obtaining. The core approaches, offline face is captured, the eyes landmarks can be obtained by the
training and online monitoring, are introduced in this section Dlib toolkit. Then, the unique SVM classifier of the specific
as well. In Section III, we conduct extensive experiments driver, which trained in offline module, is applied to judge
to verify the rationality of the proposed algorithm in both whether the eyes are open or closed in the current frame.
accuracy and speed. The conclusion is in Section IV. At last, we assess whether the driver is sleepy or not according
to the ratio of the number of sleepy frames and total frames in
II. APPROACH a period of time. Besides, if DCCNN does not detect a face,
The overall pipeline of our approach is shown in Figure 1. we judge the driver is drowsy in the current frame maybe
The algorithm consists of the following two modules: because of the improper head posture.
Offline Training: It is equivalent to an initialization
process. As usually seen, the characteristics of a driver’s face, A. DEEP CASCADED CONVOLUTIONAL
especially the size of the eyes, have obvious differences. NEURAL NETWORK
Thus, we firstly obtain two sets of data by asking the driver Face detection is one of the most important technologies
to open his/her eyes and close his/her eyes for a while on a for driving drowsiness detection based on computer vision.
simulator. For each image data, applying the deep cascaded In practical application, driving drowsy detection system
convolutional neural network we designed called DCCNN to not only requires high accuracy, but also fast speed. As we
detect the face of drive in the current frame. Further, based know, deep learning methods, especially the convolutional
on the Dlib toolkit, facial landmarks can be obtained. Then, neural network model, greatly improve the accuracy of image
two types of EAR, as shown in Figure 1, the EAR1 relates recognition. However, the complex network structure reduces
the eye-open frames and the EAR2 is the value calculated the algorithm speed. In [24], a Multi-Task Cascaded Con-
when the state of driver’s eyes are closed, can be calculated. volutional Networks called MTCNN have been designed
FIGURE 2. The architecture of DCCNN, where ‘‘conv’’ means convolutional layer, ‘‘MP’’ means max pooling layer and ‘‘fc’’ means fully
connected layer.
for face detection. The architecture consists of three sub- of the three sub-networks is cascaded. In DCCNN, the output
networks, instead of a complex network. Each sub-network of the previous sub-network acts as the input to the next sub-
has less numbers of filters but more discrimination of them, network. By building the image pyramid, the face of a driver
which effectively improves the speed of the algorithm. can be detected accurate no matter how larger proportion of
However, we noticed its performance might be limited by the the face in frames.
useless information about the five facial landmarks. Because In the training phase of DCCNN, the WIDER_FACE data
we just pay attention to the information of the facial in set [27] and the AFLW data set [28] are used as the training
video frames. Thus, we design a new convolutional neural data. Among them, the WIDER_FACE data set includes more
network to detect the face of a driver. It is more efficient than 30,000 pictures and 400,000 personal faces. It is the
thanks to the lightweight architecture and removing five largest and most complex face detection public data set in
facial landmarks. The architecture of Deep Cascaded Con- the world. The AFLW face database includes 25,000 hand-
volutional Neural Network (DCCNN) is shown in Figure 2. labeled face images. The database is widely used in face
In practical application, the proportion of the face in frames recognition, face detection, face alignment and other aspects.
is uncertain. Thus, it is an effective approach to resize the In this paper, the WIDER_FACE and AFLW datasets are
original image into different scales. Similar to MTCNN, used to crop some face images and non-face images based
we build an image pyramid by resizing the original image on manually labeled face regions. A total of 190,000 face
to different scales, which is the input for three cascaded images, 600,000 partial face images and 900,000 non-face
networks. Network-1 is a fully convolutional network [25], images are obtained.
to obtain a large number of candidate windows and their The procedure of the training of the DCCNN is aiming to
bounding box regression vectors. Then we use the estimated get the optimal model by adjusting the parameters dynami-
bounding box regression vectors to calibrate the candidates. cally. To express the influence of different parameters on the
Finally, non-maximum suppression(NMS) [26] is employed performance of the DCCNN, a loss function is introduced
to merge highly overlapped candidates. After Network-1, All during training. The loss function is an index to measure
candidates are fed to the next CNN, called network-2, which the difference between the predictive output and the actually
further reject a larger number of false candidates, performs marked label, and there are various loss functions according
calibration with bounding box regression, and NMS candi- to different tasks. The training process includes two tasks,
date merge. Network-3 is a CNN similar to network-2, but respectively, face and non-face classification task and face
it has more convolutional layers and pooling layers so that it region box bounding fitting task.
can describe the face in detail. The output of network-3 is the The first task, face and non-face is a classification task, so
facial bounding box with high confidence. The relationship we apply the cross entropy [29] loss function for the training.
the EAR of different states are approximate. In contrast, TABLE 2. Driver fatigue detection accuracy comparison.
the EAR of object 4 in two types of state are significant
different.
Therefore, the correlation of the EAR and the size of eyes is
strong as we assumed above. Generally, the lager the driver’s
eyes are, the higher the EAR is. And the smaller the eyes are,
the lower the EAR is. It is necessary to develop the driving
drowsiness detection model with consideration of driver’s
individual differences, especially the size of eyes, to improve
the performance of algorithm.
D. COMPARATIVE EXPERIMENTS
In order to verify the rationality of the driving drowsiness
detection algorithm with consideration of individual differ-
ences, a series of comparative experiments are conducted.
Firstly, we compared the accuracy on driving fatigue detec-
tion of using the hard threshold (0.8) and the method proposed
in this paper. Then, we ask a driver to simulate the normal
driving and drowsy driving state on the simulator, so that the
change of eyes state and PERCLOS is visible clearly. Finally,
we conduct another comparative experiment to compare the
performance of our methods and relevant literatures both on
accuracy and speed. For fair comparison, we use the same
data and platform for both methods.
The comparative experiment is repeated 10 times.
1) COMPARE WITH P80 Table 2 shows the accuracy of driving drowsiness detection.
We select 10 segments of simulated driving video randomly It can be seen from the table that the algorithm we proposed
from the SDVD data set. The DCCNN is applied to detect improves the accuracy effectively.
the face of driver and then obtain the landmarks of face using
the Dlib toolkit. On this basis, a comparative experiment was 2) STATE OF EYES AND PERCLOS IN DIFFERENT
conducted. DRIVING CONDITION
Experiment 1: Calculate the EAR according to the land- We have description the correlation of EAR and the size of
marks of eyes. P80 is used to judge whether the eyes of driver eyes with four types of driver. In order to visually evaluate
are open or closed. That is, the eyes of driver are assessed to the changes of driver’s eyes state and the PERCLOS values
be closed if EAR < 0.2. Then, the driving state of driver is under different driving conditions. We ask a driver (size: 1) to
determined by the ratio of closed frames for a specific time simulate two different driving states. As shown in Figure 14.
(100 frames). It can be seen from Figure.14 that when the driver is in
Experiment 2: Initializing the system firstly by training a normal driving state, the duration of the eyes opening is much
unique SVM classifier with the input of the two sets of EAR longer than the time of closure. And the values of PERCLOS
value. Next, the driving state is assessed by using the online are all below the threshold (0.4). Conversely, the duration of
monitoring model proposed in this paper. closure is longer than the time of opening when the driver is in
As mentioned above, for safety, we conduct the com- drowsy state. Most of the PERCLOS are over the threshold.
parative experiments on the simulator. The SDVD data set
contains the driving video with labeled information. For a
sample, the driving state has been marked in advance as 3) COMPARE WITH RELEVANT LITERATURES
shown in Figure 8. The comparative experiments will detect Figure 15 shows the calculation speed of each module of
the driving state and record the period of awake and drowsy. the algorithm during the experiment of the four subjects.
If the result obtained by the driving drowsiness model is The experimental video stream is 640∗480 resolution and the
consistent with the labeled information, it is considered that frame rate is 30fps. It can be seen from the figure that the
the model gets a correct evaluation of driving state. Therefore, average speed of the face detection module is 35.95ms/f,
the accuracy can be computed by: and the average speed of the landmarks obtaining module is
13.80 ms/f. The average speed of the fatigue driving detection
Nc algorithm we proposed is 49.75 ms/f.
accuracy = (11)
N In order to furtherly verify the performance of the proposed
where Nc is the number of video clips correctly evaluated and algorithm, we use the self-built data set, SDVD, to compare
N is the total number of experiment samples. our algorithm with the drowsy detection algorithms [40] [41]
FIGURE 14. Eyes state and PERCLOS under different driving conditions.
IV. CONCLUSION
Research on drowsy driving detection algorithm is one of
the most important methods to reduce traffic accidents.
As we know, there are significant individual differences,
especially the size of eyes, between different person. It is
crucial to take the individual differences into consideration
when study on the algorithm based on computer vision.
In this paper, we propose a new driving drowsiness detec-
tion algorithm with consideration of individual differences.
Firstly, we design a deep cascaded convolutional neural
network model named DCCNN, which avoids the process
of artificial feature extraction in traditional face detection
algorithms, to obtain the face of a driver in live video. The
performance of the model is tested by qualitative description
and quantitative evaluation. Experimental results show that
FIGURE 15. Algorithm speed. the accuracy of face detection can reach at 98.8%. Sec-
ondly, we propose a new parameter, EAR, based on the Dlib
TABLE 3. Comparison. toolkit, to assess the state of driver’s eyes. Compared with
the traditional methods, the EAR is more stable thanks to the
Cascaded Pose Regression algorithm. Experimental results
show that there is a very strong correlation between the EAR
and the size of a driver’s eyes, which proves the rationality
of our ideas. Finally, taking the individual differences of the
drivers into consideration, we construct the offline training
module and online monitoring module in the paper. A unique
classifier based on SVM is trained for a specific driver and the
state of eyes is judged with the application of the pre-trained
proposed in recent years. The detection accuracies and speeds classifier during driving. Experimental results demonstrate
are shown in Table 3. that our methods consistently outperform the state-of-the-art
methods on the public data set and self-built data set while [17] Z. Xiao, Z. Hu, L. Geng, F. Zhang, J. Wu, and Y. Li, ‘‘Fatigue driving
keeping real-time performance. recognition network: Fatigue driving recognition via convolutional neural
network and long short-term memory units,’’ IET Intell. Transp. Syst.,
In the future, we will focus on the following topic: vol. 13, no. 9, pp. 1410–1416, Sep. 2019.
1) explore the multi-feature such as mouth and head pos- [18] W. Liu, ‘‘Convolutional two-stream network using multi-facial feature
ture fusion methods to further improve the algorithm perfor- fusion for driver fatigue detection,’’ Future Internet, vol. 11, no. 5, p. 115,
2019.
mance. 2) conduct the driving drowsiness detection research [19] Z. Ning, P. Dong, X. Wang, J. Rodrigues, and F. Xia, ‘‘Deep reinforcement
at nighttime because it is easier to drowsy driving at night. learning for vehicular edge computing: An intelligent offloading system,’’
ACM Trans. Intell. Syst. Technol., vol. 10, no. 6, 2019, Art. no. 60.
[20] X. Wang and C. Xu, ‘‘Driver drowsiness detection based on non-intrusive
ACKNOWLEDGMENT metrics considering individual specifics,’’ Accident Anal. Prevention,
This work was partially support by "Guangdong Natural vol. 95, pp. 350–357, Oct. 2016.
Science Foundation 2020- Front Vehicles Perception and [21] L. Zhang, F. Liu, and J. Tang, ‘‘Real-time system for driver fatigue detec-
tion by RGB-D Camera,’’ ACM Trans. Intell. Syst. Technol. (TIST), vol. 6,
Cooperative Control of Safety Distance for Driverless Vehicle no. 2, pp. 1–17, 2015.
at Nighttime". [22] R. Fu, H. Wang, and W. Zhao, ‘‘Dynamic driver fatigue detection using hid-
den Markov model in real driving condition,’’ Expert Syst. Appl., vol. 63,
pp. 397–411, Nov. 2016.
REFERENCES [23] Y. Jiang, S. Guo, and S. Deng, ‘‘Denoising and chaotic feature extrac-
[1] A. Amodio, M. Ermidoro, D. Maggi, S. Formentin, and S. M. Savaresi, tion of electrocardial signals for driver fatigue detection by kolmogorov
‘‘Automatic detection of driver impairment based on pupillary light entropy,’’ J. Dyn. Syst., Meas. Control, Trans. ASME, vol. 141, no. 2, 2019,
reflex,’’ IEEE Trans. Intell. Transp. Syst., vol. 20, no. 8, pp. 3038–3048, Art. no. 0210131, doi: 10.1115/1.4041355.
Aug. 2019, doi: 10.1109/TITS.2018.2871262. [24] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao, ‘‘Joint face detection and alignment
[2] X. Li, X. Lian, and F. Liu, ‘‘Rear-end road crash characteristics analysis using multitask cascaded convolutional networks,’’ IEEE Signal Process.
based on Chinese in-depth crash study data,’’ in Proc. Cota Int. Conf. Lett., vol. 23, no. 10, pp. 1499–1503, Oct. 2016.
Transp. Prof., 2016, pp. 1536–1545. [25] Z. Ning, J. Huang, X. Wang, J. Rodrigues, and L. Guo, ‘‘Mobile edge
[3] G. Zhang, K. K. Yau, X. Zhang, and Y. Li, ‘‘Traffic accidents involving computing-enabled Internet of Vehicles: Toward energy-efficient schedul-
fatigue driving and their extent of casualties,’’ Accident Anal., Prevent., ing,’’ IEEE Netw., vol. 33, no. 5, pp. 198–205, Sep./Oct. 2019.
vol. 87, pp. 34–42, Feb. 2016, doi: 10.1016/j.aap.2015.10.033. [26] H. Xiong, X. Zhu, and R. Zhang, ‘‘Energy recovery strategy numeri-
[4] D. Mollicone, K. Kan, C. Mott, R. Bartels, S. Bruneau, M. van Wollen, cal simulation for dual axle drive pure electric vehicle based on motor
A. R. Sparrow, and H. P. A. Van Dongen, ‘‘Predicting performance and loss model and big data calculation,’’ Complexity, vol. 2018, Aug. 2018,
safety based on driver fatigue,’’ Accident Anal., Prevention, vol. 126, Art. no. 4071743, doi: 10.1155/2018/4071743.
pp. 142–145, May 2019, doi: 10.1016/j.aap.2018.03.004. [27] S. Yang, P. Luo, C. C. Loy, and X. Tang, ‘‘WIDER FACE: A face detec-
[5] X. Sun, H. Zhang, W. Meng, R. Zhang, K. Li, and T. Peng, ‘‘Primary reso- tion benchmark,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,
nance analysis and vibration suppression for the harmonically excited non- Jun. 2016, pp. 5525–5533, doi: 10.1109/CVPR.2016.596.
linear suspension system using a pair of symmetric viscoelastic buffers,’’ [28] M. Köstinger, P. Wohlhart, P. M. Roth, H. Bischof, ‘‘Annotated facial land-
Nonlinear Dyn., vol.94, no. 2, pp. 1243–1265, 2018, doi: 10.1007/s11071 marks in the wild: A large-scale, real-world database for facial landmark
-018-4421-9. localization,’’ in Proc. IEEE Int. Conf. Comput. Vis. Workshops, Nov. 2011,
[6] Z. Ning, P. Dong, X. Wang, M. S. Obaidat, X. Hu, L. Guo, Y. Guo, J. Huang, pp. 2144–2151, doi: 10.1109/ICCVW.2011.6130513.
B. Hu, and Y. Li, ‘‘When deep reinforcement learning meets 5G vehicular [29] Z. Ning, J. Huang, and X. Wang, ‘‘Vehicular fog computing: Enabling
networks: A distributed offloading framework for traffic big data,’’ IEEE real-time traffic management for smart cities,’’ IEEE Wireless Commun.,
Trans. Ind. Informat., to be published, doi: 10.1109/TII.2019.2937079. vol. 26, no. 1, pp. 87–93, Feb. 2019.
[7] F. You, Y.-H. Li, L. Huang, K. Chen, R.-H. Zhang, and J.-M. Xu, ‘‘Moni- [30] P. Dollár, P. Welinder, and P. Perona, ‘‘Cascaded pose regression,’’ in
toring drivers’ sleepy status at night based on machine vision,’’ Multimed. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., Jun. 2010,
Tools Appl., vol. 76, no. 13, pp. 14869–14886, 2017, doi: 10.1007/s11042 pp. 1078–1085, doi: 10.1109/CVPR.2010.5540094.
-016-4103-x. [31] D. E. King, ‘‘Dlib-ml: A machine learning toolkit,’’ J. Mach. Learn. Res.,
[8] G. Sikander and S. Anwar, ‘‘Driver fatigue detection systems: A review,’’ vol. 10, pp. 1755–1758, Jan. 2009.
IEEE Trans. Intell. Transp. Syst., vol. 20, no. 6, pp. 2339–2352, Jun. 2019, [32] A. Tolba, ‘‘Content accessibility preference approach for improving
doi: 10.1109/TITS.2018.2868499. service optimality in Internet of Vehicles,’’ Comput. Netw., vol. 152,
[9] X. Li, L. Hong, J. Wang, and X. Liu, ‘‘Fatigue driving detection model pp. 78–86, Apr. 2019.
based on multi-feature fusion and semi-supervised active learning,’’ IET [33] X. Kong, F. Xia, J. Li, M. Hou, M. Li, and Y. Xiang, ‘‘A shared
Intell. Transp. Syst., vol. 13, no. 9, pp. 1401–1409, Sep. 2019. bus profiling scheme for smart cities based on heterogeneous mobile
[10] G. Niu and C. Wang, ‘‘Driver fatigue features extraction,’’ Math. Problems crowdsourced data,’’ IEEE Trans. Ind. Informat., to be published, doi:
Eng., vol. 2014, pp. 1–10, Jun. 2014. 10.1109/TII.2019.2947063.
[11] Z. Ning, Y. Feng, M. Collotta, X. Kong, X. Wang, L. Guo, X. Hu, [34] J. Xu, W. Zeng, Y. Lan, J. Guo, and X. Cheng, ‘‘Modeling the parameter
and B. Hu, ‘‘Deep learning in edge of vehicles: Exploring trirelation- interactions in ranking SVM with low-rank approximation,’’ IEEE Trans.
ship for data transmission,’’ IEEE Trans. Ind. Informat., vol. 15, no. 10, Knowl. Data Eng., vol. 31, no. 6, pp. 1181–1193, Jun. 2019.
pp. 5737–5746, Oct. 2019. [35] A. Tolba, ‘‘Trust-based distributed authentication method for collision
[12] M. Khan and S. Lee, ‘‘A comprehensive survey of driving monitoring and attack avoidance in VANETs,’’ IEEE Access, vol. 6, pp. 62747–62755,
assistance systems,’’ Sensors, vol. 19, no. 11, p. 2574, 2019. 2018.
[13] A. S. Zandi, A. Quddus, L. Prest, and F. J. E. Comeau, ‘‘Non-intrusive [36] R.-H. Zhang, Z.-C. He, H.-W. Wang, F. You, and K.-N. Li, ‘‘Study on
detection of drowsy driving based on eye tracking data,’’ Transp. Res. Rec., self-tuning tyre friction control for developing main-servo loop integrated
J. Transp. Res. Board, vol. 2673, no. 6, pp. 247–257, May 2019. chassis control system,’’ IEEE Access, vol. 5, pp. 6649–6660, 2017.
[14] L. M. Bergasa, J. Nuevo, M. A. Sotelo, R. Barea, and M. E. Lopez, ‘‘Real- [37] X. Sun, P. Wu, and S. C. H. Hoi, ‘‘Face detection using deep learn-
time system for monitoring driver vigilance,’’ IEEE Trans. Intell. Transp. ing: An improved faster RCNN approach,’’ Neurocomputing, vol. 299,
Syst., vol. 7, no. 1, pp. 63–77, Mar. 2006. pp. 42–50, Jul. 2018.
[15] X. Zhao, ‘‘Eye feature point detection based on single convolutional [38] X. Kong, X. Liu, B. Jedari, M. Li, L. Wan, and F. Xia, ‘‘Mobile crowdsourc-
neural network,’’ IET Comput. Vis., vol. 12, no. 4, pp. 453–457, ing in smart cities: Technologies, applications, and future challenges,’’
2018. IEEE Internet Things J., vol. 6, no. 5, pp. 8095–8113, Oct. 2019.
[16] F. Zhang, J. Su, L. Geng, and Z. Xiao, ‘‘Driver fatigue detection based [39] A. Tolba, O. Said, and Z. Al-Makhadmeh, ‘‘MDS: Multi-level decision sys-
on eye state recognition,’’ in Proc. Int. Conf. Mach. Vis. Inf. Technol., tem for patient behavior analysis based on wearable device information,’’
Feb. 2017, pp. 105–110, doi: 10.1109/CMVIT.2017.25. Comput. Commun., vol. 147, pp. 180–187, Nov. 2019.
[40] L. Geng, ‘‘Real-time driver fatigue detection based on morphology infrared YUNBO GONG was born in Tieling, Liaoning,
features and deep learning,’’ Hongwai Yu Jiguang Gongcheng/Infrared in 1995. He received the B.E. degree from
Laser Eng., vol. 47, no. 2, 2018, Art. no. 203009. the South China University of Technology,
[41] J. Guo and H. Markoni, ‘‘Driver drowsiness detection using hybrid convo- Guangzhou, China, in 2018, where he is currently
lutional neural network and long short-term memory,’’ Multimedia Tools pursuing the master’s degree in traffic informa-
Appl., vol. 78, no. 20, pp. 29059–29087, 2019. tion engineering and control. His interests include
intelligent vehicles, computer vision, and 3D laser
radar.
FENG YOU was born in Zunyi, Guizhou, in 1977. HAIWEI WANG received the B.E. and M.S.
He received the B.S. degree in automobile trans- degrees from Jilin University, Changchun, China,
portation engineering from Guizhou University, and the Ph.D. degree from the South China Uni-
Guizhou, in 1995, and the M.S. and Ph.D. versity of Technology. She is currently a Lecturer
degrees in automobile transportation engineer- with the School of Transportation and Economic
ing from Jilin University, Changchun, in 2005. Management, Guangdong Communication Poly-
From 2005 to 2011, he was a Lecturer with technic, Guangzhou, China. She has published
the School of Civil Engineering and Transporta- four articles in international journals. Her current
tion, South China University of Technology. Since research interests include ITS and vehicle control.
2012, he has been an Associate Professor with the
School of Civil Engineering and Transportation, South China University of
Technology. He is the author of more than 50 articles, and more than ten
inventions. His research interests include connected vehicle, advanced driver
assistance, traffic information and safety, cooperative safety control, and
vehicle active safety and driver behavior. He is a Reviewer of the Journal
of Safety Science, Accident Analysis & Prevention, and Jilin University
HONGYI LI was born in Urumqi, Xinjiang, China,
Engineering and Technology Edition.
in 1986. He received the B.S. degree in mechanic
from the Beijing Institute of Technology, Beijing,
in 2008, and the Ph.D. degree in physics and chem-
istry materials from the Xinjiang Technical Insti-
tute of Physics and Chemistry, Chinese Academy
XIAOLONG LI was born in Bengbu, Anhui, of Sciences, Urumqi, in 2013. From 2011 to 2012,
in 1993. He received the B.E degree from he was a dual culture Ph.D. with Northwestern
Southeast University, Nanjing, China, in 2016. University, Chicago, USA. From 2013 to 2014,
He is currently pursuing the master’s degree he was a Senior Engineer. Since 2015, he has been
in communication and transportation engineer- a Professor with the Xinjiang Quality of Products Supervision and Inspection
ing with the South China University of Tech- Institute of Technology. Besides, he was a Visit Research Fellow with the
nology, Guangzhou, China. His interests include Xinjiang Technical Institute of Physics and Chemistry, Chinese Academy of
intelligent vehicles, computer vision, and driving Sciences. His research interests include materials science and information
behavior analysis. technology.