Abstract
Fatigue driving is one of main problems threatening driving safety. Therefore, it attracts numerous researchers interests. This paper introduces a new method based on eye feature to research the fatigue driving. Firstly, the face is detected by the model of skin-color in the YCbCr color space, which extracted face region from complex background quickly and accurately. Secondly, eye detection includes extracting eye region and detecting eye two steps. Specifically, the proposed method extracts eye region in face image based on gray-scale projection and then detect eye using Hough transform. Finally, calculate the area of the eye profile after dilation and use it as the parameter to analysis eye state. Put forward the standard to recognize fatigue base on the PERCLOS. The experiment results illustrate the efficiency and accurately of the proposed method, especially, detected face as well as extracted eye region with a high accuracy.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
In recent years, more and more researchers have been engaged in fatigue driving recognition and many important achievements have been made, such as lane-track alarm system [1], EEG method [2], heart rate detector [3], etc. Although there are many ways to realize driver fatigue detection to a certain extent, the performances of these methods are various. These methods can be divided into three categories. The first way is to detect fatigue through the analysis of the vehicle status. For example, American Ellison Research Labs monitored lane track to achieve fatigue detection in 2004. While, some researchers detected fatigue through other vehicle status such as the speed of vehicle running and the turning of the steering wheel. But those ways are hard to get a standard to judge whether the drivers are dozing. The second method is based on the physiological property such as EEG and heart rates. According to the research of Japan Pioneer Company, the heart rate will be slow when drivers are drowsy. So through physical characteristics like heart rate, EEG can make an accurate detection. However, these techniques are usually more expensive and will annoy the drivers, so they are not easily accepted by the drivers. In general, driver fatigue monitoring system should consider that whether it is easy to accept for the driver, timeliness, reliability, scalable, and the cost of these. With the development of digital image processing and computer technology, more mature techniques are provided to analyze videos and images. Using digital image processing technology to find out fatigue in time and then prevent it which is the third approach. Fatigue recognition according to facial features has following advantages: real time, reliability and little interference to the driver. Driver fatigue recognition based on facial feature is a trend. Many researchers proposed different methods in this area [4, 5]. The eye contains a lot of information, so the research and analysis of the human eye has become a very hot issue.
The paper is organized as follows. System design is presented in Sect. 2. Section 3 shows the design of fatigue detection algorithm. Section 4 contains the results of experiment experiments. The paper is concluded in Sect. 5.
2 System Design
This driver fatigue monitoring system includes Pre-treatment (video acquisition and image preprocessing), face detection, eye detection, eyes feature extraction and driving fatigue recognition. The driver fatigue monitoring system is shown in Fig. 1 and we will introduce these in detail in the following sections.
3 Fatigue Detection Algorithm Design
This driver fatigue monitoring system detects driver drowsiness based on eye features. Getting the picture from the video and using skin-color to detect face. Using skin-color way to detect face has many advantages, but it is influenced by many factors (light, clothes similar to skin color, etc.), so choosing YCbCr color space and add restricted conditions are necessary. Gray-scale projection method combines with Circular Hough transform method to extract the eye region accurately. Finally, according to eye status we can recognize driving fatigue.
3.1 Face Detection
Face detection should be done before eye detection, so the first step is face detection. The face detection ways currently used include Eigenface [6], neural networks [7], Gabor transform [8], skin-color [9–11]. Compared with mentioned method above, the face image based on skin-color method can extract face from complex background quickly and accurately. It achieves good real-time performance and strong practicability. The skin-color model is established should choose appropriate color space. RGB, NTSC, YCbCr and HSV are used commonly color space. YCbCr color space put the RGB color space divided into three components, a brightness component(Y) and the two color component(Cr and Cb). YCbCr model is commonly used in color digital video model. In this color space, Y contains brightness information, Chromaticity information is stored in the Cr and Cb. Cb shows green component relative reference value, Cr shows red component relative reference value, and they are independent relationship. YCbCr model data can be described by double precision. As brightness component and color component are separated in YCbCr color space, reduce the effects of light. Cr and Cb contain chrominance information and fit to establish skin-color model14. So we chose YCbCr color space. The RGB components and the YCbCr components can be converted by the following formulas.
The skin color distribution similar to Gaussian distribution, Gaussian model is proposed base on the theoretical. Based on this, calculate the probability of each pixel belongs to skin color which use two-dimensional Gaussian in the color image and get the probability value of skin. Those values make skin probability graph. The higher probability of the pixel color values of skin area in figure, it is the candidate skin region.
The values of the pixel probability are calculated by the principle of two-dimensional Gaussian model of skin color detection. Selecting a threshold base on those probability values, if the pixel probability values greater than the threshold, it belongs to the skin, if not, it cant belong to the skin. Most researchers choose Gaussian model as a skin color model under normal conditions. But this method needs to consider the samples which are not belong to the skin-color.
In the model based on color, getting Cb and Cr values of each pixel of the image. If the Cr value of a pixel ranges from 140 to160 and the Cb value of this pixel ranges from 140 to 195(skin color differs from person to person, the data base on this experiment), the pixel is judged skin-color and set the gray value of the pixel 255. If not, set the gray value of the pixel 0. In this way, we get the binary image. Binary image include noise and the gray-scale images converted into binary image process will inevitably increase noise, so the binary image needs to be smoothed. The techniques which are used include corrosion and expansion of images. Figure 2 contains an example of face detection procedure.
If the image background is complex, we may get multiple candidate face region. Those are candidate skin region need to judgment. And it will be affected by realistic environment, such as clothes, skin-color background, so we will get some candidate skin region as described earlier in this paper, it must be added some conditions to limit:
-
(1)
The limitation of length and width of the skin region, these values are set according to the proportion of face in the image.
-
(2)
Height and width ratio, in reality, the ratio is set range from 0.6 to 2.
-
(3)
The region should contain the eyes, in other words there are two black areas in real face region.
Determine the face region accurately by adding the three restrictions. Figure 3 shows the illustration of face detection adding the restrictions mentioned above from complicated background.
The procedures of recognizing face based on skin-color are as follows:
-
Step 1: The color images from RGB color space converted into the YCbCr color space. The function we used is YCbCr = rgb2ycbcr(RGB).
-
Step 2: The color images change into binary image base on the values of Cb and Cr which get according to the skin in YCbCr model distribution range.
-
Step 3: Binary image includes noise and the gray-scale images converted into binary image process will inevitably increase noise, so this step we eliminate the noise.
-
Step 4: Corrosion and fill hole processing.
-
Step 5: We will get some candidate skin region, we choose true face region from them. According to the three restrictions which have already been mentioned we determine the face region accurately.
3.2 Extraction the Eye Area
The extracting face region is the basis for the recognize fatigue driving. Recognize fatigue driving by the eye state. The human eye is one of vital organ to reflect the fatigue driving or not, it contains a lot of information. The researches show that recognize fatigue and eye state have a close relationship, eyes are closed or almost closed state in a long time when people tired, at the same time the blink rate is accelerated and the closed time will become long. So this system through the eye state to recognize fatigue driving.
The gray level around the eye part is lower than other parts in the face, so we can detect the human eye using this characteristic. Gray-scale projection method is used widely in image processing. Using this method to process the original image directly, the noise is relatively large and it is difficult to achieve desired results. But with the development of technology, the gray projection method are also improved, this method in the human eye feature extraction is widely applied [12, 13]. The system is based on previous studies on the application of this theoretical knowledge, Combination with the knowledge of the digital image. In order to improve the efficiency and reduce calculation, we use the binary image instead of the original image to gain the horizontal and vertical integral projection. Binary image of the face can clearly find eyebrow, eyes, nose and mouth etc. The binary image with gray-scale projection can provide a more determined accurately area of the human eye.
The binary image of the human face can clearly find eyebrows, eyes and mouth. Using gray-scale projection algorithm to deal with the binary image can get the position of eye roughly. G(x, y) represents the gray value of the pixel in(x, y), H(x) stands for the value of binary image horizontal integral projection, V(y) stands for the value of binary image vertical integral projection, which are shown as follows:
We have got coordinate of the face area in the section A and extracted face area base on the coordinate. In Fig. 4, (a) shows the extracted face region, (b) shows the binary image of the face and (c) shows the extracted eye region. It is gained by the gray-scale projection.
We get the gray value from the binary image of the face with horizontal and vertical integral projection. Figure 5 shows the gray value of horizontal integral. Figure 6 shows the gray value of vertical integration.
The black region which closes to forehead shows eyebrows and eyes. It corresponds to the Horizontal integral projection curve are two minimum.
It is easy to roughly get the ordinate of eyebrows and eyes from the figure, so that we can choose the appropriate width. The same method, two valleys in vertical gray projection curve represent left and right eyes. So we extract the eye region.
3.3 Eye Detection
Hough transform which describes the border of region, is often used to detect geometry, such as circular, oval or line in the image. In this paper, we use this method to detect eye accurately. Before using Hough transform we should detect the borders of the image. By comparing the different edge detection algorithms, we find Prewitt algorithm in this system performs best. It is easy to find that the eye is similar to oval in binary figure. But using the oval model to judge human eyes need to determine the center axislength and minor axis length and steer of the oval which similar to eye. Detecting an oval needs five parameters [14], it is a great amount of calculation. The human iris is circular. Locating the iris circle model by Hough transform to determine the human eye is open or close, if it is close at that time we can’t detect the circle [15]. It is only requires three parameters, the center (x, y) and radius r.
The basic idea of the Hough transform is according to the majority points of the boundary to determine the curve. So through describing the boundary of the curve, image space changes into curves space. So it is good tolerance and robustness to some possible noise of area boundary.
The appearance of eye model is shown in Fig. 7. We know the circle can be expressed with the following formula:
From Eq. 4 we can see that a circle requires three parameters. Judging a circle with the three parameters still exists a certain difficulty. So researchers limit the center (x, y) and radius r within a certain range, and then calculate the parameters. In this way, the parameters have reduced. It reduces much calculating obviouslly. The radius of the circle is calculated first for a very small and sealed area, then judgment all the points on the edge. Consequently, we can quickly identify the edge of the circle.
Set the parameter of the circle is one pixel, the step of radians change is 0.2 Pixel, the minimum Radius \(r_{min} =5\), the maximum Radius, \(r_{max} =8\), the threshold value is 0.685. These parameters are set according to the system. Figure 8 shows the illustration of Hough transform.
This paper presents the gray-scale projection and the Hough transform to quickly detect eye. It includes two steps which are extracting the eye area and eye detection accurately. This section is very important for the driver fatigue monitoring system, we describes two algorithms in section B and C. The procedures of recognizing eye are follows:
-
Step 1: Segmentation of face region and change it to binary image.
-
Step 2: Binary image of face with horizontal integral projection get the distribution of the vertical graph. We get the value of y that the minimum gray values which stand for the eye and eyebrow.
-
Step 3: The width adds to 2d with y values as the center. Get out the region.
-
Step 4: The region with horizontal integral projection and get two lows from the graph which can determine the eye part. The eye area can be extracted.
-
Step 5: Detect the border of the image by Prewitt algorithms.
-
Step 6: Set the parameters of the circle with \(r_{min}=5\), \(r_{max}=8\) and threshold value is 0.685. Then eye detection by Hough transform.
3.4 Eye State Analysis
We calculate the quantity of the eye profile pixel with dilation. Under the same conditions, the pixels number of the open eyes certainly more than the eyes closed pixels. So based on this we can judge whether human eyes are open or closed, and judge how much eyes are open.
PERCLOS is a classical method to determine whether human eyes are fatigue or not, the fatigue recognition model based P80 criterion [16]. The 100 % open is the eyes largest area in all images during a period of time. If the eye closed degree more than 80 % is determined closed state. In this paper, the pixels number of the eye largest is the eyes largest area.
To deal with each frame from the human eye features respectively, the number of eyes closed frames is \(CloseFrame\underline{\ } Num\) and the total number of frames that deal with is \(SumFrame\underline{\ }Num\). According to the following formula the value of PERCLOS can be calculated.
If the value of PERCLOS in experiment is greater than the threshold that we set the value of 20 %, we think the driver is fatigue, then the alarm system start warning.
4 Experimental Results
In this section, we take the video which acquired by camera in simulation driving condition as an example. Video processing using MATLAB converted to images. The video in the system is 30 frames /s and \(640 *480\) pixels per a frame. Based on the fact that movement is continuous and not too fast, there is no need for testing each frame. First of all, we extract one image per 5 frames. In order to not only accurately judge the results but also improve the efficiency of the system in this way. So we extract 6 frames per second [17].
Lots of experiments show that this method is better than other methods. It can identify the face more accurately and quickly which is the foundation for later processing. Figures 9 and 10 show the original image and the framed face image recognize face by this method.
The experiments illustrate that the face detection method can detect the faces accurately even though in different distances and positions, and the accuracy is over 95 %. It also can meet the real-time face detection. So the face detection method is the basis of the fatigue state judgment.
After extracting the face and then extract eyes from the face image based on the described algorithms for eye detection. Some examples of the detected eyes are shown in Figs. 11 and 12.
Numerous experiments demonstrate the viability of the proposed method. In this section we extract the eye region and detect eye better.
Figure 13 shows the eye edge after dilation of person who is not drowsy. And Fig. 14 illustrates the eye edge of a sleepy driver. According to Figs. 13 and 14, the area of eye edge in Fig. 13 is bigger than what in Fig. 14. Two testers with four 30 s videos which simulation of driving situations, according to the algorithm, each video includes 180 frames, Table 1 shows the experimental results.
From Table 1, the system can detect the faces accurately and the extracted eye region is exact, but using Hough transform to detect eye is imperfect. From the experimental results, the proposed algorithm is faster and has higher accuracy. The proposed algorithm can reach correct detection rate of 86.4 % on average. On the other side, the correct rate of video 4 is lower than video1, the main reason is that the system is easily influenced by illumination.
5 Conclusion
In this paper, we propose a method for fatigue driving detection base on eye feature. As can be seen from the experimental results the system detected face as well as extracted eye region with a high accuracy of more than 95 %. Using the last system, fatigue detection is not as well as expected. But the system can fully meet the real-time and accurate requirements. We will realize driver fatigue monitoring system on the hardware platform and improve the accuracy in the future study, besides, we will study more evaluation indexes or compare its performance with other similar methods, to demonstrate the effectiveness of the proposed approach.
References
Pei, Z., Song, Z.H., Zhou, Y.M.: Research status and development trend of motor vehicle driver fatigue evaluation method. J. China Agric. Univ. 6(6), 101–105 (2001)
Lampetch, S., Punsawad, Y., Wongsawat, Y.: EEG-based mental fatigue prediction for driving application. In: Biomedical Engineering International Conference (BMEICON), pp. 1–5 (2012)
Vicente, J., Laguna, P., Bartra, A., Bailon, R.: Detection of driver’s drowsiness by means of HRV analysis. In: Computing in Cardiology, pp. 89–92 (2011)
Wang, P., Shen, L.: A method detecting driver drowsiness state based on multi-features of face. In: 2012 5th International Congress on Image and Signal Processing (CISP 2012), pp. 1171–1175 (2012)
Lee, B.G., Chung, W.Y.: Driver alertness monitoring using fusion of facial features and bio-signals. IEEE Sens. J. 12(7), 2416–2422 (2012)
Watta, P., Gandhi, N., Lakshmanan, S.: An Eigenface approach for estimating driver pose. In: 2000 Proceedings Intelligent Transportation Systems, pp. 376–381. IEEE (2000)
Ni, Q.K., Guo, C., Yang, J.: Research of face image recognition based on probabilistic neural networks. In: 2012 24th Chinese Control and Decision Conference (CCDC), pp. 3885–3888 (2012)
Shan, D., Ward, R.K.: Improved face representation by nonuniform multilevel selection of gabor convolution features. IEEE Trans. Sys. Man Cybern. Part B Cybern. 39(6), 1408–1419 (2009)
Zhao, Y.L., Gao, Z., Wu, W.X.: The detection algorithm of locomotive driverss fatigue based on vision. In: 2010 3rd International Congress on Image and Signal Processing (CISP2010), pp. 2686–2690 (2010)
Devi, M.S., Choudhari, M.V., et al.: Driver drowsiness detection using skin color algorithm and circular hough transform. In: 2011 Fourth International Conference on Emerging Trends in Engineering and Technology, pp. 129–134 (2009)
Wu, C.D., Zhang, C.B.: Detecting and locating method of human face in driver fatigue surveillance. J. Shenyang Jianzhu Univ. Nat. Sci. 25(2), 386–389 (2009)
Lu, L., Yang, Y., Wang, L., Tang, B.: Eye location based on gray projection. In: 2009 Third International Symposium on Intelligent Information Technology Application, pp. 58–60 (2009)
Feng, J.Q., Liu, W.B., Yu, S.L.: Eyes location based on gray-level integration projection. Comput. Simul. 22(4), 75–76 (2005)
Yang, Q.F., Gui, W.H., et al.: Eye location novel algorithm for fatigue driver. Comput. Eng. Appl. 44(6), 20–24 (2008)
Qu, P.S., Dong, W.H.: Eye states recognition based on eyelid curvature and fuzzy logic. Comput. Eng. Sci. 29(8), 50–53 (2007)
Pan, X.D., Li, J.X.: Eye state-based fatigue drive monitoring approach. J. Tongji Univ. Nat. Sci. 39(2), 231–235 (2011)
Wang, Y., Hu, J.W.: A method for detection of driver eye fatigue state based on 3G video. Electron. Sci. Tech. 24(10), 84–85 (2011)
Acknowledgments
This work is supported by the Basic Research Program (Natural Science Foundation) of Jiangsu Province of China (No.BK20130209), the Fundamental Research Funds for the Central Universities (No.2013QNA24), the Project Funded by China Postdoctoral Science Foundation (No.2014M560460), the Project Funded by Jiangsu Postdoctoral Science Foundation (No.1302037C).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 2.5 International License (http://creativecommons.org/licenses/by-nc/2.5/), which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Xu, X., Cui, X., Wang, G., Sun, T., Feng, H. (2015). A New Method for Driver Fatigue Detection Based on Eye State. In: Ciucci, D., Wang, G., Mitra, S., Wu, WZ. (eds) Rough Sets and Knowledge Technology. RSKT 2015. Lecture Notes in Computer Science(), vol 9436. Springer, Cham. https://doi.org/10.1007/978-3-319-25754-9_45
Download citation
DOI: https://doi.org/10.1007/978-3-319-25754-9_45
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-25753-2
Online ISBN: 978-3-319-25754-9
eBook Packages: Computer ScienceComputer Science (R0)