Elsevier

Neurocomputing

Volume 126, 27 February 2014, Pages 78-94
Neurocomputing

Hybrid computer vision system for drivers' eye recognition and fatigue monitoring

https://doi.org/10.1016/j.neucom.2013.01.048Get rights and content

Abstract

This paper presents a hybrid visual system for monitoring driver's states of fatigue, sleepiness and inattention based on driver's eye recognition. Safe operation in car conditions and processing in daily and night conditions are obtained thanks to the custom setup of two cameras operating in the visible and near infra-red spectra, respectively. In each of these spectra image processing is performed by a cascade of two classifiers. The first classifier in a cascade is responsible for detection of eye regions based on the proposed eye models specific to each spectrum. The second classifier in each cascade is responsible for eye verification. It is based on the higher order singular value decomposition of the tensors of geometrically deformed versions of real eye prototypes, specific to the visible and NIR spectra. Experiments were performed in real car conditions in which four volunteer drivers participated. The obtained results show high recognition accuracy and real-time processing in software implementation. Thanks to these the system can become a part of the advanced driver’s assisting system.

Introduction

Road accidents are the leading cause of death in industrialized countries. These are mostly caused by human errors occurring in different activities related to vehicle driving, as well as low violations such as speeding. Also drivers' fatigue, sleepiness or inattention can lead to serious accidents. As reported by the European Union only driver tiredness constitutes a significant factor of about 20% of crashes involving heavy commercial vehicles [29]. These problems were recognized and countermeasures were undertaken in many countries. For instance, in the European Union there are plans to significantly reduce a number of fatal accidents in the nearest years by changing low regulations and also employing new technologies [28], [30].

Thanks to recent hardware and software developments computers can be used to monitor drivers' conditions with the potential of alerting in dangerous situations [45], [5]. In this respect hybrid intelligence offers new methods and tools which can be successful in solving such difficult problems, as reported in literature [11], [1], [20], [60], [12], [46], [8]. Such driver's supporting methods become parts of the Driver Assisting Systems (DAS). It is a matter of incoming years to have DAS practically available in modern cars. One of the key functionalities of DAS is reliable detection of driver’s fatigue and sleepiness, mostly based on observation of his/her eyes[6], [26].

Face and eye recognition belong to the fundamental tasks of computer vision [27], [56], [57]. Eye recognition for monitoring driver's behavior, fatigue, sleepiness or inattention is one of the research topics which were greatly enhanced for the last years [21], [22], [40], [41], [42], [43], [51], [52]. Eye detection can be divided into active and passive methods [58]. In the former group, a custom hardware is required which supplies special lighting, such as near IR [61], [5]. On the other hand, passive methods assume only the natural illumination spectrum. This group can be further partitioned into the methods which use color or monochrome images [7]. Reported eye recognition methods rely on detection of some characteristic eye features. These, in turn, can be obtained with template matching [33], projections [62], Hough transform [45], [62], gradients [36] or wavelets [44], to name a few. Frequently used indicator of driver's state is the percentage of eye closure (PERCLOSE) [23], [24], [54]. However, the existing solutions still lack sufficient accuracy or speed to reliably operate in real car conditions.

In this paper a visual system for driver's eye recognition is proposed which allows real-time operation in moving vehicles and under various conditions. For this purpose the original hardware setup was constructed which consists of two cameras operating in the visual and near infra-red (NIR) spectra. NIR illumination is obtained with a NIR light emitting diode (LED). This paper follows our previous work focused upon a method of processing images acquired in only daily conditions with color camera [18]. The system presented in this paper contains a new vision path processing NIR signals which allows operation in night conditions.

In the presented system there are two independent blocks of visual signal processing. Each of them is composed of two different classifiers connected in series: The front-end is responsible for the detection of eye candidates, followed by the trained classifier whose role is to refine the candidates and to respond whether the eyes are well visible. The two front detectors for the visible and NIR spectra are different, however. In the former, operation starts with the skin segmentation from color images [18]. This stage is followed by the skin region detection obtained with the adaptively window growing method (AWG) [14]. Then, eye candidate regions are selected thanks to the proposed eye model. Finally, eyes are verified by the classifier operating with the higher order singular value decomposition (HOSVD) of the tensor of geometrically deformed images of real eye prototypes [15]. On the other hand, processing of the NIR images starts with the detection of pupil candidates. This is done in the associated integral image which allows very fast computation of sums of pixel values in any rectangle within an image. Then iris and finally eye regions are detected based on the novel eye model, designed especially for the NIR signals. Eye regions are then refined to only those which fulfill predefined geometrical relations. Finally and similarly to the visible spectrum path, eye region candidates are recognized by the HOSVD classifier, this time trained with the prototypes from the NIR images, however. In this paper we focus mostly on the new NIR processing path. As already mentioned, part of the system operating in the visible spectrum is described in our previous paper [18].

Driver's safety is especially stressed in the presented system. This relates to the hardware construction and especially to the NIR lighting module. It is well known that long exposition to the NIR illumination of high power, although not directly visible, can be annoying or even dangerous for the exposed person [2]. Therefore in our system we use only one LED of low power synchronized with the shutter of the camera.

Experiments were conducted in different car conditions with the help of four volunteer drivers. Presented experimental results obtained in real car conditions show high accuracy and real-time operation of the system in software implementation.

This paper is organized as follows. Architecture of the system and the custom hardware setup are presented in Section 2. The proposed method of eye detection in NIR spectrum is dealt with in Section 3. Similarly, detection in the color images is discussed in Section 4. In Section 5 the method of eye recognition in the tensor space of deformed pattern prototypes is presented and discussed. Additionally, the Image Euclidean Distance method is presented which increases accuracy of eye recognition. Experimental results are presented in Section 6. This paper ends with conclusion in Section 7 as well as with the literature references.

Section snippets

System architecture

Block diagram of the proposed system is presented in Fig. 1. It is organized into two distinctive paths, one for processing images from the visible spectrum and the second from the NIR spectrum, respectively. Further, each processing path is organized as a cascade of specialized modules, each refining output of its predecessor. Such connections of many classifiers are characteristic of usually higher accuracy compared to complex but simple classifiers [37], [34], [48], [49]. As already

Eye detection in the near infrared images

As already mentioned, detection of eye candidates is performed with help of an eye model proposed here. However, it is important to notice that because of the cascade configuration of classifiers its role is to sieve out only the regions which definitely are not eye regions, passing at the same time all those regions which, even with some doubts, can contain the eyes. More specifically, at first we care about small false negatives (FN) ratio and then about the high true-positive (TP) one. In

Eye detection in the visible spectrum

The second processing path presented in Fig. 1 is responsible for detection and recognition of driver’s eyes in the color images acquired in the visible spectrum. More often than not daily lighting is assumed, although artificial light is also possible. Similar to the NIR processing path, at first eye candidate regions are detected. These are then refined by the HOSVD classifier, trained this time with prototype patterns corresponding to the daily light conditions. In the following sections the

Eye recognition in the space of deformed pattern prototypes

The next module in the cascade of eye classification chain consists of the classifier operating with geometrically deformed eye prototypes. Thus, for each prototype pattern, such as an eye or other object, a tensor is obtained which contains expected views of that pattern. However, to facilitate object recognition we propose two modifications. At first, each prototype pattern, which in our case can belong to the group of eyes and non-eyes objects, is additionally geometrically deformed by small

Experimental results

For experiments a number of 10–20 min long video sequences were recorded in which four volunteer drivers participated. Results of the part of the system operating in the visual spectrum were also presented in our previous paper [18]. Therefore in this paper we mostly discuss results obtained by the NIR vision module.

Fig. 11 shows results of the NIR detector operating in images containing persons with open eyes in fronto-parallel position. The three different test sequences obtained in night

Conclusions

This paper presents a vision system for driver's eye recognition, capable of operating in real car conditions and real-time. The system consists of the two cameras, one color for the visible spectrum and the second one operating in the NIR range. It contains also one NIR illuminating LED mounted on the rail attached to the windshield. Video signals from the cameras are passed with the IEEE 1394b link to the local computer which also runs recognition software which allows real-time processing

Acknowledgment

The work was financially supported by the National Center for Research and Development under Lider Program, Contract no. LIDER/16/66/L-1/09/NCBiR/2010. The authors are very grateful to Mr. Marcin Bugaj, Mr. Stanisław Groński, as well as Mr. Krzysztof Groński for their help in the experiments.

Bogusław Cyganek received his M.Sc. degree in Electronics in 1993, then in Computer Science in 1996 from the AGH University of Science and Technology, Krakow, Poland. He obtained his Ph.D. degree cum laude in 2001 with a thesis on Correlation of Stereo Images, and D.Sc. degree in 2011 with a thesis on Methods and Algorithms of Object Recognition in Digital Images.

He is currently a researcher and lecturer at the Department of Electronics, AGH University of Science and Technology. During the

References (62)

  • Z.H. Zhou et al.

    Projection functions for eye detection

    Pattern Recognition

    (2004)
  • Agilent Technologies Inc.. Application Note 1118: Compliance of Infrared Communication Products to IEC 825-1 and...
  • C. Ahlstrom et al.

    Processing of eye/head-tracking data in large-scale naturalistic driving data sets

    IEEE Transactions on Intelligent Transportation Systems

    (2012)
  • AlliedVision:...
  • L.M. Bergasa et al.

    Visual monitoring of driver inattention

  • P. Campadelli et al.

    Eye localization: a survey

    NATO Science Series

    (2006)
  • A. Cichocki et al.

    Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-way Data Analysis and Blind Source Separation

    (2009)
  • D. Comaniciu, V. Ramesh, P. Meer: Real-time tracking of non-rigid objects using mean shift, in: Proceedings of the IEEE...
  • F.C. Crow, Summed-area table for texture mapping, in: Proceedings of Computer Graphics SIGGRAPH'84, vol. 18, no. 3,...
  • B. Cyganek et al.

    An Introduction to 3D Computer Vision Techniques and Algorithms

    (2009)
  • B. Cyganek

    An Analysis of the Road Signs Classification Based on the Higher-Order Singular Value Decomposition of the Deformable Pattern Tensors, Advanced Concepts for Intelligent Vision Systems Acivs 2010, LNCS 6475

    (2010)
  • B. Cyganek, Image segmentation with a hybrid ensemble of one-class support vector machines, in: M. Graña Romay et al....
  • B. Cyganek

    A software framework for tensor representation and decompositions for object recognition

    Image Processing & Communications

    (2010)
  • B. Cyganek, S. Gruszczyński, Visual system for drivers' eye recognition, in: E. Corchado, M. Kurzyński, M. Woźniak...
  • B. Cyganek, Adding parallelism to the hybrid image processing library in multi-threading and multi-core systems, in:...
  • J. Derrac, S. García, F. Herrera, A First Study on the Use of Coevolutionary Algorithms for Instance and Feature...
  • Y. Dong et al.

    Driver inattention monitoring system for intelligent vehicles: a review

    IEEE Transactions on Intelligent Transportation Systems

    (2011)
  • Y. Du et al.

    Driver status recognition by neighborhood covering rules

    Lecture Notes on Computer Science

    (2011)
  • I. García, S. Bronte, L.M. Bergasa, N. Hernandez, B. Delgado, M. Sevillano, Vision-based drowsiness detector for a...
  • I. García, S. Bronte, L.M. Bergasa, J. Almazán, J. Yebes, Vision-based drowsiness detector for Real Driving Conditions,...
  • G.H. Golub et al.

    Matrix Computations

    (1996)
  • Cited by (120)

    View all citing articles on Scopus

    Bogusław Cyganek received his M.Sc. degree in Electronics in 1993, then in Computer Science in 1996 from the AGH University of Science and Technology, Krakow, Poland. He obtained his Ph.D. degree cum laude in 2001 with a thesis on Correlation of Stereo Images, and D.Sc. degree in 2011 with a thesis on Methods and Algorithms of Object Recognition in Digital Images.

    He is currently a researcher and lecturer at the Department of Electronics, AGH University of Science and Technology. During the period 2004–2005 he was cooperating with the University of Glasgow under the projects Racine-S and Racine-IP. His research interests include computer vision, pattern recognition, as well as development of programmable devices and embedded systems. He is an author or a co-author of over eighty conference and journal papers and three books with the latest “An Introduction to 3D Computer Vision Techniques and Algorithms” published by Wiley. Dr. Cyganek is a member of the IEEE, IAPR and SIAM.

    Sławomir Gruszczyński was born in Wroclaw, Poland, on December 14, 1976. He received the M.Sc. degree and the Ph.D. degree in Electronics and Electrical Engineering from the Wroclaw University of Technology, Poland, in 2001 and 2006, respectively.

    From 2001 to 2006 he has been with Telecommunications Research Institute, Wroclaw Division; from 2005 to 2009, he worked at the Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology. In 2009, he joined the Faculty of Electronics at AGH University of Science and Technology. He has coauthored 45 scientific papers. He is a member of the IEEE, and a member of Young Scientists' Academy at Polish Academy of Sciences (PAN) and Committee of Electronics and Telecommunications at Polish Academy of Sciences (PAN).

    View full text