510 likes | 753 Views
Automated Drowsiness Detection For Improved Driving Safety. Aytül Erçil November 13 , 2008. Outline. Problem Background and Description Technological Background Action Unit Detection Drowsiness Prediction. Objectives/Overview.
E N D
Automated Drowsiness Detection For Improved Driving Safety Aytül Erçil November 13, 2008
Outline Problem Background and Description Technological Background Action Unit Detection Drowsiness Prediction
Objectives/Overview • Statistical Inference of fatigue Using Machine Learning Techniques
In over 500.000accidents in 2005 (in Turkey): • Injured: 123,985 people • Deceased: 3,215people • Financial loss: 651,166,236 USD Driver error has been blamed as the primary cause for approximately 80% of these traffic accidents.
The US National Highway Traffic SafetyAdministration estimates that in theUS alone approximately 100,000 crashes each year are caused primarily by driverdrowsiness or fatigue
Growing Interest In Intelligent Vehicles US Department of Transportation Initiative European Transport Policy for 2010: set a target to halve road fatalities by 2010. Problem Background
Current Funding Status: • Turkish Development Agency funding of Drive-Safe (August 2005-July. 2009) • Japanese New Energy and Industrial Technology Development Organization (NEDO) (October 2005 -December 2008) • FP6SPICE Project at Sabancı University (May 2005- October 2008) • FP6 AUTOCOM Project at ITU Mekar (May 2005- April 2008).
Readiness-to-perform Mathematical models of alertness dynamics Vehicle-based performance technologies (Vehicle Speed, Lateral Position, Pedal Movement) In-vehicle, on-line, operator status monitoring technologies Fatigue Detection and Prediction Technologies
Physiological Signals (heartrate, pulse rate and Electroencephalography (EEG)) Computer Vision Systems (detect and recognize the facialmotion and appearance changes occurring during drowsiness) In-vehicle, on-line, operator status monitoring technologies
Computer Vision Systems Visual Behaviors Examples Gaze Direction Head Movement Yawning No requirement for physical contact
Facial Actions Ekman & Friesen, 1978
Proposed Work Detection Of Driver Fatigue From A Recorded Video Using Facial Appearance Changes The framework will be based on graphical models and machine learning approaches
Proposed Architecture Fatigue Fatigue Entire Face Behavior Inattentive Falling Asleep Inattentive Falling Asleep AU 61 AU 62 AU 51 AU 52 AU 61 AU 62 AU 51 AU 52 Single AU Pupil Motion Gaze Pupil Motion Gaze Partial Face Behavior Eye Tracker Gaze Tracker Eye Tracker Gaze Tracker Sensing Channels Features Time n-1 Time n
Action Unit Tracking Previous techniques Do not employ a spatially and temporally dependent structure for Action Unit Tracking Contextual information is not exploited Temporal information is not exploited
Classification- Challenges Which action units or combinations is a cue for fatigue?
Learning from real examples Different Neural pathways for posed/spontaneous expressions Posed Drowsiness Actual Drowsiness
Initial Experimental Setup Subjects played a driving video game on a windows machine using a steeringwheeland an open source multi-platform video game. At random times,a wind effect was applied that dragged the car to the right or left, forcing thesubject to correct the position of the car.
Head movement measures Head movement was measured using an accelerometer that has 3 degrees offreedom. This three dimensional accelerometer has three one dimensional accelerometers mounted at right angles measuring accelerations in the range of 5gto +5g
The one minute preceding a sleep episode or a crash was identified as a non-alertstate. There was a mean of 24 non-alert episodes with a minimum of 9 and amaximum of 35. Fourteen alert segments for each subject were collected from thefirst 20 minutes of the driving task.
Crash Distance from center Overcorrection Steering Eye opening Eyes closed 0 Seconds 20
Histograms for Eye Closure and Eye Brow Up Eye Closure: AU45 Brow Raise:AU2 Area under the ROC
Facial Action Unit Detection Machine Learning AU1 AU2 AU4 …. …. AU46 Pattern Recognition (Adaboost) (SVM) Feature Selection + +
Drowsiness Prediction The facial action outputs were passed to a classifier for predicting drowsinessbased on the automatically detected facial behavior. Two learning-based classifiers, Adaboost and multinomial logistic regression are compared. Within-subjectprediction of drowsiness and across-subject (subject independent) prediction ofdrowsiness were both tested.
31 Facial Action Channels • Continuous output for each frame AU1 Alert AU2 Multinomial Logistic Regression (MLR) AU4 60 sec Before crash : AU31 Frame Classification Task
Within subject drowsiness prediction For the within-subject prediction, 80% of the alert and non-alert episodes wereused for training and the other 20% were reserved for testing. This resulted ina mean of 19 non-alert and 11 alert episodes for training, and 5 non-alert and 3alert episodes for testing per subject.
Across Subject Drowsiness Prediction Training : 31 actions -> MLR Classifier Framewise training Cross validation: 3 subjects –> training 1 subject –> testing Crash prediction: • choose 5 best features by sequential feature selection • Sum MLR weighted features over 12 second time interval • .98 across subjects (Area under the ROC)
Predictive Performance of Individual Facial Actions More when critically drowsy Eye Closure Brow Raise Chin Raise Frown Nose Jaw Wrinkle Sideways
Predictive Performance of Individual Facial Actions Less when critically drowsy Smile Squint Nostril Brow Lower Jaw Drop Compressor A’ > .75
We observed during this study that many subjects raised their eyebrows inan attempt to keep their eyes open, and the strong association of the AU 2detector is consistent with that observation. Also of note is that action 26, jawdrop, which occurs during yawning, actually occurred less often in the critical60 seconds prior to a crash. This is consistent with the prediction that yawningdoes not tend to occur in the final moments before falling asleep.
Drowsiness detection performance, using an MLR classifierwith different feature combinations.
Effect of Temporal Window Length * A’ 12 seconds Seconds
ALERT DROWSY r=0.87 Eye Closure Brow Raise Brow Raises Brow Raise Brow Raises Eye Closure Eye Openness 0 Seconds 10 0 Seconds 10 Coupling of Facial Movements
Coupling of Steering and Head Motion ALERT DROWSY r=0.65 r=0.27 Head Acceleration Steering Head Acceleration Steering 0 Seconds 60 0 60 Seconds
New associations between facial behavior and drowsiness • Brow raise • Chin raise • More head roll • Possibly less yawning just before crash • Coupling of behaviors • Head movement and steering • Brow raise and eye opening
Future Work • Extend the graphical model so that it captures the temporal relationships using a discriminative approach
Future Work: More Data Collection in Simulator Environment Uykucu (Sleepy)