[go: up one dir, main page]

US20130057671A1 - Method for classification of eye closures - Google Patents

Method for classification of eye closures Download PDF

Info

Publication number
US20130057671A1
US20130057671A1 US13/558,780 US201213558780A US2013057671A1 US 20130057671 A1 US20130057671 A1 US 20130057671A1 US 201213558780 A US201213558780 A US 201213558780A US 2013057671 A1 US2013057671 A1 US 2013057671A1
Authority
US
United States
Prior art keywords
operator
physiological data
movement
eye
rules
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US13/558,780
Other versions
US9117358B2 (en
Inventor
Daniel Levin
Lena Westervall
Susanna Leandersson
Peter Kronberg
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Volvo Car Corp
Original Assignee
Volvo Car Corp
Volvo Technology AB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Volvo Car Corp, Volvo Technology AB filed Critical Volvo Car Corp
Assigned to VOLVO CAR CORPORATION reassignment VOLVO CAR CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LEVIN, DANIEL, Westervall, Lena
Assigned to VOLVO TECHNOLOGY CORPORATION reassignment VOLVO TECHNOLOGY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KRONBERG, PETER, Leandersson, Susanna
Publication of US20130057671A1 publication Critical patent/US20130057671A1/en
Assigned to VOLVO CAR CORPORATION reassignment VOLVO CAR CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VOLVO TECHNOLOGY CORPORATION
Application granted granted Critical
Publication of US9117358B2 publication Critical patent/US9117358B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B21/00Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
    • G08B21/02Alarms for ensuring the safety of persons
    • G08B21/06Alarms for ensuring the safety of persons indicating a condition of sleep, e.g. anti-dozing alarms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/163Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state by tracking eye movement, gaze, or pupil change
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/18Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state for vehicle drivers or machine operators
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7221Determining signal validity, reliability or quality
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/08Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
    • B60W2040/0818Inactivity or incapacity of driver
    • B60W2040/0827Inactivity or incapacity of driver due to sleepiness
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W40/00Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
    • B60W40/08Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers

Definitions

  • the present invention relates to improvements in relation to e.g. vehicle implemented methods for automatic drowsiness detection, specifically to a method and corresponding system for classification of eye closures for improving the reliability of e.g. the method for drowsiness detection.
  • Traffic accidents often occur due to driver impairment caused by, for example, drowsiness.
  • driver impairment caused by, for example, drowsiness.
  • it may be vital to provide the driver with a warning message to re-establish the attention of the driver to the surrounding traffic situation, or in a critical situation to advice the driver to take a break or switch to another driver of the vehicle.
  • drowsiness detection algorithms that are based on detection of the driver's eye closure with a monitoring sensor, e.g. a camera or the like. These algorithms may be used in an image input unit that, by use of the monitoring sensor, detects the face and eyes of the driver.
  • the driver's gaze direction can be detected, e.g. the direction in which the driver of the vehicle is currently looking.
  • US 2003/039 378 An example of a system having an image input unit can be found in US 2003/039 378.
  • the image input unit of US 2003/039 378 inputs an image including the driver's face area to specify the driver's gaze or face direction.
  • the system generates an input pattern based on the received image input, for determining the driver's gaze or face direction.
  • a dictionary pattern comprising a plurality of stored targets, i.e. positions inside and outside the vehicle which the driver have looked at in the past.
  • the targets may, for example, be the side mirrors, back mirror, infotainment system, etc.
  • the input pattern generated by the system is compared to the stored targets in order to determine the current gaze or face direction of the driver.
  • the systems for detecting the driver's eyes have become more and more sophisticated, they do not take in account that there are other factors than drowsiness that may cause the driver to reduce the eye opening size or look away in a certain direction (e.g. due to changing lighting conditions within the vehicle compartment). This may, for example, also be caused by external factors such as direct sunlight, flickering light, headlights from oncoming traffic at nights, etc. Also, if the driver is looking down or sideways, the systems may detect this as an eye closure. There is hence a risk that the driver of the vehicle is provided with a warning message intended for a drowsy driver when, in fact, the driver is fully aware of the situation. It is therefore desirable to provide a method which discriminate between the actual factor(s) of a detected eye closure before a warning message is provided to the driver.
  • the above is at least partly met by a method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the method comprising receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least one of an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
  • the invention is based on the understanding that by comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, a robust and reliable method may be provided for determining if e.g. an identified eyelid closure, eye movement and/or head movement is correctly identified or is caused by e.g. measuring data imperfection. For example, if the received information from the image sensor is a flickering movement of the eyes, which is physically impossible, the comparison with the predetermined set of rules may determine that the received information is caused by a measuring data imperfection. This imperfection may be caused by, for instance, that the image sensor at the moment is unable to locate and provide an accurate image of the operator's eyes.
  • the image sensor may lose track of the operator's e.g. eyes and instead falsely identify an eyelid closure. Comparison with the predetermined set of rules may than determine that the identified eyelid closure in fact is caused by the driver looking over his shoulder and is not a correct eyelid closure. Hereby, the result of the comparison may be filtered out since it, in this example, was false.
  • the operator of the vehicle suddenly is being blinded by e.g. sunlight or the headlights of oncoming vehicles at night (i.e.
  • the operator might turn his head or close his eyelids in a manner that may be identified by the image sensor as an eyelid closure caused by e.g. drowsiness.
  • comparison with the predetermined set of rules may than filter out the registered eyelid closure since the lighting condition in the vehicle is such that it is likely that the operator has reacted to the bright light than and is thus aware of the surrounding traffic condition.
  • physiological data should in the following be interpreted as all type of data that may be identified by an image based system that identifies e.g. the operator's eyes, face, body, as well as eye gaze direction, eyelid closures, etc.
  • the wording “identifying an indication of an eyelid closure” should be interpreted as the image based system identifies that the operator is blinking (fast or slow) or closing his eyes.
  • the lighting condition within the operator compartment may be provided to the image based system by, for example, a separate lighting sensor or the image based system itself.
  • the separate lighting sensor may, for instance, be an infrared sensor (IR), near infrared sensor (NIR), camera with light sensitive characteristics, etc.
  • comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a confidence indication comprised with the physiological data being lower than a predetermined threshold, and providing an error classification of the physiological data if the confidence level is lower than the predetermined threshold.
  • an error classification can be made if the physiological data is behaving in a way that is more or less physically impossible, i.e. it is more likely that the image sensor lost track of the positions in the operator's eyes, arms and/or body.
  • “Physically impossible” behaviour may for example be if the physiological data indicates an unnaturally high frequency in head and/or body rotation and the like.
  • comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a change in a position and/or direction of the face comprised with the physiological data being larger than a predetermined threshold, and providing an error classification of the physiological data if the change in position and/or direction of the face is larger than the predetermined threshold. If, for example, the face of the vehicle operator is directed towards one of the rear view mirrors and the image sensor identifies an eyelid closure, an error classification may be provided since it is more likely that the image sensor lost track of the eyes and that the operator is fully aware of the traffic situation.
  • comparing the physiological data with the predetermined set of rules for a current operator status comprises estimating a matching level between an eyelid movement, eye movement and/or head movement comprised with the physiological data and a predetermined eyelid movement, eye movement and/or head movement template, and providing an error classification of the physiological data if the matching level is lower than a predetermined threshold.
  • comparing the physiological data with the predetermined set of rules for a current operator status comprises identifying at least one of eye, face and body movement representing a non-drowsy state of the operator, and providing a non-drowsiness classification of the physiological data if a non-drowsiness state is identified.
  • comparing the lighting condition within the operator compartment with the predetermined set of rules for a current operator status comprises identifying at least one of an illumination level towards the operator, and a changing lighting condition within the operator compartment.
  • identifying at least one of the illumination level towards the operator, and the changing lighting condition within the operator compartment comprises analyzing at least one of a contrast level, illumination level, and/or a frequency of illumination level of image data comprised with the physiological data from the image sensor, or from light sensor arranged within the operator compartment of the vehicle.
  • a control system for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle
  • the control device comprising a control unit configured to receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identify an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, compare at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classify the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
  • control system preferably is a driver assistance system for a vehicle, such as a car, a bus or a truck.
  • vehicle such as a car, a bus or a truck.
  • a computer readable medium embodying a computer program product for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle
  • the computer program product comprising code configured to, when executed by a processor receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
  • this aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
  • FIG. 1 illustrates a perspective view of the interior of a vehicle, equipped with internal sensors and a plurality of positions within the vehicle normally being gazed by the operator,
  • FIG. 2 illustrates a coordinate system of the face of a vehicle operator
  • FIG. 3 conceptually illustrates the logical elements of a control system according to a currently preferred embodiment of the invention.
  • FIG. 1 illustrates an operating compartment 114 of a vehicle, here illustrated as a car 100 .
  • the car 100 includes, in the illustrated embodiment, a vehicle operator (not shown) and an internal sensor, here illustrated as a camera system 102 .
  • the camera system 102 is arranged to determine the behaviour of the vehicle operator during vehicle operation. Furthermore, the camera system 102 may be arranged to focus on a predetermined number of positions of the operator's face, arms and body.
  • the camera system 102 may be pre-calibrated for a specific operator normally operating the car 100 or being calibrated each time an operator enters the driver seat of the car 100 . As the camera system 102 has detected the different positions of the operator's face, an estimation of facial behaviour is possible. The camera system 102 may hence detect, e.g. head and eye direction, head pose, eye saccade, head-eye saccade, eye closure, speed of eye closure, arm motions, body movements, etc. The camera system 102 may also, by use of a coordinate system 202 in connection to the operator's face 204 , illustrated in FIG.
  • the coordinate system 202 of the face 204 is preferably a polar coordinate system with its origin positioned between the eyes of the operator.
  • the car also comprises an interior light sensor 104 , in the described embodiment located behind the interior rear view mirror 106 , in the forward direction of the car 100 .
  • the exact positioning of the interior light sensor 100 may of course depend on the type of vehicle and its specific interior design. That is, in e.g. a truck there may not be any rear view mirror present and thus the interior light sensor must be placed at another suitable position.
  • the interior light sensor 104 is arranged to detect oncoming light to the operator compartment, such as bright sunlight or light from the headlights of oncoming traffic at night, which might cause the operator of the vehicle to react. Still further, there are a number of positions in the vehicle compartment which the operator of the vehicle on a normal basis is paying attention to. These are, for example, the exterior rear view mirrors 108 , the infotainment system 110 , the speed indicator 112 , the passenger next to the driver, the gear shift lever, etc.
  • FIG. 3 illustrating an exemplary embodiment of a system 300 for providing the method according to the present invention.
  • the system 300 comprises an image sensor 302 which receives information from the camera system 102 , an interior light sensor 304 , an identification of eyelid closure module 306 , a comparison operating module 308 and a drowsiness detection module 310 .
  • the eyelid closure module 306 and the comparison operating module 308 can be one and the same module and are only for illustrating purposes divided from each other.
  • the image sensor 302 receives physiological data of the vehicle operator and provides the information to the eyelid closure module 306 and to the comparison operating module 308 .
  • the comparison operating module also receives information regarding e.g.
  • the interior light sensor 304 receives information regarding the lighting condition within the operating compartment 114 , such as e.g. bright sunlight and headlight from oncoming traffic at night.
  • the interior light sensor 304 also provides the received lighting condition to the comparison operating module 308 .
  • the comparison operating module 308 compares the information with respect to a predetermined set of rules in order to evaluate the current status of the operator. Thereafter, the drowsiness detection module 310 receives the detected eyelid closure from the eyelid closure module 306 and a classification from the comparison operating module 308 regarding the actual cause(s) of the detected eyelid closure.
  • the predetermined set of rules may include the plurality of positions within the operating compartment 114 as described above, such that the system can determine, based on the identified eye-gaze direction and/or face direction, that the operator of the vehicle is currently paying attention to one of these positions.
  • the system may provide an error classification of the identified eyelid closure, since it is more likely that the operator is fully aware of the situation and that the detection of the eyelids most probably occurred due to the rotation of the operator's head, which the image sensor was unable to correctly follow.
  • the predetermined set of rules may also include certain patterns which are physically possible or impossible to occur. For example, if the image sensor detects that the eyes of the operator is moving irregular up/down or left/right in a way that seems implausible, a predetermined rule may classify the detected behaviour of the eyes as an error.
  • the comparison operating module 306 may provide an error classification of the eyelid closure, since the actual cause is more likely that the vehicle operator squinted with his eyes to compensate for the blinding light. It should be noted that the data from the image sensor 302 may be affected not only at a short (e.g. second(s)) period but over a longer time (e.g. minute(s)) as the image sensor 302 may e.g. lose its calibration due to the changing lighting condition.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Surgery (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Medical Informatics (AREA)
  • Psychiatry (AREA)
  • Molecular Biology (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Pathology (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Developmental Disabilities (AREA)
  • Child & Adolescent Psychology (AREA)
  • Social Psychology (AREA)
  • Psychology (AREA)
  • Hospice & Palliative Care (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Emergency Management (AREA)
  • Business, Economics & Management (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physiology (AREA)
  • Traffic Control Systems (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The present invention generally relates to a method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle. An embodiment of the method includes receiving, from the image sensor, physiological data including information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data; comparing at least one of the physiological data and a lighting condition within the operator compartment with a set of rules for a current operator status; and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.

Description

    PRIORITY STATEMENT
  • This claims priority under 35 U.S.C. §119 to European Patent Application No. 11179807.0, filed on Sep. 2, 2011, the contents of each of which are hereby incorporated by reference in their entirety.
  • TECHNICAL FIELD
  • The present invention relates to improvements in relation to e.g. vehicle implemented methods for automatic drowsiness detection, specifically to a method and corresponding system for classification of eye closures for improving the reliability of e.g. the method for drowsiness detection.
  • BACKGROUND OF THE INVENTION
  • Traffic accidents often occur due to driver impairment caused by, for example, drowsiness. In order to prevent accidents caused by driver impairment, it may be vital to provide the driver with a warning message to re-establish the attention of the driver to the surrounding traffic situation, or in a critical situation to advice the driver to take a break or switch to another driver of the vehicle.
  • Recently, much progress has been made in developing drowsiness detection algorithms that are based on detection of the driver's eye closure with a monitoring sensor, e.g. a camera or the like. These algorithms may be used in an image input unit that, by use of the monitoring sensor, detects the face and eyes of the driver. Hereby, the driver's gaze direction can be detected, e.g. the direction in which the driver of the vehicle is currently looking.
  • An example of a system having an image input unit can be found in US 2003/039 378. The image input unit of US 2003/039 378 inputs an image including the driver's face area to specify the driver's gaze or face direction. The system generates an input pattern based on the received image input, for determining the driver's gaze or face direction. Moreover, there is provided a dictionary pattern comprising a plurality of stored targets, i.e. positions inside and outside the vehicle which the driver have looked at in the past. The targets may, for example, be the side mirrors, back mirror, infotainment system, etc. Furthermore, the input pattern generated by the system is compared to the stored targets in order to determine the current gaze or face direction of the driver.
  • However, even though the systems for detecting the driver's eyes have become more and more sophisticated, they do not take in account that there are other factors than drowsiness that may cause the driver to reduce the eye opening size or look away in a certain direction (e.g. due to changing lighting conditions within the vehicle compartment). This may, for example, also be caused by external factors such as direct sunlight, flickering light, headlights from oncoming traffic at nights, etc. Also, if the driver is looking down or sideways, the systems may detect this as an eye closure. There is hence a risk that the driver of the vehicle is provided with a warning message intended for a drowsy driver when, in fact, the driver is fully aware of the situation. It is therefore desirable to provide a method which discriminate between the actual factor(s) of a detected eye closure before a warning message is provided to the driver.
  • SUMMARY OF THE INVENTION
  • According to an aspect of the invention, the above is at least partly met by a method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the method comprising receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least one of an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
  • The invention is based on the understanding that by comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, a robust and reliable method may be provided for determining if e.g. an identified eyelid closure, eye movement and/or head movement is correctly identified or is caused by e.g. measuring data imperfection. For example, if the received information from the image sensor is a flickering movement of the eyes, which is physically impossible, the comparison with the predetermined set of rules may determine that the received information is caused by a measuring data imperfection. This imperfection may be caused by, for instance, that the image sensor at the moment is unable to locate and provide an accurate image of the operator's eyes. According to another example, if the operator of the vehicle suddenly looks over his shoulder to see if he is being overtaken by another vehicle, the image sensor may lose track of the operator's e.g. eyes and instead falsely identify an eyelid closure. Comparison with the predetermined set of rules may than determine that the identified eyelid closure in fact is caused by the driver looking over his shoulder and is not a correct eyelid closure. Hereby, the result of the comparison may be filtered out since it, in this example, was false. According to yet another example, if the operator of the vehicle suddenly is being blinded by e.g. sunlight or the headlights of oncoming vehicles at night (i.e. light entering the compartment from the outside), the operator might turn his head or close his eyelids in a manner that may be identified by the image sensor as an eyelid closure caused by e.g. drowsiness. Hence, comparison with the predetermined set of rules may than filter out the registered eyelid closure since the lighting condition in the vehicle is such that it is likely that the operator has reacted to the bright light than and is thus aware of the surrounding traffic condition.
  • The wording “physiological data” should in the following be interpreted as all type of data that may be identified by an image based system that identifies e.g. the operator's eyes, face, body, as well as eye gaze direction, eyelid closures, etc. Moreover, the wording “identifying an indication of an eyelid closure” should be interpreted as the image based system identifies that the operator is blinking (fast or slow) or closing his eyes. Furthermore, the lighting condition within the operator compartment may be provided to the image based system by, for example, a separate lighting sensor or the image based system itself. The separate lighting sensor may, for instance, be an infrared sensor (IR), near infrared sensor (NIR), camera with light sensitive characteristics, etc.
  • Furthermore, comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a confidence indication comprised with the physiological data being lower than a predetermined threshold, and providing an error classification of the physiological data if the confidence level is lower than the predetermined threshold. Hereby, an error classification can be made if the physiological data is behaving in a way that is more or less physically impossible, i.e. it is more likely that the image sensor lost track of the positions in the operator's eyes, arms and/or body. “Physically impossible” behaviour may for example be if the physiological data indicates an unnaturally high frequency in head and/or body rotation and the like.
  • Still further, comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a change in a position and/or direction of the face comprised with the physiological data being larger than a predetermined threshold, and providing an error classification of the physiological data if the change in position and/or direction of the face is larger than the predetermined threshold. If, for example, the face of the vehicle operator is directed towards one of the rear view mirrors and the image sensor identifies an eyelid closure, an error classification may be provided since it is more likely that the image sensor lost track of the eyes and that the operator is fully aware of the traffic situation.
  • According to a further embodiment, comparing the physiological data with the predetermined set of rules for a current operator status comprises estimating a matching level between an eyelid movement, eye movement and/or head movement comprised with the physiological data and a predetermined eyelid movement, eye movement and/or head movement template, and providing an error classification of the physiological data if the matching level is lower than a predetermined threshold.
  • According to a further embodiment, wherein comparing the physiological data with the predetermined set of rules for a current operator status comprises identifying at least one of eye, face and body movement representing a non-drowsy state of the operator, and providing a non-drowsiness classification of the physiological data if a non-drowsiness state is identified.
  • According to a further embodiment, comparing the lighting condition within the operator compartment with the predetermined set of rules for a current operator status comprises identifying at least one of an illumination level towards the operator, and a changing lighting condition within the operator compartment.
  • According to a further embodiment, identifying at least one of the illumination level towards the operator, and the changing lighting condition within the operator compartment comprises analyzing at least one of a contrast level, illumination level, and/or a frequency of illumination level of image data comprised with the physiological data from the image sensor, or from light sensor arranged within the operator compartment of the vehicle.
  • According to another aspect of the invention there is provided a control system for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the control device comprising a control unit configured to receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identify an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, compare at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classify the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison. This aspect of the invention provides similar advantages as discussed above in relation to the previous aspect of the invention.
  • It should be noted that the control system preferably is a driver assistance system for a vehicle, such as a car, a bus or a truck. A more thorough discussion will be given below in relation to the detailed description of the invention. Other types of vehicles are of course possible and within the scope of the invention.
  • According to a still further aspect of the invention there is provided a computer readable medium embodying a computer program product for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the computer program product comprising code configured to, when executed by a processor receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison. Also this aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
  • Further features of, and advantages with, the present invention will become apparent when studying the appended claims and the following description. The skilled addressee realize that different features of the present invention may be combined to create embodiments other than those described in the following, without departing from the scope of the present invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The various aspects of the invention, including its particular features and advantages, will be readily understood from the following detailed description and the accompanying drawings, in which:
  • FIG. 1 illustrates a perspective view of the interior of a vehicle, equipped with internal sensors and a plurality of positions within the vehicle normally being gazed by the operator,
  • FIG. 2 illustrates a coordinate system of the face of a vehicle operator, and
  • FIG. 3 conceptually illustrates the logical elements of a control system according to a currently preferred embodiment of the invention.
  • DETAILED DESCRIPTION
  • The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which currently preferred embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided for thoroughness and completeness, and fully convey the scope of the invention to the skilled addressee. Like reference characters refer to like elements throughout.
  • In the following, the present invention is described with reference to a method for improving classification of eye closures of an operator of a vehicle. The vehicle is preferably equipped with interior sensor(s) for retrieving information of the vehicle operator. For the sake of better understanding, reference is now made to FIG. 1, which illustrates an operating compartment 114 of a vehicle, here illustrated as a car 100. The car 100 includes, in the illustrated embodiment, a vehicle operator (not shown) and an internal sensor, here illustrated as a camera system 102. The camera system 102 is arranged to determine the behaviour of the vehicle operator during vehicle operation. Furthermore, the camera system 102 may be arranged to focus on a predetermined number of positions of the operator's face, arms and body. These positions may, for example, be the eyes, eyelids, eyebrows, nose, mouth, cheek, etc. The camera system 102 may be pre-calibrated for a specific operator normally operating the car 100 or being calibrated each time an operator enters the driver seat of the car 100. As the camera system 102 has detected the different positions of the operator's face, an estimation of facial behaviour is possible. The camera system 102 may hence detect, e.g. head and eye direction, head pose, eye saccade, head-eye saccade, eye closure, speed of eye closure, arm motions, body movements, etc. The camera system 102 may also, by use of a coordinate system 202 in connection to the operator's face 204, illustrated in FIG. 2, detect if the head of the operator is rotating to the right or left (yaw) 205, rotating up or down (pitch) 206 or leaning towards the right or left shoulder (roll) 207. The coordinate system 202 of the face 204 is preferably a polar coordinate system with its origin positioned between the eyes of the operator.
  • Furthermore, the car also comprises an interior light sensor 104, in the described embodiment located behind the interior rear view mirror 106, in the forward direction of the car 100. The exact positioning of the interior light sensor 100 may of course depend on the type of vehicle and its specific interior design. That is, in e.g. a truck there may not be any rear view mirror present and thus the interior light sensor must be placed at another suitable position.
  • The interior light sensor 104 is arranged to detect oncoming light to the operator compartment, such as bright sunlight or light from the headlights of oncoming traffic at night, which might cause the operator of the vehicle to react. Still further, there are a number of positions in the vehicle compartment which the operator of the vehicle on a normal basis is paying attention to. These are, for example, the exterior rear view mirrors 108, the infotainment system 110, the speed indicator 112, the passenger next to the driver, the gear shift lever, etc.
  • Now referring to FIG. 3, illustrating an exemplary embodiment of a system 300 for providing the method according to the present invention. The system 300 comprises an image sensor 302 which receives information from the camera system 102, an interior light sensor 304, an identification of eyelid closure module 306, a comparison operating module 308 and a drowsiness detection module 310. It should however be noted that the eyelid closure module 306 and the comparison operating module 308 can be one and the same module and are only for illustrating purposes divided from each other. Moreover, the image sensor 302 receives physiological data of the vehicle operator and provides the information to the eyelid closure module 306 and to the comparison operating module 308. The comparison operating module also receives information regarding e.g. operator eyes, arms and body motion from the image sensor 302. Furthermore, the interior light sensor 304 receives information regarding the lighting condition within the operating compartment 114, such as e.g. bright sunlight and headlight from oncoming traffic at night. The interior light sensor 304 also provides the received lighting condition to the comparison operating module 308.
  • Moreover, when the comparison operating module 308 has received information regarding the lighting condition within the operating compartment 114 as well as the physiological data regarding eyes, arms and body motion of the operator, the comparison operating module 308 compares the information with respect to a predetermined set of rules in order to evaluate the current status of the operator. Thereafter, the drowsiness detection module 310 receives the detected eyelid closure from the eyelid closure module 306 and a classification from the comparison operating module 308 regarding the actual cause(s) of the detected eyelid closure. According to an example, the predetermined set of rules may include the plurality of positions within the operating compartment 114 as described above, such that the system can determine, based on the identified eye-gaze direction and/or face direction, that the operator of the vehicle is currently paying attention to one of these positions. For instance, if the system identifies an eyelid closure of the operator, and at the same time identifies that the operator's face is directed towards the left rear view mirror 108, the system may provide an error classification of the identified eyelid closure, since it is more likely that the operator is fully aware of the situation and that the detection of the eyelids most probably occurred due to the rotation of the operator's head, which the image sensor was unable to correctly follow. Furthermore, the predetermined set of rules may also include certain patterns which are physically possible or impossible to occur. For example, if the image sensor detects that the eyes of the operator is moving irregular up/down or left/right in a way that seems implausible, a predetermined rule may classify the detected behaviour of the eyes as an error.
  • According to a further example, if the interior light sensor 304 detects a bright light towards the vehicle operator at the same time as the image sensor 302 detects an eyelid closure of the operator, the comparison operating module 306 may provide an error classification of the eyelid closure, since the actual cause is more likely that the vehicle operator squinted with his eyes to compensate for the blinding light. It should be noted that the data from the image sensor 302 may be affected not only at a short (e.g. second(s)) period but over a longer time (e.g. minute(s)) as the image sensor 302 may e.g. lose its calibration due to the changing lighting condition.
  • Even though the invention has been described with reference to specific exemplifying embodiments thereof, many different alterations, modifications and the like will become apparent for those skilled in the art. Variations to the disclosed embodiments can be understood and effected by the skilled addressee in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims. For example, the invention has mainly been described above with reference to a few embodiments. However, as is readily appreciated by the skilled addressee, other embodiments than the ones disclosed above are equally possible within the scope of the invention, as defined by the appended patent claims. For example, the invention is also applicable for trucks, buses, dumpers, wheel loaders and other type of vehicles than the above described car.
  • In the claims, the word “comprises” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. A single computer or other unit may fulfil the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measured cannot be used to advantage.

Claims (12)

1. A method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the method comprising:
receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator;
identifying an indication of at least one of an eyelid closure, eye movement and head movement of the operator based on the physiological data;
comparing at least one of the physiological data and a lighting condition within the operator compartment with a set of rules for a current operator status; and
classifying a type of at least one of eyelid closure, eye movement and head movement by correlating the identified indication of the at least one of eyelid closure, eye movement and head movement and a result of the comparison.
2. Method according to claim 1, wherein comparing the physiological data with the set of rules for a current operator status comprises identifying a confidence indication comprised with the physiological data being lower than a threshold, and providing an error classification of the physiological data if the confidence level is lower than the threshold.
3. Method according to claim 1, wherein comparing the physiological data with the set of rules for a current operator status comprises identifying a change in at least one of a position and direction of the face comprised with the physiological data being larger than a threshold, and providing an error classification of the physiological data if the change in at least one of position and direction of the face is larger than the threshold.
4. Method according to claim 1, wherein comparing the physiological data with the set of rules for a current operator status comprises estimating a matching level between at least one of an eyelid movement, eye movement and head movement comprised with the physiological data and at least one of an eyelid movement, eye movement and head movement template, and providing an error classification of the physiological data if the matching level is lower than a threshold.
5. Method according to claim 1, wherein comparing the physiological data with the set of rules for a current operator status comprises identifying at least one of eye, face and body movement representing a non-drowsy state of the operator, and providing a non-drowsiness classification of the physiological data if a non-drowsiness state is identified.
6. Method according to claim 1, wherein comparing the lighting condition within the operator compartment with the set of rules for a current operator status comprises identifying at least one of an illumination level towards the operator, and a changing lighting condition within the operator compartment.
7. Method according to claim 6, wherein identifying at least one of the illumination level towards the operator, and the changing lighting condition within the operator compartment comprises analyzing at least one of a contrast level, illumination level, and a frequency of illumination level of image data comprised with
the physiological data from the image sensor, or
from light sensor arranged within the operator compartment of the vehicle.
8. A control system for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the control device comprising a control unit configured to:
receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator;
identify an indication of at least one of an eyelid closure, eye movement, and head movement of the operator based on the physiological data;
compare at least one of the physiological data and a lighting condition within the operator compartment with a set of rules for a current operator status; and
classify the type of at least one of eyelid closure, eye movement and head movement by correlating the identified at least one of eyelid closure, eye movement and head movement and a result of the comparison.
9. Control system according to claim 8, wherein the control unit is further configured to, when comparing the physiological data with the set of rules for a current operator status, identify a confidence indication comprised with the physiological data being lower than a threshold, and providing an error classification of the physiological data if the confidence level is lower than the threshold.
10. Control system according to claim 8, wherein the control unit is further configured to, when comparing the physiological data with the set of rules for a current operator status, identify a change in at least one of a position and direction of the face comprised with the physiological data being larger than a threshold, and providing an error classification of the physiological data if the change in at least one of position and direction of the face is larger than the threshold.
11. Control system according to claim 8, wherein the control system is a driver assistance system for a vehicle.
12. A computer readable medium embodying a computer program product for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the computer program product comprising code configured to, when executed by a processor:
receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator;
identify an indication of at least one of an eyelid closure, eye movement and head movement of the operator based on the physiological data;
compare at least one of the physiological data and a lighting condition within the operator compartment with a set of rules for a current operator status; and
classify the type of at least one of eyelid closure, eye movement and head movement by correlating the identified at least one of eyelid closure, eye movement and head movement and a result of the comparison.
US13/558,780 2011-09-02 2012-07-26 Method for classification of eye closures Active 2033-10-19 US9117358B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
EP11179807.0A EP2564777B1 (en) 2011-09-02 2011-09-02 Method for classification of eye closures
EP11179807.0 2011-09-02
EP11179807 2011-09-02

Publications (2)

Publication Number Publication Date
US20130057671A1 true US20130057671A1 (en) 2013-03-07
US9117358B2 US9117358B2 (en) 2015-08-25

Family

ID=44785312

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/558,780 Active 2033-10-19 US9117358B2 (en) 2011-09-02 2012-07-26 Method for classification of eye closures

Country Status (5)

Country Link
US (1) US9117358B2 (en)
EP (1) EP2564777B1 (en)
JP (1) JP2013054735A (en)
CN (1) CN103150545B (en)
BR (1) BR102012022143A2 (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140210978A1 (en) * 2013-01-25 2014-07-31 Toyota Motor Engineering & Manufacturing North America, Inc. Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US20160167661A1 (en) * 2013-07-19 2016-06-16 Audi Ag Method for operating a driver assistance system of a motor vehicle and driver assistance system for a motor vehicle
US9771081B2 (en) * 2014-09-29 2017-09-26 The Boeing Company System for fatigue detection using a suite of physiological measurement devices
US9908530B1 (en) 2014-04-17 2018-03-06 State Farm Mutual Automobile Insurance Company Advanced vehicle operator intelligence system
WO2018042122A1 (en) 2016-08-29 2018-03-08 Renault S.A.S. Method for determining the placement of the head of a vehicle driver
US9934667B1 (en) 2014-03-07 2018-04-03 State Farm Mutual Automobile Insurance Company Vehicle operator emotion management system and method
US10118487B1 (en) * 2014-05-05 2018-11-06 State Farm Mutual Automobile Insurance Company System and method to monitor and alert vehicle operator of impairment
US10475127B1 (en) 2014-07-21 2019-11-12 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and insurance incentives
US20200103900A1 (en) * 2018-09-30 2020-04-02 Strong Force Intellectual Capital, Llc Intelligent transportation systems
CN115352450A (en) * 2022-08-12 2022-11-18 重庆长安汽车股份有限公司 Vehicle driving assistance method and device, vehicle and storage medium
CN115771387A (en) * 2021-09-06 2023-03-10 博泰车联网(南京)有限公司 Sun visor control method and device
US20230105891A1 (en) * 2020-02-06 2023-04-06 Mitsubishi Electric Corporation Sign detection device, driving assistance control device, and sign detection method
US11868127B2 (en) 2018-09-30 2024-01-09 Strong Force Tp Portfolio 2022, Llc Radial basis function neural network optimizing operating parameter of vehicle based on emotional state of rider determined by recurrent neural network
US20240010222A1 (en) * 2022-01-14 2024-01-11 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods to increase driver awareness of exterior occurrences
EP3075315B1 (en) * 2015-04-02 2024-02-07 Essilor International System and computer-implemented method for monitoring the visual behavior of a person
WO2024260940A1 (en) * 2023-06-21 2024-12-26 Bayerische Motoren Werke Aktiengesellschaft Assistance system for supporting a driver of a vehicle during a lateral guidance of the vehicle while taking into account a dazzling of the driver, vehicle, and method

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101655818B1 (en) * 2014-12-11 2016-09-08 현대자동차주식회사 Wearable glass, control method thereof and vehicle control system
EP3440592B1 (en) * 2016-04-07 2023-08-30 Seeing Machines Limited Method and system of distinguishing between a glance event and an eye closure event
US10821987B2 (en) 2016-07-20 2020-11-03 Ford Global Technologies, Llc Vehicle interior and exterior monitoring
WO2018227597A1 (en) * 2017-06-16 2018-12-20 Boe Technology Group Co., Ltd. Vision-based interactive control apparatus and method of controlling rear-view mirror for vehicle
JP7099037B2 (en) * 2018-05-07 2022-07-12 オムロン株式会社 Data processing equipment, monitoring system, awakening system, data processing method, and data processing program
JP7099036B2 (en) * 2018-05-07 2022-07-12 オムロン株式会社 Data processing equipment, monitoring system, awakening system, data processing method, and data processing program
US20220117529A1 (en) * 2020-10-20 2022-04-21 Harman Becker Automotive Systems Gmbh System and method for determining an eye movement
DE102020214910A1 (en) * 2020-11-27 2022-06-02 Robert Bosch Gesellschaft mit beschränkter Haftung Method for monitoring a vehicle interior
CN112668548B (en) * 2021-01-15 2024-07-12 重庆大学 Method and system for detecting driver's blurriness
CN114162119A (en) * 2021-10-27 2022-03-11 广州广日电气设备有限公司 Lateral control method, equipment, medium and product of automobile advanced driving auxiliary system
CN118372835B (en) * 2024-06-24 2024-09-06 江西科技学院 Fatigue driving identification method and system based on eye movement characteristics

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6097295A (en) * 1998-01-28 2000-08-01 Daimlerchrysler Ag Apparatus for determining the alertness of a driver
US20050073136A1 (en) * 2002-10-15 2005-04-07 Volvo Technology Corporation Method and arrangement for interpreting a subjects head and eye activity
US20080180235A1 (en) * 2007-01-25 2008-07-31 Hsuan Chang Method and apparatus for manipulating driver core temperature to enhance driver alertness
US20080238694A1 (en) * 2007-03-26 2008-10-02 Denso Corporation Drowsiness alarm apparatus and program
US20080252466A1 (en) * 2007-04-11 2008-10-16 Yopp W Trent System and method for implementing active safety counter measures for an impaired driver
US20090097701A1 (en) * 2007-10-11 2009-04-16 Denso Corporation Sleepiness level determination device for driver
US20100214105A1 (en) * 2009-02-24 2010-08-26 Panasonic Automotive Systems Company Of America, Division Of Panasonic Corporation Of North America Method of detecting drowsiness of a vehicle operator
US20110060496A1 (en) * 2009-08-11 2011-03-10 Certusview Technologies, Llc Systems and methods for complex event processing of vehicle information and image information relating to a vehicle

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6154559A (en) * 1998-10-01 2000-11-28 Mitsubishi Electric Information Technology Center America, Inc. (Ita) System for classifying an individual's gaze direction
JP3790680B2 (en) 2001-05-25 2006-06-28 株式会社東芝 Image processing system and driving support system using the same
US6927694B1 (en) * 2001-08-20 2005-08-09 Research Foundation Of The University Of Central Florida Algorithm for monitoring head/eye motion for driver alertness with one camera
KR100436861B1 (en) 2001-08-27 2004-06-30 나노메트릭스코리아 주식회사 Method and apparatus for inspecting defects on polishing pad to be used with chemical mechanical polishing apparatus
CN1956680A (en) * 2004-05-24 2007-05-02 日本先锋公司 Biological information detection device
CN100462047C (en) * 2007-03-21 2009-02-18 汤一平 Safe driving auxiliary device based on omnidirectional computer vision
CN100462046C (en) * 2007-04-16 2009-02-18 孙军 Online early warning system and method of weariness working
CN101030316B (en) * 2007-04-17 2010-04-21 北京中星微电子有限公司 Safety driving monitoring system and method for vehicle
CN101655907B (en) * 2009-08-28 2013-03-20 大连鑫奇辉科技有限公司 Trainman driving state monitoring intelligent alarm system

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6097295A (en) * 1998-01-28 2000-08-01 Daimlerchrysler Ag Apparatus for determining the alertness of a driver
US20050073136A1 (en) * 2002-10-15 2005-04-07 Volvo Technology Corporation Method and arrangement for interpreting a subjects head and eye activity
US20080180235A1 (en) * 2007-01-25 2008-07-31 Hsuan Chang Method and apparatus for manipulating driver core temperature to enhance driver alertness
US20080238694A1 (en) * 2007-03-26 2008-10-02 Denso Corporation Drowsiness alarm apparatus and program
US20080252466A1 (en) * 2007-04-11 2008-10-16 Yopp W Trent System and method for implementing active safety counter measures for an impaired driver
US20090097701A1 (en) * 2007-10-11 2009-04-16 Denso Corporation Sleepiness level determination device for driver
US20100214105A1 (en) * 2009-02-24 2010-08-26 Panasonic Automotive Systems Company Of America, Division Of Panasonic Corporation Of North America Method of detecting drowsiness of a vehicle operator
US20110060496A1 (en) * 2009-08-11 2011-03-10 Certusview Technologies, Llc Systems and methods for complex event processing of vehicle information and image information relating to a vehicle
US20110093162A1 (en) * 2009-08-11 2011-04-21 Certusview Technologies, Llc Systems and methods for complex event processing of vehicle-related information
US8463487B2 (en) * 2009-08-11 2013-06-11 Certusview Technologies, Llc Systems and methods for complex event processing based on a hierarchical arrangement of complex event processing engines

Cited By (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140210978A1 (en) * 2013-01-25 2014-07-31 Toyota Motor Engineering & Manufacturing North America, Inc. Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US8847771B2 (en) * 2013-01-25 2014-09-30 Toyota Motor Engineering & Manufacturing North America, Inc. Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US20140340228A1 (en) * 2013-01-25 2014-11-20 Toyota Motor Engineering & Manufacturing North America, Inc. Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US9299237B2 (en) * 2013-01-25 2016-03-29 Toyota Jidosha Kabushiki Kaisha Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US20160171322A1 (en) * 2013-01-25 2016-06-16 Toyota Jidosha Kabushiki Kaisha Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US9824286B2 (en) * 2013-01-25 2017-11-21 Toyota Jidosha Kabushiki Kaisha Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning
US20160167661A1 (en) * 2013-07-19 2016-06-16 Audi Ag Method for operating a driver assistance system of a motor vehicle and driver assistance system for a motor vehicle
US9934667B1 (en) 2014-03-07 2018-04-03 State Farm Mutual Automobile Insurance Company Vehicle operator emotion management system and method
US10593182B1 (en) 2014-03-07 2020-03-17 State Farm Mutual Automobile Insurance Company Vehicle operator emotion management system and method
US10121345B1 (en) 2014-03-07 2018-11-06 State Farm Mutual Automobile Insurance Company Vehicle operator emotion management system and method
US9908530B1 (en) 2014-04-17 2018-03-06 State Farm Mutual Automobile Insurance Company Advanced vehicle operator intelligence system
US10118488B1 (en) 2014-05-05 2018-11-06 State Farm Mutual Automobile Insurance Co. System and method to monitor and alert vehicle operator of impairment
US10569650B1 (en) 2014-05-05 2020-02-25 State Farm Mutual Automobile Insurance Company System and method to monitor and alert vehicle operator of impairment
US10118487B1 (en) * 2014-05-05 2018-11-06 State Farm Mutual Automobile Insurance Company System and method to monitor and alert vehicle operator of impairment
US11068995B1 (en) 2014-07-21 2021-07-20 State Farm Mutual Automobile Insurance Company Methods of reconstructing an accident scene using telematics data
US11634103B2 (en) 2014-07-21 2023-04-25 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US10475127B1 (en) 2014-07-21 2019-11-12 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and insurance incentives
US12365308B2 (en) 2014-07-21 2025-07-22 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US12358463B2 (en) 2014-07-21 2025-07-15 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and driving behavior identification
US10723312B1 (en) 2014-07-21 2020-07-28 State Farm Mutual Automobile Insurance Company Methods of theft prevention or mitigation
US10825326B1 (en) 2014-07-21 2020-11-03 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US10832327B1 (en) 2014-07-21 2020-11-10 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and driving behavior identification
US10974693B1 (en) 2014-07-21 2021-04-13 State Farm Mutual Automobile Insurance Company Methods of theft prevention or mitigation
US10997849B1 (en) 2014-07-21 2021-05-04 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US11030696B1 (en) 2014-07-21 2021-06-08 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and anonymous driver data
US11069221B1 (en) 2014-07-21 2021-07-20 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US12179695B2 (en) 2014-07-21 2024-12-31 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US11257163B1 (en) 2014-07-21 2022-02-22 State Farm Mutual Automobile Insurance Company Methods of pre-generating insurance claims
US12151644B2 (en) 2014-07-21 2024-11-26 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US11634102B2 (en) 2014-07-21 2023-04-25 State Farm Mutual Automobile Insurance Company Methods of facilitating emergency assistance
US11565654B2 (en) 2014-07-21 2023-01-31 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and driving behavior identification
US10540723B1 (en) 2014-07-21 2020-01-21 State Farm Mutual Automobile Insurance Company Methods of providing insurance savings based upon telematics and usage-based insurance
US9771081B2 (en) * 2014-09-29 2017-09-26 The Boeing Company System for fatigue detection using a suite of physiological measurement devices
EP3075315B1 (en) * 2015-04-02 2024-02-07 Essilor International System and computer-implemented method for monitoring the visual behavior of a person
WO2018042122A1 (en) 2016-08-29 2018-03-08 Renault S.A.S. Method for determining the placement of the head of a vehicle driver
US12298759B2 (en) 2018-09-30 2025-05-13 Strong Force Tp Portfolio 2022, Llc Using social media data of a vehicle occupant to alter a route plan of the vehicle
US12228924B2 (en) 2018-09-30 2025-02-18 Strong Force Tp Portfolio 2022, Llc Social data sources feeding a neural network to predict an emerging condition relevant to a transportation plan of at least one individual
US11868126B2 (en) 2018-09-30 2024-01-09 Strong Force Tp Portfolio 2022, Llc Wearable device determining emotional state of rider in vehicle and optimizing operating parameter of vehicle to improve emotional state of rider
US20200103900A1 (en) * 2018-09-30 2020-04-02 Strong Force Intellectual Capital, Llc Intelligent transportation systems
US11694288B2 (en) 2018-09-30 2023-07-04 Strong Force Tp Portfolio 2022, Llc Method of optimizing rider satisfaction
US12124257B2 (en) 2018-09-30 2024-10-22 Strong Force Tp Portfolio 2022, Llc Intelligent transportation systems
US11486721B2 (en) * 2018-09-30 2022-11-01 Strong Force Intellectual Capital, Llc Intelligent transportation systems
US12321169B2 (en) 2018-09-30 2025-06-03 Strong Force Tp Portfolio 2022, Llc Optimizing a vehicle operating parameter based in part on a sensed emotional state of a rider
US12321168B2 (en) 2018-09-30 2025-06-03 Strong Force Tp Portfolio 2022, Llc Transportation system to optimize an operating parameter of a vehicle based on an emotional state of an occupant of the vehicle determined from a sensor to detect a physiological condition of the occupant
US11868127B2 (en) 2018-09-30 2024-01-09 Strong Force Tp Portfolio 2022, Llc Radial basis function neural network optimizing operating parameter of vehicle based on emotional state of rider determined by recurrent neural network
US12235641B2 (en) 2018-09-30 2025-02-25 Strong Force Tp Portfolio 2022, Llc Hybrid neural networks sourcing social data sources to optimize satisfaction of rider in intelligent transportation systems
US12298760B2 (en) 2018-09-30 2025-05-13 Strong Force Tp Portfolio 2022, Llc Neural net optimization of continuously variable powertrain
US20230105891A1 (en) * 2020-02-06 2023-04-06 Mitsubishi Electric Corporation Sign detection device, driving assistance control device, and sign detection method
CN115771387A (en) * 2021-09-06 2023-03-10 博泰车联网(南京)有限公司 Sun visor control method and device
US12240483B2 (en) * 2022-01-14 2025-03-04 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods to increase driver awareness of exterior occurrences
US20240010222A1 (en) * 2022-01-14 2024-01-11 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods to increase driver awareness of exterior occurrences
CN115352450A (en) * 2022-08-12 2022-11-18 重庆长安汽车股份有限公司 Vehicle driving assistance method and device, vehicle and storage medium
WO2024260940A1 (en) * 2023-06-21 2024-12-26 Bayerische Motoren Werke Aktiengesellschaft Assistance system for supporting a driver of a vehicle during a lateral guidance of the vehicle while taking into account a dazzling of the driver, vehicle, and method

Also Published As

Publication number Publication date
EP2564777B1 (en) 2017-06-07
JP2013054735A (en) 2013-03-21
BR102012022143A2 (en) 2013-10-01
CN103150545B (en) 2017-04-12
EP2564777A1 (en) 2013-03-06
US9117358B2 (en) 2015-08-25
CN103150545A (en) 2013-06-12

Similar Documents

Publication Publication Date Title
US9117358B2 (en) Method for classification of eye closures
US9616809B1 (en) Lane change prediction and turn signal activation upon observation of head and eye movement
JP6087088B2 (en) System and method for improving vehicle operator performance estimates
CN110532976A (en) Method for detecting fatigue driving and system based on machine learning and multiple features fusion
US20060287779A1 (en) Method of mitigating driver distraction
US20180197030A1 (en) Driver abnormality detection device and driver abnormality detection method
CN101968917A (en) Vehicle-mounted warning apparatus
WO2007092512A2 (en) Driver drowsiness and distraction monitor
CN114872713A (en) Device and method for monitoring abnormal driving state of driver
CN109969195A (en) An alarm system and method for abnormal driver facial behavior and a vehicle based thereon
KR20120074820A (en) Control system for vehicle using face recognition function
KR101500016B1 (en) Lane Departure Warning System
JP2015085719A (en) Gazing object estimation device
US12179787B2 (en) Pedestrian alert system
Bergasa et al. Visual monitoring of driver inattention
US20220319200A1 (en) Device and method for determining image data of the eyes, eye positions and/or a viewing direction of a vehicle user in a vehicle
WO2023017595A1 (en) Occupant state determining device, occupant state determining method, and occupant state determining system
US20230159010A1 (en) Driver assistance device
JP7612118B2 (en) Drowsiness estimation device and drowsiness estimation method
Su et al. Image-Based Driver Status Monitoring System for Determining the Transfer of Dynamic Driving Tasks in Autonomous Vehicles
JP2023027699A (en) Driver state determination device
JP7091632B2 (en) Safe driving promotion device and safe driving promotion method
US12403818B2 (en) Illumination device for vehicle
WO2025004257A1 (en) Drowsiness estimation device and drowsiness estimation method
EP4510089A1 (en) Physical driver condition determination for vehicle safety

Legal Events

Date Code Title Description
AS Assignment

Owner name: VOLVO CAR CORPORATION, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEVIN, DANIEL;WESTERVALL, LENA;REEL/FRAME:029120/0394

Effective date: 20120817

Owner name: VOLVO TECHNOLOGY CORPORATION, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEANDERSSON, SUSANNA;KRONBERG, PETER;SIGNING DATES FROM 20120927 TO 20121002;REEL/FRAME:029120/0582

AS Assignment

Owner name: VOLVO CAR CORPORATION, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VOLVO TECHNOLOGY CORPORATION;REEL/FRAME:034350/0686

Effective date: 20141120

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8