US20130057671A1 - Method for classification of eye closures - Google Patents
Method for classification of eye closures Download PDFInfo
- Publication number
- US20130057671A1 US20130057671A1 US13/558,780 US201213558780A US2013057671A1 US 20130057671 A1 US20130057671 A1 US 20130057671A1 US 201213558780 A US201213558780 A US 201213558780A US 2013057671 A1 US2013057671 A1 US 2013057671A1
- Authority
- US
- United States
- Prior art keywords
- operator
- physiological data
- movement
- eye
- rules
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 20
- 230000004399 eye closure Effects 0.000 title description 8
- 210000000744 eyelid Anatomy 0.000 claims abstract description 47
- 230000004424 eye movement Effects 0.000 claims abstract description 26
- 230000004886 head movement Effects 0.000 claims abstract description 26
- 230000033001 locomotion Effects 0.000 claims abstract description 18
- 210000003128 head Anatomy 0.000 claims abstract description 13
- 238000012544 monitoring process Methods 0.000 claims abstract description 9
- 206010041349 Somnolence Diseases 0.000 claims description 12
- 238000005286 illumination Methods 0.000 claims description 8
- 230000008859 change Effects 0.000 claims description 6
- 238000004590 computer program Methods 0.000 claims description 4
- 238000001514 detection method Methods 0.000 description 7
- 230000008901 benefit Effects 0.000 description 5
- 230000006735 deficit Effects 0.000 description 2
- 230000004434 saccadic eye movement Effects 0.000 description 2
- 206010039203 Road traffic accident Diseases 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000004397 blinking Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 210000004709 eyebrow Anatomy 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/06—Alarms for ensuring the safety of persons indicating a condition of sleep, e.g. anti-dozing alarms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/163—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state by tracking eye movement, gaze, or pupil change
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/18—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state for vehicle drivers or machine operators
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7221—Determining signal validity, reliability or quality
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
- B60W2040/0818—Inactivity or incapacity of driver
- B60W2040/0827—Inactivity or incapacity of driver due to sleepiness
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
Definitions
- the present invention relates to improvements in relation to e.g. vehicle implemented methods for automatic drowsiness detection, specifically to a method and corresponding system for classification of eye closures for improving the reliability of e.g. the method for drowsiness detection.
- Traffic accidents often occur due to driver impairment caused by, for example, drowsiness.
- driver impairment caused by, for example, drowsiness.
- it may be vital to provide the driver with a warning message to re-establish the attention of the driver to the surrounding traffic situation, or in a critical situation to advice the driver to take a break or switch to another driver of the vehicle.
- drowsiness detection algorithms that are based on detection of the driver's eye closure with a monitoring sensor, e.g. a camera or the like. These algorithms may be used in an image input unit that, by use of the monitoring sensor, detects the face and eyes of the driver.
- the driver's gaze direction can be detected, e.g. the direction in which the driver of the vehicle is currently looking.
- US 2003/039 378 An example of a system having an image input unit can be found in US 2003/039 378.
- the image input unit of US 2003/039 378 inputs an image including the driver's face area to specify the driver's gaze or face direction.
- the system generates an input pattern based on the received image input, for determining the driver's gaze or face direction.
- a dictionary pattern comprising a plurality of stored targets, i.e. positions inside and outside the vehicle which the driver have looked at in the past.
- the targets may, for example, be the side mirrors, back mirror, infotainment system, etc.
- the input pattern generated by the system is compared to the stored targets in order to determine the current gaze or face direction of the driver.
- the systems for detecting the driver's eyes have become more and more sophisticated, they do not take in account that there are other factors than drowsiness that may cause the driver to reduce the eye opening size or look away in a certain direction (e.g. due to changing lighting conditions within the vehicle compartment). This may, for example, also be caused by external factors such as direct sunlight, flickering light, headlights from oncoming traffic at nights, etc. Also, if the driver is looking down or sideways, the systems may detect this as an eye closure. There is hence a risk that the driver of the vehicle is provided with a warning message intended for a drowsy driver when, in fact, the driver is fully aware of the situation. It is therefore desirable to provide a method which discriminate between the actual factor(s) of a detected eye closure before a warning message is provided to the driver.
- the above is at least partly met by a method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the method comprising receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least one of an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
- the invention is based on the understanding that by comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, a robust and reliable method may be provided for determining if e.g. an identified eyelid closure, eye movement and/or head movement is correctly identified or is caused by e.g. measuring data imperfection. For example, if the received information from the image sensor is a flickering movement of the eyes, which is physically impossible, the comparison with the predetermined set of rules may determine that the received information is caused by a measuring data imperfection. This imperfection may be caused by, for instance, that the image sensor at the moment is unable to locate and provide an accurate image of the operator's eyes.
- the image sensor may lose track of the operator's e.g. eyes and instead falsely identify an eyelid closure. Comparison with the predetermined set of rules may than determine that the identified eyelid closure in fact is caused by the driver looking over his shoulder and is not a correct eyelid closure. Hereby, the result of the comparison may be filtered out since it, in this example, was false.
- the operator of the vehicle suddenly is being blinded by e.g. sunlight or the headlights of oncoming vehicles at night (i.e.
- the operator might turn his head or close his eyelids in a manner that may be identified by the image sensor as an eyelid closure caused by e.g. drowsiness.
- comparison with the predetermined set of rules may than filter out the registered eyelid closure since the lighting condition in the vehicle is such that it is likely that the operator has reacted to the bright light than and is thus aware of the surrounding traffic condition.
- physiological data should in the following be interpreted as all type of data that may be identified by an image based system that identifies e.g. the operator's eyes, face, body, as well as eye gaze direction, eyelid closures, etc.
- the wording “identifying an indication of an eyelid closure” should be interpreted as the image based system identifies that the operator is blinking (fast or slow) or closing his eyes.
- the lighting condition within the operator compartment may be provided to the image based system by, for example, a separate lighting sensor or the image based system itself.
- the separate lighting sensor may, for instance, be an infrared sensor (IR), near infrared sensor (NIR), camera with light sensitive characteristics, etc.
- comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a confidence indication comprised with the physiological data being lower than a predetermined threshold, and providing an error classification of the physiological data if the confidence level is lower than the predetermined threshold.
- an error classification can be made if the physiological data is behaving in a way that is more or less physically impossible, i.e. it is more likely that the image sensor lost track of the positions in the operator's eyes, arms and/or body.
- “Physically impossible” behaviour may for example be if the physiological data indicates an unnaturally high frequency in head and/or body rotation and the like.
- comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a change in a position and/or direction of the face comprised with the physiological data being larger than a predetermined threshold, and providing an error classification of the physiological data if the change in position and/or direction of the face is larger than the predetermined threshold. If, for example, the face of the vehicle operator is directed towards one of the rear view mirrors and the image sensor identifies an eyelid closure, an error classification may be provided since it is more likely that the image sensor lost track of the eyes and that the operator is fully aware of the traffic situation.
- comparing the physiological data with the predetermined set of rules for a current operator status comprises estimating a matching level between an eyelid movement, eye movement and/or head movement comprised with the physiological data and a predetermined eyelid movement, eye movement and/or head movement template, and providing an error classification of the physiological data if the matching level is lower than a predetermined threshold.
- comparing the physiological data with the predetermined set of rules for a current operator status comprises identifying at least one of eye, face and body movement representing a non-drowsy state of the operator, and providing a non-drowsiness classification of the physiological data if a non-drowsiness state is identified.
- comparing the lighting condition within the operator compartment with the predetermined set of rules for a current operator status comprises identifying at least one of an illumination level towards the operator, and a changing lighting condition within the operator compartment.
- identifying at least one of the illumination level towards the operator, and the changing lighting condition within the operator compartment comprises analyzing at least one of a contrast level, illumination level, and/or a frequency of illumination level of image data comprised with the physiological data from the image sensor, or from light sensor arranged within the operator compartment of the vehicle.
- a control system for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle
- the control device comprising a control unit configured to receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identify an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, compare at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classify the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
- control system preferably is a driver assistance system for a vehicle, such as a car, a bus or a truck.
- vehicle such as a car, a bus or a truck.
- a computer readable medium embodying a computer program product for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle
- the computer program product comprising code configured to, when executed by a processor receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
- this aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
- FIG. 1 illustrates a perspective view of the interior of a vehicle, equipped with internal sensors and a plurality of positions within the vehicle normally being gazed by the operator,
- FIG. 2 illustrates a coordinate system of the face of a vehicle operator
- FIG. 3 conceptually illustrates the logical elements of a control system according to a currently preferred embodiment of the invention.
- FIG. 1 illustrates an operating compartment 114 of a vehicle, here illustrated as a car 100 .
- the car 100 includes, in the illustrated embodiment, a vehicle operator (not shown) and an internal sensor, here illustrated as a camera system 102 .
- the camera system 102 is arranged to determine the behaviour of the vehicle operator during vehicle operation. Furthermore, the camera system 102 may be arranged to focus on a predetermined number of positions of the operator's face, arms and body.
- the camera system 102 may be pre-calibrated for a specific operator normally operating the car 100 or being calibrated each time an operator enters the driver seat of the car 100 . As the camera system 102 has detected the different positions of the operator's face, an estimation of facial behaviour is possible. The camera system 102 may hence detect, e.g. head and eye direction, head pose, eye saccade, head-eye saccade, eye closure, speed of eye closure, arm motions, body movements, etc. The camera system 102 may also, by use of a coordinate system 202 in connection to the operator's face 204 , illustrated in FIG.
- the coordinate system 202 of the face 204 is preferably a polar coordinate system with its origin positioned between the eyes of the operator.
- the car also comprises an interior light sensor 104 , in the described embodiment located behind the interior rear view mirror 106 , in the forward direction of the car 100 .
- the exact positioning of the interior light sensor 100 may of course depend on the type of vehicle and its specific interior design. That is, in e.g. a truck there may not be any rear view mirror present and thus the interior light sensor must be placed at another suitable position.
- the interior light sensor 104 is arranged to detect oncoming light to the operator compartment, such as bright sunlight or light from the headlights of oncoming traffic at night, which might cause the operator of the vehicle to react. Still further, there are a number of positions in the vehicle compartment which the operator of the vehicle on a normal basis is paying attention to. These are, for example, the exterior rear view mirrors 108 , the infotainment system 110 , the speed indicator 112 , the passenger next to the driver, the gear shift lever, etc.
- FIG. 3 illustrating an exemplary embodiment of a system 300 for providing the method according to the present invention.
- the system 300 comprises an image sensor 302 which receives information from the camera system 102 , an interior light sensor 304 , an identification of eyelid closure module 306 , a comparison operating module 308 and a drowsiness detection module 310 .
- the eyelid closure module 306 and the comparison operating module 308 can be one and the same module and are only for illustrating purposes divided from each other.
- the image sensor 302 receives physiological data of the vehicle operator and provides the information to the eyelid closure module 306 and to the comparison operating module 308 .
- the comparison operating module also receives information regarding e.g.
- the interior light sensor 304 receives information regarding the lighting condition within the operating compartment 114 , such as e.g. bright sunlight and headlight from oncoming traffic at night.
- the interior light sensor 304 also provides the received lighting condition to the comparison operating module 308 .
- the comparison operating module 308 compares the information with respect to a predetermined set of rules in order to evaluate the current status of the operator. Thereafter, the drowsiness detection module 310 receives the detected eyelid closure from the eyelid closure module 306 and a classification from the comparison operating module 308 regarding the actual cause(s) of the detected eyelid closure.
- the predetermined set of rules may include the plurality of positions within the operating compartment 114 as described above, such that the system can determine, based on the identified eye-gaze direction and/or face direction, that the operator of the vehicle is currently paying attention to one of these positions.
- the system may provide an error classification of the identified eyelid closure, since it is more likely that the operator is fully aware of the situation and that the detection of the eyelids most probably occurred due to the rotation of the operator's head, which the image sensor was unable to correctly follow.
- the predetermined set of rules may also include certain patterns which are physically possible or impossible to occur. For example, if the image sensor detects that the eyes of the operator is moving irregular up/down or left/right in a way that seems implausible, a predetermined rule may classify the detected behaviour of the eyes as an error.
- the comparison operating module 306 may provide an error classification of the eyelid closure, since the actual cause is more likely that the vehicle operator squinted with his eyes to compensate for the blinding light. It should be noted that the data from the image sensor 302 may be affected not only at a short (e.g. second(s)) period but over a longer time (e.g. minute(s)) as the image sensor 302 may e.g. lose its calibration due to the changing lighting condition.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Surgery (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Medical Informatics (AREA)
- Psychiatry (AREA)
- Molecular Biology (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Developmental Disabilities (AREA)
- Child & Adolescent Psychology (AREA)
- Social Psychology (AREA)
- Psychology (AREA)
- Hospice & Palliative Care (AREA)
- Educational Technology (AREA)
- General Physics & Mathematics (AREA)
- Emergency Management (AREA)
- Business, Economics & Management (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physiology (AREA)
- Traffic Control Systems (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
Description
- This claims priority under 35 U.S.C. §119 to European Patent Application No. 11179807.0, filed on Sep. 2, 2011, the contents of each of which are hereby incorporated by reference in their entirety.
- The present invention relates to improvements in relation to e.g. vehicle implemented methods for automatic drowsiness detection, specifically to a method and corresponding system for classification of eye closures for improving the reliability of e.g. the method for drowsiness detection.
- Traffic accidents often occur due to driver impairment caused by, for example, drowsiness. In order to prevent accidents caused by driver impairment, it may be vital to provide the driver with a warning message to re-establish the attention of the driver to the surrounding traffic situation, or in a critical situation to advice the driver to take a break or switch to another driver of the vehicle.
- Recently, much progress has been made in developing drowsiness detection algorithms that are based on detection of the driver's eye closure with a monitoring sensor, e.g. a camera or the like. These algorithms may be used in an image input unit that, by use of the monitoring sensor, detects the face and eyes of the driver. Hereby, the driver's gaze direction can be detected, e.g. the direction in which the driver of the vehicle is currently looking.
- An example of a system having an image input unit can be found in US 2003/039 378. The image input unit of US 2003/039 378 inputs an image including the driver's face area to specify the driver's gaze or face direction. The system generates an input pattern based on the received image input, for determining the driver's gaze or face direction. Moreover, there is provided a dictionary pattern comprising a plurality of stored targets, i.e. positions inside and outside the vehicle which the driver have looked at in the past. The targets may, for example, be the side mirrors, back mirror, infotainment system, etc. Furthermore, the input pattern generated by the system is compared to the stored targets in order to determine the current gaze or face direction of the driver.
- However, even though the systems for detecting the driver's eyes have become more and more sophisticated, they do not take in account that there are other factors than drowsiness that may cause the driver to reduce the eye opening size or look away in a certain direction (e.g. due to changing lighting conditions within the vehicle compartment). This may, for example, also be caused by external factors such as direct sunlight, flickering light, headlights from oncoming traffic at nights, etc. Also, if the driver is looking down or sideways, the systems may detect this as an eye closure. There is hence a risk that the driver of the vehicle is provided with a warning message intended for a drowsy driver when, in fact, the driver is fully aware of the situation. It is therefore desirable to provide a method which discriminate between the actual factor(s) of a detected eye closure before a warning message is provided to the driver.
- According to an aspect of the invention, the above is at least partly met by a method for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the method comprising receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least one of an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison.
- The invention is based on the understanding that by comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, a robust and reliable method may be provided for determining if e.g. an identified eyelid closure, eye movement and/or head movement is correctly identified or is caused by e.g. measuring data imperfection. For example, if the received information from the image sensor is a flickering movement of the eyes, which is physically impossible, the comparison with the predetermined set of rules may determine that the received information is caused by a measuring data imperfection. This imperfection may be caused by, for instance, that the image sensor at the moment is unable to locate and provide an accurate image of the operator's eyes. According to another example, if the operator of the vehicle suddenly looks over his shoulder to see if he is being overtaken by another vehicle, the image sensor may lose track of the operator's e.g. eyes and instead falsely identify an eyelid closure. Comparison with the predetermined set of rules may than determine that the identified eyelid closure in fact is caused by the driver looking over his shoulder and is not a correct eyelid closure. Hereby, the result of the comparison may be filtered out since it, in this example, was false. According to yet another example, if the operator of the vehicle suddenly is being blinded by e.g. sunlight or the headlights of oncoming vehicles at night (i.e. light entering the compartment from the outside), the operator might turn his head or close his eyelids in a manner that may be identified by the image sensor as an eyelid closure caused by e.g. drowsiness. Hence, comparison with the predetermined set of rules may than filter out the registered eyelid closure since the lighting condition in the vehicle is such that it is likely that the operator has reacted to the bright light than and is thus aware of the surrounding traffic condition.
- The wording “physiological data” should in the following be interpreted as all type of data that may be identified by an image based system that identifies e.g. the operator's eyes, face, body, as well as eye gaze direction, eyelid closures, etc. Moreover, the wording “identifying an indication of an eyelid closure” should be interpreted as the image based system identifies that the operator is blinking (fast or slow) or closing his eyes. Furthermore, the lighting condition within the operator compartment may be provided to the image based system by, for example, a separate lighting sensor or the image based system itself. The separate lighting sensor may, for instance, be an infrared sensor (IR), near infrared sensor (NIR), camera with light sensitive characteristics, etc.
- Furthermore, comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a confidence indication comprised with the physiological data being lower than a predetermined threshold, and providing an error classification of the physiological data if the confidence level is lower than the predetermined threshold. Hereby, an error classification can be made if the physiological data is behaving in a way that is more or less physically impossible, i.e. it is more likely that the image sensor lost track of the positions in the operator's eyes, arms and/or body. “Physically impossible” behaviour may for example be if the physiological data indicates an unnaturally high frequency in head and/or body rotation and the like.
- Still further, comparing the physiological data with the predetermined set of rules for a current operator status may comprise identifying a change in a position and/or direction of the face comprised with the physiological data being larger than a predetermined threshold, and providing an error classification of the physiological data if the change in position and/or direction of the face is larger than the predetermined threshold. If, for example, the face of the vehicle operator is directed towards one of the rear view mirrors and the image sensor identifies an eyelid closure, an error classification may be provided since it is more likely that the image sensor lost track of the eyes and that the operator is fully aware of the traffic situation.
- According to a further embodiment, comparing the physiological data with the predetermined set of rules for a current operator status comprises estimating a matching level between an eyelid movement, eye movement and/or head movement comprised with the physiological data and a predetermined eyelid movement, eye movement and/or head movement template, and providing an error classification of the physiological data if the matching level is lower than a predetermined threshold.
- According to a further embodiment, wherein comparing the physiological data with the predetermined set of rules for a current operator status comprises identifying at least one of eye, face and body movement representing a non-drowsy state of the operator, and providing a non-drowsiness classification of the physiological data if a non-drowsiness state is identified.
- According to a further embodiment, comparing the lighting condition within the operator compartment with the predetermined set of rules for a current operator status comprises identifying at least one of an illumination level towards the operator, and a changing lighting condition within the operator compartment.
- According to a further embodiment, identifying at least one of the illumination level towards the operator, and the changing lighting condition within the operator compartment comprises analyzing at least one of a contrast level, illumination level, and/or a frequency of illumination level of image data comprised with the physiological data from the image sensor, or from light sensor arranged within the operator compartment of the vehicle.
- According to another aspect of the invention there is provided a control system for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the control device comprising a control unit configured to receive, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identify an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, compare at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classify the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison. This aspect of the invention provides similar advantages as discussed above in relation to the previous aspect of the invention.
- It should be noted that the control system preferably is a driver assistance system for a vehicle, such as a car, a bus or a truck. A more thorough discussion will be given below in relation to the detailed description of the invention. Other types of vehicles are of course possible and within the scope of the invention.
- According to a still further aspect of the invention there is provided a computer readable medium embodying a computer program product for improving the reliability of a portion of physiological data from an image sensor monitoring an operator positioned in an operator compartment of a vehicle, the computer program product comprising code configured to, when executed by a processor receiving, from the image sensor, physiological data comprising information relating to at least one of eye, face, head, arms and body motion of the operator, identifying an indication of at least an eyelid closure, eye movement or head movement of the operator based on the physiological data, comparing at least one of the physiological data and a lighting condition within the operator compartment with a predetermined set of rules for a current operator status, and classifying the type of eyelid closure, eye movement and/or head movement by correlating the identified eyelid closure, eye movement and/or head movement and a result of the comparison. Also this aspect of the invention provides similar advantages as discussed above in relation to the previous aspects of the invention.
- Further features of, and advantages with, the present invention will become apparent when studying the appended claims and the following description. The skilled addressee realize that different features of the present invention may be combined to create embodiments other than those described in the following, without departing from the scope of the present invention.
- The various aspects of the invention, including its particular features and advantages, will be readily understood from the following detailed description and the accompanying drawings, in which:
-
FIG. 1 illustrates a perspective view of the interior of a vehicle, equipped with internal sensors and a plurality of positions within the vehicle normally being gazed by the operator, -
FIG. 2 illustrates a coordinate system of the face of a vehicle operator, and -
FIG. 3 conceptually illustrates the logical elements of a control system according to a currently preferred embodiment of the invention. - The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which currently preferred embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided for thoroughness and completeness, and fully convey the scope of the invention to the skilled addressee. Like reference characters refer to like elements throughout.
- In the following, the present invention is described with reference to a method for improving classification of eye closures of an operator of a vehicle. The vehicle is preferably equipped with interior sensor(s) for retrieving information of the vehicle operator. For the sake of better understanding, reference is now made to
FIG. 1 , which illustrates anoperating compartment 114 of a vehicle, here illustrated as acar 100. Thecar 100 includes, in the illustrated embodiment, a vehicle operator (not shown) and an internal sensor, here illustrated as acamera system 102. Thecamera system 102 is arranged to determine the behaviour of the vehicle operator during vehicle operation. Furthermore, thecamera system 102 may be arranged to focus on a predetermined number of positions of the operator's face, arms and body. These positions may, for example, be the eyes, eyelids, eyebrows, nose, mouth, cheek, etc. Thecamera system 102 may be pre-calibrated for a specific operator normally operating thecar 100 or being calibrated each time an operator enters the driver seat of thecar 100. As thecamera system 102 has detected the different positions of the operator's face, an estimation of facial behaviour is possible. Thecamera system 102 may hence detect, e.g. head and eye direction, head pose, eye saccade, head-eye saccade, eye closure, speed of eye closure, arm motions, body movements, etc. Thecamera system 102 may also, by use of a coordinatesystem 202 in connection to the operator'sface 204, illustrated inFIG. 2 , detect if the head of the operator is rotating to the right or left (yaw) 205, rotating up or down (pitch) 206 or leaning towards the right or left shoulder (roll) 207. The coordinatesystem 202 of theface 204 is preferably a polar coordinate system with its origin positioned between the eyes of the operator. - Furthermore, the car also comprises an interior
light sensor 104, in the described embodiment located behind the interiorrear view mirror 106, in the forward direction of thecar 100. The exact positioning of the interiorlight sensor 100 may of course depend on the type of vehicle and its specific interior design. That is, in e.g. a truck there may not be any rear view mirror present and thus the interior light sensor must be placed at another suitable position. - The interior
light sensor 104 is arranged to detect oncoming light to the operator compartment, such as bright sunlight or light from the headlights of oncoming traffic at night, which might cause the operator of the vehicle to react. Still further, there are a number of positions in the vehicle compartment which the operator of the vehicle on a normal basis is paying attention to. These are, for example, the exterior rear view mirrors 108, theinfotainment system 110, thespeed indicator 112, the passenger next to the driver, the gear shift lever, etc. - Now referring to
FIG. 3 , illustrating an exemplary embodiment of asystem 300 for providing the method according to the present invention. Thesystem 300 comprises animage sensor 302 which receives information from thecamera system 102, an interiorlight sensor 304, an identification ofeyelid closure module 306, acomparison operating module 308 and adrowsiness detection module 310. It should however be noted that theeyelid closure module 306 and thecomparison operating module 308 can be one and the same module and are only for illustrating purposes divided from each other. Moreover, theimage sensor 302 receives physiological data of the vehicle operator and provides the information to theeyelid closure module 306 and to thecomparison operating module 308. The comparison operating module also receives information regarding e.g. operator eyes, arms and body motion from theimage sensor 302. Furthermore, the interiorlight sensor 304 receives information regarding the lighting condition within theoperating compartment 114, such as e.g. bright sunlight and headlight from oncoming traffic at night. The interiorlight sensor 304 also provides the received lighting condition to thecomparison operating module 308. - Moreover, when the
comparison operating module 308 has received information regarding the lighting condition within theoperating compartment 114 as well as the physiological data regarding eyes, arms and body motion of the operator, thecomparison operating module 308 compares the information with respect to a predetermined set of rules in order to evaluate the current status of the operator. Thereafter, thedrowsiness detection module 310 receives the detected eyelid closure from theeyelid closure module 306 and a classification from thecomparison operating module 308 regarding the actual cause(s) of the detected eyelid closure. According to an example, the predetermined set of rules may include the plurality of positions within theoperating compartment 114 as described above, such that the system can determine, based on the identified eye-gaze direction and/or face direction, that the operator of the vehicle is currently paying attention to one of these positions. For instance, if the system identifies an eyelid closure of the operator, and at the same time identifies that the operator's face is directed towards the leftrear view mirror 108, the system may provide an error classification of the identified eyelid closure, since it is more likely that the operator is fully aware of the situation and that the detection of the eyelids most probably occurred due to the rotation of the operator's head, which the image sensor was unable to correctly follow. Furthermore, the predetermined set of rules may also include certain patterns which are physically possible or impossible to occur. For example, if the image sensor detects that the eyes of the operator is moving irregular up/down or left/right in a way that seems implausible, a predetermined rule may classify the detected behaviour of the eyes as an error. - According to a further example, if the interior
light sensor 304 detects a bright light towards the vehicle operator at the same time as theimage sensor 302 detects an eyelid closure of the operator, thecomparison operating module 306 may provide an error classification of the eyelid closure, since the actual cause is more likely that the vehicle operator squinted with his eyes to compensate for the blinding light. It should be noted that the data from theimage sensor 302 may be affected not only at a short (e.g. second(s)) period but over a longer time (e.g. minute(s)) as theimage sensor 302 may e.g. lose its calibration due to the changing lighting condition. - Even though the invention has been described with reference to specific exemplifying embodiments thereof, many different alterations, modifications and the like will become apparent for those skilled in the art. Variations to the disclosed embodiments can be understood and effected by the skilled addressee in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims. For example, the invention has mainly been described above with reference to a few embodiments. However, as is readily appreciated by the skilled addressee, other embodiments than the ones disclosed above are equally possible within the scope of the invention, as defined by the appended patent claims. For example, the invention is also applicable for trucks, buses, dumpers, wheel loaders and other type of vehicles than the above described car.
- In the claims, the word “comprises” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. A single computer or other unit may fulfil the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measured cannot be used to advantage.
Claims (12)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11179807.0A EP2564777B1 (en) | 2011-09-02 | 2011-09-02 | Method for classification of eye closures |
EP11179807.0 | 2011-09-02 | ||
EP11179807 | 2011-09-02 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130057671A1 true US20130057671A1 (en) | 2013-03-07 |
US9117358B2 US9117358B2 (en) | 2015-08-25 |
Family
ID=44785312
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/558,780 Active 2033-10-19 US9117358B2 (en) | 2011-09-02 | 2012-07-26 | Method for classification of eye closures |
Country Status (5)
Country | Link |
---|---|
US (1) | US9117358B2 (en) |
EP (1) | EP2564777B1 (en) |
JP (1) | JP2013054735A (en) |
CN (1) | CN103150545B (en) |
BR (1) | BR102012022143A2 (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140210978A1 (en) * | 2013-01-25 | 2014-07-31 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US20160167661A1 (en) * | 2013-07-19 | 2016-06-16 | Audi Ag | Method for operating a driver assistance system of a motor vehicle and driver assistance system for a motor vehicle |
US9771081B2 (en) * | 2014-09-29 | 2017-09-26 | The Boeing Company | System for fatigue detection using a suite of physiological measurement devices |
US9908530B1 (en) | 2014-04-17 | 2018-03-06 | State Farm Mutual Automobile Insurance Company | Advanced vehicle operator intelligence system |
WO2018042122A1 (en) | 2016-08-29 | 2018-03-08 | Renault S.A.S. | Method for determining the placement of the head of a vehicle driver |
US9934667B1 (en) | 2014-03-07 | 2018-04-03 | State Farm Mutual Automobile Insurance Company | Vehicle operator emotion management system and method |
US10118487B1 (en) * | 2014-05-05 | 2018-11-06 | State Farm Mutual Automobile Insurance Company | System and method to monitor and alert vehicle operator of impairment |
US10475127B1 (en) | 2014-07-21 | 2019-11-12 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and insurance incentives |
US20200103900A1 (en) * | 2018-09-30 | 2020-04-02 | Strong Force Intellectual Capital, Llc | Intelligent transportation systems |
CN115352450A (en) * | 2022-08-12 | 2022-11-18 | 重庆长安汽车股份有限公司 | Vehicle driving assistance method and device, vehicle and storage medium |
CN115771387A (en) * | 2021-09-06 | 2023-03-10 | 博泰车联网(南京)有限公司 | Sun visor control method and device |
US20230105891A1 (en) * | 2020-02-06 | 2023-04-06 | Mitsubishi Electric Corporation | Sign detection device, driving assistance control device, and sign detection method |
US11868127B2 (en) | 2018-09-30 | 2024-01-09 | Strong Force Tp Portfolio 2022, Llc | Radial basis function neural network optimizing operating parameter of vehicle based on emotional state of rider determined by recurrent neural network |
US20240010222A1 (en) * | 2022-01-14 | 2024-01-11 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods to increase driver awareness of exterior occurrences |
EP3075315B1 (en) * | 2015-04-02 | 2024-02-07 | Essilor International | System and computer-implemented method for monitoring the visual behavior of a person |
WO2024260940A1 (en) * | 2023-06-21 | 2024-12-26 | Bayerische Motoren Werke Aktiengesellschaft | Assistance system for supporting a driver of a vehicle during a lateral guidance of the vehicle while taking into account a dazzling of the driver, vehicle, and method |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101655818B1 (en) * | 2014-12-11 | 2016-09-08 | 현대자동차주식회사 | Wearable glass, control method thereof and vehicle control system |
EP3440592B1 (en) * | 2016-04-07 | 2023-08-30 | Seeing Machines Limited | Method and system of distinguishing between a glance event and an eye closure event |
US10821987B2 (en) | 2016-07-20 | 2020-11-03 | Ford Global Technologies, Llc | Vehicle interior and exterior monitoring |
WO2018227597A1 (en) * | 2017-06-16 | 2018-12-20 | Boe Technology Group Co., Ltd. | Vision-based interactive control apparatus and method of controlling rear-view mirror for vehicle |
JP7099037B2 (en) * | 2018-05-07 | 2022-07-12 | オムロン株式会社 | Data processing equipment, monitoring system, awakening system, data processing method, and data processing program |
JP7099036B2 (en) * | 2018-05-07 | 2022-07-12 | オムロン株式会社 | Data processing equipment, monitoring system, awakening system, data processing method, and data processing program |
US20220117529A1 (en) * | 2020-10-20 | 2022-04-21 | Harman Becker Automotive Systems Gmbh | System and method for determining an eye movement |
DE102020214910A1 (en) * | 2020-11-27 | 2022-06-02 | Robert Bosch Gesellschaft mit beschränkter Haftung | Method for monitoring a vehicle interior |
CN112668548B (en) * | 2021-01-15 | 2024-07-12 | 重庆大学 | Method and system for detecting driver's blurriness |
CN114162119A (en) * | 2021-10-27 | 2022-03-11 | 广州广日电气设备有限公司 | Lateral control method, equipment, medium and product of automobile advanced driving auxiliary system |
CN118372835B (en) * | 2024-06-24 | 2024-09-06 | 江西科技学院 | Fatigue driving identification method and system based on eye movement characteristics |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6097295A (en) * | 1998-01-28 | 2000-08-01 | Daimlerchrysler Ag | Apparatus for determining the alertness of a driver |
US20050073136A1 (en) * | 2002-10-15 | 2005-04-07 | Volvo Technology Corporation | Method and arrangement for interpreting a subjects head and eye activity |
US20080180235A1 (en) * | 2007-01-25 | 2008-07-31 | Hsuan Chang | Method and apparatus for manipulating driver core temperature to enhance driver alertness |
US20080238694A1 (en) * | 2007-03-26 | 2008-10-02 | Denso Corporation | Drowsiness alarm apparatus and program |
US20080252466A1 (en) * | 2007-04-11 | 2008-10-16 | Yopp W Trent | System and method for implementing active safety counter measures for an impaired driver |
US20090097701A1 (en) * | 2007-10-11 | 2009-04-16 | Denso Corporation | Sleepiness level determination device for driver |
US20100214105A1 (en) * | 2009-02-24 | 2010-08-26 | Panasonic Automotive Systems Company Of America, Division Of Panasonic Corporation Of North America | Method of detecting drowsiness of a vehicle operator |
US20110060496A1 (en) * | 2009-08-11 | 2011-03-10 | Certusview Technologies, Llc | Systems and methods for complex event processing of vehicle information and image information relating to a vehicle |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6154559A (en) * | 1998-10-01 | 2000-11-28 | Mitsubishi Electric Information Technology Center America, Inc. (Ita) | System for classifying an individual's gaze direction |
JP3790680B2 (en) | 2001-05-25 | 2006-06-28 | 株式会社東芝 | Image processing system and driving support system using the same |
US6927694B1 (en) * | 2001-08-20 | 2005-08-09 | Research Foundation Of The University Of Central Florida | Algorithm for monitoring head/eye motion for driver alertness with one camera |
KR100436861B1 (en) | 2001-08-27 | 2004-06-30 | 나노메트릭스코리아 주식회사 | Method and apparatus for inspecting defects on polishing pad to be used with chemical mechanical polishing apparatus |
CN1956680A (en) * | 2004-05-24 | 2007-05-02 | 日本先锋公司 | Biological information detection device |
CN100462047C (en) * | 2007-03-21 | 2009-02-18 | 汤一平 | Safe driving auxiliary device based on omnidirectional computer vision |
CN100462046C (en) * | 2007-04-16 | 2009-02-18 | 孙军 | Online early warning system and method of weariness working |
CN101030316B (en) * | 2007-04-17 | 2010-04-21 | 北京中星微电子有限公司 | Safety driving monitoring system and method for vehicle |
CN101655907B (en) * | 2009-08-28 | 2013-03-20 | 大连鑫奇辉科技有限公司 | Trainman driving state monitoring intelligent alarm system |
-
2011
- 2011-09-02 EP EP11179807.0A patent/EP2564777B1/en active Active
-
2012
- 2012-07-26 US US13/558,780 patent/US9117358B2/en active Active
- 2012-08-22 CN CN201210299500.0A patent/CN103150545B/en not_active Expired - Fee Related
- 2012-08-29 JP JP2012188432A patent/JP2013054735A/en active Pending
- 2012-09-03 BR BR102012022143-8A patent/BR102012022143A2/en not_active IP Right Cessation
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6097295A (en) * | 1998-01-28 | 2000-08-01 | Daimlerchrysler Ag | Apparatus for determining the alertness of a driver |
US20050073136A1 (en) * | 2002-10-15 | 2005-04-07 | Volvo Technology Corporation | Method and arrangement for interpreting a subjects head and eye activity |
US20080180235A1 (en) * | 2007-01-25 | 2008-07-31 | Hsuan Chang | Method and apparatus for manipulating driver core temperature to enhance driver alertness |
US20080238694A1 (en) * | 2007-03-26 | 2008-10-02 | Denso Corporation | Drowsiness alarm apparatus and program |
US20080252466A1 (en) * | 2007-04-11 | 2008-10-16 | Yopp W Trent | System and method for implementing active safety counter measures for an impaired driver |
US20090097701A1 (en) * | 2007-10-11 | 2009-04-16 | Denso Corporation | Sleepiness level determination device for driver |
US20100214105A1 (en) * | 2009-02-24 | 2010-08-26 | Panasonic Automotive Systems Company Of America, Division Of Panasonic Corporation Of North America | Method of detecting drowsiness of a vehicle operator |
US20110060496A1 (en) * | 2009-08-11 | 2011-03-10 | Certusview Technologies, Llc | Systems and methods for complex event processing of vehicle information and image information relating to a vehicle |
US20110093162A1 (en) * | 2009-08-11 | 2011-04-21 | Certusview Technologies, Llc | Systems and methods for complex event processing of vehicle-related information |
US8463487B2 (en) * | 2009-08-11 | 2013-06-11 | Certusview Technologies, Llc | Systems and methods for complex event processing based on a hierarchical arrangement of complex event processing engines |
Cited By (53)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140210978A1 (en) * | 2013-01-25 | 2014-07-31 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US8847771B2 (en) * | 2013-01-25 | 2014-09-30 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US20140340228A1 (en) * | 2013-01-25 | 2014-11-20 | Toyota Motor Engineering & Manufacturing North America, Inc. | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US9299237B2 (en) * | 2013-01-25 | 2016-03-29 | Toyota Jidosha Kabushiki Kaisha | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US20160171322A1 (en) * | 2013-01-25 | 2016-06-16 | Toyota Jidosha Kabushiki Kaisha | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US9824286B2 (en) * | 2013-01-25 | 2017-11-21 | Toyota Jidosha Kabushiki Kaisha | Method and apparatus for early detection of dynamic attentive states for providing an inattentive warning |
US20160167661A1 (en) * | 2013-07-19 | 2016-06-16 | Audi Ag | Method for operating a driver assistance system of a motor vehicle and driver assistance system for a motor vehicle |
US9934667B1 (en) | 2014-03-07 | 2018-04-03 | State Farm Mutual Automobile Insurance Company | Vehicle operator emotion management system and method |
US10593182B1 (en) | 2014-03-07 | 2020-03-17 | State Farm Mutual Automobile Insurance Company | Vehicle operator emotion management system and method |
US10121345B1 (en) | 2014-03-07 | 2018-11-06 | State Farm Mutual Automobile Insurance Company | Vehicle operator emotion management system and method |
US9908530B1 (en) | 2014-04-17 | 2018-03-06 | State Farm Mutual Automobile Insurance Company | Advanced vehicle operator intelligence system |
US10118488B1 (en) | 2014-05-05 | 2018-11-06 | State Farm Mutual Automobile Insurance Co. | System and method to monitor and alert vehicle operator of impairment |
US10569650B1 (en) | 2014-05-05 | 2020-02-25 | State Farm Mutual Automobile Insurance Company | System and method to monitor and alert vehicle operator of impairment |
US10118487B1 (en) * | 2014-05-05 | 2018-11-06 | State Farm Mutual Automobile Insurance Company | System and method to monitor and alert vehicle operator of impairment |
US11068995B1 (en) | 2014-07-21 | 2021-07-20 | State Farm Mutual Automobile Insurance Company | Methods of reconstructing an accident scene using telematics data |
US11634103B2 (en) | 2014-07-21 | 2023-04-25 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US10475127B1 (en) | 2014-07-21 | 2019-11-12 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and insurance incentives |
US12365308B2 (en) | 2014-07-21 | 2025-07-22 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US12358463B2 (en) | 2014-07-21 | 2025-07-15 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and driving behavior identification |
US10723312B1 (en) | 2014-07-21 | 2020-07-28 | State Farm Mutual Automobile Insurance Company | Methods of theft prevention or mitigation |
US10825326B1 (en) | 2014-07-21 | 2020-11-03 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US10832327B1 (en) | 2014-07-21 | 2020-11-10 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and driving behavior identification |
US10974693B1 (en) | 2014-07-21 | 2021-04-13 | State Farm Mutual Automobile Insurance Company | Methods of theft prevention or mitigation |
US10997849B1 (en) | 2014-07-21 | 2021-05-04 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US11030696B1 (en) | 2014-07-21 | 2021-06-08 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and anonymous driver data |
US11069221B1 (en) | 2014-07-21 | 2021-07-20 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US12179695B2 (en) | 2014-07-21 | 2024-12-31 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US11257163B1 (en) | 2014-07-21 | 2022-02-22 | State Farm Mutual Automobile Insurance Company | Methods of pre-generating insurance claims |
US12151644B2 (en) | 2014-07-21 | 2024-11-26 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US11634102B2 (en) | 2014-07-21 | 2023-04-25 | State Farm Mutual Automobile Insurance Company | Methods of facilitating emergency assistance |
US11565654B2 (en) | 2014-07-21 | 2023-01-31 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and driving behavior identification |
US10540723B1 (en) | 2014-07-21 | 2020-01-21 | State Farm Mutual Automobile Insurance Company | Methods of providing insurance savings based upon telematics and usage-based insurance |
US9771081B2 (en) * | 2014-09-29 | 2017-09-26 | The Boeing Company | System for fatigue detection using a suite of physiological measurement devices |
EP3075315B1 (en) * | 2015-04-02 | 2024-02-07 | Essilor International | System and computer-implemented method for monitoring the visual behavior of a person |
WO2018042122A1 (en) | 2016-08-29 | 2018-03-08 | Renault S.A.S. | Method for determining the placement of the head of a vehicle driver |
US12298759B2 (en) | 2018-09-30 | 2025-05-13 | Strong Force Tp Portfolio 2022, Llc | Using social media data of a vehicle occupant to alter a route plan of the vehicle |
US12228924B2 (en) | 2018-09-30 | 2025-02-18 | Strong Force Tp Portfolio 2022, Llc | Social data sources feeding a neural network to predict an emerging condition relevant to a transportation plan of at least one individual |
US11868126B2 (en) | 2018-09-30 | 2024-01-09 | Strong Force Tp Portfolio 2022, Llc | Wearable device determining emotional state of rider in vehicle and optimizing operating parameter of vehicle to improve emotional state of rider |
US20200103900A1 (en) * | 2018-09-30 | 2020-04-02 | Strong Force Intellectual Capital, Llc | Intelligent transportation systems |
US11694288B2 (en) | 2018-09-30 | 2023-07-04 | Strong Force Tp Portfolio 2022, Llc | Method of optimizing rider satisfaction |
US12124257B2 (en) | 2018-09-30 | 2024-10-22 | Strong Force Tp Portfolio 2022, Llc | Intelligent transportation systems |
US11486721B2 (en) * | 2018-09-30 | 2022-11-01 | Strong Force Intellectual Capital, Llc | Intelligent transportation systems |
US12321169B2 (en) | 2018-09-30 | 2025-06-03 | Strong Force Tp Portfolio 2022, Llc | Optimizing a vehicle operating parameter based in part on a sensed emotional state of a rider |
US12321168B2 (en) | 2018-09-30 | 2025-06-03 | Strong Force Tp Portfolio 2022, Llc | Transportation system to optimize an operating parameter of a vehicle based on an emotional state of an occupant of the vehicle determined from a sensor to detect a physiological condition of the occupant |
US11868127B2 (en) | 2018-09-30 | 2024-01-09 | Strong Force Tp Portfolio 2022, Llc | Radial basis function neural network optimizing operating parameter of vehicle based on emotional state of rider determined by recurrent neural network |
US12235641B2 (en) | 2018-09-30 | 2025-02-25 | Strong Force Tp Portfolio 2022, Llc | Hybrid neural networks sourcing social data sources to optimize satisfaction of rider in intelligent transportation systems |
US12298760B2 (en) | 2018-09-30 | 2025-05-13 | Strong Force Tp Portfolio 2022, Llc | Neural net optimization of continuously variable powertrain |
US20230105891A1 (en) * | 2020-02-06 | 2023-04-06 | Mitsubishi Electric Corporation | Sign detection device, driving assistance control device, and sign detection method |
CN115771387A (en) * | 2021-09-06 | 2023-03-10 | 博泰车联网(南京)有限公司 | Sun visor control method and device |
US12240483B2 (en) * | 2022-01-14 | 2025-03-04 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods to increase driver awareness of exterior occurrences |
US20240010222A1 (en) * | 2022-01-14 | 2024-01-11 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods to increase driver awareness of exterior occurrences |
CN115352450A (en) * | 2022-08-12 | 2022-11-18 | 重庆长安汽车股份有限公司 | Vehicle driving assistance method and device, vehicle and storage medium |
WO2024260940A1 (en) * | 2023-06-21 | 2024-12-26 | Bayerische Motoren Werke Aktiengesellschaft | Assistance system for supporting a driver of a vehicle during a lateral guidance of the vehicle while taking into account a dazzling of the driver, vehicle, and method |
Also Published As
Publication number | Publication date |
---|---|
EP2564777B1 (en) | 2017-06-07 |
JP2013054735A (en) | 2013-03-21 |
BR102012022143A2 (en) | 2013-10-01 |
CN103150545B (en) | 2017-04-12 |
EP2564777A1 (en) | 2013-03-06 |
US9117358B2 (en) | 2015-08-25 |
CN103150545A (en) | 2013-06-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9117358B2 (en) | Method for classification of eye closures | |
US9616809B1 (en) | Lane change prediction and turn signal activation upon observation of head and eye movement | |
JP6087088B2 (en) | System and method for improving vehicle operator performance estimates | |
CN110532976A (en) | Method for detecting fatigue driving and system based on machine learning and multiple features fusion | |
US20060287779A1 (en) | Method of mitigating driver distraction | |
US20180197030A1 (en) | Driver abnormality detection device and driver abnormality detection method | |
CN101968917A (en) | Vehicle-mounted warning apparatus | |
WO2007092512A2 (en) | Driver drowsiness and distraction monitor | |
CN114872713A (en) | Device and method for monitoring abnormal driving state of driver | |
CN109969195A (en) | An alarm system and method for abnormal driver facial behavior and a vehicle based thereon | |
KR20120074820A (en) | Control system for vehicle using face recognition function | |
KR101500016B1 (en) | Lane Departure Warning System | |
JP2015085719A (en) | Gazing object estimation device | |
US12179787B2 (en) | Pedestrian alert system | |
Bergasa et al. | Visual monitoring of driver inattention | |
US20220319200A1 (en) | Device and method for determining image data of the eyes, eye positions and/or a viewing direction of a vehicle user in a vehicle | |
WO2023017595A1 (en) | Occupant state determining device, occupant state determining method, and occupant state determining system | |
US20230159010A1 (en) | Driver assistance device | |
JP7612118B2 (en) | Drowsiness estimation device and drowsiness estimation method | |
Su et al. | Image-Based Driver Status Monitoring System for Determining the Transfer of Dynamic Driving Tasks in Autonomous Vehicles | |
JP2023027699A (en) | Driver state determination device | |
JP7091632B2 (en) | Safe driving promotion device and safe driving promotion method | |
US12403818B2 (en) | Illumination device for vehicle | |
WO2025004257A1 (en) | Drowsiness estimation device and drowsiness estimation method | |
EP4510089A1 (en) | Physical driver condition determination for vehicle safety |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: VOLVO CAR CORPORATION, SWEDEN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEVIN, DANIEL;WESTERVALL, LENA;REEL/FRAME:029120/0394 Effective date: 20120817 Owner name: VOLVO TECHNOLOGY CORPORATION, SWEDEN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEANDERSSON, SUSANNA;KRONBERG, PETER;SIGNING DATES FROM 20120927 TO 20121002;REEL/FRAME:029120/0582 |
|
AS | Assignment |
Owner name: VOLVO CAR CORPORATION, SWEDEN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VOLVO TECHNOLOGY CORPORATION;REEL/FRAME:034350/0686 Effective date: 20141120 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |