US20160007910A1 - Avoidance of cognitive impairment events - Google Patents
Avoidance of cognitive impairment events Download PDFInfo
- Publication number
- US20160007910A1 US20160007910A1 US14/328,349 US201414328349A US2016007910A1 US 20160007910 A1 US20160007910 A1 US 20160007910A1 US 201414328349 A US201414328349 A US 201414328349A US 2016007910 A1 US2016007910 A1 US 2016007910A1
- Authority
- US
- United States
- Prior art keywords
- wearer
- time
- sensor device
- buffer
- readings
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 208000028698 Cognitive impairment Diseases 0.000 title claims abstract description 81
- 208000010877 cognitive disease Diseases 0.000 title claims abstract description 81
- 239000000872 buffer Substances 0.000 claims abstract description 171
- 230000006998 cognitive state Effects 0.000 claims abstract description 76
- 230000006735 deficit Effects 0.000 claims abstract description 59
- 238000000034 method Methods 0.000 claims abstract description 40
- 230000004044 response Effects 0.000 claims abstract description 34
- 230000009471 action Effects 0.000 claims abstract description 22
- 230000000694 effects Effects 0.000 claims abstract description 8
- 230000036962 time dependent Effects 0.000 claims description 94
- 239000003550 marker Substances 0.000 claims description 23
- 230000001771 impaired effect Effects 0.000 claims description 16
- 230000036961 partial effect Effects 0.000 claims description 10
- 230000007613 environmental effect Effects 0.000 claims description 7
- 230000035790 physiological processes and functions Effects 0.000 claims description 5
- 230000001960 triggered effect Effects 0.000 claims description 5
- 239000013598 vector Substances 0.000 claims description 5
- 230000008859 change Effects 0.000 claims description 4
- 238000003780 insertion Methods 0.000 claims description 4
- 230000037431 insertion Effects 0.000 claims description 4
- 230000001681 protective effect Effects 0.000 claims description 4
- 230000002547 anomalous effect Effects 0.000 claims description 3
- 230000000977 initiatory effect Effects 0.000 claims description 3
- 238000009826 distribution Methods 0.000 claims description 2
- 239000002243 precursor Substances 0.000 claims description 2
- 238000007619 statistical method Methods 0.000 claims description 2
- 230000007704 transition Effects 0.000 claims description 2
- 239000011159 matrix material Substances 0.000 description 39
- 238000003860 storage Methods 0.000 description 21
- 238000010586 diagram Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 13
- 230000033001 locomotion Effects 0.000 description 9
- 238000012545 processing Methods 0.000 description 9
- 208000019901 Anxiety disease Diseases 0.000 description 6
- 206010048909 Boredom Diseases 0.000 description 6
- 230000036506 anxiety Effects 0.000 description 6
- 238000012544 monitoring process Methods 0.000 description 6
- 238000009825 accumulation Methods 0.000 description 5
- 238000009499 grossing Methods 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 238000004590 computer program Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000003491 array Methods 0.000 description 3
- 230000008921 facial expression Effects 0.000 description 3
- WWYNJERNGUHSAO-XUDSTZEESA-N (+)-Norgestrel Chemical compound O=C1CC[C@@H]2[C@H]3CC[C@](CC)([C@](CC4)(O)C#C)[C@@H]4[C@@H]3CCC2=C1 WWYNJERNGUHSAO-XUDSTZEESA-N 0.000 description 2
- 206010012289 Dementia Diseases 0.000 description 2
- 206010041349 Somnolence Diseases 0.000 description 2
- 230000036772 blood pressure Effects 0.000 description 2
- 230000019771 cognition Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000002565 electrocardiography Methods 0.000 description 2
- 238000000537 electroencephalography Methods 0.000 description 2
- 230000001747 exhibiting effect Effects 0.000 description 2
- 230000005021 gait Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000007170 pathology Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 210000000707 wrist Anatomy 0.000 description 2
- 208000000044 Amnesia Diseases 0.000 description 1
- 208000031091 Amnestic disease Diseases 0.000 description 1
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 206010016825 Flushing Diseases 0.000 description 1
- 241001417516 Haemulidae Species 0.000 description 1
- 208000032140 Sleepiness Diseases 0.000 description 1
- 206010044565 Tremor Diseases 0.000 description 1
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 230000036626 alertness Effects 0.000 description 1
- 230000006986 amnesia Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000000875 corresponding effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000002996 emotional effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000003999 initiator Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000003205 muscle Anatomy 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 239000001301 oxygen Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000035807 sensation Effects 0.000 description 1
- 230000037321 sleepiness Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/486—Biofeedback
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/0022—Monitoring a patient using a global network, e.g. telephone networks, internet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording for evaluating the cardiovascular system, e.g. pulse, heart rate, blood pressure or blood flow
- A61B5/0205—Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
- A61B5/02055—Simultaneously evaluating both cardiovascular condition and temperature
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Measuring devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor or mobility of a limb
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4058—Detecting, measuring or recording for evaluating the nervous system for evaluating the central nervous system
- A61B5/4064—Evaluating the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
- A61B5/6803—Head-worn items, e.g. helmets, masks, headphones or goggles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7225—Details of analogue processing, e.g. isolation amplifier, gain or sensitivity adjustment, filtering, baseline or drift compensation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7246—Details of waveform analysis using correlation, e.g. template matching or determination of similarity
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/725—Details of waveform analysis using specific filters therefor, e.g. Kalman or adaptive filters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7275—Determining trends in physiological measurement data; Predicting development of a medical condition based on physiological measurements, e.g. determining a risk factor
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7282—Event detection, e.g. detecting unique waveforms indicative of a medical condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient; User input means
- A61B5/742—Details of notification to user or communication with user or patient; User input means using visual displays
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient; User input means
- A61B5/746—Alarms related to a physiological condition, e.g. details of setting alarm thresholds or avoiding false alarms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient; User input means
- A61B5/7475—User input or interface means, e.g. keyboard, pointing device, joystick
- A61B5/749—Voice-controlled interfaces
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/67—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for remote operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16Z—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS, NOT OTHERWISE PROVIDED FOR
- G16Z99/00—Subject matter not provided for in other main groups of this subclass
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/02—Operational features
- A61B2560/0242—Operational features adapted to measure environmental factors, e.g. temperature, pollution
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Measuring devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor or mobility of a limb
- A61B5/112—Gait analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4803—Speech analysis specially adapted for diagnostic purposes
Definitions
- the present disclosure relates to the field of computers, and specifically to the use of computers in evaluating cognitive states. Still more particularly, the present disclosure relates to assisting a person in avoiding an impairment event associated with one or more cognitive states.
- a person's cognitive state is also known as a person's “state of mind”. This state of mind may be normal (e.g., interested, sleepy, asleep, alert, bored, curious, doubtful, etc.), or it may be indicative of some type of pathology (e.g., amnesia, confusion, panic, etc.). Often, such states of mind will manifest themselves measurably before a person (subjectively) realizes that he/she is entering such a state of mind.
- a method guides evasive actions to avoid effects of a cognitive impairment state.
- a first buffer which is communicatively coupled to at least one sensor on a wearable sensor device, is loaded with a first set of time-dependent sensor readings.
- the wearable sensor device receives a first cognitive impairment state signal, where the first cognitive impairment state signal is sent in response to an observer, of the wearer, observing an impairment to a cognitive state of the wearer of the wearable sensor device.
- the cognitive impairment state marker is inserted at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal.
- a second buffer on the wearable sensor device, initiates loading of a second set of time-dependent sensor readings from the at least one sensor on the wearable sensor device, and time-dependent sensor readings from the first buffer and the second buffer are compared.
- an alert is issued to the wearer of the wearable sensor device.
- a method enables a guidance of evasive actions to avoid effects of a cognitive impairment state.
- a first set of time-dependent sensor readings are received from a first buffer on a wearable sensor device, where the first buffer is communicatively coupled to at least one sensor on the wearable sensor device.
- a first cognitive impairment state signal is transmitted to the wearable sensor device, where the first cognitive impairment state signal is sent in response to an observer, of the wearer, observing an impairment to a cognitive state of the wearer of the wearable sensor device.
- a cognitive impairment state marker is transmitted to the wearable sensor device, wherein the cognitive impairment state marker is inserted at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal.
- An initiation of loading of a second buffer, which is on the wearable sensor device, with a second set of time-dependent sensor readings from the at least one sensor on the wearable sensor device is detected.
- Time-dependent sensor readings from the first buffer and the second buffer are compared to one another up to the predefined position.
- an alert is issued to the wearer of the wearable sensor device.
- a wearable sensor device is integrated into a sports helmet.
- the wearable sensor device comprises: a physiological sensor, wherein the physiological sensor detects a biological state of the wearer of the sports helmet; an accelerometer sensor, wherein the accelerometer sensor detects a change in velocity of the protective sports helmet; a first buffer for storing a first set of time-dependent sensor readings, wherein the first buffer is communicatively coupled to the physiological sensor and the accelerometer sensor; a receiver for receiving a first cognitive impairment state signal, wherein an observer of a wearer of the wearable sensor device sends the first cognitive impairment state signal in response to the observer observing (in one or more embodiments, subjectively) an impairment to a cognitive state of the wearer of the wearable sensor device; a data insertion logic for inserting a cognitive impairment state marker at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal; a second buffer, wherein the second buffer initiates loading of a second set of time-dependent
- FIG. 1 depicts an exemplary system and network in which the present disclosure may be implemented
- FIG. 2 illustrates an exemplary Impaired Cognitive State Predictor (ICSP) architecture in accordance with one or more embodiments of the present invention
- FIG. 3 depicts an exemplary wearable sensor device that is integrated into a sports helmet for monitoring physiological and physical conditions related to the sports helmet and its wearer;
- FIG. 4 illustrates an exemplary wrist-wearable sensor device for sensing user physiological and/or other conditions of a wearer of the device
- FIG. 5 depicts a wearer of the wrist-wearable sensor device illustrated in FIG. 4 while speaking before an interactive audience;
- FIG. 6 illustrates a high-level block diagram of an exemplary networked system used in the embodiment depicted in FIG. 5 ;
- FIG. 7 is a high-level flowchart of one or more steps performed by one or more processors to guide evasive actions for avoiding effects of an impaired cognitive state.
- the present invention may be a system, a method, and/or a computer program product.
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the block may occur out of the order noted in the figures.
- two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- FIG. 1 there is depicted a block diagram of an exemplary system and network that may be utilized by and/or in the implementation of the present invention.
- the exemplary architecture including both depicted hardware and software, shown for and within computer 102 may be utilized by software deploying server 150 and/or a wearable sensor device 152 , as well as impairment signal transmitters 506 shown in FIG. 5 and FIG. 6 , and/or monitoring system 602 shown in FIG. 6 .
- Exemplary computer 102 includes a processor 104 that is coupled to a system bus 106 .
- Processor 104 may utilize one or more processors, each of which has one or more processor cores.
- a video adapter 108 which drives/supports a display 110 , is also coupled to system bus 106 .
- System bus 106 is coupled via a bus bridge 112 to an input/output (I/O) bus 114 .
- An I/O interface 116 is coupled to I/O bus 114 .
- I/O interface 116 affords communication with various I/O devices, including a keyboard 118 , a mouse 120 , a media tray 122 (which may include storage devices such as CD-ROM drives, multi-media interfaces, etc.), a wireless signal transceiver 124 (e.g., a near field radio frequency transceiver, a Wi-Fi transceiver, etc.), and external USB port(s) 126 . While the format of the ports connected to I/O interface 116 may be any known to those skilled in the art of computer architecture, in one embodiment some or all of these ports are universal serial bus (USB) ports.
- USB universal serial bus
- Network interface 130 is a hardware network interface, such as a network interface card (NIC), etc.
- Network 128 may be an external network such as the Internet, or an internal network such as an Ethernet or a virtual private network (VPN).
- a hard drive interface 132 is also coupled to system bus 106 .
- Hard drive interface 132 interfaces with a hard drive 134 .
- hard drive 134 populates a system memory 136 , which is also coupled to system bus 106 .
- System memory is defined as a lowest level of volatile memory in computer 102 . This volatile memory includes additional higher levels of volatile memory (not shown), including, but not limited to, cache memory, registers and buffers. Data that populates system memory 136 includes computer 102 's operating system (OS) 138 and application programs 144 .
- OS operating system
- OS 138 includes a shell 140 , for providing transparent user access to resources such as application programs 144 .
- shell 140 is a program that provides an interpreter and an interface between the user and the operating system. More specifically, shell 140 executes commands that are entered into a command line user interface or from a file.
- shell 140 also called a command processor, is generally the highest level of the operating system software hierarchy and serves as a command interpreter. The shell provides a system prompt, interprets commands entered by keyboard, mouse, or other user input media, and sends the interpreted command(s) to the appropriate lower levels of the operating system (e.g., a kernel 142 ) for processing.
- a kernel 142 the appropriate lower levels of the operating system for processing.
- shell 140 is a text-based, line-oriented user interface, the present invention will equally well support other user interface modes, such as graphical, voice, gestural, etc.
- OS 138 also includes kernel 142 , which includes lower levels of functionality for OS 138 , including providing essential services required by other parts of OS 138 and application programs 144 , including memory management, process and task management, disk management, and mouse and keyboard management.
- kernel 142 includes lower levels of functionality for OS 138 , including providing essential services required by other parts of OS 138 and application programs 144 , including memory management, process and task management, disk management, and mouse and keyboard management.
- Application programs 144 include a renderer, shown in exemplary manner as a browser 146 .
- Browser 146 includes program modules and instructions enabling a world wide web (WWW) client (i.e., computer 102 ) to send and receive network messages to the Internet using hypertext transfer protocol (HTTP) messaging, thus enabling communication with software deploying server 150 and other computer systems.
- WWW world wide web
- HTTP hypertext transfer protocol
- Application programs 144 in computer 102 's system memory also include a Cognitive Impairment Event Avoidance Logic (CIEAL) 148 .
- CIEAL 148 includes code for implementing the processes described below, including those described in FIGS. 2-7 .
- computer 102 is able to download CIEAL 148 from software deploying server 150 , including in an on-demand basis, wherein the code in CIEAL 148 is not downloaded until needed for execution.
- software deploying server 150 performs all of the functions associated with the present invention (including execution of CIEAL 148 ), thus freeing computer 102 from having to use its own internal computing resources to execute CIEAL 148 .
- power supply 151 used to power various components within computer 102 and/or connected components (e.g., elements 110 , 118 , 126 , 128 , etc.)
- power supply 151 may be a solar cell, a battery (rechargeable or non-rechargeable), a public utility power grid that is accessible via a wall socket, a local limited supply source (e.g., a local fuel-powered generator), etc.
- computer 102 may include alternate memory storage devices such as magnetic cassettes, digital versatile disks (DVDs), Bernoulli cartridges, and the like. These and other variations are intended to be within the spirit and scope of the present invention.
- the present invention takes three steps.
- the first step is to capture sensor readings that describe circumstantial environments of a wearer of a sensor device, and to store them in a buffer.
- an observer in one or more embodiments a human observer
- takes some action which results in a marker being placed in the buffer.
- the third step if the wearer of the sensor device repeats actions that lead to sensor readings matching those found in the first buffer, then an alert is given to the wearer to take evasive actions to avoid the effects of again experiencing the impaired cognitive state that resulted from the circumstantial environments captured in the first step.
- an exemplary Impaired Cognitive State Predictor (ICSP) architecture 200 is presented in accordance with one or more embodiments of the present invention.
- the ICSP architecture 200 and data generated by the ICSP architecture is secret. That is, predictions of current or future cognitive states are presented only to the user that generated the sensor readings described herein and/or experiences the specific cognitive state that follows these sensor readings. Only with the express approval of the user (i.e., wearer of the wearable sensor device described herein) will such readings/states be shared with others.
- the ICSP architecture 200 also includes multiple components found in FIG. 1 (e.g., computer 102 , wearable sensor device 152 , etc.). Furthermore, in one embodiment, the continuous circular buffer(s) 202 , the push data matrix 204 , and/or the accumulation data matrix 206 shown in FIG. 2 and/or components shown in FIG. 1 are all within the wearable sensor device 152 . In one or more other embodiments, the continuous circular buffer(s) 202 are within the wearable sensor device, but the push data matrix 204 and/or accumulation data matrix 206 are stored in a hardware storage device (e.g., system memory 136 and/or hard drive 134 shown in FIG. 1 ) on a remote computer, such as computer 102 shown in FIG. 1 .
- a hardware storage device e.g., system memory 136 and/or hard drive 134 shown in FIG. 1
- Wearable sensor device 152 includes one or more sensor(s) 208 .
- each of the sensor(s) 208 are “smart sensors”, that include processing logic that is able to detect, record, and quantify what is being sensed. That is, each of the sensor(s) 208 is 1) able to detect a particular physical event (heat, noise, biometrics, etc.); 2) quantify the level of that particular physical event (e.g., how high the heat is, what the duration/intensity of the noise is, what the specific readings of the biometric is, etc.); 3) convert that level into a digital value; and/or 4) send that digital value to the continuous circular buffer(s) 202 .
- these functions are performed by dedicated hardware logic, which takes digital readings from the sensors, compares the digital readings to known ranges in order to establish the digital value, and then transmits (e.g., by a wireless digital signal transmitter) the digital value to the continuous circular buffer(s) 202 , which then (responsive to a “push” signal) send the stored digital values from the continuous circular buffer(s) 202 to a local matrix within the wearable sensor device 152 or to a remote matrix in a remote computer (e.g., computer 102 shown in FIG. 1 ).
- dedicated hardware logic takes digital readings from the sensors, compares the digital readings to known ranges in order to establish the digital value, and then transmits (e.g., by a wireless digital signal transmitter) the digital value to the continuous circular buffer(s) 202 , which then (responsive to a “push” signal) send the stored digital values from the continuous circular buffer(s) 202 to a local matrix within the wearable sensor device 152 or to a remote matrix in a remote computer (
- the system can use a near field network to send the digital value to a local storage within the wearable sensor device 152 , or to a remote device, such as a smart phone held by the user, or to a server on a cloud, etc. (e.g., using a Wi-Fi signal).
- a near field network to send the digital value to a local storage within the wearable sensor device 152 , or to a remote device, such as a smart phone held by the user, or to a server on a cloud, etc. (e.g., using a Wi-Fi signal).
- data stored in the continuous circular buffer(s) 202 is time-based. For example, data found in buffer cell 220 a is received before data that is received and stored in buffer cell 220 b , which stores data that is received before data that is received and stored in buffer cell 220 c , etc. Thus, the data stored in continuous circular buffer(s) 202 (and thus the push data matrix 204 ) is “time-based”.
- sensor(s) 208 include physiological sensors, which are defined as sensors that are able to detect physiological states of a person. In one embodiment, these sensors are attached to the person via the wearable sensor device 152 .
- Example of such sensors include, but are not limited to, a heart monitor, a blood pressure cuff/monitor (sphygmomanometer), a galvanic skin conductance monitor, an electrocardiography (ECG) device, an electroencephalography (EEG) device, etc. That is, in one embodiment, the sensor(s) 208 are biometric sensors that measure physiological functions, of the wearer, which are not musculoskeletal.
- sensor(s) 208 detect and/or measure musculoskeletal bodily acts of the user, such as facial expressions (e.g., smiles, frowns, furrowed brows, etc.), body movements (e.g., walking gait, limps, stride length, stride speed, etc.), etc.
- Facial expressions may be detected by muscle movement sensors on eyeglasses, cameras on “smart glasses”, etc.
- Body movements may be detected by motion detectors, stride counters, strain gauges in clothing, etc.
- sensor(s) 208 include speech content analyzers.
- the sensor(s) 208 includes a speech-to-text converter, which then examines the text for certain keywords, speech pattern, etc. That is, the speech-to-text converter converts spoken words into written text, which can then be examined in order to identify certain predefined keywords, speech pattern, etc. The presence (or absence) of such keywords, speech pattern, etc. is then used by logic (e.g. CIEAL 148 in FIG. 1 ) to ascertain the nature of the speech, which may lead to a warning of a future cognitive impairment state of the user (as described herein).
- logic e.g. CIEAL 148 in FIG. 1
- sensor(s) 208 include speech content analyzers.
- the sensor(s) 208 includes a speech-to-text converter, which then examines the text for certain features.
- These features may include the construction of graphs representing structural elements of speech based on a number of alternatives, such as syntactic value (article, noun, verb, adjective, etc.), or lexical root (run/ran/running) for the nodes of the graph, and text proximity for the edges of the graph.
- Graph features such as link degree, clustering, loop density, centrality, etc., representing speech topological structure are also therefore included.
- semantic vectors may be extracted from the text as features, using systems such as that provided by a Latent Semantic Analysis, WordNet, etc.
- sensor(s) 208 include speech inflection analyzers.
- the sensor(s) 208 compare voice patterns with known voice patterns (pitch, timing, tremor, etc.) of the user, in order to identify certain emotions such as stress, relaxation, alertness, sleepiness, and other cognitive states.
- the presence (or absence) of such voice patterns is then used by logic (e.g. CIEAL 148 in FIG. 1 ) to ascertain the current emotional state of the user, which may lead to a warning of a future cognitive impairment state of the user (as described herein).
- sensor(s) 208 include environmental sensors, such as an air thermometer, a microphone, a barometer, a light sensor, a moisture sensor, etc.
- sensor(s) 208 are able to detect ambient (within the proximity of the user) environmental conditions, such as rain, various light levels, sound levels, air pressure, sound (e.g., noise, music, spoken words, etc.), etc.
- sensor(s) 208 include accelerometers, which measure acceleration and/or deceleration forces as an object accelerates (i.e., increases speed) and/or decelerates (i.e., slows down and/or stops). These acceleration/deceleration forces may be abrupt, particularly the deceleration forces that occur when a moving object strikes another object, which may be fixed, moving in a direction opposite that of the first object, or is moving in the same direction as the first object but at a slower speed.
- values stored in the continuous circular buffer(s) 202 are sent to the push data matrix 204 in response to a “push” event.
- the “push” event occurs in response to an observer observing a wearer of the wearable sensor device 152 exhibiting an impaired cognitive state. That is, as soon as the observer “feels” (i.e., subjectively determines) that the wearer is in a particular impaired cognitive state (e.g., is boring, is unfocused, is disoriented, etc.), then the observer issues a “push” command, causing the contents of the continuous circular buffer(s) 202 to be loaded into the push data matrix 204 .
- the actions taken by the observer are based on the observer's subjective impressions, which are not based on clinical evidence. That is, the observations are not on unimpeachable scientific evidence of a particular pathology (e.g., an MRI that clearly shows damage to a cognition component of the brain), but rather are the subjective observations of the observer.
- a particular pathology e.g., an MRI that clearly shows damage to a cognition component of the brain
- one observer may view the wearer (person who is wearing the wearable sensor device 152 ) as being “fascinating” or “alert”, while another observer may view the same wearer as being “boring” or “disoriented”.
- the observers' observations are purely subjective in this embodiment, and are not directly correlated to any scientific/clinical facts supporting an impression of a specific cognitive impairment state.
- sensor readings from sensor(s) 208 are buffered in the continuous circular buffer(s) 202 .
- Continuous circular buffer(s) 202 are buffers that allow data to be stored in any location/cell within the buffer. Unlike a linear buffer (such as a First In First Out—FIFO buffer), a circular buffer allows “stale” data to be replaced with “fresh” data without shifting the location of existing data in other cells within the buffer.
- continuous circular buffer(s) 202 is composed of multiple circular buffers 210 a - 210 c (where “c” is an integer). In one embodiment, each of the circular buffers 210 a - 210 c is devoted to storing readings from a specific sensor from sensor(s) 208 .
- circular buffer 210 a is devoted to storing readings from a sensor 208 that measures a heart rate of the user.
- data from circular buffer 210 a is sent to push data matrix 204 , it is stored in the unrolled buffer shown as b 1 .
- circular buffer 210 b is devoted to storing readings from a sensor 208 that measures an ambient light level where the user is located.
- data from circular buffer 210 b is sent to push data matrix 204 , it is stored in the unrolled buffer shown as b 2 .
- circular buffer 210 c is devoted to storing readings from a sensor 208 that measures speech patterns of the user.
- data is sent from the continuous circular buffer(s) 202 to the push data matrix 204 in response to a “push” being initiated by an observer of the wearer of the wearable sensor device 152 observing a particular impaired cognitive state of the wearer.
- these observations are purely subjective. That is, the perception of a particular impaired cognitive state is subjective and unique to that observer. For example, one observer may determine that the wearer of the wearable sensor device 152 is experience the cognitive state of “lucid and interesting” when sensor(s) 208 detect a particular pattern of conditions (physiological, temporal, environmental, etc.).
- a smoothing function is used to “smooth out” the observations from the observers of the wearer of the wearable sensor device 152 , in order to come to an approximation of data that represents overall patterns of the observations, while eliminating outlier (i.e., out-of-bound, anomalous) observations that are unwarranted/unsupported.
- smoothing functions/algorithms/filters include, but are not limited to, additive smoothing algorithms, Kalman filters, least-squares fitting of polynomials (representing the subjective observations) algorithms, moving averages, exponential smoothing (to reduce random fluctuations in time series data), curve fitting of observational data, numerical smoothing and differentiation, etc.
- an observer has made multiple past observations of the wearer of the wearable sensor device 152 , and that the observations have varied not only due to the various states of the wearer (i.e., impairments of the wearer's cognitive state), but also to various states of the observer (i.e., the observer is hyper-alert, is sluggish, is angry, etc.), which affect whether or not the observer determines that there is an impairment to the cognitive state of the wearer of the wearable sensor device 152 .
- the observer of the wearer of the wearable sensor device sends multiple cognitive impairment state signals in response to the observer subjectively observing multiple instances of the impairment to the cognitive state of the wearer of the wearable sensor device, and the multiple cognitive impairment state signals are generated in response to the observer making multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device.
- one or more processors apply a Kalman filter to the multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device.
- This Kalman filter uses a linear quadratic estimation to recursively remove anomalous observations from the multiple observations to generate a (in one embodiment, trusted) observation of the impairment to the cognitive state of the wearer of the wearable sensor device (i.e., wearable sensor device 152 ).
- An exemplary Kalman algorithm used for this determination is:
- x k is the (trusted) observation of the impairment to the cognitive state of the wearer of the wearable sensor device
- F k is a predefined state transition model that is applied to a previous state x k-1 of observed impairments to the cognitive state of the wearer of the wearable sensor device
- B k is a predefined control-input model that is applied to a control vector u k
- w k is erroneous observation noises that are drawn from a multivariate normal distribution Q k , wherein w k is approximately equal to the set of numbers N from zero to Q k (N(0, Q k )).
- the observer in response to observing a particular cognitive impairment state, the observer will initiate a “push” of data from the continuous circular buffer(s) 202 to the push data matrix 204 , which is stored on a hardware storage device.
- the cognitive impairment state 212 is represented by a digital value that is sent to the push data matrix 204 .
- This digital value identifies a particular cognitive impairment state of the wearer of the wearable sensor device 152 , which is defined by an observer of the wearer, such that precursive readings from the sensor(s) 208 are associated with a subsequent and specific cognitive impairment state and/or other related events (e.g., “benching” a sports player).
- the particular cognitive impairment state that is associated with specific precursive events is described by the wearer's own words and/or by the observer's own words. In another embodiment, the particular cognitive impairment state is selected from a menu or is otherwise predefined.
- accumulation data matrix 206 takes continuous readings from the continuous circular buffer(s) 202 .
- each set of data that has been pushed to the push data matrix 204 is nonetheless identified within the accumulation data matrix 206 .
- data that was pushed to push data matrix 204 at the time of a “PUSH-A” is identified by block 214 a ; data that was pushed to push data matrix 204 at the time of a “PUSH-B” is identified by block 214 b ; and data that was pushed to push data matrix 204 at the time of a “PUSH-C” is identified by block 214 c.
- Push average matrix 216 is calculated (e.g., by CIEAL 148 shown in FIG. 1 ) as one-third (assuming that three pushes occurred) of the sum of the values stored in each cell of the pushed buffers. That is, assume that a push results in three columns of nine cells.
- the values in the upper left cell in each of the blocks 214 a - 214 c are summed together, divided by three, and the quotient (i.e., average) is then stored in the upper left cell of the push average matrix 216 .
- PTA push average matrix 216
- PTA push average value 216
- Ranges around these mean values are predetermined, such that when values from the continuous circular buffer(s) 202 later fall within these ranges, a prediction can be made that the wearer will again experience (or is currently experiencing) the particular cognitive impairment state.
- a buffer data matrix 218 is generated from a single buffer in the push data matrix 204 .
- buffer b 1 contains data from continuous circular buffer 210 a that describe the heart rate of the user who is wearing the wearable sensor device 152 .
- b 1 is broken down into three rows, r 1 -r 3 , in order to create the buffer data matrix 218 .
- Buffer data matrix 218 is then used in a manner similar to that described herein for push data matrixes.
- buffer data matrix 218 a matrix
- this buffer data matrix 218 is then used to warn of an impending cognitive impairment state of the wearer by comparing this buffer data matrix 218 to known single-sensor data matrixes that are precursive to the particular cognitive impairment state of the user.
- wearable sensor device 152 / 352 / 452 / 552 is presented as a wearable sensor device, in one or more embodiments the wearable sensor device 152 / 352 / 452 / 552 is a device that is simply proximate to, although not necessarily worn by, a user, such that ambient conditions, including biophysical traits of the user (e.g., frowns, smiles, flushed skin, etc.) are still sensed by sensors, such as sensors 208 .
- biophysical traits of the user e.g., frowns, smiles, flushed skin, etc.
- the wearable sensor device is integrated into a sport helmet.
- an exemplary wearable sensor device 352 is integrated into a sports helmet 300 for monitoring physiological and physical conditions related to the sports helmet and its wearer.
- One or more components of the exemplary wearable sensor device 352 are powered by a local battery 351 and/or an equivalent power source (e.g., a solar cell).
- the wearable sensor device 352 includes one or more biometric sensor(s) 308 , which measure physiological states (i.e., perspiration, skin temperature, eye flutter, voice articulations such as grunts of pain, etc.) of the wearer, similar to the sensors 208 described above.
- the wearable sensor device 352 includes an accelerometer 306 .
- the accelerometer 306 which may be any known electro-mechanical device that measures acceleration, including rapid deceleration, detects whether the helmet 300 has been subjected to a sharp blow (as indicated by a sudden acceleration/deceleration detected by the accelerometer 306 ), initiated by the wearer or by another player.
- any blow that occurs when the player is not actually wearing the helmet 300 is irrelevant to the cognitive state of the athlete.
- the battery 351 is only connected when the athlete puts on the helmet 300 .
- This selective powering on/off may be from a sensor switch (not shown, but within the interior lining of the helmet 300 ), a manual switch (also not shown), etc.
- a processor 304 processes readings from the biometric sensor(s) 308 and/or the accelerometer 306 by loading them into buffer(s) 302 , which in one embodiment have the same architecture and function as the continuous circular buffer(s) 202 shown in FIG. 2 .
- a data I/O port 326 (which in one embodiment has a same architecture as USB port(s) 126 shown in FIG. 1 ) is able to 1) download data from the buffer(s) 302 , and 2) upload markers into the buffer(s) 302 .
- the wearable sensor device 352 includes a physiological sensor such as one or more of the biometric sensor(s) 308 . These biometric/physiological sensors detect a biological state of the wearer of the sports helmet 302 , such as his/her heart rate, perspiration level, skin temperature, EEG and/or EKG, oxygen saturation level, etc.
- the wearable sensor device 352 also includes an accelerometer sensor, such as accelerometer 306 . This accelerometer sensor detects a change in velocity (e.g., a crash into another player's helmet, striking the ground or other immovable object, etc.).
- a first buffer i.e., one of the buffer(s) 302 shown in FIG. 3 .
- This first buffer (which in one embodiment is implemented as a dedicated hardware storage device—“hardware buffer”) is communicatively coupled to the physiological sensor (biometric sensor(s) 308 ) and the accelerometer sensor (accelerometer 306 ), and thus is able to store a first set of time-dependent sensor readings (i.e., sensor readings from the biometric sensor(s) 308 and/or the accelerometer 306 which are retrieved/recorded in linear time (sequentially)).
- the wearable sensor device 352 also includes a receiver (e.g., part of transceiver 310 ) for receiving a first cognitive impairment state signal. That is, when an observer of a wearer of the wearable sensor device (wearable sensor device 352 integrated into helmet 302 ) subjectively observes an impairment to a cognitive state of the wearer of the wearable sensor device/helmet, the observer sends the first cognitive impairment state signal to the transceiver 310 .
- a data insertion logic e.g., part of processor 304 ) then inserts a cognitive impairment state marker at a predefined position in the first buffer.
- a second buffer (i.e., one of the buffer(s) 302 , and which may be the same as the first buffer if steps are taken to clear and save data from the first buffer for further use), initiates loading of a second set of time-dependent sensor readings from the physiological sensor and the accelerometer sensor.
- the first buffer may store data from the biometric sensor(s) 308 and/or the accelerometer 306 taken during a first game.
- the second buffer may store data from biometric sensor(s) 308 and/or the accelerometer 306 taken during a later game (e.g., played the following week).
- the first buffer contains the following sensor readings from the biometric sensor(s) 308 (identified as “Bx”) and the accelerometer 306 (identified as “Ax”): B1, B2, B3, A1, A2, A3.
- Bx and Ax are time-dependent, meaning that they are stored in the first buffer as they are generated by their respective sensors.
- the cognitive impairment state marker (identified as “I”) is placed after “Bx” and “Ax”, thus giving the temporal sequence of: B1, B2, B3, A1, A2, A3, I.
- a hardware comparator (e.g., part of processor 304 ) then compares time-dependent sensor readings from the first buffer up to the predefined position with time-dependent sensor readings from the second buffer. For example, assume again that the first buffer contains the sensor readings B1, B2, B3, A1, A2, A3, which are stored before the cognitive impairment state marker I. Assume also that the second buffer contains the sensor readings B1, B2, A1, A2. Although B1, B2, A1, A2 is not the same as B1, B2, B3, A1, A2, A3, the present invention recognizes that the wearer is taking actions that have a similar pattern as those recorded during the last game.
- the present invention provides an “intervention” with the wearer of the helmet, giving the wearer the opportunity to perform evasive actions (i.e., stop tackling with his helmet, being less aggressive, etc.) to avoid reaching the pattern B1, B2, B3, A1, A2, A3, which prompted the earlier signal (and thus cognitive impairment state marker I) from the observer.
- evasive actions i.e., stop tackling with his helmet, being less aggressive, etc.
- the cognitive impairment state marker I is also part of the precursive pattern that warns of an impending impaired cognitive state.
- the cognitive impairment state marker I functions as a warning during future activities.
- the cognitive impairment state marker I along with the sensor readings B1, B2, B3, A1, A2, A3, would produce a pattern of B1, B2, B3, A1, A2, A3, I.
- the pattern B1, B2, B3, A1, A2, A3 is recorded a subsequent event (meeting, sports game, etc.) for the wearer of the wearable sensor device.
- This newly-derived pattern B1, B2, B3, A1, A2, A3 will then cause the system to recognize “I” as the next data point, which causes a warning to be issued to the wearer of the wearable sensor device. That is, at this point, the wearer of the wearable sensor device may not be presenting evidence of an impaired cognitive state, either internally (i.e., the wearer of the wearable sensor device does not yet “feel” the impaired cognitive state) or to another (i.e., the observer of the wearable sensor device does not perceive that the wearer of the wearable sensor device has entered into the impaired cognitive state).
- this intervention signal (indicated by the cognitive impairment state marker I) and/or the warning derived therefrom is used as part of the B k predefined control-input model that is applied to a control vector u k in the Kalman filter described herein.
- an alert generator (e.g., part of processor 304 ), in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, issues an alert to the wearer of the wearable sensor device. That is, in the example presented above, if the second buffer stores the pattern B1, B2, A1, A2, then a predefined match level (e.g., 4 out of 6 of the sensor readings B1, B2, B3, A1, A2, A3) has been reached. This triggers the alert generator to send a signal to the wearer of the helmet, such as sending a color-coded signal to a multi-color light emitting diode (LED) device 312 mounted on the faceguard 314 of the helmet 302 .
- LED multi-color light emitting diode
- the LED device 312 may turn yellow if the alert is to let the wearer know that continuing the same style of play will result in receiving a second cognitive impairment state signal from the observer.
- this second cognitive impairment state signal may result in the player being prevented from further play, being required to take additional instruction, being required to submit to a medical examination, etc.
- the LED device 312 may turn red if the alert advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer. That is, if the player continues to play in the same manner, the red alert signal indicates to the wearer that continuing this style of play will result in the same impairment (e.g., disorientation, confusion, etc.) that was observed before. In one embodiment, the red alert signal may result in the player being immediately taken out of the game.
- the wearable sensor device is simply worn on the wrist.
- the wearable sensor device 452 which is structurally similar to the wearable device 152 shown in FIG. 2 , and includes sensors such as sensor(s) 208 , a power supply (not shown), buffers (e.g., continuous circular buffer(s) 202 ), etc. needed for a wearable sensor device as presented herein.
- wearable sensor device 452 (which may be worn on the wrist) includes a keypad 402 .
- keys on the keypad are pre-programmed for a particular cognitive state.
- one of the keys may be for “boredom”.
- the user pushing the button for “boredom” causes data from the continuous circular buffer(s) 202 , along with a flag/signal that is associated with the cognitive impairment state 212 for “boredom” (and identified by pushing the key on keypad 402 for “boredom”), to be sent to the push data matrix 204 in FIG. 2 .
- the wearable sensor device 452 shown in FIG. 4 includes both biophysical (unique to the user) and ambient environmental sensors. More specifically, wearable sensor device 452 includes biometric sensors that, depending on their structure, configuration, and/or positioning on the wearable sensor device 452 , are able to monitor biometric conditions (e.g., blood pressure, heart rate, etc.), musculoskeletal motions (e.g., cameras that track a user's facial expressions, motion sensors that track a user's walking gait, etc.) and other biophysical features/conditions of the user, but also can track ambient environmental conditions (e.g., local sounds, light, moisture, air temperature, etc.).
- biometric conditions e.g., blood pressure, heart rate, etc.
- musculoskeletal motions e.g., cameras that track a user's facial expressions, motion sensors that track a user's walking gait, etc.
- ambient environmental conditions e.g., local sounds, light, moisture, air temperature, etc.
- a wearer of the wrist-wearable sensor device is a person making a presentation (i.e., presenter 502 ) to an audience 504 .
- the presenter 502 is wearing a wearable sensor device 552 , which has the same hardware and configuration as the wearable sensor device 452 depicted in FIG. 4 .
- members of the audience are able to input cognition state impairment signals via impairment signal transmitters 506 . For example, if one or more of the members of the audience 504 perceive that the presenter 502 is boring or anxious or otherwise doing poorly, then this information may be displayed on the presenter's display 508 on his/her laptop 510 .
- the presenter 502 may still have time to salvage the presentation. However, and in accordance with a preferred embodiment of the present invention, if a threshold of the members of the audience 503 (e.g., 50%) indicate that the presenter 502 is doing poorly, then it is too late to salvage the current presentation. However, the presenter 502 still has useful data points for future presentations. That is, if the wearable sensor device 552 detects a similar pattern progression as that which ultimately led to the previous audience's negative response, then a signal on the presenter's display 508 will offer suggestions to avoid (execute avoidance actions) another poor performance. In one or more embodiments of the present invention, these suggestions are defined as part of the B k predefined control-input model that is applied to a control vector u k of the Kalman filter described herein.
- a threshold of the members of the audience 503 e.g. 50%
- these suggestions are defined as part of the B k predefined control-input model that is applied to a control vector u k of the
- the presenter's display 508 may present a suggestion to “Stand Still”. That is, in the previous presentation, excessive movement by the presenter 502 (i.e., pacing back and forth, excessive hand/body gestures, etc.) led the previous audience to subjectively determine that the presenter 502 was “frenetic”. If the presenter 502 again displays such excessive movement during a subsequent/current presentation, then the system will advise him/her to modulate his body movements (e.g., with the suggestion that he/she “Stand Still”).
- FIG. 6 illustrates a high-level block diagram of an exemplary networked system used in the embodiment depicted in FIG. 5 .
- a monitoring system 602 e.g., computer 102 shown in FIG. 1
- the monitoring system 602 using a processor 604 (analogous to the processor 104 shown in FIG. 1 ) compares sensor patterns from previous presentations given by the presenter 502 with current sensor patterns. If the monitoring system 602 detects that a similar pattern is being followed (although not up to the point of losing the audience's attention/approval), then an alert and/or suggestion is sent to the presenter's display 508 .
- FIG. 7 a high-level flowchart of one or more steps performed by one or more processors to guide evasive actions for avoiding effects of an impaired cognitive state is presented. Again, note that one or more of the steps depicted may be performed by one or more processors (e.g., processor 104 in FIG. 1 , processor 304 in FIG. 3 , processor 604 in FIG. 6 , etc.)
- processors e.g., processor 104 in FIG. 1 , processor 304 in FIG. 3 , processor 604 in FIG. 6 , etc.
- a first buffer is loaded on a wearable sensor device with a first set of time-dependent sensor readings (block 704 ).
- the first buffer is communicatively coupled to at least one sensor on the wearable sensor device.
- the wearable sensor device receives a first cognitive impairment state signal.
- this first cognitive impairment state signal is generated by an observer of a wearer of the wearable sensor device in response to the observer subjectively observing an impairment to a cognitive state of the wearer of the wearable sensor device. For example, if the observer thinks that the wearer of the wearable sensor device appears to be confused, boring, anxious, etc., then the observer will generate the first cognitive impairment state signal.
- this first cognitive impairment state signal is transmitted from a transmitter (e.g., one or more of the impairment signal transmitters 506 shown in FIG. 5 ).
- this first cognitive impairment state signal is an automatic consequence of taking the step of downloading data from the sensor data buffer (e.g., buffer(s) 302 in FIG. 3 ) via a data I/O port (e.g., data I/O port 326 in FIG. 3 ).
- a cognitive impairment state marker is then inserted at a predefined position in the first buffer.
- This cognitive impairment state marker (e.g., “I” presented above) is inserted at the predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal.
- sensor reading B1 may be loaded, followed by sensor readings B1, A1, followed by sensor readings B1, A1, A2, etc.
- time-dependent sensor readings from the first buffer up to the predefined position are then compared with time-dependent sensor readings from the second buffer.
- an alert is issued to the wearer of the wearable sensor device (block 716 ).
- the predefined match level may be numeric, weighted, etc.
- the first buffer contains the sensor readings B1, A1, A2, B2, B3, B4 (where “B” indicates a biometric sensor reading and “A” indicates an accelerometer reading).
- the system After storing B1, A1, A2, B2, B3, B4 in the first buffer, the system receives a cognitive impairment state marker “I”, which indicates that there is an observed impairment to the cognitive state of the wearer of the wearable sensor device, and thus “I” is placed after the sensor readings B1, A1, A2, B2, B3, B4 in the first buffer.
- the second buffer (or the first buffer after being cleared and its contents stored in a local memory for use in comparison to new sensor data) contains the sensor readings B1, A1, A2, B2.
- the “predefined match level” needed to initiate the alert may simply be a percentage, such as 50%.
- 50% of the six sensor readings B1, A1, A2, B2, B3, B4 would be three, regardless of which sensor readings are subsequently taken and stored in the second buffer.
- each of the sensor readings B1, A1, A2, B2, B3, B4 may be weighted.
- sensor readings from the accelerometer (A1, A2, A3, etc.) may be predetermined to be more important, and thus weighted more heavily, than sensor readings form the biometric sensors (B1, B2, B3, etc.).
- sensor readings from the accelerometer A1, A2, A3, etc.
- biometric sensors B1, B2, B3, etc.
- the flow-chart of FIG. 7 ends at terminator block 718 .
- the alert to the wearer of the wearable sensor device advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid receiving a second cognitive impairment state signal from the observer. That is, the alert may advise the wearer to take corrective/ameliorative steps to avoid being told again that his/her style of play is dangerous, he/she is boring, etc.
- the alert to the wearer of the wearable sensor device advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer. That is, the alert may advise the wearer of the wearable sensor device that his/her style of play will actually cause him/her to be disoriented, that his/her presentation style will be boring, etc., even if nobody tells him/her.
- the first set of time-dependent sensor readings are analyzed, in order to identify a cause of the impairment to the cognitive state of the wearer. For example, assume that historical records show that when events that result in sensor readings A1, A1, A3, B1, B2, B3 ultimately lead to a state of dementia, a conclusion is reached that these events are the cause of the state of dementia. Assuming that such records and conclusions are available, then in one embodiment they are used to verify the observations of the observer of the wearer of the wearable sensor device.
- the first buffer and the second buffer in the wearable sensor device are both continuous circular buffers, in which each stores data from a different sensor in the wearable sensor device.
- a cause of the impairment to the cognitive state of the wearer of the wearable sensor device is predicted by a probability formula:
- E) is a probability that the impairment to the cognitive state will occur (M) given that (
- M) is a probability that data from the continuous circular buffers falls within the predefined PTA of previously pushed data from the continuous circular buffers (E) given that (
- P(M) is a probability that the impairment to the cognitive state of the wearer will occur regardless of any other information; and
- ⁇ m is a sum of all occurrences m, for the probability P(E
- predicting whether the impairment to the cognitive state of the wearer of the wearable sensor device will occur is based on a statistical analysis of the subsequent set of sensor readings compared to the pushed sensor readings, wherein a match within a predefined statistical range between the subsequent set of sensor readings and the pushed sensor readings leads to the prediction of the impairment to the cognitive state of the wearer of the wearable sensor device. That is, if the previous sensor readings align with current sensor readings within a statistically significant range, then an assumption/prediction is made that a recurrence of the cognitive impairment state is likely.
- the predefined position in the first buffer at which the cognitive impairment state marker is inserted is at an end of the first set of time-dependent sensor readings.
- a determination is made that sensor readings stored prior to the end of the first set of time-dependent sensor readings are precursors to the impairment to the cognitive state of the wearer of the wearable sensor device. That is if the sensor readings B1, B2, B3, A1, A2, A3 lead to the insertion of the cognitive impairment state marker I at their end, then a conclusion is reached that the events that caused sensor readings B1, B2, B3, A1, A2, A3 lead to the cognitive impairment state reflected by “I”.
- the first set of time-dependent sensor readings are made up of a first subset of time-dependent sensor readings and a second subset of time-dependent sensor readings.
- the first subset of time-dependent sensor readings record event states that occur before event states that are represented by the second subset of time-dependent sensor readings.
- a set of sensor readings B1, A1, A2, B2, B3, A3 may be made up of a first subset of sensor readings B1, A1, A2, which occur before a second subset of sensor readings B2, B3, A3 (i.e., the events that caused sensor readings B1, A1, A2 occurred before events that caused sensor readings B2, B3, A3).
- the second set of time-dependent sensor readings comprise a third subset of time-dependent sensor readings and a fourth subset of time-dependent sensor readings, where the third subset of time-dependent sensor readings record event states that occur before event states that are represented by the fourth subset of time-dependent sensor readings.
- the third subset found at the beginning of the second set of time-dependent sensors may be B1, A1, A2. Since the first subset and the third subset are identical (B1, A1, A2), then a conclusion is reached that the wearer of the wearable sensor device is headed for a specific cognitive impairment state, regardless of the contents of the fourth subset of time-dependent sensor readings in the second set of time-dependent sensor readings, and an alert is issued.
- the sensors in the wearable sensor device detect physiological states of the user, musculoskeletal bodily acts of the user, keywords spoken by the user, a quality of a voice pattern from the user, and ambient environmental conditions around the user.
- specific patterns of sensor readings from all of these sensors are used to provide a warning of a recurrence of a particular impairment to the cognitive state of the wearer of the wearable sensor device.
- VHDL VHSIC Hardware Description Language
- VHDL is an exemplary design-entry language for Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), and other similar electronic devices.
- FPGA Field Programmable Gate Arrays
- ASIC Application Specific Integrated Circuits
- any software-implemented method described herein may be emulated by a hardware-based VHDL program, which is then applied to a VHDL chip, such as a FPGA.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Physics & Mathematics (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Veterinary Medicine (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Physiology (AREA)
- Cardiology (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Computer Networks & Wireless Communication (AREA)
- Pulmonology (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- General Business, Economics & Management (AREA)
- Business, Economics & Management (AREA)
- Neurology (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Evolutionary Computation (AREA)
- Power Engineering (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Dentistry (AREA)
- Neurosurgery (AREA)
- Psychology (AREA)
- Biodiversity & Conservation Biology (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
A method guides evasive actions to avoid effects of a cognitive impairment state. A first buffer, which is communicatively coupled to at least one sensor on a wearable sensor device, is loaded with a first set of sensor readings. The wearable sensor device receives a first cognitive impairment state signal based on an observer of a wearer of the wearable sensor device observing an impairment to a cognitive state of the wearer of the wearable sensor device. Subsequently, a second buffer on the wearable sensor device initiates loading of a second set of sensor readings, and the first buffer and the second buffer are compared. In response to sensors readings from the first and second buffers matching, an alert is issued to the wearer of the wearable sensor device, thus prompting the wearer to take evasive steps to avoid a recurrence of the impairment.
Description
- The present disclosure relates to the field of computers, and specifically to the use of computers in evaluating cognitive states. Still more particularly, the present disclosure relates to assisting a person in avoiding an impairment event associated with one or more cognitive states.
- A person's cognitive state is also known as a person's “state of mind”. This state of mind may be normal (e.g., interested, sleepy, asleep, alert, bored, curious, doubtful, etc.), or it may be indicative of some type of pathology (e.g., amnesia, confusion, panic, etc.). Often, such states of mind will manifest themselves measurably before a person (subjectively) realizes that he/she is entering such a state of mind.
- In one embodiment of the present invention, a method guides evasive actions to avoid effects of a cognitive impairment state. A first buffer, which is communicatively coupled to at least one sensor on a wearable sensor device, is loaded with a first set of time-dependent sensor readings. The wearable sensor device receives a first cognitive impairment state signal, where the first cognitive impairment state signal is sent in response to an observer, of the wearer, observing an impairment to a cognitive state of the wearer of the wearable sensor device. The cognitive impairment state marker is inserted at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal. A second buffer, on the wearable sensor device, initiates loading of a second set of time-dependent sensor readings from the at least one sensor on the wearable sensor device, and time-dependent sensor readings from the first buffer and the second buffer are compared. In response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, an alert is issued to the wearer of the wearable sensor device.
- In one embodiment of the present invention, a method enables a guidance of evasive actions to avoid effects of a cognitive impairment state. A first set of time-dependent sensor readings are received from a first buffer on a wearable sensor device, where the first buffer is communicatively coupled to at least one sensor on the wearable sensor device. A first cognitive impairment state signal is transmitted to the wearable sensor device, where the first cognitive impairment state signal is sent in response to an observer, of the wearer, observing an impairment to a cognitive state of the wearer of the wearable sensor device. A cognitive impairment state marker is transmitted to the wearable sensor device, wherein the cognitive impairment state marker is inserted at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal. An initiation of loading of a second buffer, which is on the wearable sensor device, with a second set of time-dependent sensor readings from the at least one sensor on the wearable sensor device is detected. Time-dependent sensor readings from the first buffer and the second buffer are compared to one another up to the predefined position. In response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, an alert is issued to the wearer of the wearable sensor device.
- In one embodiment of the present invention, a wearable sensor device is integrated into a sports helmet. The wearable sensor device comprises: a physiological sensor, wherein the physiological sensor detects a biological state of the wearer of the sports helmet; an accelerometer sensor, wherein the accelerometer sensor detects a change in velocity of the protective sports helmet; a first buffer for storing a first set of time-dependent sensor readings, wherein the first buffer is communicatively coupled to the physiological sensor and the accelerometer sensor; a receiver for receiving a first cognitive impairment state signal, wherein an observer of a wearer of the wearable sensor device sends the first cognitive impairment state signal in response to the observer observing (in one or more embodiments, subjectively) an impairment to a cognitive state of the wearer of the wearable sensor device; a data insertion logic for inserting a cognitive impairment state marker at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal; a second buffer, wherein the second buffer initiates loading of a second set of time-dependent sensor readings from the physiological sensor and the accelerometer sensor; a hardware comparator for comparing time-dependent sensor readings from the first buffer and the second buffer up to the predefined position; and an alert generator that issues an alert to the wearer of the wearable sensor device in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level.
-
FIG. 1 depicts an exemplary system and network in which the present disclosure may be implemented; -
FIG. 2 illustrates an exemplary Impaired Cognitive State Predictor (ICSP) architecture in accordance with one or more embodiments of the present invention; -
FIG. 3 depicts an exemplary wearable sensor device that is integrated into a sports helmet for monitoring physiological and physical conditions related to the sports helmet and its wearer; -
FIG. 4 illustrates an exemplary wrist-wearable sensor device for sensing user physiological and/or other conditions of a wearer of the device; -
FIG. 5 depicts a wearer of the wrist-wearable sensor device illustrated inFIG. 4 while speaking before an interactive audience; -
FIG. 6 illustrates a high-level block diagram of an exemplary networked system used in the embodiment depicted inFIG. 5 ; and -
FIG. 7 is a high-level flowchart of one or more steps performed by one or more processors to guide evasive actions for avoiding effects of an impaired cognitive state. - The present invention may be a system, a method, and/or a computer program product. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
- With reference now to the figures, and in particular to
FIG. 1 , there is depicted a block diagram of an exemplary system and network that may be utilized by and/or in the implementation of the present invention. Note that some or all of the exemplary architecture, including both depicted hardware and software, shown for and withincomputer 102 may be utilized bysoftware deploying server 150 and/or awearable sensor device 152, as well asimpairment signal transmitters 506 shown inFIG. 5 andFIG. 6 , and/ormonitoring system 602 shown inFIG. 6 . -
Exemplary computer 102 includes aprocessor 104 that is coupled to a system bus 106.Processor 104 may utilize one or more processors, each of which has one or more processor cores. Avideo adapter 108, which drives/supports adisplay 110, is also coupled to system bus 106. System bus 106 is coupled via abus bridge 112 to an input/output (I/O)bus 114. An I/O interface 116 is coupled to I/O bus 114. I/O interface 116 affords communication with various I/O devices, including akeyboard 118, amouse 120, a media tray 122 (which may include storage devices such as CD-ROM drives, multi-media interfaces, etc.), a wireless signal transceiver 124 (e.g., a near field radio frequency transceiver, a Wi-Fi transceiver, etc.), and external USB port(s) 126. While the format of the ports connected to I/O interface 116 may be any known to those skilled in the art of computer architecture, in one embodiment some or all of these ports are universal serial bus (USB) ports. - As depicted,
computer 102 is able to communicate with asoftware deploying server 150, using anetwork interface 130.Network interface 130 is a hardware network interface, such as a network interface card (NIC), etc.Network 128 may be an external network such as the Internet, or an internal network such as an Ethernet or a virtual private network (VPN). - A
hard drive interface 132 is also coupled to system bus 106.Hard drive interface 132 interfaces with ahard drive 134. In one embodiment,hard drive 134 populates asystem memory 136, which is also coupled to system bus 106. System memory is defined as a lowest level of volatile memory incomputer 102. This volatile memory includes additional higher levels of volatile memory (not shown), including, but not limited to, cache memory, registers and buffers. Data that populatessystem memory 136 includescomputer 102's operating system (OS) 138 andapplication programs 144. -
OS 138 includes ashell 140, for providing transparent user access to resources such asapplication programs 144. Generally,shell 140 is a program that provides an interpreter and an interface between the user and the operating system. More specifically,shell 140 executes commands that are entered into a command line user interface or from a file. Thus,shell 140, also called a command processor, is generally the highest level of the operating system software hierarchy and serves as a command interpreter. The shell provides a system prompt, interprets commands entered by keyboard, mouse, or other user input media, and sends the interpreted command(s) to the appropriate lower levels of the operating system (e.g., a kernel 142) for processing. Note that whileshell 140 is a text-based, line-oriented user interface, the present invention will equally well support other user interface modes, such as graphical, voice, gestural, etc. - As depicted,
OS 138 also includeskernel 142, which includes lower levels of functionality forOS 138, including providing essential services required by other parts ofOS 138 andapplication programs 144, including memory management, process and task management, disk management, and mouse and keyboard management. -
Application programs 144 include a renderer, shown in exemplary manner as abrowser 146.Browser 146 includes program modules and instructions enabling a world wide web (WWW) client (i.e., computer 102) to send and receive network messages to the Internet using hypertext transfer protocol (HTTP) messaging, thus enabling communication withsoftware deploying server 150 and other computer systems. -
Application programs 144 incomputer 102's system memory (as well assoftware deploying server 150's system memory) also include a Cognitive Impairment Event Avoidance Logic (CIEAL) 148.CIEAL 148 includes code for implementing the processes described below, including those described inFIGS. 2-7 . In one embodiment,computer 102 is able to downloadCIEAL 148 fromsoftware deploying server 150, including in an on-demand basis, wherein the code inCIEAL 148 is not downloaded until needed for execution. Note further that, in one embodiment of the present invention,software deploying server 150 performs all of the functions associated with the present invention (including execution of CIEAL 148), thus freeingcomputer 102 from having to use its own internal computing resources to executeCIEAL 148. - Note that
computer 102 is connected to apower supply 151 used to power various components withincomputer 102 and/or connected components (e.g.,elements power supply 151 may be a solar cell, a battery (rechargeable or non-rechargeable), a public utility power grid that is accessible via a wall socket, a local limited supply source (e.g., a local fuel-powered generator), etc. - Note that the hardware elements depicted in
computer 102 are not intended to be exhaustive, but rather are representative to highlight essential components required by the present invention. For instance,computer 102 may include alternate memory storage devices such as magnetic cassettes, digital versatile disks (DVDs), Bernoulli cartridges, and the like. These and other variations are intended to be within the spirit and scope of the present invention. - As a high-level overview of one or more embodiments of the present invention, the present invention takes three steps. The first step is to capture sensor readings that describe circumstantial environments of a wearer of a sensor device, and to store them in a buffer. Thereafter (the second step), an observer (in one or more embodiments a human observer) of the wearer of the sensor device, in response to observing an impairment to the cognitive state of the wearer of the sensor device, takes some action, which results in a marker being placed in the buffer. Thereafter (the third step), if the wearer of the sensor device repeats actions that lead to sensor readings matching those found in the first buffer, then an alert is given to the wearer to take evasive actions to avoid the effects of again experiencing the impaired cognitive state that resulted from the circumstantial environments captured in the first step.
- With reference now to
FIG. 2 , an exemplary Impaired Cognitive State Predictor (ICSP)architecture 200 is presented in accordance with one or more embodiments of the present invention. Note that theICSP architecture 200 and data generated by the ICSP architecture is secret. That is, predictions of current or future cognitive states are presented only to the user that generated the sensor readings described herein and/or experiences the specific cognitive state that follows these sensor readings. Only with the express approval of the user (i.e., wearer of the wearable sensor device described herein) will such readings/states be shared with others. - Note that in one embodiment the
ICSP architecture 200 also includes multiple components found inFIG. 1 (e.g.,computer 102,wearable sensor device 152, etc.). Furthermore, in one embodiment, the continuous circular buffer(s) 202, thepush data matrix 204, and/or theaccumulation data matrix 206 shown inFIG. 2 and/or components shown inFIG. 1 are all within thewearable sensor device 152. In one or more other embodiments, the continuous circular buffer(s) 202 are within the wearable sensor device, but thepush data matrix 204 and/oraccumulation data matrix 206 are stored in a hardware storage device (e.g.,system memory 136 and/orhard drive 134 shown inFIG. 1 ) on a remote computer, such ascomputer 102 shown inFIG. 1 . -
Wearable sensor device 152 includes one or more sensor(s) 208. In one embodiment, each of the sensor(s) 208 are “smart sensors”, that include processing logic that is able to detect, record, and quantify what is being sensed. That is, each of the sensor(s) 208 is 1) able to detect a particular physical event (heat, noise, biometrics, etc.); 2) quantify the level of that particular physical event (e.g., how high the heat is, what the duration/intensity of the noise is, what the specific readings of the biometric is, etc.); 3) convert that level into a digital value; and/or 4) send that digital value to the continuous circular buffer(s) 202. In one embodiment, these functions are performed by dedicated hardware logic, which takes digital readings from the sensors, compares the digital readings to known ranges in order to establish the digital value, and then transmits (e.g., by a wireless digital signal transmitter) the digital value to the continuous circular buffer(s) 202, which then (responsive to a “push” signal) send the stored digital values from the continuous circular buffer(s) 202 to a local matrix within thewearable sensor device 152 or to a remote matrix in a remote computer (e.g.,computer 102 shown inFIG. 1 ). In either embodiment, the system can use a near field network to send the digital value to a local storage within thewearable sensor device 152, or to a remote device, such as a smart phone held by the user, or to a server on a cloud, etc. (e.g., using a Wi-Fi signal). - Note that in one or more embodiments of the present invention, data stored in the continuous circular buffer(s) 202 is time-based. For example, data found in
buffer cell 220 a is received before data that is received and stored inbuffer cell 220 b, which stores data that is received before data that is received and stored inbuffer cell 220 c, etc. Thus, the data stored in continuous circular buffer(s) 202 (and thus the push data matrix 204) is “time-based”. - In one embodiment, sensor(s) 208 include physiological sensors, which are defined as sensors that are able to detect physiological states of a person. In one embodiment, these sensors are attached to the person via the
wearable sensor device 152. Example of such sensors include, but are not limited to, a heart monitor, a blood pressure cuff/monitor (sphygmomanometer), a galvanic skin conductance monitor, an electrocardiography (ECG) device, an electroencephalography (EEG) device, etc. That is, in one embodiment, the sensor(s) 208 are biometric sensors that measure physiological functions, of the wearer, which are not musculoskeletal. - In one embodiment, sensor(s) 208 detect and/or measure musculoskeletal bodily acts of the user, such as facial expressions (e.g., smiles, frowns, furrowed brows, etc.), body movements (e.g., walking gait, limps, stride length, stride speed, etc.), etc. Facial expressions may be detected by muscle movement sensors on eyeglasses, cameras on “smart glasses”, etc. Body movements may be detected by motion detectors, stride counters, strain gauges in clothing, etc.
- In one embodiment, sensor(s) 208 include speech content analyzers. In this embodiment, the sensor(s) 208 includes a speech-to-text converter, which then examines the text for certain keywords, speech pattern, etc. That is, the speech-to-text converter converts spoken words into written text, which can then be examined in order to identify certain predefined keywords, speech pattern, etc. The presence (or absence) of such keywords, speech pattern, etc. is then used by logic (
e.g. CIEAL 148 inFIG. 1 ) to ascertain the nature of the speech, which may lead to a warning of a future cognitive impairment state of the user (as described herein). - In one embodiment, sensor(s) 208 include speech content analyzers. In this embodiment, the sensor(s) 208 includes a speech-to-text converter, which then examines the text for certain features. These features may include the construction of graphs representing structural elements of speech based on a number of alternatives, such as syntactic value (article, noun, verb, adjective, etc.), or lexical root (run/ran/running) for the nodes of the graph, and text proximity for the edges of the graph. Graph features such as link degree, clustering, loop density, centrality, etc., representing speech topological structure are also therefore included. Similarly, semantic vectors may be extracted from the text as features, using systems such as that provided by a Latent Semantic Analysis, WordNet, etc. These methods allow the computation of a distance between words and specific concepts (e.g. introspection, anxiety, depression), such that the text can be transformed into features representing a field of distances to a concept, a field of fields of distances to the entire lexicon, or a field of distances to other texts including books, essays, chapters and textbooks. The syntactic and semantic features may then be combined either as a “bag of features” or as integrated fields, such as the Potts model. Similarly, locally embedded graphs may be constructed, so that a trajectory in a high-dimensional feature space is computed for each text. This trajectory is used as a measure of coherence of the speech, as well as a measure of distance between speech trajectories using methods such as Dynamic Time Warping.
- In one embodiment, sensor(s) 208 include speech inflection analyzers. In this embodiment, the sensor(s) 208 compare voice patterns with known voice patterns (pitch, timing, tremor, etc.) of the user, in order to identify certain emotions such as stress, relaxation, alertness, sleepiness, and other cognitive states. The presence (or absence) of such voice patterns is then used by logic (
e.g. CIEAL 148 inFIG. 1 ) to ascertain the current emotional state of the user, which may lead to a warning of a future cognitive impairment state of the user (as described herein). - In one embodiment, sensor(s) 208 include environmental sensors, such as an air thermometer, a microphone, a barometer, a light sensor, a moisture sensor, etc. In this embodiment, sensor(s) 208 are able to detect ambient (within the proximity of the user) environmental conditions, such as rain, various light levels, sound levels, air pressure, sound (e.g., noise, music, spoken words, etc.), etc.
- In one embodiment, sensor(s) 208 include accelerometers, which measure acceleration and/or deceleration forces as an object accelerates (i.e., increases speed) and/or decelerates (i.e., slows down and/or stops). These acceleration/deceleration forces may be abrupt, particularly the deceleration forces that occur when a moving object strikes another object, which may be fixed, moving in a direction opposite that of the first object, or is moving in the same direction as the first object but at a slower speed.
- As described herein, values stored in the continuous circular buffer(s) 202 are sent to the
push data matrix 204 in response to a “push” event. In one embodiment, the “push” event occurs in response to an observer observing a wearer of thewearable sensor device 152 exhibiting an impaired cognitive state. That is, as soon as the observer “feels” (i.e., subjectively determines) that the wearer is in a particular impaired cognitive state (e.g., is boring, is unfocused, is disoriented, etc.), then the observer issues a “push” command, causing the contents of the continuous circular buffer(s) 202 to be loaded into thepush data matrix 204. Note that in one or more embodiments of the present invention, the actions taken by the observer are based on the observer's subjective impressions, which are not based on clinical evidence. That is, the observations are not on unimpeachable scientific evidence of a particular pathology (e.g., an MRI that clearly shows damage to a cognition component of the brain), but rather are the subjective observations of the observer. Thus, one observer may view the wearer (person who is wearing the wearable sensor device 152) as being “fascinating” or “alert”, while another observer may view the same wearer as being “boring” or “disoriented”. Thus, the observers' observations are purely subjective in this embodiment, and are not directly correlated to any scientific/clinical facts supporting an impression of a specific cognitive impairment state. - Thus, in one or more embodiments of the present invention, sensor readings from sensor(s) 208 are buffered in the continuous circular buffer(s) 202. Continuous circular buffer(s) 202 are buffers that allow data to be stored in any location/cell within the buffer. Unlike a linear buffer (such as a First In First Out—FIFO buffer), a circular buffer allows “stale” data to be replaced with “fresh” data without shifting the location of existing data in other cells within the buffer. In one embodiment, continuous circular buffer(s) 202 is composed of multiple circular buffers 210 a-210 c (where “c” is an integer). In one embodiment, each of the circular buffers 210 a-210 c is devoted to storing readings from a specific sensor from sensor(s) 208.
- For example, assume that
circular buffer 210 a is devoted to storing readings from asensor 208 that measures a heart rate of the user. When data fromcircular buffer 210 a is sent to pushdata matrix 204, it is stored in the unrolled buffer shown as b1. Assume further thatcircular buffer 210 b is devoted to storing readings from asensor 208 that measures an ambient light level where the user is located. When data fromcircular buffer 210 b is sent to pushdata matrix 204, it is stored in the unrolled buffer shown as b2. Assume further thatcircular buffer 210 c is devoted to storing readings from asensor 208 that measures speech patterns of the user. When data fromcircular buffer 210 c is sent to pushdata matrix 204, it is stored in the unrolled buffer shown as b3. Thus, readings from a particular sensor are stored in a particular circular buffer as well as a particular unrolled (linear) buffer in a buffer matrix. - Note that while the present disclosure presents continuous circular buffer(s) 202 as a single circle, other circular buffers having multiple interlocking circular buffers are contemplated as being within the scope of the present invention.
- As described herein, data is sent from the continuous circular buffer(s) 202 to the
push data matrix 204 in response to a “push” being initiated by an observer of the wearer of thewearable sensor device 152 observing a particular impaired cognitive state of the wearer. Note that in one embodiment, these observations are purely subjective. That is, the perception of a particular impaired cognitive state is subjective and unique to that observer. For example, one observer may determine that the wearer of thewearable sensor device 152 is experience the cognitive state of “lucid and interesting” when sensor(s) 208 detect a particular pattern of conditions (physiological, temporal, environmental, etc.). However, another observer may determine that the wearer of thewearable sensor device 152 is being “disorganized and boring” when sensor(s) 208 detect this same particular pattern of conditions for this same wearer. Thus, each observer may respond differently to the wearer when the same set/pattern of conditions occurs. In order to address this subjective variation among multiple observers, in one embodiment a smoothing function is used to “smooth out” the observations from the observers of the wearer of thewearable sensor device 152, in order to come to an approximation of data that represents overall patterns of the observations, while eliminating outlier (i.e., out-of-bound, anomalous) observations that are unwarranted/unsupported. Examples of such smoothing functions/algorithms/filters include, but are not limited to, additive smoothing algorithms, Kalman filters, least-squares fitting of polynomials (representing the subjective observations) algorithms, moving averages, exponential smoothing (to reduce random fluctuations in time series data), curve fitting of observational data, numerical smoothing and differentiation, etc. - For example, assume that an observer has made multiple past observations of the wearer of the
wearable sensor device 152, and that the observations have varied not only due to the various states of the wearer (i.e., impairments of the wearer's cognitive state), but also to various states of the observer (i.e., the observer is hyper-alert, is sluggish, is angry, etc.), which affect whether or not the observer determines that there is an impairment to the cognitive state of the wearer of thewearable sensor device 152. Thus, in this embodiment, the observer of the wearer of the wearable sensor device sends multiple cognitive impairment state signals in response to the observer subjectively observing multiple instances of the impairment to the cognitive state of the wearer of the wearable sensor device, and the multiple cognitive impairment state signals are generated in response to the observer making multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device. In one embodiment, one or more processors apply a Kalman filter to the multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device. This Kalman filter uses a linear quadratic estimation to recursively remove anomalous observations from the multiple observations to generate a (in one embodiment, trusted) observation of the impairment to the cognitive state of the wearer of the wearable sensor device (i.e., wearable sensor device 152). An exemplary Kalman algorithm used for this determination is: -
x k =F k x k-1 +B k u k +w k - where xk is the (trusted) observation of the impairment to the cognitive state of the wearer of the wearable sensor device,
Fk is a predefined state transition model that is applied to a previous state xk-1 of observed impairments to the cognitive state of the wearer of the wearable sensor device,
Bk is a predefined control-input model that is applied to a control vector uk, and
wk is erroneous observation noises that are drawn from a multivariate normal distribution Qk, wherein wk is approximately equal to the set of numbers N from zero to Qk (N(0, Qk)). - Continuing now with reference to
FIG. 2 , in response to observing a particular cognitive impairment state, the observer will initiate a “push” of data from the continuous circular buffer(s) 202 to thepush data matrix 204, which is stored on a hardware storage device. As shown inFIG. 2 , thecognitive impairment state 212 is represented by a digital value that is sent to thepush data matrix 204. This digital value identifies a particular cognitive impairment state of the wearer of thewearable sensor device 152, which is defined by an observer of the wearer, such that precursive readings from the sensor(s) 208 are associated with a subsequent and specific cognitive impairment state and/or other related events (e.g., “benching” a sports player). - In one embodiment, the particular cognitive impairment state that is associated with specific precursive events (detected by the sensor(s) 208) is described by the wearer's own words and/or by the observer's own words. In another embodiment, the particular cognitive impairment state is selected from a menu or is otherwise predefined.
- Continuing now with
FIG. 2 , assume that data from continuous circular buffer(s) 202 is continuously sent toaccumulation data matrix 206. In this embodiment,accumulation data matrix 206 takes continuous readings from the continuous circular buffer(s) 202. However, each set of data that has been pushed to thepush data matrix 204 is nonetheless identified within theaccumulation data matrix 206. For example, data that was pushed to pushdata matrix 204 at the time of a “PUSH-A” is identified byblock 214 a; data that was pushed to pushdata matrix 204 at the time of a “PUSH-B” is identified byblock 214 b; and data that was pushed to pushdata matrix 204 at the time of a “PUSH-C” is identified byblock 214 c. - Data from blocks 214 a-214 c are then used to determine a Push Triggered Average (PTA), shown as push
average matrix 216. Pushaverage matrix 216 is calculated (e.g., byCIEAL 148 shown inFIG. 1 ) as one-third (assuming that three pushes occurred) of the sum of the values stored in each cell of the pushed buffers. That is, assume that a push results in three columns of nine cells. The values in the upper left cell in each of the blocks 214 a-214 c are summed together, divided by three, and the quotient (i.e., average) is then stored in the upper left cell of the pushaverage matrix 216. Other cells in the 214 a-214 c are similarly summed together, divided by three, and their quotients (i.e., averages) are then stored in the corresponding cell of the pushaverage matrix 216. This PTA (push average matrix 216) is then used as a “fuzzy” reference for new values pushed from the continuous circular buffer(s) 202. That is, PTA (push average value 216) provides a mean average for each of the sensed parameters. Ranges around these mean values (above and below) are predetermined, such that when values from the continuous circular buffer(s) 202 later fall within these ranges, a prediction can be made that the wearer will again experience (or is currently experiencing) the particular cognitive impairment state. - In one embodiment, a
buffer data matrix 218 is generated from a single buffer in thepush data matrix 204. For example, consider buffer b1 frompush data matrix 204. Assume that buffer b1 contains data from continuouscircular buffer 210 a that describe the heart rate of the user who is wearing thewearable sensor device 152. As depicted, b1 is broken down into three rows, r1-r3, in order to create thebuffer data matrix 218.Buffer data matrix 218 is then used in a manner similar to that described herein for push data matrixes. That is, rather than require a push data matrix from multiple sensors, pushed (or alternatively, non-pushed but rather continuously streamed) data from a single sensor is converted into a matrix (buffer data matrix 218), which is then used to warn of an impending cognitive impairment state of the wearer by comparing thisbuffer data matrix 218 to known single-sensor data matrixes that are precursive to the particular cognitive impairment state of the user. - Note that while, as the name suggests,
wearable sensor device 152/352/452/552 is presented as a wearable sensor device, in one or more embodiments thewearable sensor device 152/352/452/552 is a device that is simply proximate to, although not necessarily worn by, a user, such that ambient conditions, including biophysical traits of the user (e.g., frowns, smiles, flushed skin, etc.) are still sensed by sensors, such assensors 208. - In one embodiment, the wearable sensor device is integrated into a sport helmet. For example, as shown in
FIG. 3 , an exemplarywearable sensor device 352 is integrated into asports helmet 300 for monitoring physiological and physical conditions related to the sports helmet and its wearer. One or more components of the exemplarywearable sensor device 352 are powered by alocal battery 351 and/or an equivalent power source (e.g., a solar cell). - In one embodiment, the
wearable sensor device 352 includes one or more biometric sensor(s) 308, which measure physiological states (i.e., perspiration, skin temperature, eye flutter, voice articulations such as grunts of pain, etc.) of the wearer, similar to thesensors 208 described above. In addition, thewearable sensor device 352 includes anaccelerometer 306. Theaccelerometer 306, which may be any known electro-mechanical device that measures acceleration, including rapid deceleration, detects whether thehelmet 300 has been subjected to a sharp blow (as indicated by a sudden acceleration/deceleration detected by the accelerometer 306), initiated by the wearer or by another player. Note that any blow that occurs when the player is not actually wearing the helmet 300 (e.g., while being transported to the game, if jostled within a gym bag, etc.) is irrelevant to the cognitive state of the athlete. Thus, in one embodiment thebattery 351 is only connected when the athlete puts on thehelmet 300. This selective powering on/off may be from a sensor switch (not shown, but within the interior lining of the helmet 300), a manual switch (also not shown), etc. - A
processor 304 processes readings from the biometric sensor(s) 308 and/or theaccelerometer 306 by loading them into buffer(s) 302, which in one embodiment have the same architecture and function as the continuous circular buffer(s) 202 shown inFIG. 2 . - A data I/O port 326 (which in one embodiment has a same architecture as USB port(s) 126 shown in
FIG. 1 ) is able to 1) download data from the buffer(s) 302, and 2) upload markers into the buffer(s) 302. - Thus, in the embodiment of the present invention in which a
wearable sensor device 352 is integrated into asports helmet 302, thewearable sensor device 352 includes a physiological sensor such as one or more of the biometric sensor(s) 308. These biometric/physiological sensors detect a biological state of the wearer of thesports helmet 302, such as his/her heart rate, perspiration level, skin temperature, EEG and/or EKG, oxygen saturation level, etc. In this embodiment, thewearable sensor device 352 also includes an accelerometer sensor, such asaccelerometer 306. This accelerometer sensor detects a change in velocity (e.g., a crash into another player's helmet, striking the ground or other immovable object, etc.). - Also part of the
wearable sensor device 352 is a first buffer (i.e., one of the buffer(s) 302 shown inFIG. 3 ). This first buffer (which in one embodiment is implemented as a dedicated hardware storage device—“hardware buffer”) is communicatively coupled to the physiological sensor (biometric sensor(s) 308) and the accelerometer sensor (accelerometer 306), and thus is able to store a first set of time-dependent sensor readings (i.e., sensor readings from the biometric sensor(s) 308 and/or theaccelerometer 306 which are retrieved/recorded in linear time (sequentially)). - The
wearable sensor device 352 also includes a receiver (e.g., part of transceiver 310) for receiving a first cognitive impairment state signal. That is, when an observer of a wearer of the wearable sensor device (wearable sensor device 352 integrated into helmet 302) subjectively observes an impairment to a cognitive state of the wearer of the wearable sensor device/helmet, the observer sends the first cognitive impairment state signal to the transceiver 310. A data insertion logic (e.g., part of processor 304) then inserts a cognitive impairment state marker at a predefined position in the first buffer. - Thereafter, a second buffer (i.e., one of the buffer(s) 302, and which may be the same as the first buffer if steps are taken to clear and save data from the first buffer for further use), initiates loading of a second set of time-dependent sensor readings from the physiological sensor and the accelerometer sensor. For example assume that the wearer of the
helmet 302 is a contact sport (e.g., American football) player. The first buffer may store data from the biometric sensor(s) 308 and/or theaccelerometer 306 taken during a first game. Thereafter, the second buffer may store data from biometric sensor(s) 308 and/or theaccelerometer 306 taken during a later game (e.g., played the following week). - That is, assume that the first buffer contains the following sensor readings from the biometric sensor(s) 308 (identified as “Bx”) and the accelerometer 306 (identified as “Ax”): B1, B2, B3, A1, A2, A3. As described herein, “Bx” and “Ax” are time-dependent, meaning that they are stored in the first buffer as they are generated by their respective sensors. For purposes of illustration, assume then that the cognitive impairment state marker (identified as “I”) is placed after “Bx” and “Ax”, thus giving the temporal sequence of: B1, B2, B3, A1, A2, A3, I.
- A hardware comparator (e.g., part of processor 304) then compares time-dependent sensor readings from the first buffer up to the predefined position with time-dependent sensor readings from the second buffer. For example, assume again that the first buffer contains the sensor readings B1, B2, B3, A1, A2, A3, which are stored before the cognitive impairment state marker I. Assume also that the second buffer contains the sensor readings B1, B2, A1, A2. Although B1, B2, A1, A2 is not the same as B1, B2, B3, A1, A2, A3, the present invention recognizes that the wearer is taking actions that have a similar pattern as those recorded during the last game. Thus, the present invention provides an “intervention” with the wearer of the helmet, giving the wearer the opportunity to perform evasive actions (i.e., stop tackling with his helmet, being less aggressive, etc.) to avoid reaching the pattern B1, B2, B3, A1, A2, A3, which prompted the earlier signal (and thus cognitive impairment state marker I) from the observer.
- Note that in one embodiment of the present invention, the cognitive impairment state marker I is also part of the precursive pattern that warns of an impending impaired cognitive state. In this embodiment, the cognitive impairment state marker I functions as a warning during future activities. For example, the cognitive impairment state marker I, along with the sensor readings B1, B2, B3, A1, A2, A3, would produce a pattern of B1, B2, B3, A1, A2, A3, I. Assume that the pattern B1, B2, B3, A1, A2, A3 is recorded a subsequent event (meeting, sports game, etc.) for the wearer of the wearable sensor device. This newly-derived pattern B1, B2, B3, A1, A2, A3 will then cause the system to recognize “I” as the next data point, which causes a warning to be issued to the wearer of the wearable sensor device. That is, at this point, the wearer of the wearable sensor device may not be presenting evidence of an impaired cognitive state, either internally (i.e., the wearer of the wearable sensor device does not yet “feel” the impaired cognitive state) or to another (i.e., the observer of the wearable sensor device does not perceive that the wearer of the wearable sensor device has entered into the impaired cognitive state). However, a warning, triggered by the “I” after the newly-recorded pattern of B1, B2, B3, A1, A2, A3, will give the wearer additional warning that an impaired cognitive state is impending. In one or more embodiments, this intervention signal (indicated by the cognitive impairment state marker I) and/or the warning derived therefrom is used as part of the Bk predefined control-input model that is applied to a control vector uk in the Kalman filter described herein.
- Thus, an alert generator (e.g., part of processor 304), in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, issues an alert to the wearer of the wearable sensor device. That is, in the example presented above, if the second buffer stores the pattern B1, B2, A1, A2, then a predefined match level (e.g., 4 out of 6 of the sensor readings B1, B2, B3, A1, A2, A3) has been reached. This triggers the alert generator to send a signal to the wearer of the helmet, such as sending a color-coded signal to a multi-color light emitting diode (LED)
device 312 mounted on thefaceguard 314 of thehelmet 302. - If capable of selectively displaying different colors, the
LED device 312 may turn yellow if the alert is to let the wearer know that continuing the same style of play will result in receiving a second cognitive impairment state signal from the observer. For example, this second cognitive impairment state signal may result in the player being prevented from further play, being required to take additional instruction, being required to submit to a medical examination, etc. - Similarly, the
LED device 312 may turn red if the alert advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer. That is, if the player continues to play in the same manner, the red alert signal indicates to the wearer that continuing this style of play will result in the same impairment (e.g., disorientation, confusion, etc.) that was observed before. In one embodiment, the red alert signal may result in the player being immediately taken out of the game. - While the present invention has been described above in the context of a sporting event, in another embodiment the present invention is applied to a non-sporting event. For example, in another embodiment, the wearable sensor device is simply worn on the wrist. Thus, and with reference now to
FIG. 4 , consider thewearable sensor device 452, which is structurally similar to thewearable device 152 shown inFIG. 2 , and includes sensors such as sensor(s) 208, a power supply (not shown), buffers (e.g., continuous circular buffer(s) 202), etc. needed for a wearable sensor device as presented herein. As shown, wearable sensor device 452 (which may be worn on the wrist) includes akeypad 402. In one embodiment, keys on the keypad are pre-programmed for a particular cognitive state. For example, one of the keys may be for “boredom”. Thus, if the user is experiencing “boredom”, then the user pushing the button for “boredom” causes data from the continuous circular buffer(s) 202, along with a flag/signal that is associated with thecognitive impairment state 212 for “boredom” (and identified by pushing the key onkeypad 402 for “boredom”), to be sent to thepush data matrix 204 inFIG. 2 . Similarly, if the user is experiencing “anxiety”, then data from the continuous circular buffer(s) 202, along with a flag/signal that is associated with thecognitive impairment state 212 for “anxiety” (and identified by pushing the key onkeypad 402 for “anxiety”), is sent to thepush data matrix 204 inFIG. 2 when the user pushes the “anxiety” button on thekeypad 402. Thus, if an observer of the wearer of thewearable sensor device 452 depicted inFIG. 4 decides that the wearer is exhibiting boredom or anxiety, this observation can be confirmed by the wearer's own personal sensation. - Note that in one embodiment, the
wearable sensor device 452 shown inFIG. 4 includes both biophysical (unique to the user) and ambient environmental sensors. More specifically,wearable sensor device 452 includes biometric sensors that, depending on their structure, configuration, and/or positioning on thewearable sensor device 452, are able to monitor biometric conditions (e.g., blood pressure, heart rate, etc.), musculoskeletal motions (e.g., cameras that track a user's facial expressions, motion sensors that track a user's walking gait, etc.) and other biophysical features/conditions of the user, but also can track ambient environmental conditions (e.g., local sounds, light, moisture, air temperature, etc.). - With reference now to
FIG. 5 , consider the scenario in which a wearer of the wrist-wearable sensor device is a person making a presentation (i.e., presenter 502) to anaudience 504. Note that thepresenter 502 is wearing awearable sensor device 552, which has the same hardware and configuration as thewearable sensor device 452 depicted inFIG. 4 . During the presentation given bypresenter 502, members of the audience are able to input cognition state impairment signals viaimpairment signal transmitters 506. For example, if one or more of the members of theaudience 504 perceive that thepresenter 502 is boring or anxious or otherwise doing poorly, then this information may be displayed on the presenter'sdisplay 508 on his/herlaptop 510. If only a few members of the audience feel that thepresenter 502 is doing poorly, then thepresenter 502 may still have time to salvage the presentation. However, and in accordance with a preferred embodiment of the present invention, if a threshold of the members of the audience 503 (e.g., 50%) indicate that thepresenter 502 is doing poorly, then it is too late to salvage the current presentation. However, thepresenter 502 still has useful data points for future presentations. That is, if thewearable sensor device 552 detects a similar pattern progression as that which ultimately led to the previous audience's negative response, then a signal on the presenter'sdisplay 508 will offer suggestions to avoid (execute avoidance actions) another poor performance. In one or more embodiments of the present invention, these suggestions are defined as part of the Bk predefined control-input model that is applied to a control vector uk of the Kalman filter described herein. - For example, at the subsequent presentation, if the
presenter 502 is once again moving about too much (as detected by sensors in the wearable sensor device 552), the presenter'sdisplay 508 may present a suggestion to “Stand Still”. That is, in the previous presentation, excessive movement by the presenter 502 (i.e., pacing back and forth, excessive hand/body gestures, etc.) led the previous audience to subjectively determine that thepresenter 502 was “frenetic”. If thepresenter 502 again displays such excessive movement during a subsequent/current presentation, then the system will advise him/her to modulate his body movements (e.g., with the suggestion that he/she “Stand Still”). -
FIG. 6 illustrates a high-level block diagram of an exemplary networked system used in the embodiment depicted inFIG. 5 . That is, a monitoring system 602 (e.g.,computer 102 shown inFIG. 1 ) receives inputs from the audience (via their impairment signal transmitters 506) and the presenter (via his/her wearable sensor device 552). Themonitoring system 602, using a processor 604 (analogous to theprocessor 104 shown inFIG. 1 ) compares sensor patterns from previous presentations given by thepresenter 502 with current sensor patterns. If themonitoring system 602 detects that a similar pattern is being followed (although not up to the point of losing the audience's attention/approval), then an alert and/or suggestion is sent to the presenter'sdisplay 508. - With reference now to
FIG. 7 , a high-level flowchart of one or more steps performed by one or more processors to guide evasive actions for avoiding effects of an impaired cognitive state is presented. Again, note that one or more of the steps depicted may be performed by one or more processors (e.g.,processor 104 inFIG. 1 ,processor 304 inFIG. 3 ,processor 604 inFIG. 6 , etc.) - After
initiator block 702, a first buffer is loaded on a wearable sensor device with a first set of time-dependent sensor readings (block 704). As described herein, the first buffer is communicatively coupled to at least one sensor on the wearable sensor device. - As described in
block 706, the wearable sensor device receives a first cognitive impairment state signal. As described herein, this first cognitive impairment state signal is generated by an observer of a wearer of the wearable sensor device in response to the observer subjectively observing an impairment to a cognitive state of the wearer of the wearable sensor device. For example, if the observer thinks that the wearer of the wearable sensor device appears to be confused, boring, anxious, etc., then the observer will generate the first cognitive impairment state signal. In one embodiment, this first cognitive impairment state signal is transmitted from a transmitter (e.g., one or more of theimpairment signal transmitters 506 shown inFIG. 5 ). In one embodiment, this first cognitive impairment state signal is an automatic consequence of taking the step of downloading data from the sensor data buffer (e.g., buffer(s) 302 inFIG. 3 ) via a data I/O port (e.g., data I/O port 326 inFIG. 3 ). - As described in
block 708, a cognitive impairment state marker is then inserted at a predefined position in the first buffer. This cognitive impairment state marker (e.g., “I” presented above) is inserted at the predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal. - As described in
block 710, loading of a second buffer on the wearable sensor device with a second set of time-dependent sensor readings from at least one sensor on the wearable sensor device is initiated. Thus, as described above, sensor reading B1 may be loaded, followed by sensor readings B1, A1, followed by sensor readings B1, A1, A2, etc. - As described in
block 712, time-dependent sensor readings from the first buffer up to the predefined position are then compared with time-dependent sensor readings from the second buffer. In response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings (now being taken in real time) reaching a predefined match level (query block 714), then an alert is issued to the wearer of the wearable sensor device (block 716). - The predefined match level may be numeric, weighted, etc. For example, assume that the first buffer contains the sensor readings B1, A1, A2, B2, B3, B4 (where “B” indicates a biometric sensor reading and “A” indicates an accelerometer reading). After storing B1, A1, A2, B2, B3, B4 in the first buffer, the system receives a cognitive impairment state marker “I”, which indicates that there is an observed impairment to the cognitive state of the wearer of the wearable sensor device, and thus “I” is placed after the sensor readings B1, A1, A2, B2, B3, B4 in the first buffer.
- Assume now that the second buffer (or the first buffer after being cleared and its contents stored in a local memory for use in comparison to new sensor data) contains the sensor readings B1, A1, A2, B2. The “predefined match level” needed to initiate the alert may simply be a percentage, such as 50%. Thus, 50% of the six sensor readings B1, A1, A2, B2, B3, B4 would be three, regardless of which sensor readings are subsequently taken and stored in the second buffer.
- However, in another embodiment, each of the sensor readings B1, A1, A2, B2, B3, B4 may be weighted. For example, sensor readings from the accelerometer (A1, A2, A3, etc.) may be predetermined to be more important, and thus weighted more heavily, than sensor readings form the biometric sensors (B1, B2, B3, etc.). Thus, as few as one or two accelerometer events (A1, A2) may be enough to trigger the alert, even without any biometric events (B1, B2, etc.).
- The flow-chart of
FIG. 7 ends atterminator block 718. - As described herein, in one embodiment of the present invention the alert to the wearer of the wearable sensor device advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid receiving a second cognitive impairment state signal from the observer. That is, the alert may advise the wearer to take corrective/ameliorative steps to avoid being told again that his/her style of play is dangerous, he/she is boring, etc.
- As described herein, in one embodiment of the present invention the alert to the wearer of the wearable sensor device advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer. That is, the alert may advise the wearer of the wearable sensor device that his/her style of play will actually cause him/her to be disoriented, that his/her presentation style will be boring, etc., even if nobody tells him/her.
- In one embodiment of the present invention, the first set of time-dependent sensor readings are analyzed, in order to identify a cause of the impairment to the cognitive state of the wearer. For example, assume that historical records show that when events that result in sensor readings A1, A1, A3, B1, B2, B3 ultimately lead to a state of dementia, a conclusion is reached that these events are the cause of the state of dementia. Assuming that such records and conclusions are available, then in one embodiment they are used to verify the observations of the observer of the wearer of the wearable sensor device.
- As described herein, in one embodiment of the present invention the first buffer and the second buffer in the wearable sensor device are both continuous circular buffers, in which each stores data from a different sensor in the wearable sensor device. In this embodiment, a cause of the impairment to the cognitive state of the wearer of the wearable sensor device is predicted by a probability formula:
-
- where:
P(M|E) is a probability that the impairment to the cognitive state will occur (M) given that (|) data from the continuous circular buffers falls within a predefined Push Triggered Average (PTA—described above) of previously pushed data from the continuous circular buffers (E);
P(E|M) is a probability that data from the continuous circular buffers falls within the predefined PTA of previously pushed data from the continuous circular buffers (E) given that (|) the impairment to the cognitive state of the wearer is actually occurring (M);
P(M) is a probability that the impairment to the cognitive state of the wearer will occur regardless of any other information; and
Σm is a sum of all occurrences m, for the probability P(E|M) times the probability P(M). - In one embodiment of the present invention, predicting whether the impairment to the cognitive state of the wearer of the wearable sensor device will occur is based on a statistical analysis of the subsequent set of sensor readings compared to the pushed sensor readings, wherein a match within a predefined statistical range between the subsequent set of sensor readings and the pushed sensor readings leads to the prediction of the impairment to the cognitive state of the wearer of the wearable sensor device. That is, if the previous sensor readings align with current sensor readings within a statistically significant range, then an assumption/prediction is made that a recurrence of the cognitive impairment state is likely.
- As described herein, in one embodiment of the present invention the predefined position in the first buffer at which the cognitive impairment state marker is inserted is at an end of the first set of time-dependent sensor readings. In this embodiment, a determination is made that sensor readings stored prior to the end of the first set of time-dependent sensor readings are precursors to the impairment to the cognitive state of the wearer of the wearable sensor device. That is if the sensor readings B1, B2, B3, A1, A2, A3 lead to the insertion of the cognitive impairment state marker I at their end, then a conclusion is reached that the events that caused sensor readings B1, B2, B3, A1, A2, A3 lead to the cognitive impairment state reflected by “I”.
- In one embodiment of the present invention, the first set of time-dependent sensor readings are made up of a first subset of time-dependent sensor readings and a second subset of time-dependent sensor readings. The first subset of time-dependent sensor readings record event states that occur before event states that are represented by the second subset of time-dependent sensor readings. For example, a set of sensor readings B1, A1, A2, B2, B3, A3 may be made up of a first subset of sensor readings B1, A1, A2, which occur before a second subset of sensor readings B2, B3, A3 (i.e., the events that caused sensor readings B1, A1, A2 occurred before events that caused sensor readings B2, B3, A3). Similarly, the second set of time-dependent sensor readings comprise a third subset of time-dependent sensor readings and a fourth subset of time-dependent sensor readings, where the third subset of time-dependent sensor readings record event states that occur before event states that are represented by the fourth subset of time-dependent sensor readings. For example, the third subset found at the beginning of the second set of time-dependent sensors may be B1, A1, A2. Since the first subset and the third subset are identical (B1, A1, A2), then a conclusion is reached that the wearer of the wearable sensor device is headed for a specific cognitive impairment state, regardless of the contents of the fourth subset of time-dependent sensor readings in the second set of time-dependent sensor readings, and an alert is issued.
- As noted herein, in one or more embodiments of the present invention the sensors in the wearable sensor device detect physiological states of the user, musculoskeletal bodily acts of the user, keywords spoken by the user, a quality of a voice pattern from the user, and ambient environmental conditions around the user. Thus, specific patterns of sensor readings from all of these sensors are used to provide a warning of a recurrence of a particular impairment to the cognitive state of the wearer of the wearable sensor device.
- Note that any methods described in the present disclosure may be implemented through the use of a VHDL (VHSIC Hardware Description Language) program and a VHDL chip. VHDL is an exemplary design-entry language for Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), and other similar electronic devices. Thus, any software-implemented method described herein may be emulated by a hardware-based VHDL program, which is then applied to a VHDL chip, such as a FPGA.
- Having thus described embodiments of the present invention of the present application in detail and by reference to illustrative embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the present invention defined in the appended claims.
Claims (20)
1. A method of guiding evasive actions to avoid effects of an impaired cognitive state, the method comprising:
loading a first buffer on a wearable sensor device with a first set of time-dependent sensor readings, wherein the first buffer is communicatively coupled to at least one sensor on the wearable sensor device;
receiving, by the wearable sensor device, a first cognitive impairment state signal, wherein an observer of a wearer of the wearable sensor device sends the first cognitive impairment state signal in response to observing an impairment to a cognitive state of the wearer of the wearable sensor device;
inserting a cognitive impairment state marker at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal;
initiating loading of a second buffer on the wearable sensor device with a second set of time-dependent sensor readings from said at least one sensor on the wearable sensor device;
comparing time-dependent sensor readings from the first buffer up to the predefined position with time-dependent sensor readings from the second buffer; and
in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, issuing an alert to the wearer of the wearable sensor device.
2. The method of claim 1 , wherein the alert advises the wearer of the wearable sensor device to take an action that has been predetermined to avoid receiving a second cognitive impairment state signal from the observer.
3. The method of claim 2 , further comprising:
loading the alert into the second buffer as a member of the second set of time-dependent sensor readings.
4. The method of claim 1 , wherein the alert advises the wearer of the wearable sensor device to take an action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer.
5. The method of claim 1 , further comprising:
identifying a cause of the impairment to the cognitive state of the wearer by analyzing, by one or more processors, the first set of time-dependent sensor readings.
6. The method of claim 1 , wherein the first buffer and the second buffer are both continuous circular buffers, wherein each of the continuous circular buffers stores data from a different sensor in the wearable sensor device, and wherein the method further comprises:
predicting a cause of the impairment to the cognitive state of the wearer based on a probability formula:
where:
P(M|E) is a probability that the impairment to the cognitive state will occur (M) given that (|) data from the continuous circular buffers falls within a predefined Push Triggered Average (PTA) of previously pushed data from the continuous circular buffers (E);
P(E|M) is a probability that data from the continuous circular buffers falls within the predefined PTA of previously pushed data from the continuous circular buffers (E) given that (|) the impairment to the cognitive state of the wearer is actually occurring (M);
P(M) is a probability that the impairment to the cognitive state of the wearer will occur regardless of any other information; and
Σm is a sum of all occurrences m, for the probability P(E|M) times the probability P(M).
7. The method of claim 1 , further comprising:
predicting whether the impairment to the cognitive state of the wearer of the wearable sensor device will occur based on a statistical analysis of the second set of time-dependent sensor readings compared to the first set of time-dependent sensor readings, wherein a match within a predefined statistical range between the second set of time-dependent sensor readings and the first set of time-dependent sensor readings leads to a prediction of the impairment to the cognitive state of the wearer of the wearable sensor device.
8. The method of claim 1 , wherein the wearable sensor device is a protective sports helmet, wherein said at least one sensor comprises a physiological sensor and an accelerometer sensor, wherein the physiological sensor detects a biological state of the wearer of the wearable sensor device, wherein the accelerometer sensor detects a change in velocity of the protective sports helmet, and wherein the method further comprises:
loading the first buffer and the second buffer with sensor readings from a combination of the physiological sensor and the accelerometer sensor.
9. The method of claim 8 , further comprising:
detecting a download of the first set of time-dependent sensor readings from the first buffer; and
in response to detecting the download of the first set of time-dependent sensor readings from the first buffer, generating the first cognitive impairment state signal.
10. The method of claim 1 , wherein the predefined position in the first buffer at which the cognitive impairment state marker is inserted is at an end of the first set of time-dependent sensor readings, and wherein the method further comprises:
determining that sensor readings stored prior to the end of the first set of time-dependent sensor readings are precursors to the impairment to the cognitive state of the wearer of the wearable sensor device.
11. The method of claim 1 , wherein the first set of time-dependent sensor readings comprise a first subset of time-dependent sensor readings and a second subset of time-dependent sensor readings, wherein the first subset of time-dependent sensor readings record event states that occur before event states that are represented by the second subset of time-dependent sensor readings, wherein the second set of time-dependent sensor readings comprise a third subset of time-dependent sensor readings and a fourth subset of time-dependent sensor readings, wherein the third subset of time-dependent sensor readings record event states that occur before event states that are represented by the fourth subset of time-dependent sensor readings, and wherein the method further comprises:
defining the partial match as a match of sensor readings from the first subset and the third subset of time-dependent sensor readings.
12. The method of claim 1 , wherein the said at least one sensor are multiple sensors that detect physiological states of the user, musculoskeletal bodily acts of the user, keywords spoken by the user, a quality of a voice pattern from the user, and ambient environmental conditions around the user.
13. A method of enabling a guidance of evasive actions to avoid an impaired cognitive state, the method comprising:
receiving a first set of time-dependent sensor readings from a first buffer on a wearable sensor device, wherein the first buffer is communicatively coupled to at least one sensor on the wearable sensor device;
transmitting, to the wearable sensor device, a first cognitive impairment state signal, wherein an observer of a wearer of the wearable sensor device sends the first cognitive impairment state signal in response to observing an impairment to a cognitive state of the wearer of the wearable sensor device;
transmitting a cognitive impairment state marker to the wearable sensor device, wherein the cognitive impairment state marker is inserted at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal;
detecting an initiation of loading of a second buffer on the wearable sensor device with a second set of time-dependent sensor readings from said at least one sensor on the wearable sensor device;
comparing time-dependent sensor readings from the first buffer up to the predefined position with time-dependent sensor readings from the second buffer; and
in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level, issuing an alert to the wearer of the wearable sensor device.
14. The method of claim 13 , wherein the alert advises the wearer of the wearable sensor device to take an evasive action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer.
15. The method of claim 13 , wherein the observer of the wearer of the wearable sensor device sends multiple cognitive impairment state signals in response to observing multiple instances of the impairment to the cognitive state of the wearer of the wearable sensor device, wherein the multiple cognitive impairment state signals are generated in response to the observer making multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device, and wherein the method further comprises:
applying, by one or more processors, a Kalman filter to the multiple observations of the impairment to the cognitive state of the wearer of the wearable sensor device, wherein the Kalman filter uses a linear quadratic estimation to recursively remove anomalous observations from the multiple observations to generate an observation of the impairment to the cognitive state of the wearer of the wearable sensor device using an algorithm:
x k =F k x k-1 +B k u k +w k
x k =F k x k-1 +B k u k +w k
where xk is the observation of the impairment to the cognitive state of the wearer of the wearable sensor device,
Fk is a predefined state transition model that is applied to a previous state xk-1 of observed impairments to the cognitive state of the wearer of the wearable sensor device,
Bk is a predefined control-input model that is applied to a control vector uk, and
wk is erroneous observation noises that are drawn from a multivariate normal distribution Qk, wherein wk is approximately equal to the set of numbers N from zero to Qk (N(0, Qk)).
16. The method of claim 13 , wherein the first buffer and the second buffer are both continuous circular buffers, wherein each of the continuous circular buffers stores data from a different sensor in the wearable sensor device, and wherein the method further comprises:
predicting a cause of the impairment to the cognitive state of the wearer based on a probability formula:
where:
P(M|E) is a probability that the impairment to the cognitive state will occur (M) given that (|) data from the continuous circular buffers falls within a predefined Push Triggered Average (PTA) of previously pushed data from the continuous circular buffers (E);
P(E|M) is a probability that data from the continuous circular buffers falls within the predefined PTA of previously pushed data from the continuous circular buffers (E) given that (|) the impairment to the cognitive state of the wearer is actually occurring (M);
P(M) is a probability that the impairment to the cognitive state of the wearer will occur regardless of any other information; and
Σm is a sum of all occurrences m, for the probability P(E|M) times the probability P(M).
17. The method of claim 13 , wherein the first set of time-dependent sensor readings comprise a first subset of time-dependent sensor readings and a second subset of time-dependent sensor readings, wherein the first subset of time-dependent sensor readings record event states that occur before event states that are represented by the second subset of time-dependent sensor readings, wherein the second set of time-dependent sensor readings comprise a third subset of time-dependent sensor readings and a fourth subset of time-dependent sensor readings, wherein the third subset of time-dependent sensor readings record event states that occur before event states that are represented by the fourth subset of time-dependent sensor readings, and wherein the method further comprises:
defining the partial match as a match of sensor readings from the first subset and the third subset of time-dependent sensor readings.
18. A sports helmet, wherein a wearable sensor device is integrated into the sports helmet, and wherein the wearable sensor device comprises:
a physiological sensor, wherein the physiological sensor detects a biological state of the wearer of the sports helmet;
an accelerometer sensor, wherein the accelerometer sensor detects a change in velocity of the protective sports helmet;
a first buffer for storing a first set of time-dependent sensor readings, wherein the first buffer is communicatively coupled to the physiological sensor and the accelerometer sensor;
a receiver for receiving a first cognitive impairment state signal, wherein an observer of a wearer of the wearable sensor device sends the first cognitive impairment state signal in response to the observer subjectively observing an impairment to a cognitive state of the wearer of the wearable sensor device;
a data insertion logic for inserting a cognitive impairment state marker at a predefined position in the first buffer in response to the wearable sensor device receiving the first cognitive impairment state signal;
a second buffer, wherein the second buffer initiates loading of a second set of time-dependent sensor readings from the physiological sensor and the accelerometer sensor;
a hardware comparator for comparing time-dependent sensor readings from the first buffer up to the predefined position with time-dependent sensor readings from the second buffer; and
an alert generator that issues an alert to the wearer of the wearable sensor device in response to a partial match of the first set of time-dependent sensors readings up to the predefined position and the second set of time-dependent sensors readings sensor readings reaching a predefined match level.
19. The sport helmet of claim 18 , wherein the alert advises the wearer of the wearable sensor device to take an action that has been predetermined to avoid receiving a second cognitive impairment state signal from the observer.
20. The sports helmet of claim 18 , wherein the alert advises the wearer of the wearable sensor device to take an action that has been predetermined to avoid experiencing an impairment to the cognitive state of the wearer.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/328,349 US20160007910A1 (en) | 2014-07-10 | 2014-07-10 | Avoidance of cognitive impairment events |
US15/848,092 US10827927B2 (en) | 2014-07-10 | 2017-12-20 | Avoidance of cognitive impairment events |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/328,349 US20160007910A1 (en) | 2014-07-10 | 2014-07-10 | Avoidance of cognitive impairment events |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/848,092 Continuation US10827927B2 (en) | 2014-07-10 | 2017-12-20 | Avoidance of cognitive impairment events |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160007910A1 true US20160007910A1 (en) | 2016-01-14 |
Family
ID=55066090
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/328,349 Abandoned US20160007910A1 (en) | 2014-07-10 | 2014-07-10 | Avoidance of cognitive impairment events |
US15/848,092 Expired - Fee Related US10827927B2 (en) | 2014-07-10 | 2017-12-20 | Avoidance of cognitive impairment events |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/848,092 Expired - Fee Related US10827927B2 (en) | 2014-07-10 | 2017-12-20 | Avoidance of cognitive impairment events |
Country Status (1)
Country | Link |
---|---|
US (2) | US20160007910A1 (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160100794A1 (en) * | 2014-10-08 | 2016-04-14 | Landon C.G. Miller | Real time brain trauma treatment |
US20160278686A1 (en) * | 2015-03-23 | 2016-09-29 | International Business Machines Corporation | Monitoring a person for indications of a brain injury |
US20170319063A1 (en) * | 2016-05-09 | 2017-11-09 | Neurovision Imaging Llc | Apparatus and method for recording and analysing lapses in memory and function |
US20180146876A1 (en) * | 2016-11-29 | 2018-05-31 | The General Hospital Corporation | Systems and methods for analyzing electrophysiological data from patients undergoing medical treatments |
US20180299973A1 (en) * | 2016-03-16 | 2018-10-18 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US10380882B1 (en) * | 2018-06-28 | 2019-08-13 | International Business Machines Corporation | Reconfigurable hardware platform for processing of classifier outputs |
CN111067537A (en) * | 2019-11-11 | 2020-04-28 | 珠海格力电器股份有限公司 | Sleeping posture monitoring method, monitoring terminal and storage medium |
US20200249771A1 (en) * | 2016-03-16 | 2020-08-06 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US11039761B2 (en) * | 2018-12-14 | 2021-06-22 | At&T Intellectual Property I, L.P. | Fall prediction based on electroencephalography and gait analysis data |
US20210369115A1 (en) * | 2017-07-18 | 2021-12-02 | Facecontrol, Inc. | Telemetry monitoring and assessment system of parameters of human vital functions and clinical signs of possible functional disorders and determination of methods of control |
US11324421B2 (en) * | 2014-09-15 | 2022-05-10 | 3M Innovative Properties Company | Impairment detection with environmental considerations |
JP2022133183A (en) * | 2021-03-01 | 2022-09-13 | パラマウントベッド株式会社 | Information processing device and information processing method |
US11463605B2 (en) | 2016-02-12 | 2022-10-04 | Contrast, Inc. | Devices and methods for high dynamic range video |
US11910099B2 (en) | 2016-08-09 | 2024-02-20 | Contrast, Inc. | Real-time HDR video for vehicle control |
US11985316B2 (en) | 2018-06-04 | 2024-05-14 | Contrast, Inc. | Compressed high dynamic range video |
US12309427B2 (en) | 2018-08-14 | 2025-05-20 | Contrast, Inc. | Image compression |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110103077B (en) * | 2019-05-08 | 2020-09-22 | 北京理工大学 | An intelligent monitoring method of boring state with multi-sensor integration of boring bar |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100217097A1 (en) * | 2005-06-29 | 2010-08-26 | National Ict Australia Limited | Measuring cognitive load |
US20100324427A1 (en) * | 2008-02-22 | 2010-12-23 | Koninklijke Philips Electronics N.V. | System and kit for stress and relaxation management |
US20110301436A1 (en) * | 2009-04-22 | 2011-12-08 | Teixeira Rodrigo E | Apparatus for processing physiological sensor data using a physiological model and method of operation therefor |
US20120029311A1 (en) * | 2010-07-27 | 2012-02-02 | Carefusion 303, Inc. | System and method for storing and forwarding data from a vital-signs monitor |
US20120289789A1 (en) * | 2011-05-13 | 2012-11-15 | Fujitsu Limited | Continuous Monitoring of Stress Using Environmental Data |
US20150226621A1 (en) * | 2014-02-13 | 2015-08-13 | Evigia Systems, Inc. | System and method for head acceleration measurement in helmeted activities |
US20150305426A1 (en) * | 2014-04-25 | 2015-10-29 | Ford Global Technologies, Llc | Bicycle helmet with integrated electronics |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5621922A (en) * | 1992-01-10 | 1997-04-22 | Rush, Iii; Gus A. | Sports helmet capable of sensing linear and rotational forces |
US6497658B2 (en) | 1999-12-19 | 2002-12-24 | Michael F. Roizen | Alarm upon detection of impending sleep state |
KR100580618B1 (en) | 2002-01-23 | 2006-05-16 | 삼성전자주식회사 | Apparatus and method for recognizing user emotion through short time monitoring of physiological signals |
US7874983B2 (en) | 2003-01-27 | 2011-01-25 | Motorola Mobility, Inc. | Determination of emotional and physiological states of a recipient of a communication |
US7403815B2 (en) | 2004-06-04 | 2008-07-22 | Drexel University | Brain state recognition system |
WO2006050523A1 (en) | 2004-11-02 | 2006-05-11 | Medtronic, Inc. | Apparatus for data retention in an implantable medical device |
US7653605B1 (en) | 2005-04-15 | 2010-01-26 | Science Applications International Corporation | Method of and apparatus for automated behavior prediction |
JP4783181B2 (en) | 2006-03-13 | 2011-09-28 | 株式会社東芝 | Behavior prediction device |
US20090002178A1 (en) | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Dynamic mood sensing |
US20110184319A1 (en) * | 2010-01-22 | 2011-07-28 | X2Impact, Inc. | Mouth guard with sensor |
US20110263946A1 (en) | 2010-04-22 | 2011-10-27 | Mit Media Lab | Method and system for real-time and offline analysis, inference, tagging of and responding to person(s) experiences |
US20130262182A1 (en) | 2012-03-31 | 2013-10-03 | Affectiva, Inc. | Predicting purchase intent based on affect |
EP2593015B1 (en) * | 2010-07-15 | 2018-03-14 | The Cleveland Clinic Foundation | Classification of impacts from sensor data |
US8702516B2 (en) * | 2010-08-26 | 2014-04-22 | Blast Motion Inc. | Motion event recognition system and method |
EP2609533A2 (en) | 2010-08-27 | 2013-07-03 | Smartex S.r.l. | Monitoring method and system for assessment or prediction of mood trends |
US8412665B2 (en) | 2010-11-17 | 2013-04-02 | Microsoft Corporation | Action prediction and identification temporal user behavior |
MX352036B (en) * | 2011-09-01 | 2017-11-07 | Riddell | Systems and methods for monitoring a physiological parameter of persons engaged in physical activity. |
US20130297536A1 (en) | 2012-05-01 | 2013-11-07 | Bernie Almosni | Mental health digital behavior monitoring support system and method |
US20140073993A1 (en) * | 2012-08-02 | 2014-03-13 | University Of Notre Dame Du Lac | Systems and methods for using isolated vowel sounds for assessment of mild traumatic brain injury |
US9024770B2 (en) * | 2013-02-12 | 2015-05-05 | David Isidore Reuben | Electronic football concussion avoidance training module |
US9615797B2 (en) * | 2013-03-16 | 2017-04-11 | Jaison C. John | Method, apparatus and system for determining a health risk using a wearable housing for sensors |
US8621673B1 (en) * | 2013-03-20 | 2014-01-07 | Antonio Pietrantonio | Concussion indicator |
US20150040685A1 (en) * | 2013-08-08 | 2015-02-12 | Headcase Llc | Impact sensing, evaluation & tracking system |
US20150206053A1 (en) | 2014-01-21 | 2015-07-23 | Samsung Electronics Co., Ltd. | Computing system with comprehensive sensor mechanism and method of operation thereof |
US20160018278A1 (en) * | 2014-07-18 | 2016-01-21 | II Perry Jeter | Impact detection |
-
2014
- 2014-07-10 US US14/328,349 patent/US20160007910A1/en not_active Abandoned
-
2017
- 2017-12-20 US US15/848,092 patent/US10827927B2/en not_active Expired - Fee Related
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100217097A1 (en) * | 2005-06-29 | 2010-08-26 | National Ict Australia Limited | Measuring cognitive load |
US20100324427A1 (en) * | 2008-02-22 | 2010-12-23 | Koninklijke Philips Electronics N.V. | System and kit for stress and relaxation management |
US20110301436A1 (en) * | 2009-04-22 | 2011-12-08 | Teixeira Rodrigo E | Apparatus for processing physiological sensor data using a physiological model and method of operation therefor |
US20120029311A1 (en) * | 2010-07-27 | 2012-02-02 | Carefusion 303, Inc. | System and method for storing and forwarding data from a vital-signs monitor |
US20120289789A1 (en) * | 2011-05-13 | 2012-11-15 | Fujitsu Limited | Continuous Monitoring of Stress Using Environmental Data |
US20150226621A1 (en) * | 2014-02-13 | 2015-08-13 | Evigia Systems, Inc. | System and method for head acceleration measurement in helmeted activities |
US20150305426A1 (en) * | 2014-04-25 | 2015-10-29 | Ford Global Technologies, Llc | Bicycle helmet with integrated electronics |
Cited By (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11324421B2 (en) * | 2014-09-15 | 2022-05-10 | 3M Innovative Properties Company | Impairment detection with environmental considerations |
US20160100794A1 (en) * | 2014-10-08 | 2016-04-14 | Landon C.G. Miller | Real time brain trauma treatment |
US10292643B2 (en) * | 2014-10-08 | 2019-05-21 | Astrocytical, Inc. | Real time brain trauma treatment |
US10653353B2 (en) * | 2015-03-23 | 2020-05-19 | International Business Machines Corporation | Monitoring a person for indications of a brain injury |
US20160278686A1 (en) * | 2015-03-23 | 2016-09-29 | International Business Machines Corporation | Monitoring a person for indications of a brain injury |
US20160278666A1 (en) * | 2015-03-23 | 2016-09-29 | International Business Machines Corporation | Monitoring a person for indications of a brain injury |
US10667737B2 (en) * | 2015-03-23 | 2020-06-02 | International Business Machines Corporation | Monitoring a person for indications of a brain injury |
US12250357B2 (en) | 2016-02-12 | 2025-03-11 | Contrast, Inc. | Combined HDR/LDR video streaming |
US11785170B2 (en) | 2016-02-12 | 2023-10-10 | Contrast, Inc. | Combined HDR/LDR video streaming |
US11463605B2 (en) | 2016-02-12 | 2022-10-04 | Contrast, Inc. | Devices and methods for high dynamic range video |
US20180299973A1 (en) * | 2016-03-16 | 2018-10-18 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US11307682B2 (en) * | 2016-03-16 | 2022-04-19 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US20200249771A1 (en) * | 2016-03-16 | 2020-08-06 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US10664075B2 (en) * | 2016-03-16 | 2020-05-26 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
WO2017196785A1 (en) * | 2016-05-09 | 2017-11-16 | Neurovision Imaging Llc | Apparatus and method for recording and analysing lapses in memory and function |
JP2019523027A (en) * | 2016-05-09 | 2019-08-22 | ニューロビジョン イメージング,インコーポレーテッド | Apparatus and method for recording and analysis of memory and function decline |
US20170319063A1 (en) * | 2016-05-09 | 2017-11-09 | Neurovision Imaging Llc | Apparatus and method for recording and analysing lapses in memory and function |
US11910099B2 (en) | 2016-08-09 | 2024-02-20 | Contrast, Inc. | Real-time HDR video for vehicle control |
WO2018102402A1 (en) * | 2016-11-29 | 2018-06-07 | The General Hospital Corporation | Systems and methods for analyzing electrophysiological data from patients undergoing medical treatments |
US20180146876A1 (en) * | 2016-11-29 | 2018-05-31 | The General Hospital Corporation | Systems and methods for analyzing electrophysiological data from patients undergoing medical treatments |
US10786168B2 (en) * | 2016-11-29 | 2020-09-29 | The General Hospital Corporation | Systems and methods for analyzing electrophysiological data from patients undergoing medical treatments |
US20210369115A1 (en) * | 2017-07-18 | 2021-12-02 | Facecontrol, Inc. | Telemetry monitoring and assessment system of parameters of human vital functions and clinical signs of possible functional disorders and determination of methods of control |
US11985316B2 (en) | 2018-06-04 | 2024-05-14 | Contrast, Inc. | Compressed high dynamic range video |
US10380882B1 (en) * | 2018-06-28 | 2019-08-13 | International Business Machines Corporation | Reconfigurable hardware platform for processing of classifier outputs |
US12309427B2 (en) | 2018-08-14 | 2025-05-20 | Contrast, Inc. | Image compression |
US11039761B2 (en) * | 2018-12-14 | 2021-06-22 | At&T Intellectual Property I, L.P. | Fall prediction based on electroencephalography and gait analysis data |
CN111067537A (en) * | 2019-11-11 | 2020-04-28 | 珠海格力电器股份有限公司 | Sleeping posture monitoring method, monitoring terminal and storage medium |
JP2022133183A (en) * | 2021-03-01 | 2022-09-13 | パラマウントベッド株式会社 | Information processing device and information processing method |
JP7689837B2 (en) | 2021-03-01 | 2025-06-09 | パラマウントベッド株式会社 | Information processing device and information processing method |
Also Published As
Publication number | Publication date |
---|---|
US10827927B2 (en) | 2020-11-10 |
US20180110412A1 (en) | 2018-04-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10827927B2 (en) | Avoidance of cognitive impairment events | |
JP6892233B2 (en) | Two-way remote patient monitoring and state management intervention system | |
US10524711B2 (en) | Cognitive event predictor | |
CN107427716B (en) | Method and system for optimizing and training human performance | |
US20210015417A1 (en) | Emotion data training method and system | |
US9763571B2 (en) | Monitoring a person for indications of a brain injury | |
Rehg et al. | Mobile health | |
Arif et al. | Classification of anxiety disorders using machine learning methods: a literature review | |
JP4481682B2 (en) | Information processing apparatus and control method thereof | |
US20140170609A1 (en) | Personalized compliance feedback via model-driven sensor data assessment | |
US11612341B2 (en) | Systems and methods for assessment and measurement of reaction time in virtual/augmented reality | |
US10653353B2 (en) | Monitoring a person for indications of a brain injury | |
US10877444B1 (en) | System and method for biofeedback including relevance assessment | |
WO2014052506A2 (en) | Devices and methods to facilitate affective feedback using wearable computing devices | |
Guthier et al. | Affective computing in games | |
US9968287B2 (en) | Monitoring a person for indications of a brain injury | |
CN108209946A (en) | An emotion control device and method | |
Cheng et al. | Predicting transitions in oxygen saturation using phone sensors | |
Abdessalem et al. | Toward real-time system adaptation using excitement detection from eye tracking | |
Mahmood | A package of smartphone and sensor-based objective measurement tools for physical and social exertional activities for patients with illness-limiting capacities | |
Mateos-García et al. | Driver Stress Detection in Simulated Driving Scenarios with Photoplethysmography | |
Jyotsna et al. | A Personalized Healthcare Platform for Monitoring Mental Health of a Person During COVID | |
Limbani et al. | WEARS: Wearable Emotion AI with Real-time Sensor data | |
Eio et al. | Design of data collection and analysis method for a pleasant and safe user experience of personal mobility device | |
Adithya et al. | A Deep Learning Framework for Comprehensive Sleep Analysis and Quality Tracking |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BOSS, GREGORY J.;DHILLON, JILL S.;HAMILTON, RICK A., II;AND OTHERS;REEL/FRAME:033289/0982 Effective date: 20140710 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |