WO2014066703A2 - Smart contextual display for a wearable device - Google Patents
Smart contextual display for a wearable device Download PDFInfo
- Publication number
- WO2014066703A2 WO2014066703A2 PCT/US2013/066716 US2013066716W WO2014066703A2 WO 2014066703 A2 WO2014066703 A2 WO 2014066703A2 US 2013066716 W US2013066716 W US 2013066716W WO 2014066703 A2 WO2014066703 A2 WO 2014066703A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- display
- context
- motion data
- modifying
- time points
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G04—HOROLOGY
- G04C—ELECTROMECHANICAL CLOCKS OR WATCHES
- G04C3/00—Electromechanical clocks or watches independent of other time-pieces and in which the movement is maintained by electric means
- G04C3/001—Electromechanical switches for setting or display
- G04C3/002—Position, e.g. inclination dependent switches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G04—HOROLOGY
- G04G—ELECTRONIC TIME-PIECES
- G04G9/00—Visual time or date indication means
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2200/00—Indexing scheme relating to G06F1/04 - G06F1/32
- G06F2200/16—Indexing scheme relating to G06F1/16 - G06F1/18
- G06F2200/163—Indexing scheme relating to constructional details of the computer
- G06F2200/1637—Sensing arrangement for detection of housing movement or orientation, e.g. for controlling scrolling or cursor movement on the display of an handheld computer
Definitions
- the disclosure generally relates to the field of modifying and controlling a display for a wearable device based on a context of the device.
- Wearable devices such as watches, music players and the like enable users to interact with technology in a convenient and continuous manner, since they can be present on the body in the context of all lifestyle activities.
- Devices now have more and more functions. While additional functionality is supposed to add utility, increased functionality also means interacting with the device more to activate the various functions. Additionally, wearable devices are preferably small and thus any controls are also small. These aspects result in devices that have less than desired utility.
- FIG. 1 illustrates one embodiment of a wearable device with a display.
- FIG. 2 illustrates another view of an embodiment of a wearable device.
- FIG. 3 illustrates a flow chart for activating a display according to one embodiment.
- FIG. 4 illustrates a wearable device and axes around which motion is determined according to one embodiment.
- One embodiment of the disclosed device and method includes a motion detection system to identify the context of a device.
- the context of the device is a function of its motion and relative position as well as time.
- the context of the device can be a position of the device relative to the user or an activity in which the user is engaged.
- the operation of a screen (or display) of the device is modified in response to the identified context. Modifying operation of the display includes, but is not limited to, activating or deactivating the display, increasing or decreasing a brightness of the display, turning on a backlight on the display, and providing information on the display.
- the identified context is the device facing the user and in response to that context, a screen (or display) of the device is activated. Activating the display includes turning on the display or turning on a backlight to make the display more visible.
- the context for the device results in specified content being provided on the display of the device.
- FIG. 1 it illustrates an embodiment of a wearable device 100.
- the exemplary device 100 is worn on the wrist attached through a fastening system 101.
- the fastening system 101 may be removable, exchangeable or customizable.
- the device 100 includes a display 102 and user interaction points 103.
- FIG. 2 illustrates another view of an embodiment of the wearable device 100.
- the view includes the fastening system 101, display 102. Additionally the motion system 207 and processor 205 are shown.
- Example processors 205 include the TIMSP430 from TEXAS INSTRUMENTS and ARM Cortex-M class microcontrollers.
- the processor 205 receives data from the position system 207 and determines when to activate display 102.
- the motion system 207 is any kind of motion sensor.
- Example motion sensors include an accelerometer, a gyroscope, a pressure sensor, a compass and a magnetometer. EXAMPLE PROCESSING CONFIGURATION
- FIGS. 3 and 4 An example of the disclosed device and method is described in reference to FIGS. 3 and 4.
- the context identified is the device 100 being positioned to face the user and in response the display 102 is activated or deactivated.
- the processor 205 receives 309 data from the motion system 207 in three axes.
- the motion system 207 in this embodiment is an accelerometer.
- FIG. 4 illustrates another view of an embodiment of the device 100 and shows one example of the axes relative to which motion data and position data is determined by the motion system 207.
- the processor 205 compares 311 the data for each axis and determines when each axis is within a predetermined range. The predetermined ranges are indicative that the device has been turned to face the user.
- the processor activates 313 the display 102.
- the processor 205 determines the device has been turned to face the user based on data from one or more axes being within its predetermined range for a threshold period of time.
- predetermined ranges for motion data are within an X-Y-Z Cartesian coordinate system for a user wearing a device 100 while in an upright position, for example:
- the processor activates the display when the data along one or more axes falls within its predetermined threshold for more than 300 milliseconds (ms). In other embodiments, the data along the one or more axes for 250 or 500 ms.
- the optimal time period can be determined such that the display does not activate inadvertently and still turns on quickly enough to be responsive to the user's expectation.
- the processor 205 continues to monitor motion and relative position data and deactivates the display 102 when the motion and relative position data fall outside the range that triggered activation. In some embodiments, the processor 205 deactivates the display after data from the motion system 207 for just one of the axes is no longer in the threshold range. The processor 205 can deactivate the display 102 instantaneously or after the data for the one axis remains outside the threshold for 500 ms.
- processor 205 continuously monitors the motion system 207 and activates the display 102 when there is a rotational motion of the device 100 with a principal rotational axis coincident with that of the user's wrist. If the user is upright, the wrist is along the Y-axis (referring to FIG. 4) and rotation in the negative Y direction yielding an increasing X acceleration due to increased coincidence with gravity indicates display 102 is being rotated toward user. For example, a change in the positive X acceleration of +0.25g over the course of a predetermined time period, e.g., 1 second, may indicate such a rotation.
- the processor 205 When the condition is met within a certain predefined timeframe, and when rotation stops with the device oriented with the display 102 (defined by the X/Y plane as in the example above) the processor 205 activates display 102.
- the processor 205 applies timing windows and orientation range limits to protect against the activation of the backlight during similar gesticulations such as running or showering.
- the processor 205 activates the display 102 when there is complex rotational motion of the device 100 with multiple axes of rotation.
- the first rotational component corresponding to rotation of the user's forearm about their elbow (primarily observed as a rotation about the Z axis), thus indicating that the user's forearm is being swung or lifted up toward the user's face.
- a second rotational component corresponding to rotation of the user's wrist (see the rotation about the "Y-axis", as in the example above, and in FIG. 4), in the negative Y direction, thus indicating display 102 is being rotated toward the user.
- the following operations may be detected:
- the device 100 is then observed to remain in the final orientation observed at the end of the previous step for a period of time, indicating that the user is viewing the device 100 and the display 102 should be activated.
- This period of time could be 300ms, as in the above example.
- the device could disable the display 102 with a change in orientation, a rotation out of this position or a timeout. This timeout could also be 500ms, as in the example above.
- the processor 205 applies an algorithm to the orientation of the gravity vector, relative to the device's reference frame.
- the output of the algorithm identifies complex rotations that indicate a change in user context. For example, (referring to FIG. 4) when device 100 rotation causes the gravity vector to traverse the X/Y plane and indicate rotation in the negative Y direction, with rotation stopping at a device orientation indicating user viewing, and when these conditions are met within a certain predefined timeframe, the processor 205 will activate the display 102.
- a window of rotational components may also be defined, whereby the timeframe to execute the rotations is 1 second.
- the algorithm uses timing windows and orientation range limits to protect against the activation of the display 102 during similar gesticulations such as running or showering.
- the display 102 cannot be activated for a predetermined amount of time after it has been deactivated.
- reactivation is blocked only after a threshold number of activations and deactivations within a given time period. This further protects against inadvertent activation of the display 102. For example, if the display has been activated and deactivated a number of times in a minute, it is likely that the activation is in error and thus it is beneficial to prevent the display 102 from reactivating for a period of time such as 5 seconds or 10 seconds.
- the processor 205 can provide specified content for display on display 102 based on an identified context of the device 100.
- the identification of context includes learning from the user's interactions with the device 100. For example, if a user accesses data from the device 100 around the same time every morning, the processor 205 can display the usual data in response to the device 100 facing the user at that time. If the device 100 is turned to face the user at another time of day, the processor 205 merely activates the display 102 but does not provide any particular content.
- the device 100 may use context information to modify other display parameters based on context, including the contrast, brightness, orientation or displayed content.
- the device may also use other feedback such as an audio cue or a physical feedback such as vibration.
- a context may also be used to trigger input.
- the device 100 may activate other functions such as a microphone to enable voice recording, a speaker to activate sound, atelephony or voice- activated feature, or the connection to a wireless network to synchronize its data or new display content.
- Example wireless networks include a LAN, MAN, WAN, mobile network, and a telecommunication network.
- the device 100 communicates via the wireless network to modify operation of a display on a remote device in the same way that operation of a display on the device 100 is modified. Operation of the remote display may be modified in addition to or in place of modifying operation of the display 104 on the device 100.
- a combination of contexts may be detected in sequence to create additional contextual information. For example, if the motion system 207 detects that a user is stepping (e.g., walking, running), and the device 100 is facing the user, a recent step count may be displayed as well as the backlight being activated. This is an example of two detected contexts providing additional opportunity for customization of the user experience based on more than one detected context in parallel or in sequence. Another example would be using a detection of sleep to disable a backlight being activated. For example, if the motion system 207 detects a period of low motion, the device 100 may require a period of high motion before the automatic backlight would again be activated when the device is positioned to face the user.
- An additional identified context collected may also represent social gestures such as a handshake, "fist bump", or "high five". Since this action has a particular orientation and acceleration pattern associated with it, the context of this social gesture may be detected and used to modify the nature of the data displayed or stored around the event. For example, a precise timestamp, location, data signature or other information may be saved as a result of this event. The information could also be compared to other users' saved data as a means of identifying social gestures between users. For example, if another user from a similar location had engaged in a similar gesture at the same time, these two users could be linked on a network, share some information or have the event recorded on the device 100 for subsequent processing.
- Another example of multiple contexts being used to generate behavior or user information would be the automatic detection of which hand the device 100 is being worn on.
- the accelerometer signals to detect step events, as well as orientation, the device 100 can infer where on the body or in which orientation the device 100 is being worn. For example, if the device 100 is being worn on the wrist, a first context of slowly stepping and a second context of the orientation of the device 100 could be used to determine which wrist the device 100 is being worn on.
- the device 100 may incorporate other sensors and use them to improve the mechanism described above. These may be to provide additional contextual information, or to provide information for display based on contextual conditions. Examples of each of these sensors are summarized in Table 1 below.
- Ultraviolet light visible light, moisture/humidity, air content
- Motion sensing can provide additional contextual information via the detection and recognition of signature motion environments, actions, and/or contexts such as: in-car, in-airplane, walking, running, swimming, exercising, brushing teeth, washing car, eating, drinking, shaking hands, arm wrestling.
- signature motion environments, actions, and/or contexts such as: in-car, in-airplane, walking, running, swimming, exercising, brushing teeth, washing car, eating, drinking, shaking hands, arm wrestling.
- an accelerometer is used to detect the gesture corresponding to the wearer looking at the device 100 is another example.
- Other gestures such as a fist-bump, hand shake, wave, or signature are further examples.
- Motion and relative position analytics may also be calculated for display, such as step count, activity type, context specific analysis of recent activity, summaries for the day, week, month or other time period to date.
- multiple motion sensors can be used. For example both an accelerometer and gyroscope could be used. Additional types of motion sensors provides for more detailed inputs allowing for determination of additional contexts of the device 100.
- a detected context is an activity in which a user is engaged.
- An example activity is exercise.
- Exercises that involve repeated motion are detected by the processor 205 by identifying from motion data received from the motion system 207 repeated motions within a predetermined time period.
- Lifting weights and jumping jacks are examples of exercises that involve repeating the same motion.
- the device 100 may contain non-invasive, internal physiological parameter sensing such as the detection of blood flow, respiration or other parameters. These could be used to detect context, or as inputs to the display triggered by a context detected. For example, an accelerometer could detect the context of running, followed by the context of looking at the device 100, and the device 100 could display both a detected distance for the run and the maximum heart rate. Similarly, the device 100 could detect the blood alcohol level of the wearer and, if above a threshold, this context could trigger a visual alert to the user.
- non-invasive, internal physiological parameter sensing such as the detection of blood flow, respiration or other parameters. These could be used to detect context, or as inputs to the display triggered by a context detected. For example, an accelerometer could detect the context of running, followed by the context of looking at the device 100, and the device 100 could display both a detected distance for the run and the maximum heart rate. Similarly, the device 100 could detect the blood alcohol level of the wearer and, if above a threshold, this context could trigger
- Thermal sensors could be used to detect body, skin and ambient temperature for the purpose of context detection or display parameters. For example, a sensor able to detect skin temperature could use this information to infer the context of the user exerting himself physically and change the display to reflect parameters relevant to physical effort. Thermal information may also be relevant to display based on other contexts. For example, if the wearer is sweating due to physical exertion, the difference between environmental temperature and skin temperature could provide a parameter to inform the time to recovery from the physical exertion. In this case the context could be detected by an accelerometer, but the displayed metric be derived (at least in part) from thermal sensors.
- a chance in temperature identifies the context that the wearer of the device has changed locations.
- a drop in ambient temperature from 90 degrees F to 75 degrees F indicates the user has left outside and gone into a building.
- Skin surface sensors detect parameters purely via measuring non-invasively from the skin surface.
- a galvanic skin response or electrodermal activity sensor may be used to detect small changes in skin surface conductivity and thereby detect events such as emotional arousal.
- This can be used as context for an adaptive display modification, such as triggering the backlight or vibrating the device 100 to alert the user to the stress event and help them take action to mitigate it.
- a device 100 could vibrate more strongly if a stress event was more pronounced.
- a perspiration sensor could also be used to generate a display of workout parameters such as intensity if the context of physical exertion was detected. This context could be detected by the perspiration sensor itself, or by another sensor such as the accelerometer.
- Sensors that detect signals pertaining to the environment around the wearer provide both a unique reference point for signals sourced from the wearer himself, as well as additional context to inform adaptive processing and display.
- a device 100 that included an ultraviolet light sensor could modify the display when the user had received the recommended maximum exposure to ultraviolet light for the day.
- a user whose context has been determined to be sweating via a skin surface perspiration sensor may be exposed to a display of the humidity in the air around them as a result of this contextual awareness.
- Environmental sensors can also be used to identify a change from an indoor to outdoor context (or vice versa).
- An ambient light sensor would sense the change from generally lower light indoors to generally brighter light outdoors.
- a humidity sensor an also identify this context. Heating and cooling a building often results in less humidity than outdoors. Thus an increase or decrease in humidity is indicative of changing context from indoors to outdoors or the reverse.
- the disclosed embodiments beneficially allow for making a device more intuitive and therefore useful for the user.
- any reference to "one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment.
- the appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
- Coupled and “connected” along with their derivatives. It should be understood that these terms are not intended as synonyms for each other. For example, some embodiments may be described using the term “connected” to indicate that two or more elements are in direct physical or electrical contact with each other. In another example, some embodiments may be described using the term “coupled” to indicate that two or more elements are in direct physical or electrical contact. The term “coupled,” however, may also mean that two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other. The embodiments are not limited in this context.
- the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion.
- a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.
- "or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Hardware Design (AREA)
- User Interface Of Digital Computer (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/438,207 US20150277572A1 (en) | 2012-10-24 | 2013-10-24 | Smart contextual display for a wearable device |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261717642P | 2012-10-24 | 2012-10-24 | |
US61/717,642 | 2012-10-24 | ||
US201261727074P | 2012-11-15 | 2012-11-15 | |
US61/727,074 | 2012-11-15 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2014066703A2 true WO2014066703A2 (en) | 2014-05-01 |
WO2014066703A3 WO2014066703A3 (en) | 2014-06-19 |
Family
ID=50545490
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2013/066716 WO2014066703A2 (en) | 2012-10-24 | 2013-10-24 | Smart contextual display for a wearable device |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2014066703A2 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016097381A1 (en) * | 2014-12-19 | 2016-06-23 | Koninklijke Philips N.V. | Dynamic wearable device behavior based on schedule detection |
WO2018111614A1 (en) * | 2016-12-15 | 2018-06-21 | Holor, Llc | Wearable multi-functional personal security device |
US10088866B2 (en) | 2015-03-18 | 2018-10-02 | Motorola Mobility Llc | Controlling the orientation of a device display based on usage context |
CN110638422A (en) * | 2014-09-23 | 2020-01-03 | 飞比特公司 | Method, system and device for updating screen content in response to user gesture |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050212760A1 (en) * | 2004-03-23 | 2005-09-29 | Marvit David L | Gesture based user interface supporting preexisting symbols |
US8462109B2 (en) * | 2007-01-05 | 2013-06-11 | Invensense, Inc. | Controlling and accessing content using motion processing on mobile devices |
EP2347321B1 (en) * | 2008-09-25 | 2013-09-18 | Movea S.A | Command by gesture interface |
US20110224564A1 (en) * | 2010-03-10 | 2011-09-15 | Sotera Wireless, Inc. | Body-worn vital sign monitor |
US20110304531A1 (en) * | 2010-06-10 | 2011-12-15 | Peter Brooks | Method and system for interfacing and interaction with location-aware devices |
-
2013
- 2013-10-24 WO PCT/US2013/066716 patent/WO2014066703A2/en active Application Filing
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110638422A (en) * | 2014-09-23 | 2020-01-03 | 飞比特公司 | Method, system and device for updating screen content in response to user gesture |
CN110638422B (en) * | 2014-09-23 | 2022-07-22 | 飞比特有限公司 | Method, system and device for updating screen content in response to user gesture |
WO2016097381A1 (en) * | 2014-12-19 | 2016-06-23 | Koninklijke Philips N.V. | Dynamic wearable device behavior based on schedule detection |
US20180000414A1 (en) * | 2014-12-19 | 2018-01-04 | Koninklijke Philips N.V. | Dynamic wearable device behavior based on schedule detection |
US11484261B2 (en) | 2014-12-19 | 2022-11-01 | Koninklijke Philips N.V. | Dynamic wearable device behavior based on schedule detection |
US10088866B2 (en) | 2015-03-18 | 2018-10-02 | Motorola Mobility Llc | Controlling the orientation of a device display based on usage context |
WO2018111614A1 (en) * | 2016-12-15 | 2018-06-21 | Holor, Llc | Wearable multi-functional personal security device |
Also Published As
Publication number | Publication date |
---|---|
WO2014066703A3 (en) | 2014-06-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20150277572A1 (en) | Smart contextual display for a wearable device | |
US10990187B2 (en) | Methods, systems, and apparatuses to update screen content responsive to user gestures | |
US12114959B2 (en) | Sedentary period detection using a wearable electronic device | |
US8751194B2 (en) | Power consumption management of display in portable device based on prediction of user input | |
US8768648B2 (en) | Selection of display power mode based on sensor data | |
US8781791B2 (en) | Touchscreen with dynamically-defined areas having different scanning modes | |
JP6434144B2 (en) | Raise gesture detection on devices | |
KR102463383B1 (en) | Method for measuring bio-signal and wearable electronic device | |
EP3030952B1 (en) | Wrist-worn athletic device with gesture recognition and power management | |
CN112512411B (en) | Context-aware respiration rate determination using an electronic device | |
US20170109990A1 (en) | Method and system for fall detection | |
US20150182160A1 (en) | Function operating method based on biological signals and electronic device supporting the same | |
CN110151137A (en) | Sleep state monitoring method, device, equipment and medium based on data fusion | |
CN105511750B (en) | switching method and electronic equipment | |
WO2019132771A1 (en) | Method and system for indicating a breathing pattern | |
CN114388122A (en) | Parameter fusion processing method and device, wearable device and storage medium | |
WO2014066703A2 (en) | Smart contextual display for a wearable device | |
US20150148923A1 (en) | Wearable device that infers actionable events | |
Luan et al. | HANDIO: a wireless hand gesture recognizer based on muscle-tension and inertial sensing | |
Han et al. | Mobile Activity Sensor Logger for Monitoring Chronic Disease Patients |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13848646 Country of ref document: EP Kind code of ref document: A2 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14438207 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13848646 Country of ref document: EP Kind code of ref document: A2 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13848646 Country of ref document: EP Kind code of ref document: A2 |