WO2015162946A1 - 頭部位置検出装置及び頭部位置検出方法、画像処理装置及び画像処理方法、表示装置、並びにコンピューター・プログラム - Google Patents
頭部位置検出装置及び頭部位置検出方法、画像処理装置及び画像処理方法、表示装置、並びにコンピューター・プログラム Download PDFInfo
- Publication number
- WO2015162946A1 WO2015162946A1 PCT/JP2015/051279 JP2015051279W WO2015162946A1 WO 2015162946 A1 WO2015162946 A1 WO 2015162946A1 JP 2015051279 W JP2015051279 W JP 2015051279W WO 2015162946 A1 WO2015162946 A1 WO 2015162946A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- head
- user
- posture
- coordinate system
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/211—Input arrangements for video game devices characterised by their sensors, purposes or types using inertial sensors, e.g. accelerometers or gyroscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/50—Controlling the output signals based on the game progress
- A63F13/52—Controlling the output signals based on the game progress involving aspects of the displayed game scene
- A63F13/525—Changing parameters of virtual cameras
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/212—Input arrangements for video game devices characterised by their sensors, purposes or types using sensors worn by the player, e.g. for measuring heart beat or leg activity
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B21/00—Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant
- G01B21/22—Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant for measuring angles or tapers; for testing the alignment of axes
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/10—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
- G01C21/12—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
- G01C21/16—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
- G01C21/165—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
- G01C21/1656—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments with passive imaging devices, e.g. cameras
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/36—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
- H04N13/376—Image reproducers using viewer tracking for tracking left-right translational head movements, i.e. lateral movements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/64—Constructional details of receivers, e.g. cabinets or dust covers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
- H04N5/7475—Constructional details of television projection apparatus
- H04N5/7491—Constructional details of television projection apparatus of head mounted projectors
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/80—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
- A63F2300/8082—Virtual reality
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/275—Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
- H04N21/42222—Additional components integrated in the remote control device, e.g. timer, speaker, sensors for detecting position, direction or movement of the remote control, microphone or battery charging device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/001—Constructional or mechanical details
Definitions
- the technology disclosed in this specification includes a head position detection device and a head position detection method for detecting the position of the user's head, an image processing device and an image for processing an image following the position and posture of the user's head
- the present invention relates to a processing method, a display device, and a computer program.
- the head-mounted display has an image display unit for each of the left and right eyes, for example, and is configured to be used in combination with headphones to control vision and hearing.
- the head-mounted display can also project different images to the left and right eyes, and can display a 3D image by displaying an image with parallax for the left and right eyes.
- an image obtained by cutting out a part of a wide-angle image can be observed.
- Examples of the wide-angle image mentioned here include an image generated by 3D graphics such as a game in addition to an image taken by a camera.
- a head-mounted display that has a head motion tracking device consisting of a gyro sensor, etc. attached to the head so that it can follow the movement of the user's head so that 360 degree images of the entire space can be realized.
- a head motion tracking device consisting of a gyro sensor, etc. attached to the head so that it can follow the movement of the user's head so that 360 degree images of the entire space can be realized.
- Patent Document 1 and Patent Document 2 By moving the display area in the wide-angle image so as to cancel the head movement detected by the gyro sensor, the image following the head movement can be reproduced, and the user looks over the entire space. Have an experience.
- Motion parallax refers to a phenomenon in which an image on the retina changes when the object and the observer move relatively (in the left-right direction) while observing a deep object. Specifically, objects that are farther than the object that is gazing appear to change position in the same direction as the direction of movement, but the object that is gazing appears to change position in the direction opposite to the direction of travel. .
- an image that does not express motion parallax has an unnatural feeling of depth and stereoscopic effect, and causes a user to experience VR (virtual reality) sickness.
- An object of the technology disclosed in the present specification is to provide an excellent head position detection apparatus and head position detection method capable of easily detecting the position of the user's head.
- a further object of the technology disclosed in the present specification is to provide an excellent image processing device and image processing method, a display device, and an image processing device that can easily detect the position of the user's head and present an image of motion parallax. To provide computer programs.
- a detection unit for detecting the posture of the user's head A detection unit for detecting the posture of the user's head; A conversion unit for converting the posture of the head into the position of the head in the user coordinate system; Is a head position detecting device.
- the detection unit of the head position detection device includes a gyro sensor mounted on a user's head, and the gyro sensor detects the gyro sensor.
- the head posture is calculated by integrating the angular velocity.
- the detection unit of the head position detection device further includes an acceleration sensor, and a drift of the posture obtained from the gyro sensor with respect to the gravitational direction. Is compensated based on the direction of gravity detected by the acceleration sensor.
- the conversion unit of the head position detection device separates the change in the angle of the user's head from the head by a predetermined arm length r.
- the position is converted to the position of the head viewed from the user coordinate system in which the origin is set at a predetermined site on the user's body.
- the conversion unit of the head position detection device is configured on a spherical surface in which the user's head is fixed at a predetermined radius r from a predetermined rotation center. It is configured to convert the change in the angle of the head into the position of the head viewed from the user coordinate system.
- the conversion unit of the head position detection device is configured such that the user's head has the arm length r with the origin on the user coordinate system as the rotation center. As moving on a spherical surface having a radius, a change in the angle of the head is converted into the position of the head viewed from the user coordinate system.
- the user's waist position is set as the origin of the user coordinate system.
- the conversion unit of the head position detection device according to claim 4 is configured such that the user's head moves on a spherical surface having the user's waist position as a rotation center and the arm length r as a radius. It is configured to convert the change in angle into the position of the head viewed from the waist position of the user.
- the conversion unit of the head position detection device is configured such that the user's head is only the first arm length r 1 shorter than the arm length r. It is configured to convert a change in the angle of the head into a position of the head viewed from the user coordinate system, as moving on a spherical surface fixed at a radius r 1 from a separated rotation center.
- the waist position of the user is set to the origin of the user coordinate system.
- the conversion unit of the head position detection device according to claim 4 is fixed to a radius r 1 from a neck at which a user's head is separated by a first arm length r 1 shorter than the arm length r.
- a change in the angle of the head is converted into the position of the head viewed from the user's waist position.
- the head position detection device further includes a second detection unit that detects the posture of the upper body part other than the user's head. . Then, based on the posture of the head detected by the detection unit and the posture of the upper body detected by the second detection unit, the conversion unit converts the posture of the head to the position of the head in the user coordinate system. Configured to convert.
- the conversion unit of the head position detection device adjusts the arm length r according to the application to which the position of the head is applied. It is configured.
- the conversion unit of the head position detection device detects the head detected by the detection unit according to an application to which the position of the head is applied.
- the position of the head is obtained by limiting at least a part of the angle component of the posture.
- the conversion unit of the head position detection apparatus estimates the arm length r at each time and obtains the position of the head at each time. It is configured as follows.
- the detection unit of the head position detection device includes a sensor that detects acceleration of the user's head. And the said conversion part is comprised so that the arm length r may be estimated based on the acceleration detected at each time, and the position of the head in each time may be calculated
- the technique described in claim 15 of the present application is: A detection step for detecting the posture of the user's head; A conversion step of converting the posture of the head into the position of the head in the user coordinate system; Is a head position detecting method.
- the technique described in claim 16 of the present application is: A detection unit for detecting the posture of the user's head; A conversion unit for converting the posture of the head into the position of the head in the user coordinate system; A drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head; Is an image processing apparatus.
- the drawing processing unit of the image processing device is configured to apply motion parallax only for a value within a predetermined angle change of the head. Yes.
- the technique described in claim 18 of the present application is: A detection step for detecting the posture of the user's head; A conversion step of converting the posture of the head into the position of the head in the user coordinate system; A drawing processing step for generating an image presenting motion parallax corresponding to the position of the head; Is an image processing method.
- the technology described in claim 19 of the present application is: A detection unit for detecting the posture of the user's head; A conversion unit for converting the posture of the head into the position of the head in the user coordinate system; A drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head; A display unit; Is a display device.
- the technique described in claim 20 of the present application is: A conversion unit that converts the posture of the head detected by the detection unit mounted on the user's head into the position of the head in the user coordinate system; A drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head; As a computer program written in a computer-readable format to make the computer function.
- the computer program according to claim 20 of the present application defines a computer program described in a computer-readable format so as to realize predetermined processing on a computer.
- a cooperative operation is exhibited on the computer, and the same effect as the image processing apparatus according to claim 16 of the present application is obtained. be able to.
- an excellent head position detection device and a head position detection method capable of easily detecting the position of a user's head using an inexpensive sensor are provided. be able to.
- an excellent image processing apparatus and image that can detect the position of the user's head and present an image of motion parallax using an inexpensive sensor.
- a processing method, a display device, and a computer program can be provided.
- FIG. 1 is a diagram schematically illustrating a configuration example of an image display system 100 to which the technology disclosed in this specification is applied.
- FIG. 2 is a diagram schematically showing a modification of the image display system 100.
- FIG. 3 is a diagram (perspective view) showing an external configuration of the display device 400.
- FIG. 4 is a diagram (left side view) showing an external configuration of the display device 400.
- FIG. 5 is a diagram showing the relationship between the coordinate systems used when detecting the posture angle of the head and calculating the position of the head from the posture of the head.
- FIG. 6A is a diagram illustrating a head position obtained based on the posture of a seated user (when the posture is substantially vertical) and the head posture of the user.
- FIG. 1 is a diagram schematically illustrating a configuration example of an image display system 100 to which the technology disclosed in this specification is applied.
- FIG. 2 is a diagram schematically showing a modification of the image display system 100.
- FIG. 3 is a diagram (
- FIG. 6B is a diagram showing the head position obtained based on the posture of the seated user (when the upper body rolls in the left direction around the waist position) and the user's head posture.
- FIG. 6C is a diagram illustrating a head position obtained based on the posture of the seated user (when the upper body is tilted forward with the waist position as the center) and the user's head posture.
- FIG. 7A is a diagram exemplifying an image observed when a seated user looks at the front in a substantially vertical posture with a plurality of balls arranged in the depth direction.
- FIG. 7B exemplifies an image observed when a seated user tilts his upper body to the left (rolls around the waist position) and looks into a plurality of balls arranged in the depth direction from the side.
- FIG. 8A is a diagram illustrating a 3D VR image that is observed when a seated user looks at the front in a substantially vertical posture.
- 8B is a diagram illustrating an image observed when a seated user tilts his upper body to the left and looks at the same VR image as FIG. 8A from the side.
- FIG. 9 is a diagram illustrating a model in which the upper body of a seated user rotates around the waist position (when tilted to the right).
- FIG. 10 is a diagram showing a model in which the upper body of a seated user rotates around the waist position (when tilted forward).
- FIG. 11 is a diagram illustrating a model in which the head of a seated user rotates around the neck (when tilted to the right).
- FIG. 12 is a diagram showing a model in which the head of a seated user rotates around the neck (when tilted forward).
- FIG. 13 is a diagram for explaining an error in a method for obtaining the position of the head from the change in the angle of the head of the user.
- FIG. 14 is a diagram showing a model in which the upper body of the seated user rotates around the waist and the head rotates around the neck.
- FIG. 15 is a diagram illustrating a game image when passing a curve that turns to the right.
- FIG. 16A is a diagram illustrating an operation in which the upper body of the seated user rolls leftward about the waist position.
- FIG. 16B is a diagram showing an operation in which only the head rotates in the left direction around the base of the neck while the trunk of the seated user is substantially stationary.
- FIG. 16C is a diagram illustrating an operation in which only the head is tilted forward around the base of the neck while the torso of the seated user is substantially stationary.
- FIG. 16D is a diagram illustrating an operation in which the upper body of the seated user tilts forward about the waist position.
- FIG. 16A is a diagram illustrating an operation in which the upper body of the seated user rolls leftward about the waist position.
- FIG. 16B is a diagram showing an operation in which only the head rotates in the left direction around the base of the neck while the trunk of the seated user is substantially stationary.
- FIG. 16C
- FIG. 17 is a diagram showing an example in which the user coordinate system XYZ is expressed by the polar coordinate system r ⁇ .
- FIG. 18 is a diagram illustrating the arm length r and the centripetal force applied to the head when the user's head is rotating around the waist.
- FIG. 19 is a diagram illustrating the arm length r and the centripetal force applied to the head when the user's head is rotating around the neck.
- the posture of the user's head can be detected using, for example, a gyroscope.
- an expensive sensor is generally required to detect the position of the head. If the position information of the head cannot be used, the AR object can only be rotated according to the posture of the head and cannot correspond to the parallel movement of the head. Therefore, motion parallax cannot be reproduced (things farther than the object being watched appear to change position in the same direction as the moving direction, and the object being watched is in the direction opposite to the direction of travel) You ca n’t change their position.)
- a method for detecting the position of an object existing in the environment using an infrared camera, a depth camera, an ultrasonic sensor, a magnetic sensor or the like installed in the environment is known in the art.
- Such a method is useful for detecting the position of the head-mounted display, but it is necessary to install a sensor outside the head-mounted display (in other words, away from the head-mounted display). And the price tends to be high.
- a sensor in the same room, if you take the head-mounted display outside and use it at the destination, it is necessary to install a sensor in the environment, which greatly impedes usage. become.
- a method of detecting the self-position by performing image processing on an image taken by the camera mounted on the head-mounted display and photographing the surrounding environment is conceivable.
- a method of installing a marker in the environment and detecting the position of the marker reflected in the captured image it is necessary to install the marker on the environment side.
- by tracking feature points such as edges on the captured image it is possible to detect the self-position without installing a marker.
- the latter is useful in that position detection can be realized only with the sensor in the head-mounted display, but the arithmetic processing for performing image processing and the camera increase the cost.
- the position of the head can be detected by mounting a gyro sensor or an acceleration sensor, which is applied in an inertial aviation system, on a head-mounted display.
- the position can be obtained by second-order integration of the motion acceleration obtained by subtracting the gravitational acceleration component from the acceleration component detected by the acceleration sensor.
- This method is useful in that the position can be detected only by a sensor in the head mounted display.
- the position drift error x at time t is expressed by the following equation (1). That is, the drift error x increases in proportion to the square of time t.
- FIG. 15 illustrates a game image when passing a curve that turns to the right.
- the game image shown corresponds to the field of view of the driver's seat.
- the driver In actual driving of a car, the driver generally tries to confirm the tip of a curve with poor visibility by tilting his body to the left.
- an image obtained by changing the viewpoint of the camera of the game from the posture of the vehicle body can be presented, but the movement of the head of the game player cannot be reflected in the game.
- it is possible to detect a change in the position of the head of the player of the sitting game it is possible to present an image ahead of a curve with poor visibility according to the movement of the head.
- FIGS. 16A to 16D illustrate an operation including movement of the head (change in position) accompanying viewpoint movement of a seated user (such as a wearer of a head mount display).
- a seated user such as a wearer of a head mount display.
- FIG. 16A the upper body of the seated user rolls leftward about the waist position, and the head moves as indicated by reference numeral 1601.
- FIG. 16B the torso of the seated user remains almost stationary, and only the head is rolled in the left direction around the base of the neck. The head moves as indicated by reference numeral 1602. To do.
- FIG. 16C the torso of the seated user is almost stationary, and only the head is tilted forward around the base of the neck, and the head moves as indicated by reference numeral 1603.
- FIG. 16A the upper body of the seated user rolls leftward about the waist position, and the head moves as indicated by reference numeral 1601.
- FIG. 16B the torso of the seated user remains almost stationary, and only the head
- FIGS. 16A to 16D shows that the upper body of the seated user is tilted forward about the waist position, and the head moves as indicated by reference numeral 1604.
- the user's head movements 1601 to 1604 are minute, and the motion parallax caused by the minute head movements 1601 to 1604 is presented. It is considered sufficient to do just that.
- yaw rotation (panning) of the seated user's head or upper body is an exercise that does not involve the movement of the head, and is not shown.
- the movement of the head of the seated user is very small, and the change in the position of the head accompanying the movement of the viewpoint is accompanied by a rotational movement of the head. Therefore, simple motion parallax can be presented by detecting the rotational movement of the head using an inexpensive posture / angle sensor such as a gyro sensor and deriving the change in the position of the head based on the detection result. can do.
- the technology disclosed in this specification detects the rotational movement of the head from a posture / angle sensor such as a gyro sensor installed on the head of an image user (such as a wearer of a head mount display). Based on the detection result, motion parallax due to minute movement of the head is simply presented.
- the technology disclosed in this specification cannot accurately detect the position of the head, but in a use case where the movement of the head is accompanied by a rotational movement, such as when the user is seated, the rotational movement of the head. Therefore, the position of the head can be easily obtained, and motion parallax can be presented sufficiently effectively.
- FIG. 1 schematically shows a configuration example of an image display system 100 to which the technology disclosed in this specification is applied.
- the illustrated image display system 100 includes a head motion tracking device 200, a drawing device 300, and a display device 400.
- the head movement tracking device 200 is used by being mounted on the head of a user who observes an image displayed on the display device 400, and outputs posture information of the user's head to the drawing device 200 at a predetermined transmission cycle.
- the head movement tracking device 200 includes a sensor unit 201, a posture angle calculation unit 202, and a transmission unit 203 that transmits the calculation result of the posture angle calculation unit 202 to the drawing device 300.
- the sensor unit 201 includes a sensor element that detects the posture of the head of the user wearing the head motion tracking device 200.
- the sensor unit 201 basically includes a gyro sensor mounted on the user's head as a sensor element.
- the gyro sensor is inexpensive, has a very low processing load on the attitude angle calculation unit 202 of the detection signal of the sensor, and is easy to mount.
- S / N is better than other sensors such as a camera.
- since the amount of movement of the head is obtained from the posture angle detected by the gyro sensor with a high sampling rate, it is possible to contribute to the presentation of extremely smooth motion parallax including the head movement from low speed to high speed.
- an acceleration sensor may be used in combination with the Jero sensor in the sensor unit 201.
- the drift of the posture obtained from the gyro sensor with respect to the gravitational direction can be easily compensated from the gravitational direction detected by the acceleration sensor, and the drift of the viewpoint movement with time can also be suppressed.
- a magnetic sensor may be used in combination as necessary to correct the drift of the posture of the head around the yaw axis.
- the sensor unit 201 may be configured to detect a change in the posture of the head using a sensor element other than the gyro sensor.
- the posture may be detected from the direction of gravitational acceleration applied to the acceleration sensor.
- a change in the posture of the head may be detected by performing image processing on a surrounding image captured by a camera mounted on the user's head (or mounted on a head-mounted display).
- the posture angle calculation unit 202 calculates the posture angle of the user's head based on the detection result by the sensor unit 201. Specifically, the posture angle calculation unit 202 integrates the angular velocity obtained from the gyro sensor and calculates the posture of the head.
- head posture information may be handled as a quaternion.
- a quaternion is a quaternion consisting of a rotation axis (vector) and a rotation angle (scalar).
- the posture information of the head may be described in other formats such as Euler angles and polar coordinates.
- the posture angle calculation unit 202 calculates the movement amount of the head from the posture angle by a method described later. Then, the transmission unit 203 transmits the head position information obtained by the posture angle calculation unit 202 to the drawing apparatus 300. Alternatively, the posture angle calculation unit 202 only calculates the posture angle, transmits the head posture information from the transmission unit 203 to the drawing device 300, and changes the head posture information into the head position information on the drawing device 300 side. You may make it convert.
- the head movement tracking device 200 and the drawing device 300 are interconnected by wireless communication such as Bluetooth (registered trademark) communication.
- wireless communication such as Bluetooth (registered trademark) communication.
- the head movement tracking device 200 and the drawing device 300 may be connected via a high-speed wired interface such as a USB (Universal Serial Bus).
- the drawing device 300 performs a rendering process of an image displayed on the display device 400.
- the drawing apparatus 300 is configured as an Android (registered trademark) terminal such as a smartphone or a tablet, a personal computer, or a game machine, but is not limited to these apparatuses.
- the drawing device 300 may be a server device on the Internet.
- the head movement tracking device 200 transmits the user's head posture / position information to a server as the drawing device 300, and the drawing device 300 generates a moving image stream corresponding to the received head posture / position information. The stream is transmitted to the display device 400.
- the drawing device 300 includes a receiving unit 301 that receives position information of the user's head from the head motion tracking device 200, a drawing processing unit 302 that performs image rendering processing, and a display device that displays the rendered image.
- a transmission unit 302 that transmits data to 400 and an image source 304 that serves as a supply source of image data are provided.
- the receiving unit 301 receives the position information or posture information of the user's head from the head movement tracking device 200 via Bluetooth (registered trademark) communication or the like.
- the posture information is expressed in the form of, for example, a rotation matrix or quaternion.
- the image source 304 is, for example, a storage device such as HDD (Hard Disc Drive) or SSD (Solid State Drive) that records image content, a media playback device that plays back recording media such as Blu-ray (registered trademark), and a digital broadcast signal. It consists of a broadcast tuner for selecting and receiving a channel, a communication interface for receiving a moving image stream from a streaming server installed on the Internet, and the like.
- the drawing processing unit 302 executes a game for generating 3D graphics and an application for displaying an image captured by the camera, and renders an image to be displayed on the display device 400 side from the image data of the image source 304.
- the rendering processing unit 302 presents motion parallax corresponding to the position of the head from the original image supplied from the image source 304 based on the position information of the user's head received by the receiving unit 301. The rendered image. If head posture information is sent from the head motion tracking device 200 instead of the user's head position information, the drawing processing unit 302 converts the head posture information into position information. Processing shall also be performed.
- the drawing device 300 and the display device 400 are connected by a wired cable such as HDMI (registered trademark) (High Definition Multimedia Interface) or MHL (Mobile High-definition Link). Or you may connect by wireless communication like wirelessHD or Miracast.
- the transmission unit 303 transmits the image data rendered by the drawing processing unit 302 to the display device 400 without compression using any of the communication paths.
- the display device 400 includes a receiving unit 401 that receives an image from the drawing device 300 and a display unit 402 that displays the received image.
- the display device 400 is configured as, for example, a head-mounted display that is fixed to the head or face of a user who observes an image.
- the display device 400 may be a normal TV monitor, a large screen display, or a projection display device.
- the receiving unit 401 receives uncompressed image data from the drawing apparatus 300 via a communication path such as HDMI (registered trademark) or MHL.
- the display unit 402 displays the received image data on the screen.
- the display unit 402 When the display device 400 is configured as a head-mounted display, for example, the display unit 402 includes left and right screens fixed to the left and right eyes of the user, and displays a left-eye image and a right-eye image. .
- the screen of the display unit 402 is composed of a display panel such as an organic EL (Electro-Luminescence) element or a micro display such as a liquid crystal display, or a laser scanning display such as a direct retina display.
- a virtual image optical unit that magnifies and projects the display image of the display unit 402 and forms a magnified virtual image having a predetermined angle of view on the user's pupil is provided.
- FIG. 2 schematically shows a modification of the image display system 100.
- the image display system 100 includes three independent devices, ie, a head motion tracking device 200, a drawing device 300, and a display device 400.
- functions are mounted in the display device 400.
- the same components as those in FIG. 1 are given the same reference numerals. Here, description of each component is abbreviate
- the display device 400 can be reduced in size, weight, and cost.
- FIG. 3 and 4 show an external configuration of the display device 400.
- the display device 400 is configured as a head-mounted display that is used by being fixed to the head or face of a user who observes an image.
- 3 is a perspective view of the head mounted display
- FIG. 4 is a left side view of the head mounted display.
- the illustrated display device 400 has a hat shape or a belt-like structure that wraps the entire circumference of the head, and can be mounted with a load that is distributed over the entire head to reduce the burden on the user. It is a mount display.
- the display device 400 includes a main body part 41 including most parts including a display system, a forehead support part 42 protruding from the upper surface of the main body part 41, and a head band that branches into an upper band 44 and a lower band 45. And left and right headphones.
- a display unit and a circuit board are accommodated in the main body 41.
- a nose pad 43 is formed below the main body 41 so as to follow the back of the nose.
- the forehead pad 42 comes into contact with the user's forehead, and the upper band 44 and the lower band 45 of the head band come into contact with the back of the head. That is, the display device 400 is mounted on the user's head by supporting the forehead support 42, the upper band 44, and the lower band 45 at three points. Therefore, unlike the structure of normal glasses that mainly supports the weight with the nose pad portion, the display device 400 can be mounted with the load distributed over the entire head to reduce the burden on the user. .
- the illustrated display device 400 is also provided with a nose pad portion 43, but only auxiliary support contributes. Further, by tightening the forehead support 42 with a head band, it is possible to support the movement in the rotation direction so as not to rotate from the head of the user to whom the display device 400 is mounted.
- the head movement tracking device 200 can be mounted in the main body 41 of the display device 400 configured as a head-mounted display. However, in the present embodiment, the head movement tracking device 200 is provided as an optional product externally attached to the display device 400 in order to reduce the size, weight, and cost of the display device 400.
- the head movement tracking device 200 is used as an accessory attached to any place such as the upper band 44, the lower band 45, and the forehead support portion 42 of the display device 400, for example.
- the posture angle calculation unit 202 calculates the head posture by integrating the angular velocity obtained from the sensor unit 201 (hereinafter, simply referred to as “gyro sensor”).
- FIG. 5 shows the relationship between the coordinate systems used when detecting the posture angle of the head and calculating the position of the head from the posture of the head in the present embodiment.
- the user's front direction is the Z axis
- the gravity direction is the Y axis
- the direction orthogonal to the Z axis and the Y axis is the X axis
- the coordinate system with the user's waist position as the origin is the world coordinate system.
- this XYZ coordinate system is referred to as a “user coordinate system”.
- the head coordinate system xyz is set at a position separated from the origin of the user coordinate system by the arm length r.
- the position of the head coordinate system is a position obtained by rotating the head posture obtained from the gyro sensor mounted on the user's head with respect to the arm length r.
- the posture of the head is a posture obtained by integrating the angular velocity obtained from the gyro sensor. Even if the user rotates about the y-axis of the head coordinate system, the position of the head does not change. On the other hand, when the user's head rotates about the x-axis or the z-axis, the position of the head changes.
- the position is calculated by second-order integration of the motion acceleration detected by the acceleration sensor, there is a problem that the position drifts with time, but the position calculation method according to the present embodiment does not have such a problem.
- FIG. 6A to 6C show the posture of the seated user on the right side and the position of the head converted from the head posture on the left side.
- the posture of the user's head can be obtained by integrating the angular velocity detected by the gyro sensor attached to the head. Then, the seated user's head moves on a spherical surface with the arm length r as a radius centered on the user's waist position, and the user's head posture is set to the head position on the user coordinate system. It can be converted.
- the right side of FIG. 6A shows a situation in which the seated user 611 is in a substantially vertical posture
- the left side of FIG. 6A shows a head position 601 converted from the posture of the head at that time.
- FIG. 6B shows a state where the upper body of the seated user 612 rolls leftward about the waist position, and the left side of FIG. 6B shows the head position 602 at that time. Yes.
- the right side of FIG. 6C shows a situation in which the upper body of the seated user 613 is tilted forward about the waist position, and the head position 603 at that time is shown on the left side of FIG. 6C.
- the motion parallax can be presented by adding the position change of the head coordinate system obtained in this way to the camera position set on the application side for rendering the image.
- FIG. 7A illustrates an image observed when a seated user 701 looks at the front in a substantially vertical posture with a plurality of balls arranged in the depth direction.
- the balls arranged at the rear are hidden behind the balls arranged at the front and cannot be seen.
- FIG. 7B is observed when a seated user 702 tilts his upper body to the left (rolls around the waist position) and looks into a plurality of balls arranged in the depth direction from the side. The image is illustrated.
- the user 702 can see the rear ball overlapping the front ball and the depth direction from the side (left side) by tilting the upper body to the left, and motion parallax is presented. .
- the distant ball appears to change its position in the same direction as the head movement direction, but the front ball appears to change its position in the direction opposite to the head movement direction. Therefore, it becomes a natural image that allows the user to perceive a sense of depth or a three-dimensional effect, and an immersive feeling increases.
- the ground appears to rotate because it is an image for a head-mounted display. That is, the ground in the image is rotated in a direction that cancels the inclination of the head of the user wearing the head-mounted display, so that it appears to the user that the ground is not rotating.
- the VR image shown in FIG. 7A simply rotates according to the head posture, that is, a plurality of balls arranged in the depth direction are integrated. Since the images change their positions in the same direction, the sense of depth and three-dimensionality become unnatural, which causes the user to get VR sickness.
- FIG. 8A illustrates an example of a 3D VR image observed when the seated user 801 looks at the front in a substantially vertical posture.
- FIG. 8B shows an image observed when the seated user 802 tilts the upper body to the right (by rotating around the waist position) and looks at the same VR image as in FIG. 8A from the side. Illustrated.
- the VR image presenting motion parallax when the head position of the user 802 moves to the right, the scenery outside the door 812 of the room has moved to the right.
- the front room appears to change its position in the direction opposite to the direction of head movement, but the outside scenery seen through the door appears to change its position in the same direction as the movement direction of the head.
- the user 802 can peep at the outside scenery hidden on the left side of the door 812 by tilting the upper body to the right. Therefore, the image becomes a natural image that allows the user 802 to perceive a sense of depth or a three-dimensional effect, and the immersive feeling increases.
- the VR image shown in FIG. 8A simply rotates according to the head posture even when the user tilts the upper body to the right. Since the images are integrally changed in the same direction, the sense of depth and the stereoscopic effect become unnatural, which causes the user to get VR sickness.
- motion parallax according to the change in head position can be presented based on the posture information of the user's head, for example, FPS (First Person Shoting) game , The player can dodge enemy attacks by moving his body (upper body).
- FPS First Person Shoting
- the method for obtaining the position of the head based on the posture information of the head for the seated user will be described in detail.
- the user coordinate system XYZ will be described using the polar coordinate system r ⁇ (see FIG. 17).
- the head angle changes ⁇ and ⁇ are obtained by the posture angle calculation unit 202, and the process of obtaining the head position based on the head angle changes ⁇ and ⁇ is executed in the drawing processing unit 302.
- FIGS. 9 and 10 illustrate a model in which the upper body of the seated user rotates around the waist position.
- FIG. 9 illustrates the case where the upper body of the seated user 901 is tilted to the left (right side of the page) centered on the waist position
- FIG. 10 illustrates the upper body of the seated user 1001 centered on the waist position. The case where it inclines forward is illustrated.
- r is the distance (arm length) from the user's waist position to the head position where the gyro sensor is mounted.
- the head moves at a position fixed at a radius r from the center of rotation.
- the position of the head (X, Y , Z) is expressed by the following equation (2).
- FIGS. 11 and 12 illustrate a model in which the head of a seated user rotates around the neck.
- FIG. 11 illustrates the case where the head of the user 1101 is tilted to the left (right side of the page) around the neck
- FIG. 12 illustrates the case where the head of the user 1201 is tilted forward about the neck. Illustrated.
- the distance from the user's neck to the head position where the gyro sensor is mounted is r 1
- the distance from the user's waist position to the neck is r 2
- the head moves from the neck, which is the center of rotation, at a position fixed at a radius r 1, and when the head angle changes are ⁇ and ⁇ , the position of the head viewed from the user's coordinate system with the waist position as the origin (X, Y, Z) is represented by the following formula (3).
- the arm lengths r, r 1 and r 2 described above are values set based on the size of the human body, but the arm lengths may be freely set by an application for rendering an image.
- the arm length may be set according to the assumed size of the robot.
- a linear or nonlinear expression is further applied to the amount of change in the position of the head calculated using the above formula (2) or (3) from the detected head posture, and the value is adjusted. Also good.
- the position of the head may be obtained (in other words, only the amount of change in the X direction of the head may be used).
- the above formula (2) or (3) does not accurately determine the position of the user's head, but is a method for simply determining the position of the head from the angle change of the user's head, and includes an error. Should be considered.
- a user 1301 who is seated has a model in which the upper body of the user 1301 rotates around the waist position, but the actual user tilts his / her head around the neck in the left-right direction. To consider.
- the position of the head as seen from the user's coordinate system with the waist position as the origin is expressed by the following equation (5). Accordingly, the position of the head is calculated according to the model shown in FIG. 13 will contain an error (e x, e y, e z) , as shown by the following formula (6).
- the angle change in the upper body part other than the head of the seated user is further detected to obtain the head position more accurately.
- a model is set in which the upper body of the seated user 1410 rotates around the waist and the head rotates around the neck.
- the sensor unit 201 is equipped with a second gyro sensor 1402 attached to the neck of the user 1410 in addition to the first gyro sensor 1401 attached to the head of the user 1410.
- the posture angle calculation unit 202 integrates the angular velocity detected by the first gyro sensor 1401 to calculate the rotation amounts ⁇ 1 and ⁇ 1 of the head around the waist position, and the second gyro sensor.
- the angular velocities detected by 1402 are integrated to calculate the rotation amounts ⁇ 2 and ⁇ 2 of the neck around the waist position. Then, the posture angle calculation unit 202 (or the drawing processing unit 302) determines the position (X, Y, Z) of the head viewed from the coordinate system of the user 1410 with the waist position as the origin as shown in the following equation (7). To calculate.
- the position of the head of the user 1410 can be obtained in consideration of the amount of rotation of the seated user 1410 around the neck and waist.
- the gyro sensors 1401 and 1402 are installed at two places, the neck and waist positions of the user 1410 who is seated, but the gyro sensors 1401 and 1402 are also rotated at other parts of the upper body of the user 1410. In some cases, gyro sensors can be installed in three or more places to determine the position of the head of the user 1410 more accurately.
- the arm length r from the origin of the user coordinate system set to the user's waist position to the head position of the user on which the gyro sensor is mounted (that is, posture detection is performed)
- the position (X, Y, Z) of the head is obtained from the change in the angle of the head according to the above equation (2).
- the user's head position may be obtained by estimating the arm length r at each time.
- the sensor unit 201 uses the acceleration sensor together with the gyro sensor.
- the gyro sensor can detect the angular velocity ⁇ of the user's head
- the acceleration sensor can detect the acceleration a y of the head.
- the head acceleration a y is a centripetal acceleration
- the acceleration sensor can observe different values of acceleration ay .
- the arm length r is large and the centripetal force applied to the head is large.
- the arm length is short and the centripetal force applied to the head is small.
- the technology disclosed in this specification it is possible to detect a change in the position of the user's head using only an inexpensive sensor such as a gyro sensor.
- an inexpensive sensor such as a gyro sensor.
- the technology disclosed in this specification is particularly effective when the head movement tracking device 200 is provided as an optional product externally attached to the display device 400 configured as a head-mounted display. Even when the head movement tracking device 200 is mounted in the main body 41 of the display device 400, the technology disclosed in the present specification can be similarly applied. Further, even when the display device 400 is a product other than the head-mounted display, the technology disclosed in the present specification can be similarly applied when reproducing an image following the movement of the user's head. .
- the embodiment in which the motion parallax is presented on the head mounted display has been mainly described, but the technology disclosed in the present specification can be applied to other use cases.
- a user who is sitting in front of a large screen display such as a television and playing a game can wear the head motion tracking device 200 to present motion parallax on the game screen in the television.
- the head position change detected by applying the technology disclosed in this specification can be reflected in the camera viewpoint of 3D graphics to present motion parallax, but can also be used for other purposes. For example, in a 2D graphics game, an enemy attack may be avoided according to a change in the position of the head.
- a detection unit that detects the posture of the user's head
- a conversion unit for converting the posture of the head into the position of the head in the user coordinate system
- a head position detection apparatus comprising: (2)
- the detection unit includes a gyro sensor mounted on a user's head, integrates the angular velocity detected by the gyro sensor, and calculates the posture of the head.
- the head position detection device according to (1) above.
- the detection unit further includes an acceleration sensor, and compensates for a drift of the posture obtained from the gyro sensor with respect to the gravity direction based on the gravity direction detected by the acceleration sensor.
- the head position detection device according to (2) above.
- the conversion unit is a head viewed from a user coordinate system in which a change in the angle of the user's head is set at a predetermined site on the user's body spaced from the head by a predetermined arm length r. Part position, The head position detection device according to any one of (1 to (3) above. (5) The conversion unit assumes that the user's head moves on a spherical surface fixed at a predetermined radius r from a predetermined center of rotation, and the head position when a change in the angle of the head is viewed from the user coordinate system. Convert to The head position detection device according to (4) above.
- the conversion unit assumes that the user's head moves on a spherical surface having the origin on the user coordinate system as the rotation center and the arm length r as the radius, and sees the change in head angle from the user coordinate system. Convert to the position of the head The head position detection device according to (4) above. (7) Set the user's waist position to the origin of the user coordinate system, The conversion unit is configured such that the user's head moves on a spherical surface having the user's waist position as a rotation center and the arm length r as a radius, and changes in the angle of the head viewed from the user's waist position. Convert to position, The head position detection device according to (4) above.
- the conversion unit is assumed to move on a spherical surface fixed to a radius r 1 from a rotation center separated from the user's head by a first arm length r 1 shorter than the arm length r. Convert the change in angle into the position of the head as seen from the user coordinate system, The head position detection device according to (4) above.
- (9) Set the user's waist position to the origin of the user coordinate system, The conversion unit assumes that the user's head moves on a spherical surface fixed at a radius r 1 from a neck separated by a first arm length r 1 shorter than the arm length r, and changes the angle of the head. Convert from the user's waist position to the head position seen, The head position detection device according to (4) above.
- a second detection unit that detects the posture of the upper body part other than the user's head;
- the conversion unit converts the posture of the head into the position of the head in the user coordinate system based on the posture of the head detected by the detection unit and the posture of the upper body detected by the second detection unit.
- the head position detection device according to (1) above.
- the conversion unit adjusts the arm length r according to an application to which the position of the head is applied.
- the conversion unit obtains the position of the head by limiting at least a part of the angle component of the posture of the head detected by the detection unit according to an application to which the position of the head is applied.
- the head position detection device according to any one of (1 to (11) above.
- the conversion unit estimates the arm length r at each time and obtains the position of the head at each time.
- the head position detection device according to (4) above.
- the detection unit includes a sensor that detects acceleration of a user's head, The conversion unit estimates the arm length r based on the acceleration detected at each time, and obtains the position of the head at each time.
- the head position detection apparatus according to (13) above.
- a detection unit for detecting the posture of the user's head;
- a conversion unit for converting the posture of the head into the position of the head in the user coordinate system;
- a drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head;
- An image processing apparatus comprising: (16-1)
- the detection unit includes a gyro sensor mounted on a user's head, calculates the posture of the head by integrating the angular velocity detected by the gyro sensor, The image processing apparatus according to (16) above.
- the detection unit further includes an acceleration sensor, and compensates for a drift of the posture obtained from the gyro sensor with respect to the gravity direction based on the gravity direction detected by the acceleration sensor.
- the image processing apparatus according to (16-1) above.
- the conversion unit is configured to view a change in the angle of the user's head from a user coordinate system in which an origin is set at a predetermined site on the user's body that is separated from the head by a predetermined arm length r. Convert to the position of the head The image processing apparatus according to any one of (16-1) to (16-2) above. (16-4) The conversion unit assumes that the user's head moves on a spherical surface fixed at a predetermined radius r from a predetermined center of rotation, and the change of the angle of the head is viewed from the user coordinate system. Convert to position, The image processing apparatus according to (16-3) above.
- the conversion unit assumes that the user's head moves on a spherical surface having the origin on the user coordinate system as the rotation center and the arm length r as the radius, and changes the angle of the head to the user coordinate system.
- (16-6) Set the user's waist position to the origin of the user coordinate system, The conversion unit is configured such that the user's head moves on a spherical surface having the user's waist position as a rotation center and the arm length r as a radius, and changes in the angle of the head viewed from the user's waist position. Convert to position, The image processing apparatus according to (16-3) above.
- (16-7) It is assumed that the conversion unit moves on a spherical surface fixed at a radius r 1 from a rotation center where the user's head is separated by a first arm length r 1 shorter than the arm length r. Convert the change in the angle of the head to the position of the head as seen from the user coordinate system, The image processing apparatus according to (16-3) above. (16-8) Set the user's waist position to the origin of the user coordinate system, The conversion unit assumes that the user's head moves on a spherical surface fixed at a radius r 1 from a neck separated by a first arm length r 1 shorter than the arm length r, and changes the angle of the head.
- the image processing apparatus Convert from the user's waist position to the head position seen, The image processing apparatus according to (16-3) above. (16-9) further comprising a second detection unit for detecting the posture of the upper body part other than the user's head, The conversion unit converts the posture of the head into the position of the head in the user coordinate system based on the posture of the head detected by the detection unit and the posture of the upper body detected by the second detection unit. , The image processing apparatus according to (16) above. (16-10) The conversion unit adjusts the arm length r according to an application to which the position of the head is applied. The image processing apparatus according to (16-3) above.
- the conversion unit obtains the head position by limiting at least a part of the angle component of the head posture detected by the detection unit according to an application to which the head position is applied. , The image processing device according to any one of (16) to (16-10).
- the conversion unit estimates the arm length r at each time and obtains the position of the head at each time.
- the detection unit includes a sensor that detects acceleration of a user's head, The conversion unit estimates the arm length r based on the acceleration detected at each time, and obtains the position of the head at each time.
- the drawing processing unit applies motion parallax only for a value within a predetermined angle change of the head.
- the image processing apparatus according to (16) above. (18) a detection step of detecting the posture of the user's head; A conversion step of converting the posture of the head into the position of the head in the user coordinate system; A drawing processing step for generating an image presenting motion parallax corresponding to the position of the head; An image processing method.
- a detection unit for detecting the posture of the user's head a detection unit for detecting the posture of the user's head; A conversion unit for converting the posture of the head into the position of the head in the user coordinate system; A drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head; A display unit; A display device comprising: (20) a conversion unit that converts the posture of the head detected by the detection unit mounted on the user's head into the position of the head in the user coordinate system; A drawing processing unit for generating an image presenting motion parallax corresponding to the position of the head; A computer program written in a computer-readable format to make a computer function as
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Optics & Photonics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Cardiology (AREA)
- General Health & Medical Sciences (AREA)
- Heart & Thoracic Surgery (AREA)
- Automation & Control Theory (AREA)
- Computer Hardware Design (AREA)
- Controls And Circuits For Display Device (AREA)
- User Interface Of Digital Computer (AREA)
- Length Measuring Devices With Unspecified Measuring Means (AREA)
Abstract
Description
ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
を具備する頭部位置検出装置である。
ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
を有する頭部位置検出方法である。
ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
を具備する画像処理装置である。
ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理ステップと、
を有する画像処理方法である。
ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
表示部と、
を具備する表示装置である。
ユーザーの頭部に装着した検出部が検出した頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部、
としてコンピューターを機能させるようにコンピューター可読形式で記述されたコンピューター・プログラムである。
(1)ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
を具備する頭部位置検出装置。
(2)前記検出部は、ユーザーの頭部に装着したジャイロ・センサーを備え、前記ジャイロ・センサーが検出した角速度を積分して、頭部の姿勢を算出する、
上記(1)に記載の頭部位置検出装置。
(3)前記検出部は、加速度センサーをさらに備え、前記ジャイロ・センサーから得られた姿勢の重力方向に対するドリフトを前記加速度センサーが検出する重力方向に基づいて補償する、
上記(2)に記載の頭部位置検出装置。
(4)前記変換部は、ユーザーの頭部の角度の変化を、頭部から所定のアーム長rだけ離間したユーザーの身体上の所定の部位に原点が設定されたユーザー座標系から見た頭部の位置に変換する、
上記(1乃至(3)のいずれかに記載の頭部位置検出装置。
(5)前記変換部は、ユーザーの頭部が所定の回転中心から所定の半径rに固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(4)に記載の頭部位置検出装置。
(6)前記変換部は、ユーザーの頭部がユーザー座標系上の原点を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(4)に記載の頭部位置検出装置。
(7)ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部がユーザーの腰位置を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
上記(4)に記載の頭部位置検出装置。
(8)前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した回転中心から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(4)に記載の頭部位置検出装置。
(9)ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した首から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
上記(4)に記載の頭部位置検出装置。
(10)ユーザーの頭部以外の上半身の部位の姿勢を検出する第2の検出部をさらに備え、
前記変換部は、前記検出部が検出した頭部の姿勢と前記第2の検出部が検出した上半身の部位の姿勢に基づいて、頭部の姿勢をユーザー座標系における頭部の位置に変換する、
上記(1)に記載の頭部位置検出装置。
(11)前記変換部は、頭部の位置を適用するアプリケーションに応じて、アーム長rを調整する、
上記(4)に記載の頭部位置検出装置。
(12)前記変換部は、頭部の位置を適用するアプリケーションに応じて、前記検出部が検出した頭部の姿勢の少なくとも一部の角度成分を制限して、頭部の位置を求める、
上記(1乃至(11)のいずれかに記載の頭部位置検出装置。
(13)前記変換部は、各時刻のアーム長rを推定して、各時刻における頭部の位置を求める、
上記(4)に記載の頭部位置検出装置。
(14)前記検出部は、ユーザーの頭部の加速度を検出するセンサーを含み、
前記変換部は、各時刻で検出された加速度に基づいてアーム長rを推定して、各時刻における頭部の位置を求める、
上記(13)に記載の頭部位置検出装置。
(15)ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
を有する頭部位置検出方法。
(16)ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
を具備する画像処理装置。
(16-1)前記検出部は、ユーザーの頭部に装着したジャイロ・センサーを備え、前記ジャイロ・センサーが検出した角速度を積分して、頭部の姿勢を算出する、
上記(16)に記載の画像処理装置。
(16-2)前記検出部は、加速度センサーをさらに備え、前記ジャイロ・センサーから得られた姿勢の重力方向に対するドリフトを前記加速度センサーが検出する重力方向に基づいて補償する、
上記(16-1)に記載の画像処理装置。
(16-3)前記変換部は、ユーザーの頭部の角度の変化を、頭部から所定のアーム長rだけ離間したユーザーの身体上の所定の部位に原点が設定されたユーザー座標系から見た頭部の位置に変換する、
上記(16-1)乃至(16-2)のいずれか)に記載の画像処理装置。
(16-4)前記変換部は、ユーザーの頭部が所定の回転中心から所定の半径rに固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(16-3)に記載の画像処理装置。
(16-5)前記変換部は、ユーザーの頭部がユーザー座標系上の原点を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(16-3)に記載の画像処理装置。
(16-6)ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部がユーザーの腰位置を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
上記(16-3)に記載の画像処理装置。
(16-7)前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した回転中心から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
上記(16-3)に記載の画像処理装置。
(16-8)ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した首から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
上記(16-3)に記載の画像処理装置。
(16-9)ユーザーの頭部以外の上半身の部位の姿勢を検出する第2の検出部をさらに備え、
前記変換部は、前記検出部が検出した頭部の姿勢と前記第2の検出部が検出した上半身の部位の姿勢に基づいて、頭部の姿勢をユーザー座標系における頭部の位置に変換する、
上記(16)に記載の画像処理装置。
(16-10)前記変換部は、頭部の位置を適用するアプリケーションに応じて、アーム長rを調整する、
上記(16-3)に記載の画像処理装置。
(16-11)前記変換部は、頭部の位置を適用するアプリケーションに応じて、前記検出部が検出した頭部の姿勢の少なくとも一部の角度成分を制限して、頭部の位置を求める、
上記(16)乃至(16-10)のいずれかに記載の画像処理装置。
(16-12)前記変換部は、各時刻のアーム長rを推定して、各時刻における頭部の位置を求める、
上記(16-3)に記載の画像処理装置。
(16-13)前記検出部は、ユーザーの頭部の加速度を検出するセンサーを含み、
前記変換部は、各時刻で検出された加速度に基づいてアーム長rを推定して、各時刻における頭部の位置を求める、
上記(16-12)に記載の画像処理装置。
(17)前記描画処理部は、頭部の角度変化が所定以内の値のみ運動視差を適用する、
上記(16)に記載の画像処理装置。
(18)ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理ステップと、
を有する画像処理方法。
(19)ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
表示部と、
を具備する表示装置。
(20)ユーザーの頭部に装着した検出部が検出した頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部、
としてコンピューターを機能させるようにコンピューター可読形式で記述されたコンピューター・プログラム。
44…上バンド、45…下バンド
100…画像表示システム
200…頭部動作追跡装置、201…センサー部、202…姿勢角演算部
203…送信部
300…描画装置、301…受信部、302…描画処理部、303…送信部
304…画像ソース
400…表示装置、401…受信部、402…表示部
Claims (20)
- ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
を具備する頭部位置検出装置。 - 前記検出部は、ユーザーの頭部に装着したジャイロ・センサーを備え、前記ジャイロ・センサーが検出した角速度を積分して、頭部の姿勢を算出する、
請求項1に記載の頭部位置検出装置。 - 前記検出部は、加速度センサーをさらに備え、前記ジャイロ・センサーから得られた姿勢の重力方向に対するドリフトを前記加速度センサーが検出する重力方向に基づいて補償する、
請求項2に記載の頭部位置検出装置。 - 前記変換部は、ユーザーの頭部の角度の変化を、頭部から所定のアーム長rだけ離間したユーザーの身体上の所定の部位に原点が設定されたユーザー座標系から見た頭部の位置に変換する、
請求項1に記載の頭部位置検出装置。 - 前記変換部は、ユーザーの頭部が所定の回転中心から所定の半径rに固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
請求項4に記載の頭部位置検出装置。 - 前記変換部は、ユーザーの頭部がユーザー座標系上の原点を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
請求項4に記載の頭部位置検出装置。 - ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部がユーザーの腰位置を回転中心とし前記アーム長rを半径とする球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
請求項4に記載の頭部位置検出装置。 - 前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した回転中心から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザー座標系から見た頭部の位置に変換する、
請求項4に記載の頭部位置検出装置。 - ユーザーの腰位置をユーザー座標系の原点に設定し、
前記変換部は、ユーザーの頭部が前記アーム長rよりも短い第1のアーム長r1だけ離間した首から半径r1に固定された球面上で動くものとして、頭部の角度の変化をユーザーの腰位置から見た頭部の位置に変換する、
請求項4に記載の頭部位置検出装置。 - ユーザーの頭部以外の上半身の部位の姿勢を検出する第2の検出部をさらに備え、
前記変換部は、前記検出部が検出した頭部の姿勢と前記第2の検出部が検出した上半身の部位の姿勢に基づいて、頭部の姿勢をユーザー座標系における頭部の位置に変換する、
請求項1に記載の頭部位置検出装置。 - 前記変換部は、頭部の位置を適用するアプリケーションに応じて、アーム長rを調整する、
請求項4に記載の頭部位置検出装置。 - 前記変換部は、頭部の位置を適用するアプリケーションに応じて、前記検出部が検出した頭部の姿勢の少なくとも一部の角度成分を制限して、頭部の位置を求める、
請求項1に記載の頭部位置検出装置。 - 前記変換部は、各時刻のアーム長rを推定して、各時刻における頭部の位置を求める、
請求項4に記載の頭部位置検出装置。 - 前記検出部は、ユーザーの頭部の加速度を検出するセンサーを含み、
前記変換部は、各時刻で検出された加速度に基づいてアーム長rを推定して、各時刻における頭部の位置を求める、
請求項13に記載の頭部位置検出装置。 - ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
を有する頭部位置検出方法。 - ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
を具備する画像処理装置。 - 前記描画処理部は、頭部の角度変化が所定以内の値のみ運動視差を適用する、
請求項16に記載の画像処理装置。 - ユーザーの頭部の姿勢を検出する検出ステップと、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換ステップと、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理ステップと、
を有する画像処理方法。 - ユーザーの頭部の姿勢を検出する検出部と、
頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部と、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部と、
表示部と、
を具備する表示装置。 - ユーザーの頭部に装着した検出部が検出した頭部の姿勢をユーザー座標系における頭部の位置に変換する変換部、
頭部の位置に対応した運動視差を提示した画像を生成する描画処理部、
としてコンピューターを機能させるようにコンピューター可読形式で記述されたコンピューター・プログラム。
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2016514733A JP6540691B2 (ja) | 2014-04-22 | 2015-01-19 | 頭部位置検出装置及び頭部位置検出方法、画像処理装置及び画像処理方法、表示装置、並びにコンピューター・プログラム |
| US15/304,081 US20170036111A1 (en) | 2014-04-22 | 2015-01-19 | Head position detecting apparatus and head position detecting method, image processing apparatus and image processing method, display apparatus, and computer program |
| KR1020167027885A KR20160147735A (ko) | 2014-04-22 | 2015-01-19 | 헤드부 위치 검출 장치 및 헤드부 위치 검출 방법, 화상 처리 장치 및 화상 처리 방법, 표시 장치, 및 컴퓨터 프로그램 |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2014-087849 | 2014-04-22 | ||
| JP2014087849 | 2014-04-22 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2015162946A1 true WO2015162946A1 (ja) | 2015-10-29 |
Family
ID=54332120
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2015/051279 Ceased WO2015162946A1 (ja) | 2014-04-22 | 2015-01-19 | 頭部位置検出装置及び頭部位置検出方法、画像処理装置及び画像処理方法、表示装置、並びにコンピューター・プログラム |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US20170036111A1 (ja) |
| JP (1) | JP6540691B2 (ja) |
| KR (1) | KR20160147735A (ja) |
| WO (1) | WO2015162946A1 (ja) |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2017129858A1 (en) * | 2016-01-29 | 2017-08-03 | Nokia Technologies Oy | Method and apparatus for processing video information |
| CN107754307A (zh) * | 2017-12-05 | 2018-03-06 | 野草莓影业(北京)有限公司 | 转动座椅的控制方法、控制装置以及转动座椅 |
| FR3059415A1 (fr) * | 2016-11-29 | 2018-06-01 | Airbus Operations | Procede et systeme de determination de l'attitude et de la position de la tete d'un pilote d'aeronef. |
| CN109243595A (zh) * | 2017-07-03 | 2019-01-18 | 上银科技股份有限公司 | 校正控制系统、控制装置及驱动端 |
| US10393490B2 (en) | 2015-03-17 | 2019-08-27 | Bagel Labs Co., Ltd. | Length measuring device and length measuring system |
| JP2019220059A (ja) * | 2018-06-22 | 2019-12-26 | 任天堂株式会社 | プログラム、情報処理装置、情報処理システム及び情報処理方法 |
| JP2024153738A (ja) * | 2020-05-07 | 2024-10-29 | ライカ インストゥルメンツ (シンガポール) プライヴェット リミテッド | 頭部装着可能ディスプレイデバイスおよび顕微鏡から受信された画像を表示する方法 |
Families Citing this family (13)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10583358B1 (en) * | 2017-01-23 | 2020-03-10 | Pixar | Headset for simulating accelerations |
| DE102017208365A1 (de) * | 2017-05-18 | 2018-11-22 | Robert Bosch Gmbh | Verfahren zur Orientierungsschätzung eines tragbaren Gerätes |
| US10444827B2 (en) * | 2017-09-18 | 2019-10-15 | Fujitsu Limited | Platform for virtual reality movement |
| KR20190118040A (ko) | 2018-04-09 | 2019-10-17 | 삼성전자주식회사 | 웨어러블 디스플레이 장치 및 그의 입체 영상 표시 방법 |
| US20190324283A1 (en) * | 2018-04-19 | 2019-10-24 | Htc Corporation | Display device and display method |
| CN109753155B (zh) * | 2019-01-02 | 2021-01-22 | 京东方科技集团股份有限公司 | 头戴显示设备、其驱动方法及虚拟现实显示装置 |
| US12111223B2 (en) * | 2019-03-22 | 2024-10-08 | Essilor International | Device and method for evaluating a performance of a visual equipment for a visual task |
| CN111723624B (zh) * | 2019-03-22 | 2023-12-05 | 京东方科技集团股份有限公司 | 一种头部运动跟踪方法和系统 |
| US11061469B2 (en) * | 2019-11-20 | 2021-07-13 | XRSpace CO., LTD. | Head mounted display system and rotation center correcting method thereof |
| CN111796682B (zh) * | 2020-07-09 | 2021-11-16 | 联想(北京)有限公司 | 一种控制方法、装置电子设备 |
| KR102862768B1 (ko) * | 2021-01-13 | 2025-09-23 | 삼성전자주식회사 | 웨어러블 전자 장치의 가속도 센서를 이용하여 사용자의 자세를 판단하는 방법 및 그 전자 장치 |
| CN112791381B (zh) * | 2021-01-21 | 2022-12-27 | 深圳市瑞立视多媒体科技有限公司 | 虚拟现实中腰带跟随玩家的移动方法、装置和计算机设备 |
| CN115546292B (zh) * | 2022-12-02 | 2023-03-17 | 首都医科大学附属北京同仁医院 | 头位判读方法、体位验证方法、计算设备和存储介质 |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH08107600A (ja) * | 1994-10-04 | 1996-04-23 | Yamaha Corp | 音像定位装置 |
| JPH11338457A (ja) * | 1998-05-26 | 1999-12-10 | Nippon Telegr & Teleph Corp <Ntt> | 両眼視差式立体表示方法及び記録媒体 |
| JP2009031295A (ja) * | 2008-08-05 | 2009-02-12 | Olympus Corp | 移動体姿勢検出装置 |
| JP2012052904A (ja) * | 2010-09-01 | 2012-03-15 | Microstone Corp | 慣性計測装置 |
| WO2013132885A1 (ja) * | 2012-03-07 | 2013-09-12 | ソニー株式会社 | 情報処理装置、情報処理方法及びプログラム |
| JP2013225042A (ja) * | 2012-04-23 | 2013-10-31 | Seiko Epson Corp | 虚像表示装置 |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH09106322A (ja) | 1995-10-09 | 1997-04-22 | Data Tec:Kk | ヘッドマウントディスプレイにおける姿勢角検出装置 |
| JP2000308092A (ja) * | 1999-04-16 | 2000-11-02 | Toshiba Mach Co Ltd | 立体視眼鏡装置 |
| NL1018198C2 (nl) * | 2001-06-01 | 2002-12-03 | Tno | Head mounted display inrichting. |
| US20080211768A1 (en) * | 2006-12-07 | 2008-09-04 | Randy Breen | Inertial Sensor Input Device |
| JP2010256534A (ja) | 2009-04-23 | 2010-11-11 | Fujifilm Corp | 全方位画像表示用ヘッドマウントディスプレイ装置 |
| JP5809779B2 (ja) * | 2010-03-09 | 2015-11-11 | 淳 久池井 | 態様を取り込むための有関節構造体用の関節パーツ、有関節構造体および態様取込システム |
| CN101861013B (zh) * | 2010-04-30 | 2013-06-05 | 鸿富锦精密工业(深圳)有限公司 | 智能灯及其控制方法 |
| US9436286B2 (en) * | 2011-01-05 | 2016-09-06 | Qualcomm Incorporated | Method and apparatus for tracking orientation of a user |
| US8784206B1 (en) * | 2011-04-15 | 2014-07-22 | Wms Gaming, Inc. | Modifying presentation of three-dimensional, wagering-game content |
-
2015
- 2015-01-19 WO PCT/JP2015/051279 patent/WO2015162946A1/ja not_active Ceased
- 2015-01-19 KR KR1020167027885A patent/KR20160147735A/ko not_active Ceased
- 2015-01-19 JP JP2016514733A patent/JP6540691B2/ja active Active
- 2015-01-19 US US15/304,081 patent/US20170036111A1/en not_active Abandoned
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH08107600A (ja) * | 1994-10-04 | 1996-04-23 | Yamaha Corp | 音像定位装置 |
| JPH11338457A (ja) * | 1998-05-26 | 1999-12-10 | Nippon Telegr & Teleph Corp <Ntt> | 両眼視差式立体表示方法及び記録媒体 |
| JP2009031295A (ja) * | 2008-08-05 | 2009-02-12 | Olympus Corp | 移動体姿勢検出装置 |
| JP2012052904A (ja) * | 2010-09-01 | 2012-03-15 | Microstone Corp | 慣性計測装置 |
| WO2013132885A1 (ja) * | 2012-03-07 | 2013-09-12 | ソニー株式会社 | 情報処理装置、情報処理方法及びプログラム |
| JP2013225042A (ja) * | 2012-04-23 | 2013-10-31 | Seiko Epson Corp | 虚像表示装置 |
Cited By (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10393490B2 (en) | 2015-03-17 | 2019-08-27 | Bagel Labs Co., Ltd. | Length measuring device and length measuring system |
| WO2017129858A1 (en) * | 2016-01-29 | 2017-08-03 | Nokia Technologies Oy | Method and apparatus for processing video information |
| US10616548B2 (en) | 2016-01-29 | 2020-04-07 | Nokia Technologies Oy | Method and apparatus for processing video information |
| FR3059415A1 (fr) * | 2016-11-29 | 2018-06-01 | Airbus Operations | Procede et systeme de determination de l'attitude et de la position de la tete d'un pilote d'aeronef. |
| CN109243595A (zh) * | 2017-07-03 | 2019-01-18 | 上银科技股份有限公司 | 校正控制系统、控制装置及驱动端 |
| CN109243595B (zh) * | 2017-07-03 | 2022-03-01 | 上银科技股份有限公司 | 校正控制系统、控制装置及驱动端 |
| CN107754307A (zh) * | 2017-12-05 | 2018-03-06 | 野草莓影业(北京)有限公司 | 转动座椅的控制方法、控制装置以及转动座椅 |
| JP2019220059A (ja) * | 2018-06-22 | 2019-12-26 | 任天堂株式会社 | プログラム、情報処理装置、情報処理システム及び情報処理方法 |
| JP2024153738A (ja) * | 2020-05-07 | 2024-10-29 | ライカ インストゥルメンツ (シンガポール) プライヴェット リミテッド | 頭部装着可能ディスプレイデバイスおよび顕微鏡から受信された画像を表示する方法 |
| JP7802873B2 (ja) | 2020-05-07 | 2026-01-20 | ライカ インストゥルメンツ (シンガポール) プライヴェット リミテッド | 頭部装着可能ディスプレイデバイスおよび顕微鏡から受信された画像を表示する方法 |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20160147735A (ko) | 2016-12-23 |
| US20170036111A1 (en) | 2017-02-09 |
| JP6540691B2 (ja) | 2019-07-10 |
| JPWO2015162946A1 (ja) | 2017-04-13 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP6540691B2 (ja) | 頭部位置検出装置及び頭部位置検出方法、画像処理装置及び画像処理方法、表示装置、並びにコンピューター・プログラム | |
| US10310595B2 (en) | Information processing apparatus, information processing method, computer program, and image processing system | |
| US9703100B2 (en) | Change nature of display according to overall motion | |
| CN105103034B (zh) | 显示器 | |
| EP3008548B1 (en) | Head-mountable apparatus and systems | |
| EP3241088B1 (en) | Methods and systems for user interaction within virtual or augmented reality scene using head mounted display | |
| JP6074525B1 (ja) | 仮想空間における視界領域調整方法、およびプログラム | |
| JP6642430B2 (ja) | 情報処理装置及び情報処理方法、並びに画像表示システム | |
| JP6447514B2 (ja) | 姿勢測定装置及び姿勢測定方法、画像処理装置及び画像処理方法、表示装置及び表示方法、コンピューター・プログラム、並びに画像表示システム | |
| JP7002648B2 (ja) | 車両内において乗物酔いを伴わずにデジタルコンテンツを見ること | |
| US20170111636A1 (en) | Information processing apparatus, information processing method, computer program, and image display system | |
| US20180361240A1 (en) | Information processing apparatus and user guide presentation method | |
| JP2021512402A (ja) | マルチビューイング仮想現実ユーザインターフェース | |
| US9372346B2 (en) | Directional light beams for angle detection | |
| US11187895B2 (en) | Content generation apparatus and method | |
| WO2018123074A1 (ja) | 撮影装置 | |
| WO2014069090A1 (ja) | 画像表示装置及び画像表示方法、並びにコンピューター・プログラム | |
| EP3673348B1 (en) | Data processing device, method and non-transitory machine-readable medium for detecting motion of the data processing device | |
| US20240329666A1 (en) | Apparatus for holding a display screen | |
| WO2023242981A1 (ja) | ヘッドマウントディスプレイ、ヘッドマウントディスプレイシステム、および、ヘッドマウントディスプレイの表示方法 | |
| JP2018190078A (ja) | コンテンツ表示プログラム、コンピュータ装置、コンテンツ表示方法、及びコンテンツ表示システム |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15783383 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 2016514733 Country of ref document: JP Kind code of ref document: A |
|
| ENP | Entry into the national phase |
Ref document number: 20167027885 Country of ref document: KR Kind code of ref document: A |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 15304081 Country of ref document: US |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 15783383 Country of ref document: EP Kind code of ref document: A1 |