US20120165084A1 - Method and apparatus for tracking locations using webcams - Google Patents
Method and apparatus for tracking locations using webcams Download PDFInfo
- Publication number
- US20120165084A1 US20120165084A1 US13/332,638 US201113332638A US2012165084A1 US 20120165084 A1 US20120165084 A1 US 20120165084A1 US 201113332638 A US201113332638 A US 201113332638A US 2012165084 A1 US2012165084 A1 US 2012165084A1
- Authority
- US
- United States
- Prior art keywords
- user
- tracking
- point
- location
- head
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 210000004209 hair Anatomy 0.000 claims abstract description 7
- 239000000284 extract Substances 0.000 claims abstract description 6
- 238000000605 extraction Methods 0.000 claims abstract description 5
- 238000013179 statistical model Methods 0.000 claims description 11
- 238000001914 filtration Methods 0.000 claims description 10
- 238000010801 machine learning Methods 0.000 claims description 5
- 238000009499 grossing Methods 0.000 claims description 3
- 230000008859 change Effects 0.000 description 11
- 230000008901 benefit Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000011410 subtraction method Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/22—Setup operations, e.g. calibration, key configuration or button assignment
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
- A63F13/426—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving on-screen location information, e.g. screen coordinates of an area at which the player is aiming with a light gun
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/837—Shooting of targets
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/213—Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1087—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
- A63F2300/1093—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera using visible light
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/6045—Methods for processing data by generating or executing the game program for mapping control signals received from the input arrangement into game commands
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/80—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
- A63F2300/8076—Shooting
Definitions
- the present invention relates generally to a method and an apparatus for tracking locations using webcams and, more particularly, to a method and an apparatus for tracking the locations of a user and a laser pointer using webcams.
- Human Computer Interaction relates to technology for ascertaining several events generated when a human uses a computer to perform operations and designing a method for enabling a human to use a computer more conveniently and safely.
- Object tracking in HCI means predicting the trajectory of an object of interest from consecutive images.
- Object tracking includes three types of auxiliary techniques: a modeling technique for enabling a computer to recognize the characteristics of an object, an object detection technique for detecting an object on an input image, and a location prediction technique for estimating the location of an object on a subsequent frame from the location of the object, obtained from a current frame, by taking the correlation between the image frames, received from a camera, into account.
- auxiliary techniques a modeling technique for enabling a computer to recognize the characteristics of an object, an object detection technique for detecting an object on an input image, and a location prediction technique for estimating the location of an object on a subsequent frame from the location of the object, obtained from a current frame, by taking the correlation between the image frames, received from a camera, into account.
- an object model on an image which is used in the modeling technique includes an articulated shape model, including points, a basic geometric shape, a silhouette/contour, and joints. Segmentation and background subtraction methods are used to determine whether a corresponding object exists in an image using the model (i.e., a feature point).
- the object is tracked by training the appearance of an object model in advance using a machine learning algorithm based on a statistical method.
- the location of the object is tracked across a plurality of frames.
- a location prediction technique for estimating the location of an object of a subsequent frame using information about a previous frame is used.
- the location prediction technique may use the most representative tracking and training algorithm, such as meanshift, kalman filtering, particle filtering, and haar-like training techniques.
- the location prediction technique is problematic in that a change in the user's viewpoint and a change in the target which are frequently used in a primary viewpoint game, such as a shooting game, cannot be predicted at the same time.
- an object of the present invention is to provide a method and an apparatus for tracking the locations of a user and a laser pointer, which predict a change in the user's viewpoint and a change in the target using webcams at the same time.
- the present invention provides a method in which an apparatus operating in conjunction with a First-Person Shooter (FPS) game tracks a user performing the game and a location of a laser pointer held by the user, the method including extracting a location, a 3-Degrees Of Freedom (3DOF) and an orientation of a head of the user by capturing the user from a screen on which the game is displayed using a first webcam, and tracking the user's gaze based on results of the extraction; and capturing the screen using a second webcam, and tracking a point, indicated by the laser pointer and projected onto the screen, based on the results of the capturing.
- FPS First-Person Shooter
- the above method of tracking the location is characterized in that the user's gaze and the point of the laser pointer can be tracked simultaneously.
- the tracking the user's gaze may include extracting at least one feature point based on the results of the capturing of the head of the user, performed by the first webcam disposed above the screen; constructing a statistical model by applying the feature point to a machine learning algorithm; and extracting the location, 3DOF, and orientation of the head of the user based on the statistical model.
- the feature point may include at least one of a face color, an eye, a nose, and a mouth of the user.
- the tracking the point indicated by the laser pointer may include performing calibration on a frame including the point; measuring a pixel intensity of the calibrated frame; and comparing the pixel intensity and a critical value, and making a determination of whether to track the point depending on results of the comparison.
- the method may further include grouping pixels whose intensities are equal to or higher than the critical value; performing Kalman filtering on the grouped pixels; and smoothing results of the Kalman filtering and tracking the point.
- the method may further include measuring the pixel intensity of the calibrated frame again.
- the tracked point may be designated as cross hairs corresponding to a target in the game to be shot at.
- the present invention provides a location tracking apparatus, including a head tracking unit for extracting a location, 3DOF and orientation of a head of a user by capturing the user from a screen on which a game is displayed using a first webcam and tracking the user's gaze based on results of the extraction; and a pointer tracking unit for tracking a point indicated by a laser pointer projected onto the screen by capturing the screen using a second webcam, and designating the tracked point as cross hairs corresponding to a target to be shot at; wherein the head tracking unit and the pointer tracking unit track the user's gaze and the point simultaneously.
- the head tracking unit may construct a statistical model based on results of the capturing performed by the first webcam disposed above the screen, and extract the location, 3DOF and orientation of the head of the user by recognizing the results of capturing the head of the user based on the statistical model.
- the pointer tracking unit may perform calibration on a frame including the point, measure the pixel intensity of the calibrated frame, compare the pixel intensity with a critical value, and determine whether to track the point depending on the results of the comparison.
- the pointer tracking unit may group pixels whose intensities are equal to or greater than the critical value, perform Kalman filtering on the grouped pixels, smooth the Kalman filtered result, and track the point.
- the pointer tracking unit may measure the pixel intensity of the calibrated frame again.
- FIG. 1 is a diagram showing a service environment to which an apparatus for tracking locations using webcams is applied according to an embodiment of the present invention
- FIG. 2 shows the configuration of a location tracking apparatus 200 according to an embodiment of the present invention
- FIG. 3 is a flowchart illustrating a method of tracking locations using webcams according to an embodiment of the present invention
- FIG. 4 is a flowchart illustrating a method of tracking a user's gaze according to an embodiment of the present invention.
- FIG. 5 is a flowchart illustrating a method of tracking a laser pointer according to an embodiment of the present invention.
- the method and an apparatus for tracking locations using webcams may belong to a technical field for intuitively sensing the intention of a user and transferring it to an application, such as a game, without using the HCI technique (i.e., a user interface using a mouse and a keyboard), but is not limited thereto.
- HCI technique i.e., a user interface using a mouse and a keyboard
- FIG. 1 is a diagram showing a service environment to which an apparatus for tracking locations using webcams is applied according to an embodiment of the present invention.
- the apparatus for tracking locations using webcams corresponds to an apparatus, such as an interface which operates in conjunction with a First-Person Shooter (FPS) game, but is not limited thereto.
- FPS First-Person Shooter
- the service environment includes two webcams 10 and 20 , a screen 30 where the two webcams 10 and 20 are placed, a projector 40 , and a laser pointer 50 .
- the projector 40 may be coupled to, for example, a computer for executing a game, and projects a game screen onto the screen 30 .
- the webcam 10 disposed above the screen 30 consecutively captures a game participant (i.e., the face of a user) in front of the screen 30 .
- the webcam 20 in front of the projector 40 consecutively captures the screen 30 on which a game screen is being executed. That is, the webcam 20 consecutively captures a point at which the laser pointer 50 held by the user overlays on the screen 30 .
- the location tracking apparatus 200 can incorporate the intention of the user into the game in real time by tracking the head of the user and the laser pointer 50 at the same time using the results captured by the two webcams 10 and 20 .
- the location tracking apparatus 200 for tracking the locations of a user and the laser pointer using the webcams will now be described in detail with reference to FIG. 2 .
- FIG. 2 shows the configuration of the location tracking apparatus 200 according to an embodiment of the present invention.
- the location tracking apparatus 200 includes a head tracking unit 210 and a pointer tracking unit 220 .
- the head tracking unit 210 tracks a user's gaze in a FPS game using the webcam 10 disposed above the screen 30 .
- the pointer tracking unit 220 tracks a point indicated by the laser pointer 50 using the webcam 20 in front of the projector 40 .
- the location tracking apparatus 200 can incorporate a change in the viewpoint of a user in the game by tracking the change using head tracking, without using a mouse or a keyboard in order to change the viewpoint of the user in a primary viewpoint game, such as a shooting game. Furthermore, the location tracking apparatus 200 tracks a point, indicated by the laser pointer 50 , simultaneously with head tracking, and designates the point as cross hairs corresponding to a target to be shot at.
- FIG. 3 is a flowchart illustrating the method of tracking locations using webcams according to an embodiment of the present invention
- FIG. 4 is a flowchart illustrating a method of tracking a user's gaze according to an embodiment of the present invention
- FIG. 5 is a flowchart illustrating a method of tracking a laser pointer according to an embodiment of the present invention.
- an environment to which the method of tracking locations using the webcams according to the embodiment of the present invention is applied includes the two webcams 10 and 20 , the screen 30 where the two webcams 10 and 20 are placed, the projector 40 , and the laser pointer 50 .
- the location tracking apparatus 200 tracks the head of a user and the location of the laser pointer using the two webcams 10 and 20 .
- the projector 40 may project a game screen onto the screen 30 .
- the location tracking apparatus 200 tracks a user's gaze in a FPS game using the webcam 10 disposed above the screen 30 at step S 100 .
- the webcam 10 disposed above the screen 30 consecutively captures the head of the user in front of the screen.
- the location tracking apparatus 200 extracts one or more feature points of the head based on the results of the capturing of the head of the user at step S 110 .
- the feature points of the head include points corresponding to characteristics, such as a face color, an eye, a nose, and a mouth.
- the location tracking apparatus 200 constructs a statistical model by applying the extracted feature points of the head to a machine learning algorithm at step S 120 .
- the location tracking apparatus 200 includes the training data of users before constructing the statistical model. Thereafter, the location tracking apparatus 200 constructs a statistical model by training based on the training data using the machine learning algorithm.
- the location tracking apparatus 200 extracts a location, the 3-Degrees Of Freedom (hereinafter referred to as the “3DOF”), and the orientation of the head of the user by recognizing the results at step S 130 .
- 3DOF 3-Degrees Of Freedom
- the location tracking apparatus 200 checks the location of the face by recognizing the face image, recognizes the locations and orientations of an eye, a nose, and a mouth, and then extracts the orientation of the face.
- the location tracking apparatus 200 may track the user's gaze based on the location, the 3DOF, and the orientation of the head of the user extracted using the webcam 10 .
- the 3DOF includes X (i.e., a horizontal direction), Y (i.e., a vertical direction), and Z (i.e., depth).
- the location tracking apparatus 200 tracks a point indicated by the laser pointer 50 using the webcam 20 in front of the projector 40 at step S 200 .
- the webcam 20 in front of the projector 40 can capture 30 or more frames per second.
- the location tracking apparatus 200 performs calibration on a frame (640 ⁇ 480) including a point that is indicated by the laser pointer 50 projected onto the screen at step S 210 .
- the location tracking apparatus 200 measures the pixel intensity for the frame at step S 220 , and determines whether the measured pixel intensity is equal to or greater than a critical value at step S 230 .
- the location tracking apparatus 200 measures the pixel intensity for the frame again. If the pixel intensity is equal to or greater than the critical value, the location tracking apparatus 200 groups pixels whose intensities are equal to or greater than the critical value at step S 240 .
- the location tracking apparatus 200 performs Kalman filtering, corresponding to a prediction-correction algorithm, on the grouped pixels at step S 250 .
- the location tracking apparatus 200 tracks a point indicated by the laser pointer 50 by smoothing the results of Kalman filtering at step S 260 . Furthermore, the location tracking apparatus 200 designates the tracked point as cross hairs corresponding to a target to be shot at.
- the location tracking apparatus 200 can predict a change in the viewpoint of a user and a change in the target of the laser pointer using the webcam 10 above the screen 30 and the webcam 20 in front of the projector 40 at the same time.
- the location tracking apparatus can predict a change in the viewpoint of a user and a change in the target of the laser pointer using webcams at the same time.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Disclosed herein are a location tracking apparatus and method. The location tracking apparatus includes a head tracking unit, and a pointer tracking unit. The head tracking unit extracts a location, 3DOF and orientation of a head of a user by capturing the user from a screen on which a game is displayed using a first webcam and tracks the user's gaze based on the results of the extraction. The pointer tracking unit tracks a point indicated by a laser pointer projected onto the screen by capturing the screen using a second webcam, and designates the tracked point as cross hairs corresponding to a target to be shot at. The head tracking unit and the pointer tracking unit tracks the user's gaze and the point simultaneously.
Description
- This application claims the benefit of Korean Patent Application Nos. 10-2010-0132875 and 10-2011-0037422, filed on Dec. 22, 2010 and Apr. 21, 2011, respectively, which are hereby incorporated by reference in their entirety into this application.
- 1. Technical Field
- The present invention relates generally to a method and an apparatus for tracking locations using webcams and, more particularly, to a method and an apparatus for tracking the locations of a user and a laser pointer using webcams.
- 2. Description of the Related Art
- Human Computer Interaction (HO) relates to technology for ascertaining several events generated when a human uses a computer to perform operations and designing a method for enabling a human to use a computer more conveniently and safely.
- Object tracking in HCI means predicting the trajectory of an object of interest from consecutive images.
- Object tracking includes three types of auxiliary techniques: a modeling technique for enabling a computer to recognize the characteristics of an object, an object detection technique for detecting an object on an input image, and a location prediction technique for estimating the location of an object on a subsequent frame from the location of the object, obtained from a current frame, by taking the correlation between the image frames, received from a camera, into account.
- In general, an object model on an image which is used in the modeling technique includes an articulated shape model, including points, a basic geometric shape, a silhouette/contour, and joints. Segmentation and background subtraction methods are used to determine whether a corresponding object exists in an image using the model (i.e., a feature point).
- If the corresponding object exists, the object is tracked by training the appearance of an object model in advance using a machine learning algorithm based on a statistical method. Here, if new data enters into the trained model, the location of the object is tracked across a plurality of frames. In this case, a location prediction technique for estimating the location of an object of a subsequent frame using information about a previous frame is used. The location prediction technique may use the most representative tracking and training algorithm, such as meanshift, kalman filtering, particle filtering, and haar-like training techniques.
- The location prediction technique is problematic in that a change in the user's viewpoint and a change in the target which are frequently used in a primary viewpoint game, such as a shooting game, cannot be predicted at the same time.
- Accordingly, the present invention has been made keeping in mind the above problems occurring in the prior art, and an object of the present invention is to provide a method and an apparatus for tracking the locations of a user and a laser pointer, which predict a change in the user's viewpoint and a change in the target using webcams at the same time.
- In order to accomplish the above object, the present invention provides a method in which an apparatus operating in conjunction with a First-Person Shooter (FPS) game tracks a user performing the game and a location of a laser pointer held by the user, the method including extracting a location, a 3-Degrees Of Freedom (3DOF) and an orientation of a head of the user by capturing the user from a screen on which the game is displayed using a first webcam, and tracking the user's gaze based on results of the extraction; and capturing the screen using a second webcam, and tracking a point, indicated by the laser pointer and projected onto the screen, based on the results of the capturing.
- The above method of tracking the location is characterized in that the user's gaze and the point of the laser pointer can be tracked simultaneously.
- The tracking the user's gaze may include extracting at least one feature point based on the results of the capturing of the head of the user, performed by the first webcam disposed above the screen; constructing a statistical model by applying the feature point to a machine learning algorithm; and extracting the location, 3DOF, and orientation of the head of the user based on the statistical model.
- The feature point may include at least one of a face color, an eye, a nose, and a mouth of the user.
- The tracking the point indicated by the laser pointer may include performing calibration on a frame including the point; measuring a pixel intensity of the calibrated frame; and comparing the pixel intensity and a critical value, and making a determination of whether to track the point depending on results of the comparison.
- If the pixel intensity is equal to or higher than the critical value, the method may further include grouping pixels whose intensities are equal to or higher than the critical value; performing Kalman filtering on the grouped pixels; and smoothing results of the Kalman filtering and tracking the point.
- If the pixel intensity is smaller than the critical value, the method may further include measuring the pixel intensity of the calibrated frame again.
- In the tracking the point indicated by the laser pointer, the tracked point may be designated as cross hairs corresponding to a target in the game to be shot at.
- In order to accomplish the above object, the present invention provides a location tracking apparatus, including a head tracking unit for extracting a location, 3DOF and orientation of a head of a user by capturing the user from a screen on which a game is displayed using a first webcam and tracking the user's gaze based on results of the extraction; and a pointer tracking unit for tracking a point indicated by a laser pointer projected onto the screen by capturing the screen using a second webcam, and designating the tracked point as cross hairs corresponding to a target to be shot at; wherein the head tracking unit and the pointer tracking unit track the user's gaze and the point simultaneously.
- The head tracking unit may construct a statistical model based on results of the capturing performed by the first webcam disposed above the screen, and extract the location, 3DOF and orientation of the head of the user by recognizing the results of capturing the head of the user based on the statistical model.
- The pointer tracking unit may perform calibration on a frame including the point, measure the pixel intensity of the calibrated frame, compare the pixel intensity with a critical value, and determine whether to track the point depending on the results of the comparison.
- If the pixel intensity is equal to or higher than the critical value, the pointer tracking unit may group pixels whose intensities are equal to or greater than the critical value, perform Kalman filtering on the grouped pixels, smooth the Kalman filtered result, and track the point.
- If the pixel intensity is smaller than the critical value, the pointer tracking unit may measure the pixel intensity of the calibrated frame again.
- The above and other objects, features and advantages of the present invention will be more clearly understood from the following detailed description taken in conjunction with the accompanying drawings, in which:
-
FIG. 1 is a diagram showing a service environment to which an apparatus for tracking locations using webcams is applied according to an embodiment of the present invention; -
FIG. 2 shows the configuration of alocation tracking apparatus 200 according to an embodiment of the present invention; -
FIG. 3 is a flowchart illustrating a method of tracking locations using webcams according to an embodiment of the present invention; -
FIG. 4 is a flowchart illustrating a method of tracking a user's gaze according to an embodiment of the present invention; and -
FIG. 5 is a flowchart illustrating a method of tracking a laser pointer according to an embodiment of the present invention. - Reference now should be made to the drawings, throughout which the same reference numerals are used to designate the same or similar components.
- The present invention will be described in detail below with reference to the accompanying drawings. Repetitive descriptions and descriptions of known functions and constructions which have been deemed to make the gist of the present invention unnecessarily vague will be omitted below. The embodiments of the present invention are provided in order to fully describe the present invention to a person having ordinary skill in the art. Accordingly, the shapes, sizes, etc. of elements in the drawings may be exaggerated to make the description clear.
- A method and an apparatus for tracking locations using webcams according to the embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
- First, the method and an apparatus for tracking locations using webcams according to the embodiments of the present invention may belong to a technical field for intuitively sensing the intention of a user and transferring it to an application, such as a game, without using the HCI technique (i.e., a user interface using a mouse and a keyboard), but is not limited thereto.
-
FIG. 1 is a diagram showing a service environment to which an apparatus for tracking locations using webcams is applied according to an embodiment of the present invention. - First, the apparatus for tracking locations using webcams (hereinafter referred to as the “location tracking apparatus”) according to the embodiment of the present invention corresponds to an apparatus, such as an interface which operates in conjunction with a First-Person Shooter (FPS) game, but is not limited thereto.
- Referring to
FIG. 1 , the service environment according to the embodiment of the present invention includes two 10 and 20, awebcams screen 30 where the two 10 and 20 are placed, awebcams projector 40, and alaser pointer 50. - The
projector 40 may be coupled to, for example, a computer for executing a game, and projects a game screen onto thescreen 30. - The
webcam 10 disposed above thescreen 30 consecutively captures a game participant (i.e., the face of a user) in front of thescreen 30. - The
webcam 20 in front of theprojector 40 consecutively captures thescreen 30 on which a game screen is being executed. That is, thewebcam 20 consecutively captures a point at which thelaser pointer 50 held by the user overlays on thescreen 30. - In the service environment according to the embodiment of the present invention, the
location tracking apparatus 200 can incorporate the intention of the user into the game in real time by tracking the head of the user and thelaser pointer 50 at the same time using the results captured by the two 10 and 20.webcams - The
location tracking apparatus 200 for tracking the locations of a user and the laser pointer using the webcams will now be described in detail with reference toFIG. 2 . -
FIG. 2 shows the configuration of thelocation tracking apparatus 200 according to an embodiment of the present invention. - Referring to
FIG. 2 , thelocation tracking apparatus 200 includes ahead tracking unit 210 and apointer tracking unit 220. - The
head tracking unit 210 tracks a user's gaze in a FPS game using thewebcam 10 disposed above thescreen 30. - The
pointer tracking unit 220 tracks a point indicated by thelaser pointer 50 using thewebcam 20 in front of theprojector 40. - As described above, the
location tracking apparatus 200 can incorporate a change in the viewpoint of a user in the game by tracking the change using head tracking, without using a mouse or a keyboard in order to change the viewpoint of the user in a primary viewpoint game, such as a shooting game. Furthermore, thelocation tracking apparatus 200 tracks a point, indicated by thelaser pointer 50, simultaneously with head tracking, and designates the point as cross hairs corresponding to a target to be shot at. - A method of tracking locations using webcams will now be described in detail with reference to
FIGS. 3 to 5 . -
FIG. 3 is a flowchart illustrating the method of tracking locations using webcams according to an embodiment of the present invention,FIG. 4 is a flowchart illustrating a method of tracking a user's gaze according to an embodiment of the present invention, andFIG. 5 is a flowchart illustrating a method of tracking a laser pointer according to an embodiment of the present invention. - First, an environment to which the method of tracking locations using the webcams according to the embodiment of the present invention is applied includes the two
10 and 20, thewebcams screen 30 where the two 10 and 20 are placed, thewebcams projector 40, and thelaser pointer 50. Here, thelocation tracking apparatus 200 tracks the head of a user and the location of the laser pointer using the two 10 and 20.webcams - For example, the
projector 40 may project a game screen onto thescreen 30. - Referring to
FIG. 3 , thelocation tracking apparatus 200 tracks a user's gaze in a FPS game using thewebcam 10 disposed above thescreen 30 at step S100. Here, thewebcam 10 disposed above thescreen 30 consecutively captures the head of the user in front of the screen. - Referring to
FIG. 4 , thelocation tracking apparatus 200 extracts one or more feature points of the head based on the results of the capturing of the head of the user at step S110. The feature points of the head include points corresponding to characteristics, such as a face color, an eye, a nose, and a mouth. - The
location tracking apparatus 200 constructs a statistical model by applying the extracted feature points of the head to a machine learning algorithm at step S120. Thelocation tracking apparatus 200 includes the training data of users before constructing the statistical model. Thereafter, thelocation tracking apparatus 200 constructs a statistical model by training based on the training data using the machine learning algorithm. - If the results captured by the
webcam 10 disposed above thescreen 30 are received, thelocation tracking apparatus 200 extracts a location, the 3-Degrees Of Freedom (hereinafter referred to as the “3DOF”), and the orientation of the head of the user by recognizing the results at step S130. - More particularly, if a face image of the user is received from the
webcam 10, thelocation tracking apparatus 200 checks the location of the face by recognizing the face image, recognizes the locations and orientations of an eye, a nose, and a mouth, and then extracts the orientation of the face. - That is, the
location tracking apparatus 200 may track the user's gaze based on the location, the 3DOF, and the orientation of the head of the user extracted using thewebcam 10. Here, the 3DOF includes X (i.e., a horizontal direction), Y (i.e., a vertical direction), and Z (i.e., depth). - Thereafter, the
location tracking apparatus 200 tracks a point indicated by thelaser pointer 50 using thewebcam 20 in front of theprojector 40 at step S200. Here, thewebcam 20 in front of theprojector 40 can capture 30 or more frames per second. - Referring to
FIG. 5 , thelocation tracking apparatus 200 performs calibration on a frame (640×480) including a point that is indicated by thelaser pointer 50 projected onto the screen at step S210. - The
location tracking apparatus 200 measures the pixel intensity for the frame at step S220, and determines whether the measured pixel intensity is equal to or greater than a critical value at step S230. - If the pixel intensity is smaller than the critical value, the
location tracking apparatus 200 measures the pixel intensity for the frame again. If the pixel intensity is equal to or greater than the critical value, thelocation tracking apparatus 200 groups pixels whose intensities are equal to or greater than the critical value at step S240. - The
location tracking apparatus 200 performs Kalman filtering, corresponding to a prediction-correction algorithm, on the grouped pixels at step S250. - The
location tracking apparatus 200 tracks a point indicated by thelaser pointer 50 by smoothing the results of Kalman filtering at step S260. Furthermore, thelocation tracking apparatus 200 designates the tracked point as cross hairs corresponding to a target to be shot at. - Accordingly, the
location tracking apparatus 200 according to the embodiment of the present invention can predict a change in the viewpoint of a user and a change in the target of the laser pointer using thewebcam 10 above thescreen 30 and thewebcam 20 in front of theprojector 40 at the same time. - As described above, according to the embodiments of the present invention, the location tracking apparatus can predict a change in the viewpoint of a user and a change in the target of the laser pointer using webcams at the same time.
- Although the preferred embodiments of the present invention have been disclosed for illustrative purposes, those skilled in the art will appreciate that various modifications, additions and substitutions are possible, without departing from the scope and spirit of the invention as disclosed in the accompanying claims.
Claims (13)
1. A method in which an apparatus operating in conjunction with a First-Person Shooter (FPS) game tracks a user performing the game and a location of a laser pointer held by the user, the method comprising:
extracting a location, 3-Degrees Of Freedom (3DOF) and orientation of a head of the user by capturing the user from a screen on which the game is displayed using a first webcam, and tracking the user's gaze based on results of the extraction; and
capturing the screen using a second webcam, and tracking a point, indicated by the laser pointer and projected onto the screen, based on the results of the capturing.
2. The method as set forth in claim 1 , wherein the user's gaze and the point of the laser pointer are tracked simultaneously.
3. The method as set forth in claim 1 , wherein the tracking the user's gaze comprises:
extracting at least one feature point based on the results of the capturing of the head of the user, performed by the first webcam disposed above the screen;
constructing a statistical model by applying the feature point to a machine learning algorithm; and
extracting the location, 3DOF, and orientation of the head of the user based on the statistical model.
4. The method as set forth in claim 3 , wherein the feature point comprises at least one of a face color, an eye, a nose, and a mouth of the user.
5. The method as set forth in claim 1 , wherein the tracking the point indicated by the laser pointer comprises:
performing calibration on a frame including the point;
measuring a pixel intensity of the calibrated frame; and
comparing the pixel intensity and a critical value, and making a determination of whether to track the point depending on results of the comparison.
6. The method as set forth in claim 5 , further comprising, if the pixel intensity is equal to or higher than the critical value:
grouping pixels whose intensities are equal to or higher than the critical value;
performing Kalman filtering on the grouped pixels; and
smoothing results of the Kalman filtering and tracking the point.
7. The method as set forth in claim 5 , further comprising, if the pixel intensity is smaller than the critical value, measuring the pixel intensity of the calibrated frame again.
8. The method as set forth in claim 1 , wherein in the tracking the point indicated by the laser pointer, the tracked point is designated as cross hairs corresponding to a target in the game to be shot at.
9. A location tracking apparatus, comprising:
a head tracking unit for extracting a location, 3DOF and orientation of a head of a user by capturing the user from a screen on which a game is displayed using a first webcam and tracking the user's gaze based on results of the extraction; and
a pointer tracking unit for tracking a point indicated by a laser pointer projected onto the screen by capturing the screen using a second webcam, and designating the tracked point as cross hairs corresponding to a target to be shot at;
wherein the head tracking unit and the pointer tracking unit track the user's gaze and the point simultaneously.
10. The location tracking apparatus as set forth in claim 9 , wherein the head tracking unit constructs a statistical model based on the results of the capturing performed by the first webcam disposed above the screen, and extracts the location, 3DOF, and orientation of the head of the user by recognizing the results of capturing the head of the user based on the statistical model.
11. The location tracking apparatus as set forth in claim 9 , wherein the pointer tracking unit performs calibration on a frame including the point, measures a pixel intensity of the calibrated frame, compares the pixel intensity with a critical value, and determines whether to track the point depending on results of the comparison.
12. The location tracking apparatus as set forth in claim 11 , wherein if the pixel intensity is equal to or higher than the critical value, the pointer tracking unit groups pixels whose intensities are equal to or greater than the critical value, performs Kalman filtering on the grouped pixels, smoothes the Kalman filtered result, and tracks the point.
13. The location tracking apparatus as set forth in claim 11 , wherein if the pixel intensity is smaller than the critical value, the pointer tracking unit measures the pixel intensity of the calibrated frame again.
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR20100132875 | 2010-12-22 | ||
| KR10-2010-0132875 | 2010-12-22 | ||
| KR1020110037422A KR20120071287A (en) | 2010-12-22 | 2011-04-21 | Apparatus and method for tracking position using webcam |
| KR10-2011-0037422 | 2011-04-21 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20120165084A1 true US20120165084A1 (en) | 2012-06-28 |
Family
ID=46317805
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/332,638 Abandoned US20120165084A1 (en) | 2010-12-22 | 2011-12-21 | Method and apparatus for tracking locations using webcams |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US20120165084A1 (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104455983A (en) * | 2014-11-28 | 2015-03-25 | 海信集团有限公司 | Screen levelness adjusting mechanism and projection display device |
| JP2023108779A (en) * | 2022-01-26 | 2023-08-07 | 富士フイルムビジネスイノベーション株式会社 | Information processing device and information processing program |
| US12541864B2 (en) * | 2020-10-28 | 2026-02-03 | Kyocera Corporation | Object tracking device and object tracking method |
-
2011
- 2011-12-21 US US13/332,638 patent/US20120165084A1/en not_active Abandoned
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104455983A (en) * | 2014-11-28 | 2015-03-25 | 海信集团有限公司 | Screen levelness adjusting mechanism and projection display device |
| US12541864B2 (en) * | 2020-10-28 | 2026-02-03 | Kyocera Corporation | Object tracking device and object tracking method |
| JP2023108779A (en) * | 2022-01-26 | 2023-08-07 | 富士フイルムビジネスイノベーション株式会社 | Information processing device and information processing program |
| JP7757810B2 (en) | 2022-01-26 | 2025-10-22 | 富士フイルムビジネスイノベーション株式会社 | Information processing device and information processing program |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN110427905B (en) | Pedestrian tracking method, device and terminal | |
| Shreve et al. | Macro-and micro-expression spotting in long videos using spatio-temporal strain | |
| Ahmed et al. | Vision based hand gesture recognition using dynamic time warping for Indian sign language | |
| US10445887B2 (en) | Tracking processing device and tracking processing system provided with same, and tracking processing method | |
| JP5203281B2 (en) | Person detection device, person detection method, and person detection program | |
| CN103677274B (en) | A kind of interaction method and system based on active vision | |
| CN103997624B (en) | Overlapping domains dual camera Target Tracking System and method | |
| Pan et al. | A real-time multi-cue hand tracking algorithm based on computer vision | |
| CN113608663B (en) | A fingertip tracking method based on deep learning and K-curvature method | |
| CN102831439A (en) | Gesture tracking method and gesture tracking system | |
| CN104298964B (en) | Method and device for quickly identifying human behavior and action | |
| US20120163661A1 (en) | Apparatus and method for recognizing multi-user interactions | |
| Wu et al. | Robust fingertip detection in a complex environment | |
| CN103376890A (en) | Gesture remote control system based on vision | |
| CN105868574A (en) | Human face tracking optimization method for camera and intelligent health monitoring system based on videos | |
| US10496874B2 (en) | Facial detection device, facial detection system provided with same, and facial detection method | |
| KR101737430B1 (en) | A method of detecting objects in the image with moving background | |
| Tan et al. | Dynamic hand gesture recognition using motion trajectories and key frames | |
| KR100692526B1 (en) | Gesture Recognition Device and System for Automatic System Control | |
| US20120165084A1 (en) | Method and apparatus for tracking locations using webcams | |
| Teng et al. | Facial expressions recognition based on convolutional neural networks for mobile virtual reality | |
| JP5464661B2 (en) | Information terminal equipment | |
| Xu et al. | Bare hand gesture recognition with a single color camera | |
| JP6350331B2 (en) | TRACKING DEVICE, TRACKING METHOD, AND TRACKING PROGRAM | |
| KR20120071287A (en) | Apparatus and method for tracking position using webcam |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTIT Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SUNG, MAN-KYU;REEL/FRAME:027424/0987 Effective date: 20111214 |
|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |