CN104123008A - Man-machine interaction method and system based on static gestures - Google Patents
Man-machine interaction method and system based on static gestures Download PDFInfo
- Publication number
- CN104123008A CN104123008A CN201410371319.5A CN201410371319A CN104123008A CN 104123008 A CN104123008 A CN 104123008A CN 201410371319 A CN201410371319 A CN 201410371319A CN 104123008 A CN104123008 A CN 104123008A
- Authority
- CN
- China
- Prior art keywords
- gesture
- skin color
- model
- tracking
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
本发明提供了一种基于静态手势的人机交互方法及系统,人机交互方法包括手势识别方法,手势识别方法中包括建立实时肤色模型步骤、建立手势几何模型步骤、建立跟踪模型步骤、识别步骤。本发明的有益效果是:本发明可以让人们通过做手势实现人与机器的交互、给机器下达指令。从交互方式层面上说,本发明提供了一种新颖的、简洁的、更为人性化的人机交互方式。从系统实现方法层面说,该系统中用到的手势识别单元有效地克服了传统手势识别的稳定性差、指令量少不可扩充、严重依赖PC平台等问题。
The invention provides a human-computer interaction method and system based on static gestures. The human-computer interaction method includes a gesture recognition method. The gesture recognition method includes the steps of establishing a real-time skin color model, establishing a gesture geometric model, establishing a tracking model, and identifying. . The beneficial effects of the invention are: the invention can allow people to realize the interaction between man and machine and give instructions to the machine by making gestures. From the aspect of interaction mode, the present invention provides a novel, concise and more humanized human-computer interaction mode. From the level of system implementation method, the gesture recognition unit used in the system effectively overcomes the problems of poor stability of traditional gesture recognition, small amount of instructions, non-expandable, and heavy dependence on the PC platform.
Description
技术领域 technical field
本发明涉及人工智能领域,尤其涉及一种基于静态手势的人机交互方法及系统。 The invention relates to the field of artificial intelligence, in particular to a static gesture-based human-computer interaction method and system. the
背景技术 Background technique
手势交互是理想的人机交互方式之一,而手势识别技术是手势交互的重要技术手段。手势识别技术一般包括肤色分割、手势区域提取、手势特征提取、手势特征分类(识别)这四部分。在传统的手势识别技术中这几个技术模块实现各有差异,但总体存在改进空间。传统手势识别技术一般具有以下缺点: Gesture interaction is one of the ideal human-computer interaction methods, and gesture recognition technology is an important technical means of gesture interaction. Gesture recognition technology generally includes four parts: skin color segmentation, gesture region extraction, gesture feature extraction, and gesture feature classification (recognition). In the traditional gesture recognition technology, the realization of these technical modules is different, but there is room for improvement in general. Traditional gesture recognition technology generally has the following disadvantages:
一、传统手势识别方法中肤色分割部分一般采用将RGB颜色转换到Ycbcr或HSV颜色空间,然后通过限定阈值得到肤色分割模块。这样的肤色分割处理方法通常会因为环境光照的影响,效果变的不稳定。而且不能处理好环境中类肤色区域的干扰,对使用环境要求比较高。 1. In the traditional gesture recognition method, the skin color segmentation part generally uses RGB color conversion to Ycbcr or HSV color space, and then obtains the skin color segmentation module by limiting the threshold. Such a skin color segmentation processing method usually becomes unstable due to the influence of ambient light. Moreover, it cannot handle the interference of skin-like areas in the environment well, and has relatively high requirements for the use environment. the
二、传统手势识别方法中手势区域提取部分,很多时候直接将肤色区域认定为手势区域,即使加入手势区域判断,也一般是采用最大肤色连通区域作为手势区域,这种弱特征判断在实际应用中很难将人脸肤色区域同手势肤色区域做区分,最终导致手势识别失败 2. In the gesture area extraction part of the traditional gesture recognition method, the skin color area is directly identified as the gesture area in many cases. Even if the gesture area judgment is added, the largest connected area of skin color is generally used as the gesture area. This weak feature judgment is used in practical applications. It is difficult to distinguish the skin color area of the face from the skin color area of the gesture, which eventually leads to the failure of gesture recognition
三、传统手势识别方法中手势特征提取部分,一般采用凸包面积占用比率作为手势类型的关键特征、另外还有手指个数、手势质心等也经常被用做手势类型特征。这些特征能在一定程度上为手势分类提供特征数据,但当手势种类多的时候则会体现出明显的局限性、误识率明显提高。 3. In the gesture feature extraction part of the traditional gesture recognition method, the convex hull area occupancy ratio is generally used as the key feature of the gesture type. In addition, the number of fingers and the gesture centroid are also often used as gesture type features. These features can provide characteristic data for gesture classification to a certain extent, but when there are many types of gestures, it will show obvious limitations and the misrecognition rate will increase significantly. the
四、在手势特征分类这一块,传统手势识别方法中一般着重于单张图像的静态手势识别,忽视了视频数据的连续性,没能充分利用数据以进一步提高识别的可靠性。 4. In the category of gesture feature classification, traditional gesture recognition methods generally focus on static gesture recognition of a single image, ignoring the continuity of video data, and failing to make full use of the data to further improve the reliability of recognition. the
发明内容 Contents of the invention
为了解决现有技术中的问题,本发明提供了一种基于静态手势的人机交互方法。 In order to solve the problems in the prior art, the present invention provides a static gesture-based human-computer interaction method. the
本发明提供了一种基于静态手势的人机交互方法,包括手势识别方法, 在所述手势识别方法中包括: The present invention provides a static gesture-based human-computer interaction method, including a gesture recognition method, including:
建立实时肤色模型步骤:从图像中提取肤色块; Steps of establishing a real-time skin color model: extracting skin color blocks from the image;
建立手势几何模型步骤:从肤色图像中提取手势特征,定义静态手势指令; Steps for establishing a gesture geometric model: extract gesture features from skin color images, and define static gesture instructions;
建立跟踪模型步骤:对手势图像做跟踪; Steps to establish a tracking model: track the gesture image;
识别步骤:用于手势指令识别; Recognition step: used for gesture command recognition;
在所述实时肤色模型步骤中包括: Include in described real-time skin color model step:
初始肤色获取步骤:利用严格肤色阈值限制和动态帧差来获取初始肤色数据块; Initial skin color acquisition step: use strict skin color threshold limit and dynamic frame difference to get initial skin color data block;
肤色模型计算步骤:利用已有的肤色数据库计算基于亮度索引的多高斯肤色模型,并且在手势识别过程中根据当前所获得肤色图像实时更新模型参数; Skin color model calculation steps: use the existing skin color database to calculate the multi-Gaussian skin color model based on brightness index, and update the model parameters in real time according to the currently obtained skin color image during the gesture recognition process;
肤色判断步骤:根据计算好的肤色模型,对像素点进行肤色判断,当概率大于设定阈值时给予肤色判断,否则做非肤色判断。 Skin color judgment step: According to the calculated skin color model, the skin color judgment is performed on the pixel point, and when the probability is greater than the set threshold, the skin color judgment is given, otherwise, the non-skin color judgment is made. the
作为本发明的进一步改进,在所述手势几何模型步骤中包括: As a further improvement of the present invention, include in described gesture geometry model step:
手势模型构建步骤:对所有肤色区域利用线段和圆在几何上重构手势手型; Gesture model construction steps: use line segments and circles to geometrically reconstruct gesture hand shapes for all skin color regions;
手型肤色区域判断步骤:在已构建的手势几何模型基础上,判断该模型是否满足手的实际特征,如若合理则做手势肤色判断,否则做非手势肤色判断; Judgment step of hand skin color area: Based on the constructed gesture geometric model, judge whether the model satisfies the actual characteristics of the hand. If it is reasonable, make a gesture skin color judgment, otherwise make a non-gesture skin color judgment;
静态手势指令录入步骤:满足用户自定义手势指令。 Static gesture instruction entry steps: meet user-defined gesture instructions. the
作为本发明的进一步改进,所述跟踪模型步骤能够完成对用户1~2个手的跟踪,所述跟踪模型步骤从实时肤色模型步骤中获取输入数据,从手势几何模型步骤获取初始跟踪窗口,最终完成对特定手势肤色块的跟踪,在跟踪过程中,利用帧间信息为手势几何模型步骤、识别步骤提供手势肤色位置信息。 As a further improvement of the present invention, the tracking model step can complete the tracking of 1 to 2 hands of the user, the tracking model step obtains input data from the real-time skin color model step, obtains the initial tracking window from the gesture geometry model step, and finally Complete the tracking of the specific gesture skin color block. During the tracking process, use the inter-frame information to provide gesture skin color position information for the gesture geometric model step and recognition step. the
作为本发明的进一步改进,在所述识别步骤中,根据手势几何模型所得到的几何特征和跟踪模型得到的跟踪信息识别出特定手势指令。 As a further improvement of the present invention, in the identifying step, a specific gesture instruction is identified according to the geometric features obtained from the gesture geometric model and the tracking information obtained from the tracking model. the
作为本发明的进一步改进,该人机交互方法包括: As a further improvement of the present invention, the human-computer interaction method includes:
视频采集步骤:采集用户手势数据,并传输给核心处理步骤; Video collection step: collect user gesture data and transmit it to the core processing step;
核心处理步骤:通过手势识别方法分析出视频中的手势指令,再将手势指令命令下达给指令执行步骤; Core processing steps: analyze the gesture commands in the video through the gesture recognition method, and then issue the gesture commands to the command execution step;
指令执行步骤:执行与手势指令相对应的指令程序。 Instruction execution step: Execute the instruction program corresponding to the gesture instruction. the
本发明还提供了一种基于静态手势的人机交互系统,包括手势识别单元,在所述手势识别单元中包括: The present invention also provides a human-computer interaction system based on static gestures, including a gesture recognition unit, including:
实时肤色模型模块:用于从图像中提取肤色块; Real-time skin color model module: used to extract skin color blocks from images;
手势几何模型模块:用于从肤色图像中提取手势特征,定义静态手势指令; Gesture geometric model module: used to extract gesture features from skin color images and define static gesture instructions;
跟踪模型模块:用于对手势图像做跟踪; Tracking model module: used to track gesture images;
识别模块:用于手势指令识别; Recognition module: used for gesture command recognition;
在所述实时肤色模型模块中包括: Include in described real-time skin color model module:
初始肤色获取模块:利用严格肤色阈值限制和动态帧差来获取初始肤色数据块; Initial skin color acquisition module: use strict skin color threshold limit and dynamic frame difference to obtain initial skin color data block;
肤色模型计算模块:利用已有的肤色数据库计算基于亮度索引的多高斯肤色模型,并且在手势识别过程中根据当前所获得肤色图像实时更新模型参数; Skin color model calculation module: use the existing skin color database to calculate the multi-Gaussian skin color model based on brightness index, and update the model parameters in real time according to the currently obtained skin color image during the gesture recognition process;
肤色判断模块:根据计算好的肤色模型,对像素点进行肤色判断,当概率大于设定阈值时给予肤色判断,否则做非肤色判断。 Skin color judgment module: According to the calculated skin color model, the skin color judgment is performed on the pixels. When the probability is greater than the set threshold, the skin color judgment is given, otherwise the non-skin color judgment is made. the
作为本发明的进一步改进,在所述手势几何模型模块中包括: As a further improvement of the present invention, include in described gesture geometry model module:
手势模型构建模块:用于对所有肤色区域利用线段和圆在几何上重构手势手型; Gesture model building blocks: for geometrically reconstructing gesture hand shapes using line segments and circles for all skin tone regions;
手型肤色区域判断模块:用于在已构建的手势几何模型基础上,判断该模型是否满足手的实际特征,如若合理则做手势肤色判断,否则做非手势肤色判断; Judgment module of hand skin color area: it is used to judge whether the model meets the actual characteristics of the hand on the basis of the constructed gesture geometric model. If it is reasonable, it will judge the skin color of the gesture, otherwise it will judge the skin color of the non-gesture;
静态手势指令录入模块:用于满足用户自定义手势指令。 Static gesture instruction input module: used to meet user-defined gesture instructions. the
作为本发明的进一步改进,所述跟踪模型模块能够完成对用户1~2个手的跟踪,所述跟踪模型模块从实时肤色模型模块中获取输入数据,从手势几何模型模块获取初始跟踪窗口,最终完成对特定手势肤色块的跟踪,在跟踪过程中,利用帧间信息为手势几何模型模块、识别模块提供手势肤色位置信息。 As a further improvement of the present invention, the tracking model module can complete the tracking of 1 to 2 hands of the user, the tracking model module obtains input data from the real-time skin color model module, obtains the initial tracking window from the gesture geometry model module, and finally Complete the tracking of the specific gesture skin color block. During the tracking process, use the inter-frame information to provide gesture skin color position information for the gesture geometric model module and recognition module. the
作为本发明的进一步改进,在所述识别模块中,根据手势几何模型模块所得到的几何特征和跟踪模型模块得到的跟踪信息识别出特定手势指令。 As a further improvement of the present invention, in the recognition module, a specific gesture instruction is recognized according to the geometric features obtained by the gesture geometric model module and the tracking information obtained by the tracking model module. the
作为本发明的进一步改进,该人机交互系统包括: As a further improvement of the present invention, the human-computer interaction system includes:
视频采集单元:用于采集用户手势数据,并传输给核心处理单元; Video acquisition unit: used to collect user gesture data and transmit it to the core processing unit;
核心处理单元:用于通过手势识别单元分析出视频中的手势指令,再 将手势指令命令下达给指令执行单元; Core processing unit: used to analyze the gesture commands in the video through the gesture recognition unit, and then issue the gesture commands to the command execution unit;
指令执行单元:用于执行与手势指令相对应的指令程序。 Instruction execution unit: used to execute instruction programs corresponding to gesture instructions. the
本发明的有益效果是:本发明可以让人们通过做手势实现人与机器的交互、给机器下达指令。从交互方式层面上说,本发明提供了一种新颖的、简洁的、更为人性化的人机交互方式。从系统实现方法层面说,该系统中用到的手势识别单元有效地克服了传统手势识别的稳定性差、指令量少不可扩充、严重依赖PC平台等问题。 The beneficial effects of the invention are: the invention can allow people to realize the interaction between man and machine and give instructions to the machine by making gestures. From the aspect of interaction mode, the present invention provides a novel, concise and more humanized human-computer interaction mode. From the perspective of system implementation method, the gesture recognition unit used in the system effectively overcomes the problems of poor stability of traditional gesture recognition, small number of instructions, non-expandable, and heavy dependence on the PC platform. the
附图说明 Description of drawings
图1是本发明的静态手势指令示意图; Fig. 1 is a schematic diagram of a static gesture instruction of the present invention;
图2是本发明的人机交互系统原理框图; Fig. 2 is a functional block diagram of the human-computer interaction system of the present invention;
图3是本发明的人机交互系统一实施例的原理框图。 Fig. 3 is a functional block diagram of an embodiment of the human-computer interaction system of the present invention. the
具体实施方式 Detailed ways
本发明公开了一种基于静态手势的人机交互方法,包括手势识别方法,在所述手势识别方法中包括: The invention discloses a static gesture-based human-computer interaction method, including a gesture recognition method, which includes:
建立实时肤色模型步骤:从图像中提取肤色块; Steps of establishing a real-time skin color model: extracting skin color blocks from the image;
建立手势几何模型步骤:从肤色图像中提取手势特征,定义静态手势指令; Steps for establishing a gesture geometric model: extract gesture features from skin color images, and define static gesture instructions;
建立跟踪模型步骤:对手势图像做跟踪; Steps to establish a tracking model: track the gesture image;
识别步骤:用于手势指令识别; Recognition step: used for gesture command recognition;
在所述实时肤色模型步骤中包括: Include in described real-time skin color model step:
初始肤色获取步骤:利用严格肤色阈值限制和动态帧差来获取初始肤色数据块; Initial skin color acquisition step: use strict skin color threshold limit and dynamic frame difference to get initial skin color data block;
肤色模型计算步骤:利用已有的肤色数据库计算基于亮度索引的多高斯肤色模型,并且在手势识别过程中根据当前所获得肤色图像实时更新模型参数; Skin color model calculation steps: use the existing skin color database to calculate the multi-Gaussian skin color model based on brightness index, and update the model parameters in real time according to the currently obtained skin color image during the gesture recognition process;
肤色判断步骤:根据计算好的肤色模型,对像素点进行肤色判断,当概率大于设定阈值时给予肤色判断,否则做非肤色判断。 Skin color judgment step: According to the calculated skin color model, the skin color judgment is performed on the pixel point, and when the probability is greater than the set threshold, the skin color judgment is given, otherwise, the non-skin color judgment is made. the
作为本发明的一个实施例,在该实时肤色模型步骤中,首先需要建立实时肤色数据库,为多高斯概率模型提供初始数据。而实时肤色数据库建立在帧差法和严格肤色数据模型的基础上。在完成了实时肤色数据库后便可按照以下原理建立基于亮度索引的多高斯概率肤色模型。建立基于亮度索引的多高斯概率肤色模型的原理为: As an embodiment of the present invention, in the step of real-time skin color model, it is first necessary to establish a real-time skin color database to provide initial data for the multi-Gaussian probability model. The real-time skin color database is based on frame difference method and strict skin color data model. After completing the real-time skin color database, a multi-Gaussian probability skin color model based on brightness index can be established according to the following principles. The principle of establishing a multi-Gaussian probability skin color model based on brightness index is:
计算亮度索引:Y=0.299×r+0.587×g+0.114×b Calculate brightness index: Y=0.299×r+0.587×g+0.114×b
计算特征向量: Compute the eigenvectors:
I1=(r+g+b)/3 I 1 =(r+g+b)/3
I2=r-b I 2 =rb
I3=(2×g-r-b)/2 I 3 =(2×grb)/2
I4=0.492×(b-Y) I 4 =0.492×(bY)
I5=0.877×(r-Y) I 5 =0.877×(rY)
高斯概率计算模型: Gaussian probability calculation model:
在所述手势几何模型步骤中包括: Include in described gesture geometry model step:
手势模型构建步骤:对所有肤色区域利用线段和圆在几何上重构手势手型; Gesture model construction steps: use line segments and circles to geometrically reconstruct gesture hand shapes for all skin color regions;
手型肤色区域判断步骤:在已构建的手势几何模型基础上,判断该模型是否满足手的实际特征(如手指和手掌有交点,手指长和手掌半径长存在比例关系等),如若合理则做手势肤色判断,否则做非手势肤色判断; Judgment step of hand skin color area: Based on the constructed gesture geometric model, judge whether the model satisfies the actual characteristics of the hand (such as the intersection point between the fingers and the palm, the proportional relationship between the length of the finger and the radius of the palm, etc.), and if it is reasonable, do Gesture skin color judgment, otherwise non-gesture skin color judgment;
静态手势指令录入步骤:满足用户自定义手势指令。 Static gesture instruction entry steps: meet user-defined gesture instructions. the
作为本发明的一个实施例,在手势几何模型步骤中,首先需要通过边缘曲率计算出指尖点所在位置。指尖点为满足以下条件的边界点pi: As an embodiment of the present invention, in the gesture geometric modeling step, it is first necessary to calculate the position of the fingertip through the edge curvature. The fingertip point is the boundary point p i satisfying the following conditions:
pipi-k×pipi-k≥0 p i p ik ×p i p ik ≥0
其中pi为边缘连续的边界点。Ω为满足一定曲率阈值的指尖点曲率取值范围。 Among them, p i is the boundary point where the edge is continuous. Ω is the value range of the curvature of the fingertip point satisfying a certain curvature threshold.
在指尖点确定后,需要再确定掌心位置,在本发明中采用距离变换掌心定位法。具体原理如下: After the fingertip point is determined, the position of the center of the palm needs to be determined again. In the present invention, the distance transformation palm center positioning method is adopted. The specific principles are as follows:
计算距离图像: Compute the distance image:
得到有效距离图像模版: Get the effective distance image template:
在有效距离图像模版中计算质心即为所求掌心: Calculate the center of mass in the effective distance image template to obtain the center of the palm:
通过计算所得指尖点和掌心点再加以圆和直线的相交原理,用圆来模拟手掌区域、用直线模拟手指,从而建立手势几何模型。 Through the calculation of fingertip points and palm points and the intersection principle of circles and straight lines, circles are used to simulate the palm area, and straight lines are used to simulate fingers, thereby establishing a gesture geometric model. the
所述跟踪模型步骤能够完成对用户1~2个手的跟踪,所述跟踪模型步骤从实时肤色模型步骤中获取输入数据,从手势几何模型步骤获取初始跟踪窗口,最终完成对特定手势肤色块的跟踪,在跟踪过程中,利用帧间信息为手势几何模型步骤、识别步骤提供手势肤色位置信息。 The tracking model step can complete the tracking of 1 to 2 hands of the user, the tracking model step obtains input data from the real-time skin color model step, obtains the initial tracking window from the gesture geometric model step, and finally completes the tracking of the specific gesture skin color block Tracking, in the tracking process, use the inter-frame information to provide gesture skin color position information for the gesture geometric model step and recognition step. the
作为本发明的一个实施例,在跟踪模型步骤中,通过利用Camshift算法,用来完成对手势图像的跟踪。Camshift算法需要初始搜索窗,在本发明中可以直接利用手势几何模型中的掌心区域作为初始搜索窗,而之后的跟踪过程中搜索窗的确定可以按以下流程来得到。 As an embodiment of the present invention, in the step of tracking the model, the Camshift algorithm is used to complete the tracking of the gesture image. The Camshift algorithm requires an initial search window. In the present invention, the palm area in the gesture geometric model can be directly used as the initial search window, and the determination of the search window in the subsequent tracking process can be obtained according to the following process. the
计算零阶距 Calculate the zero step distance
计算一阶矩 Calculate the first moment
计算搜索窗质心 Calculate search window centroid
调整搜索窗大小 Adjust the size of the search window
宽:
长:1.2s Length: 1.2s
通过这种搜索窗的不断迭代最终实现对手势图像的跟踪。 Through the continuous iteration of the search window, the gesture image tracking is finally realized. the
在所述识别步骤中,根据手势几何模型所得到的几何特征和跟踪模型得到的跟踪信息识别出特定手势指令;具体为从手势几何模型中获得每帧图像中的手势几何模型特征参数,再综合跟踪得到的帧间连续信息做特定手势判断,即手势识别。 In the recognition step, a specific gesture command is identified according to the geometric features obtained from the gesture geometric model and the tracking information obtained from the tracking model; specifically, the gesture geometric model feature parameters in each frame of image are obtained from the gesture geometric model, and then synthesized The continuous information between frames obtained by tracking is used for specific gesture judgment, that is, gesture recognition. the
该人机交互方法包括: The human-computer interaction method includes:
视频采集步骤:采集用户手势数据,并传输给核心处理步骤; Video collection step: collect user gesture data and transmit it to the core processing step;
核心处理步骤:通过手势识别方法分析出视频中的手势指令,再将手势指令命令下达给指令执行步骤; Core processing steps: analyze the gesture commands in the video through the gesture recognition method, and then issue the gesture commands to the command execution step;
指令执行步骤:执行与手势指令相对应的指令程序。 Instruction execution step: Execute the instruction program corresponding to the gesture instruction. the
如图1至图3所示,本发明还公开了一种基于静态手势的人机交互系统,包括手势识别单元,在所述手势识别单元中包括: As shown in Figures 1 to 3, the present invention also discloses a human-computer interaction system based on static gestures, including a gesture recognition unit, which includes:
实时肤色模型模块:用于从图像中提取肤色块; Real-time skin color model module: used to extract skin color blocks from images;
手势几何模型模块:用于从肤色图像中提取手势特征,定义静态手势指令; Gesture geometric model module: used to extract gesture features from skin color images and define static gesture instructions;
跟踪模型模块:用于对手势图像做跟踪; Tracking model module: used to track gesture images;
识别模块:用于手势指令识别; Recognition module: used for gesture command recognition;
在所述实时肤色模型模块中包括: Include in described real-time skin color model module:
初始肤色获取模块:利用严格肤色阈值限制和动态帧差来获取初始肤色数据块; Initial skin color acquisition module: use strict skin color threshold limit and dynamic frame difference to obtain initial skin color data block;
肤色模型计算模块:利用已有的肤色数据库计算基于亮度索引的多高斯肤色模型,并且在手势识别过程中根据当前所获得肤色图像实时更新模型参数; Skin color model calculation module: use the existing skin color database to calculate the multi-Gaussian skin color model based on brightness index, and update the model parameters in real time according to the currently obtained skin color image during the gesture recognition process;
肤色判断模块:根据计算好的肤色模型,对像素点进行肤色判断,当概率大于设定阈值时给予肤色判断,否则做非肤色判断。 Skin color judgment module: According to the calculated skin color model, the skin color judgment is performed on the pixels. When the probability is greater than the set threshold, the skin color judgment is given, otherwise the non-skin color judgment is made. the
在所述手势几何模型模块中包括: Include in described gesture geometry model module:
手势模型构建模块:用于对所有肤色区域利用线段和圆在几何上重构手势手型; Gesture model building blocks: for geometrically reconstructing gesture hand shapes using line segments and circles for all skin tone regions;
手型肤色区域判断模块:用于在已构建的手势几何模型基础上,判断该模型是否满足手的实际特征(如手指和手掌有交点,手指长和手掌半径长存在比例关系等),如若合理则做手势肤色判断,否则做非手势肤色判断; Judgment module of hand skin color area: it is used to judge whether the model satisfies the actual characteristics of the hand (such as the intersection of fingers and palm, the proportional relationship between finger length and palm radius, etc.) on the basis of the constructed gesture geometric model, if reasonable Then make a gesture skin color judgment, otherwise make a non-gesture skin color judgment;
静态手势指令录入模块:用于满足用户自定义手势指令。 Static gesture instruction input module: used to meet user-defined gesture instructions. the
所述跟踪模型模块能够完成对用户1~2个手的跟踪,所述跟踪模型模块从实时肤色模型模块中获取输入数据,从手势几何模型模块获取初始跟踪窗口,最终完成对特定手势肤色块的跟踪,在跟踪过程中,利用帧间信息为手势几何模型模块、识别模块提供手势肤色位置信息,进一步优化了几何模型的计算。 The tracking model module can complete the tracking of 1 to 2 hands of the user. The tracking model module obtains input data from the real-time skin color model module, obtains the initial tracking window from the gesture geometry model module, and finally completes the tracking of the specific gesture skin color block. Tracking, during the tracking process, the inter-frame information is used to provide gesture skin color position information for the gesture geometric model module and recognition module, which further optimizes the calculation of the geometric model. the
在所述识别模块中,根据手势几何模型模块所得到的几何特征和跟踪模型模块得到的跟踪信息识别出特定手势指令。 In the identification module, a specific gesture command is identified according to the geometric features obtained by the gesture geometric model module and the tracking information obtained by the tracking model module. the
该人机交互系统包括: The human-computer interaction system includes:
视频采集单元:用于采集用户手势数据,并传输给核心处理单元; Video acquisition unit: used to collect user gesture data and transmit it to the core processing unit;
核心处理单元:用于通过手势识别单元分析出视频中的手势指令,再将手势指令命令下达给指令执行单元; Core processing unit: used to analyze the gesture commands in the video through the gesture recognition unit, and then issue the gesture commands to the command execution unit;
指令执行单元:用于执行与手势指令相对应的指令程序。 Instruction execution unit: used to execute instruction programs corresponding to gesture instructions. the
其中跟踪模型模块和实时肤色模型模块、手势几何模型模块存在一种对数据的相互修正的关系。识别从手势几何模型模块获得手型特征数据,结合跟踪模型得到的帧间连续信息做特定手势判断。 Among them, the tracking model module, the real-time skin color model module, and the gesture geometry model module have a relationship of mutual correction of data. Recognition obtains hand shape feature data from the gesture geometric model module, and combines the inter-frame continuous information obtained by the tracking model to make specific gesture judgments. the
当用户在设备视频采集单元的视觉范围内做出特定手势指令后,视频采集单元通过视频获取模块得到视频数据,交由核心处理单元上的手势识别单元,处理流程为,首先从图像数据中分离出肤色数据块,在从肤色数据块中根据手势几何模型的可靠性得到手势肤色数据,同时跟踪模型模块也从肤色分割处获取了相应数据然后再根据几何模型确定手势肤色数据库进行特定手势命令跟踪。最终处理数据都汇总到手势命令识别器处做特定手势指令识别。 When the user makes a specific gesture instruction within the visual range of the video acquisition unit of the device, the video acquisition unit obtains video data through the video acquisition module, and hands it over to the gesture recognition unit on the core processing unit. The processing flow is as follows: The skin color data block is obtained, and the gesture skin color data is obtained from the skin color data block according to the reliability of the gesture geometric model. At the same time, the tracking model module also obtains the corresponding data from the skin color segmentation, and then determines the gesture skin color database according to the geometric model to track specific gesture commands. . The final processing data is aggregated to the gesture command recognizer for specific gesture command recognition. the
本发明的应用广泛,例如: The present invention is widely used, for example:
方案一、将特定手势指令和机器人动作指令相对应,如做出特定手势命令1的时候对应机器人向前走的动作指令。用户在机器人视觉范围内做出手势指令1,机器人视频获取单元得到带有特定手势指令1的视频数据,传到核心处理单元后,将获取的数据作为手势识别程序的输入数据,程序输出特定手势命令1。然后根据之前定义的指令对应关系,机器人获得向前走的命令,开始向前走。 Solution 1: Correspond the specific gesture command with the robot action command, for example, when the specific gesture command 1 is made, it corresponds to the motion command of the robot to move forward. The user makes a gesture command 1 within the vision range of the robot, and the video acquisition unit of the robot obtains video data with a specific gesture command 1, and after it is transmitted to the core processing unit, the acquired data is used as the input data of the gesture recognition program, and the program outputs a specific gesture order 1. Then, according to the previously defined command correspondence, the robot gets the command to move forward and starts to move forward. the
方案二、将特定指令和PC上的PPT展示指令对应,如做出特定手势1的时候对应PPT切换到下一页。用户在PC的摄像头视觉范围内做出手势指令1,PC视频获取单元得到带有特定手势指令1的视频数据,传到核心处理单元后,将获取的数据作为手势识别程序的输入数据,程序输出特 定手势命令1。然后根据之前定义的指令对应关系,PPT应用将当前展示的PPT页面切换到下一页。 Solution 2: Correspond the specific command with the PPT display command on the PC, for example, when the specific gesture 1 is made, the corresponding PPT is switched to the next page. The user makes a gesture command 1 within the visual range of the PC camera, and the PC video acquisition unit obtains video data with a specific gesture command 1, and after transmitting it to the core processing unit, the acquired data is used as the input data of the gesture recognition program, and the program outputs Specific gesture commands1. Then, according to the previously defined command correspondence, the PPT application switches the currently displayed PPT page to the next page. the
方案三、将特定指令和智能电视平台上的指令相对应,如做出特定手势1的时候对应电视频道的切换。用户在智能电视的摄像头视觉范围内做出手势指令1,智能电视的视频获取单元得到带有特定手势指令1的视频数据,传到核心处理单元后,将获取的数据作为手势识别程序的输入数据,程序输出特定手势命令1。然后根据之前定义的指令对应关系,电视频道实现切换。 Solution 3: Corresponding specific commands to commands on the smart TV platform, for example, when a specific gesture 1 is made, it corresponds to switching of TV channels. The user makes a gesture instruction 1 within the visual range of the camera of the smart TV, and the video acquisition unit of the smart TV obtains video data with a specific gesture instruction 1, and after transmitting it to the core processing unit, the acquired data is used as the input data of the gesture recognition program , the program outputs a specific gesture command 1. Then, according to the previously defined command correspondence, the TV channel is switched. the
本发明一方面改进了现有手势识别技术方案的一些缺陷,另一方面将手势识别方法应用到具体实际应用场景,为人与机器交互提供了一种更为方便、有效的交互模式、指令下达方式。对比于传统手势识别方法本发明提供了更为稳健的识别模型。能够有效降低光照带来肤色阈值难以确定的影响。在指令集选取上,提供用户自定义指令接口,在必要的情况下用户可以根据需求在合理范围内自定义手势指令。另一方面对比与传统人与机器的交互方式,本发明不需要额外的控制终端,只需要用裸手在机器视觉范围内做出相应手势指令即可实现与机器的交互、指令下达。 On the one hand, the present invention improves some defects of the existing gesture recognition technical solutions, on the other hand, it applies the gesture recognition method to specific practical application scenarios, and provides a more convenient and effective interaction mode and instruction issuing method for human-machine interaction . Compared with traditional gesture recognition methods, the present invention provides a more robust recognition model. It can effectively reduce the influence of light on the skin color threshold that is difficult to determine. In terms of instruction set selection, a user-defined instruction interface is provided. If necessary, users can customize gesture instructions within a reasonable range according to their needs. On the other hand, compared with the traditional human-machine interaction method, the present invention does not require an additional control terminal, and only needs to use bare hands to make corresponding gesture commands within the scope of machine vision to realize the interaction with the machine and issue instructions. the
本发明可以让人们通过做手势实现人与机器的交互、给机器下达指令。从交互方式层面上说,本发明提供了一种新颖的、简洁的、更为人性化的人机交互方式。从系统实现方法层面说,该系统中用到的手势识别单元有效地克服了传统手势识别的稳定性差、指令量少不可扩充、严重依赖PC平台等问题。 The invention allows people to realize the interaction between man and machine and give instructions to the machine by making gestures. From the aspect of interaction mode, the present invention provides a novel, concise and more humanized human-computer interaction mode. From the perspective of system implementation method, the gesture recognition unit used in the system effectively overcomes the problems of poor stability of traditional gesture recognition, small number of instructions, non-expandable, and heavy dependence on the PC platform. the
本发明还具有如有益效果: The present invention also has beneficial effects such as:
一、有效地改善了人与机器的交互体验。为人与机器的交互方式提供了更为人性化的体验。在使用本发明的情况下,人们可以在脱离类似遥控器这种额外控制终端的情况下,实现和机器的互动、指令下达。 1. Effectively improve the interactive experience between humans and machines. It provides a more humanized experience for the way humans and machines interact. In the case of using the present invention, people can realize the interaction with the machine and issue instructions without being separated from an additional control terminal such as a remote controller. the
二、本发明中采用的实时肤色数据库处理方案,可以有效克服光照强度变化带来的影响。通过建立多高斯概率模型,有效实现对肤色数据的判断,将该方法应用到复杂场景中,有效地降低了对类肤色数据的误判。 2. The real-time skin color database processing scheme adopted in the present invention can effectively overcome the influence brought by the change of light intensity. By establishing a multi-Gaussian probability model, the judgment of skin color data is effectively realized, and the method is applied to complex scenes, which effectively reduces the misjudgment of similar skin color data. the
三、通过建立手势几何模型,可以有效做出对手势肤色区域和非手势肤色区域的判断。减弱应用环境中多肤色区域的干扰。 3. By establishing a gesture geometric model, it is possible to effectively judge the gesture skin color area and the non-gesture skin color area. Reduce the interference of multi-skinned areas in the application environment. the
四、引入手势跟踪模型,可以有效增强手势识别结果的稳定性。 4. The introduction of gesture tracking model can effectively enhance the stability of gesture recognition results. the
以上内容是结合具体的优选实施方式对本发明所作的进一步详细说 明,不能认定本发明的具体实施只局限于这些说明。对于本发明所属技术领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干简单推演或替换,都应当视为属于本发明的保护范围。 The above content is a further detailed description of the present invention in conjunction with specific preferred embodiments, and it cannot be assumed that the specific implementation of the present invention is only limited to these descriptions. For those of ordinary skill in the technical field of the present invention, without departing from the concept of the present invention, some simple deduction or replacement can be made, which should be regarded as belonging to the protection scope of the present invention. the
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201410371319.5A CN104123008B (en) | 2014-07-30 | 2014-07-30 | A kind of man-machine interaction method and system based on static gesture |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201410371319.5A CN104123008B (en) | 2014-07-30 | 2014-07-30 | A kind of man-machine interaction method and system based on static gesture |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN104123008A true CN104123008A (en) | 2014-10-29 |
| CN104123008B CN104123008B (en) | 2017-11-03 |
Family
ID=51768445
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201410371319.5A Active CN104123008B (en) | 2014-07-30 | 2014-07-30 | A kind of man-machine interaction method and system based on static gesture |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN104123008B (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106020433A (en) * | 2015-12-09 | 2016-10-12 | 展视网(北京)科技有限公司 | 3D vehicle terminal man-machine interactive system and interaction method |
| CN108900820A (en) * | 2018-05-14 | 2018-11-27 | 河南大学 | A kind of control method and device of projector |
| CN109584507A (en) * | 2018-11-12 | 2019-04-05 | 深圳佑驾创新科技有限公司 | Driver behavior modeling method, apparatus, system, the vehicles and storage medium |
| CN111967404A (en) * | 2020-08-20 | 2020-11-20 | 苏州凝眸物联科技有限公司 | Automatic snapshot method for specific scene |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2008038096A1 (en) * | 2006-09-28 | 2008-04-03 | Nokia Corporation | Improved user interface |
| US20110304541A1 (en) * | 2010-06-11 | 2011-12-15 | Navneet Dalal | Method and system for detecting gestures |
| CN102520790A (en) * | 2011-11-23 | 2012-06-27 | 中兴通讯股份有限公司 | Character input method based on image sensing module, device and terminal |
| US20140106735A1 (en) * | 2012-10-12 | 2014-04-17 | Crestron Electronics, Inc. | User Identification and Location Determination in Control Applications |
| CN103745193A (en) * | 2013-12-17 | 2014-04-23 | 小米科技有限责任公司 | Skin color detection method and skin color detection device |
-
2014
- 2014-07-30 CN CN201410371319.5A patent/CN104123008B/en active Active
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2008038096A1 (en) * | 2006-09-28 | 2008-04-03 | Nokia Corporation | Improved user interface |
| US20110304541A1 (en) * | 2010-06-11 | 2011-12-15 | Navneet Dalal | Method and system for detecting gestures |
| CN102520790A (en) * | 2011-11-23 | 2012-06-27 | 中兴通讯股份有限公司 | Character input method based on image sensing module, device and terminal |
| US20140106735A1 (en) * | 2012-10-12 | 2014-04-17 | Crestron Electronics, Inc. | User Identification and Location Determination in Control Applications |
| CN103745193A (en) * | 2013-12-17 | 2014-04-23 | 小米科技有限责任公司 | Skin color detection method and skin color detection device |
Non-Patent Citations (2)
| Title |
|---|
| 王辉: "基于视觉的实时手势跟踪与识别及其在人机交互中的应用", 《中国优秀硕士学位论文全文数据库》 * |
| 许杏: "基于隐马尔可夫模型的手势识别研究", 《中国优秀硕士学位论文全文数据库》 * |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106020433A (en) * | 2015-12-09 | 2016-10-12 | 展视网(北京)科技有限公司 | 3D vehicle terminal man-machine interactive system and interaction method |
| CN108900820A (en) * | 2018-05-14 | 2018-11-27 | 河南大学 | A kind of control method and device of projector |
| CN109584507A (en) * | 2018-11-12 | 2019-04-05 | 深圳佑驾创新科技有限公司 | Driver behavior modeling method, apparatus, system, the vehicles and storage medium |
| CN111967404A (en) * | 2020-08-20 | 2020-11-20 | 苏州凝眸物联科技有限公司 | Automatic snapshot method for specific scene |
| CN111967404B (en) * | 2020-08-20 | 2024-05-31 | 苏州凝眸物联科技有限公司 | Automatic snapshot method for specific scene |
Also Published As
| Publication number | Publication date |
|---|---|
| CN104123008B (en) | 2017-11-03 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN103294996B (en) | A kind of 3D gesture identification method | |
| CN106598227B (en) | Gesture identification method based on Leap Motion and Kinect | |
| Kim et al. | Simultaneous gesture segmentation and recognition based on forward spotting accumulative HMMs | |
| CN102200834B (en) | Television control-oriented finger-mouse interaction method | |
| CN103353935B (en) | A kind of 3D dynamic gesture identification method for intelligent domestic system | |
| CN103208002B (en) | Based on gesture identification control method and the system of hand contour feature | |
| CN103136986B (en) | Sign Language Recognition Method and system | |
| CN102831404B (en) | Gesture detecting method and system | |
| CN114792443B (en) | Intelligent device gesture recognition control method based on image recognition | |
| CN104331158B (en) | The man-machine interaction method and device of a kind of gesture control | |
| CN111124117B (en) | Augmented reality interaction method and device based on hand-drawn sketches | |
| CN101807114A (en) | Natural interactive method based on three-dimensional gestures | |
| CN102854983A (en) | Man-machine interaction method based on gesture recognition | |
| CN103150019A (en) | Handwriting input system and method | |
| CN103984928A (en) | Finger gesture recognition method based on field depth image | |
| CN102831439A (en) | Gesture tracking method and gesture tracking system | |
| CN103207709A (en) | Multi-touch system and method | |
| CN104063039A (en) | Human-computer interaction method of wearable computer intelligent terminal | |
| CN109800676A (en) | Gesture identification method and system based on depth information | |
| CN104298354A (en) | Man-machine interaction gesture recognition method | |
| CN103488294A (en) | Non-contact gesture control mapping adjustment method based on user interactive habits | |
| CN103995595A (en) | Game somatosensory control method based on hand gestures | |
| CN106200971A (en) | Man-machine interactive system device based on gesture identification and operational approach | |
| CN109145802A (en) | More manpower gesture man-machine interaction methods and device based on Kinect | |
| CN103092437A (en) | Portable touch interactive system based on image processing technology |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |