CN106228111A - A kind of method based on skeleton sequential extraction procedures key frame - Google Patents
A kind of method based on skeleton sequential extraction procedures key frame Download PDFInfo
- Publication number
- CN106228111A CN106228111A CN201610539455.XA CN201610539455A CN106228111A CN 106228111 A CN106228111 A CN 106228111A CN 201610539455 A CN201610539455 A CN 201610539455A CN 106228111 A CN106228111 A CN 106228111A
- Authority
- CN
- China
- Prior art keywords
- information entropy
- frame
- skeleton
- original frame
- entropy value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
Abstract
本发明涉及一种基于骨骼序列提取关键帧的方法,包括:通过Kinect摄像机捕捉人类动作,得到包含多个骨骼节点的三维骨骼序列;将相邻帧的骨骼坐标相减,求出所有骨骼的三维骨骼运动矢量;将所有骨骼的三维骨骼运动矢量分别在笛卡尔正交系的三个平面上进行投影,在每个投影面上,根据方向与幅值对骨骼运动矢量进行概率统计,得到直方图;根据信息熵公式,对相邻帧的骨骼运动矢量直方图求信息熵值,将信息熵具有局部极大值的帧定义为原始帧;对于整个三维骨骼序列的各个原始帧,计算出交织系数,并对原始帧的信息熵值进行加权;得到该骨骼序列人类动作的关键帧。本发明能够准确可靠并高效的提取人类动作关键帧。
The invention relates to a method for extracting key frames based on a skeleton sequence, comprising: capturing human actions through a Kinect camera to obtain a three-dimensional skeleton sequence containing a plurality of skeleton nodes; subtracting the skeleton coordinates of adjacent frames to obtain the three-dimensional Skeleton motion vector: project the three-dimensional bone motion vectors of all bones on the three planes of the Cartesian orthogonal system, and perform probability statistics on the bone motion vectors according to the direction and amplitude on each projection plane to obtain a histogram ;According to the information entropy formula, calculate the information entropy value for the bone motion vector histogram of the adjacent frame, and define the frame with the local maximum value of the information entropy as the original frame; for each original frame of the whole three-dimensional skeleton sequence, calculate the interleaving coefficient , and weight the information entropy value of the original frame; get the key frame of the skeleton sequence human action. The invention can accurately, reliably and efficiently extract key frames of human actions.
Description
技术领域technical field
本发明属于多媒体信息处理领域,涉及一种提取关键帧的方法。The invention belongs to the field of multimedia information processing and relates to a method for extracting key frames.
背景技术Background technique
随着网络时代的到来,以及计算机行业的迅速发展,计算机智能的市场也蓬勃发展。机器学习、模式识别、数据挖掘等领域在当前社会具有广阔的发展空间。从属于模式识别领域的计算机的人类动作检测识别在当今世界上具有很多应用,例如人机交互的体感游戏、智能监控、视频检索等等。然而在计算机处理视频的角度,视频信息量往往过于庞大。为了提高视频处理速度,令基于视频的机器学习算法具有更强的可应用性,在视频中筛选出包含动作信息更加丰富的关键帧来进行处理,近年来变得非常流行。本发明提出一种针对人类动作,基于三维骨骼序列,在视频序列中提取动作关键帧的方法。With the advent of the Internet age and the rapid development of the computer industry, the market for computer intelligence is also booming. Machine learning, pattern recognition, data mining and other fields have broad development space in the current society. The human action detection and recognition of computers belonging to the field of pattern recognition has many applications in the world today, such as somatosensory games for human-computer interaction, intelligent monitoring, video retrieval, and so on. However, from the perspective of computer processing video, the amount of video information is often too large. In order to improve the speed of video processing and make video-based machine learning algorithms more applicable, it has become very popular in recent years to filter out key frames that contain more action information in videos for processing. The invention proposes a method for extracting action key frames in a video sequence based on a three-dimensional skeleton sequence for human actions.
近年来,相机行业发展迅速,能够捕捉深度信息的相机具有越来越广泛的应用。在2010年微软发布Kinect摄像机后,深度摄像机走进千家万户,国际上大量的视频以及图片方向的学者将研究方向逐渐转向基于RGB-D的信息处理。随着在深度视频序列中跟踪人类骨骼算法的不断改进,骨骼信息作为更加抽象且高层次的人体特征,有广泛的应用,因其具有光线不敏感特性,以及更加全面的三维特性。然而,目前还没有基于骨骼序列的关键帧提取技术。In recent years, the camera industry has developed rapidly, and cameras capable of capturing depth information have become more and more widely used. After Microsoft released the Kinect camera in 2010, the depth camera has entered thousands of households, and a large number of scholars in the field of video and pictures in the world have gradually shifted their research direction to information processing based on RGB-D. With the continuous improvement of human skeleton tracking algorithms in depth video sequences, skeleton information, as a more abstract and high-level human body feature, has a wide range of applications because of its light insensitivity and more comprehensive three-dimensional characteristics. However, there is currently no keyframe extraction technique based on skeletal sequences.
发明内容Contents of the invention
为了对视频序列进行更加便捷的处理,让计算机快速有效的识别人类的动作,本发明基于骨骼序列,提出一种人类动作关键帧的提取方法。该方法具有相比二维信息更加鲁棒的空间特性。同时,得利于骨骼信息的简略特性,具有很高的运行效率。发明内容如下:In order to process video sequences more conveniently and allow computers to quickly and effectively recognize human actions, the present invention proposes a method for extracting key frames of human actions based on skeleton sequences. This method has more robust spatial characteristics than two-dimensional information. At the same time, thanks to the simplicity of bone information, it has high operating efficiency. The content of the invention is as follows:
一种基于骨骼序列提取关键帧的方法,包括下列步骤:A method for extracting key frames based on a skeleton sequence, comprising the following steps:
1)通过Kinect摄像机捕捉人类动作,在捕获的数据流中进行骨骼跟踪,得到包含多个骨骼节点的三维骨骼序列;1) Capture human actions through the Kinect camera, perform skeleton tracking in the captured data stream, and obtain a 3D skeleton sequence containing multiple skeleton nodes;
2)针对每一个骨骼节点,将相邻帧的骨骼坐标相减,求出各个该骨骼节点在相邻帧间的骨骼运动矢量,进而计算出所有骨骼的三维骨骼运动矢量;2) For each skeletal node, subtract the skeletal coordinates of adjacent frames to obtain the skeletal motion vectors of each skeletal node between adjacent frames, and then calculate the three-dimensional skeletal motion vectors of all bones;
3)将所有骨骼的三维骨骼运动矢量分别在笛卡尔正交系的三个平面上进行投影,在每个投影面上,根据方向与幅值对骨骼运动矢量进行概率统计,得到直方图,将其定义为骨骼运动矢量直方图;3) Project the three-dimensional bone motion vectors of all bones on the three planes of the Cartesian orthogonal system. On each projection plane, carry out probability statistics on the bone motion vectors according to the direction and amplitude, and obtain a histogram. It is defined as a histogram of bone motion vectors;
4)根据信息熵公式,对相邻帧的骨骼运动矢量直方图求信息熵值;将整个视频序列中,所有的骨骼运动矢量直方图信息熵值按照视频先后顺序依次排列,并将其绘制成曲线图,将此曲线定义为熵曲线,在熵曲线中,求出局部极大值,将信息熵具有局部极大值的帧定义为原始帧;4) According to the information entropy formula, the information entropy value is calculated for the bone motion vector histogram of adjacent frames; in the entire video sequence, all the bone motion vector histogram information entropy values are arranged in sequence according to the video sequence, and drawn as Curve diagram, this curve is defined as an entropy curve, in the entropy curve, the local maximum value is obtained, and the frame with the information entropy having a local maximum value is defined as the original frame;
5)对于整个三维骨骼序列的各个原始帧i,根据其自身的信息熵值以及相邻帧的信息熵值,通过下述公式的交织方式能够计算出交织系数HI:5) For each original frame i of the entire 3D skeleton sequence, according to its own information entropy value and the information entropy value of adjacent frames, the interleaving coefficient HI can be calculated by the interleaving method of the following formula:
其中,原始帧的信息熵值为H(i),并令H(i±x)表示与原始帧i相邻x帧的信息熵值,+号代表之后的帧,-号代表之前的帧;令其与原始帧信息熵值进行乘积,从而对原始帧的信息熵值进行加权;Wherein, the information entropy value of the original frame is H(i), and let H(i±x) represent the information entropy value of x frames adjacent to the original frame i, the + sign represents the subsequent frame, and the - sign represents the previous frame; Let it be multiplied by the information entropy value of the original frame, so as to weight the information entropy value of the original frame;
6)将每个原始帧的信息熵值H(i),与其交织系数HI相乘进行加权,得到加权后的原始帧信息熵值;6) The information entropy value H(i) of each original frame is multiplied by its interleaving coefficient HI for weighting, and the weighted original frame information entropy value is obtained;
7)根据加权后的原始帧信息熵值,绘制新的信息熵值曲线,求出局部极大值对应的帧,作为该骨骼序列人类动作的关键帧。7) According to the weighted original frame information entropy value, draw a new information entropy value curve, and find the frame corresponding to the local maximum value as the key frame of the skeleton sequence human action.
本发明能够准确可靠并高效的提取人类动作关键帧。The invention can accurately, reliably and efficiently extract key frames of human actions.
附图说明Description of drawings
图1为整个关键帧提取框架Figure 1 is the whole key frame extraction framework
图2为本发明在MSRAction-3D数据集上对挥手动作所提取的关键帧,采用灰度图可视化Figure 2 is the key frame extracted by the present invention for the waving action on the MSRAction-3D data set, which is visualized by grayscale image
具体实施方式detailed description
1)本发明采用Windows8的32位操作系统,开发IDE为VS2010,配置好Kinect forWindows SDK v1.6以及OpenCV2.3.0或者更高版本,采用NUI骨骼跟踪方式将在Kinect捕获的数据流中进行骨骼跟踪,并将人类骨骼动作序列进行输出。1) The present invention adopts the 32-bit operating system of Windows8, the development IDE is VS2010, Kinect for Windows SDK v1.6 and OpenCV2.3.0 or higher versions are configured, and the NUI bone tracking method is used to carry out bone tracking in the data stream captured by Kinect , and output the sequence of human skeleton actions.
2)骨骼序列的每一帧内,包含20个人类骨骼节点的三维坐标。对于每一个骨骼节点,每相邻两帧的骨骼三维坐标的绝对差值即为该骨骼节点在该相邻两帧间的骨骼运动矢量,进而能够求出所有20个骨骼节点的三维骨骼运动矢量。2) Each frame of the skeleton sequence contains the three-dimensional coordinates of 20 human skeleton nodes. For each skeletal node, the absolute difference between the three-dimensional coordinates of the bones in every two adjacent frames is the bone motion vector of the bone node between the two adjacent frames, and then the three-dimensional bone motion vectors of all 20 bone nodes can be obtained .
3)对两帧间所有骨骼节点的骨骼运动矢量在笛卡尔正交系内进行三个方向的投影,每个骨骼节点的运动矢量在二维平面上的投影都具有不同的方向和矢量大小。在每个二维平面上,以x轴正方向为基准,沿逆时针方向,每旋转45°定义为一个方向,至此平面可被分为8个方向。根据实验结果,本发明以每个视频序列中所有骨骼运动矢量的最大幅度值为标准,将所有二维平面上的骨骼运动矢量分成5个大小范围。由此,根据骨骼运动矢量的大小和方向,依次定义了40个类别(类别顺序及编号不影响结果),每个二维投影面的骨骼运动矢量都可以根据方向及大小被归分为一个类别。3) The bone motion vectors of all bone nodes between two frames are projected in three directions in the Cartesian orthogonal system, and the projections of the motion vectors of each bone node on a two-dimensional plane have different directions and vector sizes. On each two-dimensional plane, with the positive direction of the x-axis as the reference, every 45° rotation in the counterclockwise direction is defined as a direction, so far the plane can be divided into 8 directions. According to the experimental results, the present invention uses the maximum amplitude value of all bone motion vectors in each video sequence as a standard, and divides all bone motion vectors on two-dimensional planes into five size ranges. Thus, according to the size and direction of the skeletal motion vector, 40 categories are defined in turn (the order and number of the categories do not affect the result), and the skeletal motion vector of each two-dimensional projection surface can be classified into a category according to the direction and size .
在每个投影面上,统计出每个类别包含的骨骼数目,可以得到一个维度为40的向量(即直方图),将三个投影面上分别统计得到的向量连接起来,得到维度为120的向量(即直方图),将其定义为骨骼运动矢量直方图。On each projection surface, the number of bones contained in each category is counted, and a vector with a dimension of 40 (that is, a histogram) can be obtained. The vectors obtained by statistics on the three projection surfaces are connected to obtain a dimension of 120. Vector (i.e. histogram), which is defined as a histogram of bone motion vectors.
4)对每一个骨骼运动矢量直方图,根据信息熵公式:可以求出每个骨骼运动矢量对应的熵值。其中H为信息熵值,pi为120维向量中第i个类别在整个直方图中所占的比例,n为直方图的长度,在本发明中取n=120。4) For each bone motion vector histogram, according to the information entropy formula: The entropy value corresponding to each bone motion vector can be calculated. Wherein H is the information entropy value, p i is the proportion of the i-th category in the entire histogram in the 120-dimensional vector, n is the length of the histogram, and n=120 is taken in the present invention.
对于整个骨骼序列,将所有信息熵连接起来,得到由信息熵组成的曲线,将其定义为熵曲线。抽取出熵曲线中的局部极大值,即熵曲线中熵值满足同时大于左右两帧熵值的点,将骨骼序列熵曲线中具有局部极大值的帧定义为原始帧。For the entire bone sequence, all information entropies are connected to obtain a curve composed of information entropy, which is defined as an entropy curve. Extract the local maximum value in the entropy curve, that is, the point in the entropy curve whose entropy value is greater than the entropy values of the left and right frames at the same time, and define the frame with the local maximum value in the bone sequence entropy curve as the original frame.
5)在熵曲线中,对每一个原始帧,假设该帧在整个视频中为第i帧,则其熵值为H(i),并另H(i±x)表示与该帧相邻的x帧(之前或之后)的熵值。根据如下交织公式计算出该帧的交织系数HI。该交织系数反映了原始帧的骨骼运动与其相邻帧的运动差别大小。用交织系数与原始帧的熵值相乘,从而实现对原始帧熵值进行加权。交织系数公式如下:5) In the entropy curve, for each original frame, assuming that the frame is the i-th frame in the entire video, its entropy value is H(i), and H(i±x) represents the adjacent frame Entropy value for x frames (before or after). Calculate the interleaving coefficient HI of the frame according to the following interleaving formula. The interleaving coefficient reflects the difference between the skeletal motion of the original frame and the motion of its adjacent frames. The entropy value of the original frame is multiplied by the interleaving coefficient, so as to realize the weighting of the entropy value of the original frame. The formula for the interleaving coefficient is as follows:
6)将进行加权后的原始帧熵值,根据视频顺序依次相连接,得到新的熵曲线。在新熵曲线上取出局部极大值所对应的帧,作为该视频中人类动作序列的关键帧。6) The weighted original frame entropy values are sequentially connected according to the video sequence to obtain a new entropy curve. Take the frame corresponding to the local maximum value on the new entropy curve as the key frame of the human action sequence in the video.
下面对本发明在MSRAction-3D数据集上进行测试的结果进行说明:The result that the present invention is tested on the MSRAction-3D data set is described below:
MSRAction-3D为当今具有影响力的人类动作检测识别数据集,该数据集包含20类动作,提供深度信息数据以及骨骼数据。本发明依据以上叙述的关键帧提取方法,对挥手动作进行了关键帧提取,该挥手动作共包含58帧,通过本发明方法,共提取出8个关键帧。图2为将关键帧对应的深度信息进行可视化后依次排列的结果。通过结果我们能够看到,原本包含多帧的动作序列通过本发明方法,提取出了能够表征整个动作的少数帧。通过本发明,对整个视频序列的处理可以转换为处理关键帧序列,从而大大减少处理数据的冗余,减少算法的运行时间、空间成本,提高了在处理视频方面复杂算法的实用性。MSRAction-3D is an influential human action detection and recognition dataset, which contains 20 types of actions, providing depth information data and bone data. According to the key frame extraction method described above, the present invention extracts the key frames of the waving action. The hand waving action contains a total of 58 frames. Through the method of the present invention, a total of 8 key frames are extracted. Figure 2 is the result of visualizing the depth information corresponding to the key frames and then arranging them sequentially. From the results, we can see that the method of the present invention extracts a few frames that can represent the entire action from an action sequence that originally included multiple frames. Through the present invention, the processing of the entire video sequence can be converted into the processing of key frame sequences, thereby greatly reducing the redundancy of processing data, reducing the running time and space cost of algorithms, and improving the practicability of complex algorithms in video processing.
Claims (1)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201610539455.XA CN106228111A (en) | 2016-07-08 | 2016-07-08 | A kind of method based on skeleton sequential extraction procedures key frame |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201610539455.XA CN106228111A (en) | 2016-07-08 | 2016-07-08 | A kind of method based on skeleton sequential extraction procedures key frame |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN106228111A true CN106228111A (en) | 2016-12-14 |
Family
ID=57520339
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201610539455.XA Pending CN106228111A (en) | 2016-07-08 | 2016-07-08 | A kind of method based on skeleton sequential extraction procedures key frame |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN106228111A (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109190474A (en) * | 2018-08-01 | 2019-01-11 | 南昌大学 | Human body animation extraction method of key frame based on posture conspicuousness |
| CN109934183A (en) * | 2019-03-18 | 2019-06-25 | 北京市商汤科技开发有限公司 | Image processing method and device, detection device and storage medium |
| CN111402290A (en) * | 2020-02-29 | 2020-07-10 | 华为技术有限公司 | Action restoration method and device based on skeleton key points |
| CN114724058A (en) * | 2022-03-14 | 2022-07-08 | 山东大学 | Method for extracting key frames of fusion characteristic motion video based on human body posture recognition |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102395984A (en) * | 2009-04-14 | 2012-03-28 | 皇家飞利浦电子股份有限公司 | Key frames extraction for video content analysis |
| WO2012078702A1 (en) * | 2010-12-10 | 2012-06-14 | Eastman Kodak Company | Video key frame extraction using sparse representation |
| CN102749993A (en) * | 2012-05-30 | 2012-10-24 | 无锡掌游天下科技有限公司 | Motion recognition method based on skeleton node data |
| CN103020648A (en) * | 2013-01-09 | 2013-04-03 | 北京东方艾迪普科技发展有限公司 | Method and device for identifying action types, and method and device for broadcasting programs |
| US20150125045A1 (en) * | 2013-11-04 | 2015-05-07 | Steffen Gauglitz | Environment Mapping with Automatic Motion Model Selection |
-
2016
- 2016-07-08 CN CN201610539455.XA patent/CN106228111A/en active Pending
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102395984A (en) * | 2009-04-14 | 2012-03-28 | 皇家飞利浦电子股份有限公司 | Key frames extraction for video content analysis |
| WO2012078702A1 (en) * | 2010-12-10 | 2012-06-14 | Eastman Kodak Company | Video key frame extraction using sparse representation |
| CN102749993A (en) * | 2012-05-30 | 2012-10-24 | 无锡掌游天下科技有限公司 | Motion recognition method based on skeleton node data |
| CN103020648A (en) * | 2013-01-09 | 2013-04-03 | 北京东方艾迪普科技发展有限公司 | Method and device for identifying action types, and method and device for broadcasting programs |
| US20150125045A1 (en) * | 2013-11-04 | 2015-05-07 | Steffen Gauglitz | Environment Mapping with Automatic Motion Model Selection |
Non-Patent Citations (2)
| Title |
|---|
| LING SHAO等: "Motion Histogram Analysis Based Key Frame Extraction for Human Action/Activity Representation", 《2009 CANADIAN CONFERENCE ON COMPUTER AND ROBOT VISION》 * |
| 邹维嘉: "基于多示例学习的动作识别与显著性检测研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109190474A (en) * | 2018-08-01 | 2019-01-11 | 南昌大学 | Human body animation extraction method of key frame based on posture conspicuousness |
| CN109190474B (en) * | 2018-08-01 | 2021-07-20 | 南昌大学 | Human animation key frame extraction method based on pose saliency |
| CN109934183A (en) * | 2019-03-18 | 2019-06-25 | 北京市商汤科技开发有限公司 | Image processing method and device, detection device and storage medium |
| CN111402290A (en) * | 2020-02-29 | 2020-07-10 | 华为技术有限公司 | Action restoration method and device based on skeleton key points |
| CN111402290B (en) * | 2020-02-29 | 2023-09-12 | 华为技术有限公司 | An action restoration method and device based on skeleton key points |
| CN114724058A (en) * | 2022-03-14 | 2022-07-08 | 山东大学 | Method for extracting key frames of fusion characteristic motion video based on human body posture recognition |
| CN114724058B (en) * | 2022-03-14 | 2024-11-15 | 山东大学 | A method for extracting key frames from motion videos based on fusion features of human posture recognition |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Nada et al. | Pushing the limits of unconstrained face detection: a challenge dataset and baseline results | |
| Ahmad et al. | Human action recognition using shape and CLG-motion flow from multi-view image sequences | |
| Ding et al. | Violence detection in video by using 3D convolutional neural networks | |
| Baraldi et al. | Gesture recognition in ego-centric videos using dense trajectories and hand segmentation | |
| Sudhakaran et al. | Convolutional long short-term memory networks for recognizing first person interactions | |
| CN103500340B (en) | Human body behavior identification method based on thematic knowledge transfer | |
| CN103336835B (en) | Image retrieval method based on weight color-sift characteristic dictionary | |
| Tekin et al. | Predicting people’s 3D poses from short sequences | |
| Chen et al. | Gradient local auto-correlations and extreme learning machine for depth-based activity recognition | |
| Wang et al. | Exploring hybrid spatio-temporal convolutional networks for human action recognition | |
| Wani et al. | Deep learning-based video action recognition: a review | |
| CN103699874A (en) | Crowd abnormal behavior identification method based on SURF (Speed-Up Robust Feature) stream and LLE (Locally Linear Embedding) sparse representation | |
| Caetano et al. | Optical Flow Co-occurrence Matrices: A novel spatiotemporal feature descriptor | |
| CN106228111A (en) | A kind of method based on skeleton sequential extraction procedures key frame | |
| Ming et al. | Uniform local binary pattern based texture-edge feature for 3D human behavior recognition | |
| Elias et al. | Motion images: An effective representation of motion capture data for similarity search | |
| CN113822134B (en) | A video-based instance tracking method, device, equipment and storage medium | |
| Hadfield et al. | Natural action recognition using invariant 3D motion encoding | |
| Liu et al. | Synthesis of spatio-temporal descriptors for dynamic hand gesture recognition using genetic programming | |
| Huang et al. | Extracting refined low-rank features of robust PCA for human action recognition | |
| Brun et al. | HAck: A system for the recognition of human actions by kernels of visual strings | |
| Gao et al. | A new dataset and evaluation for infrared action recognition | |
| Liu et al. | Genetic Programming-Evolved Spatio-Temporal Descriptor for Human Action Recognition. | |
| Wang et al. | Detecting action-relevant regions for action recognition using a three-stage saliency detection technique | |
| Huynh-The et al. | PAM-based flexible generative topic model for 3D interactive activity recognition |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20161214 |
|
| RJ01 | Rejection of invention patent application after publication |