CN117912116B - Wild animal posture estimation method in real scene - Google Patents
Wild animal posture estimation method in real scene Download PDFInfo
- Publication number
- CN117912116B CN117912116B CN202410101315.9A CN202410101315A CN117912116B CN 117912116 B CN117912116 B CN 117912116B CN 202410101315 A CN202410101315 A CN 202410101315A CN 117912116 B CN117912116 B CN 117912116B
- Authority
- CN
- China
- Prior art keywords
- heat map
- model
- formula
- wild animal
- animal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0495—Quantised networks; Sparse networks; Compressed networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及野生动物监测保护技术领域,特别是涉及一种真实场景下野生动物姿态估计方法。The present invention relates to the technical field of wildlife monitoring and protection, and in particular to a method for estimating the posture of a wild animal in a real scene.
背景技术Background technique
野生动物监测和保护工作具有重要意义,通过对动物姿态进行估计,可以更好的实现野生动物行为和习性的理解,推动野生动物保护工作的进行。目前野生动物姿态估计方法受限于场景和物种的单一性,在实际的复杂环境场景下,动物姿态估计模型的性能严重下降,使得该项技术的实际应用情况受到了一些限制,阻碍了野生动物保护工作的高效率开展。Wildlife monitoring and protection are of great significance. By estimating animal posture, we can better understand the behavior and habits of wild animals and promote the protection of wild animals. At present, the method of estimating wild animal posture is limited by the single scene and species. In the actual complex environment, the performance of the animal posture estimation model is seriously reduced, which limits the actual application of this technology and hinders the efficient implementation of wildlife protection.
目前,生态研究领域依然依靠野外工作人员来收集野生动物的相关数据,并通过专家对数据的分析来了解当前区域的野生动物生存现状。随着技术的发展,通过在野外部署红外触发相机和无线传感器网络对野生动物进行监测,使得数据的收集能力大大提升。同时,随着基于深度学习的图像识别技术发展,野生动物监测的及时性大大提升,对于动物多样性的保护能力也得到了极大的改善。但是,目前市面上常见的野生动物监测保护方法仍未完全发挥出视觉挖掘的能力,缺乏对有关生物多样性特征更详尽的信息探索;鉴于此,本发明提出一种真实场景下野生动物姿态估计方法。At present, the field of ecological research still relies on field workers to collect relevant data on wild animals, and to understand the survival status of wild animals in the current area through expert analysis of the data. With the development of technology, the ability to collect data has been greatly improved by deploying infrared triggered cameras and wireless sensor networks in the wild to monitor wild animals. At the same time, with the development of image recognition technology based on deep learning, the timeliness of wild animal monitoring has been greatly improved, and the ability to protect animal diversity has also been greatly improved. However, the common wildlife monitoring and protection methods currently on the market have not yet fully exerted the ability of visual mining, and lack more detailed information exploration on the characteristics of biodiversity; in view of this, the present invention proposes a method for estimating the posture of wild animals in real scenarios.
发明内容Summary of the invention
本发明的目的在于提供一种真实场景下野生动物姿态估计方法以解决背景技术中所提出的技术问题,实现不同野外真实场景下野生动物姿态自动估计,修正野生动物姿态估计模型的量化误差,设计野生动物姿态估计模型轻量化策略。The purpose of the present invention is to provide a method for estimating the posture of wild animals in real scenarios to solve the technical problems raised in the background technology, realize automatic estimation of the posture of wild animals in different real outdoor scenarios, correct the quantization error of the wild animal posture estimation model, and design a lightweight strategy for the wild animal posture estimation model.
为了实现上述目的,本发明采用了如下技术方案:In order to achieve the above object, the present invention adopts the following technical solutions:
一种真实场景下野生动物姿态估计方法,包括以下步骤:A method for estimating the posture of wild animals in a real scene comprises the following steps:
S1、构建动物姿态估计图像的数据集,对所构建的数据集进行处理;S1, constructing a data set of animal posture estimation images, and processing the constructed data set;
S2、基于组白化操作,构建基于热图生成的自由简单基线姿态估计模型,利用所述模型生成热图,利用热图完成模型训练;S2. Based on the group whitening operation, construct a free simple baseline pose estimation model based on heat map generation, generate a heat map using the model, and complete model training using the heat map;
S3、对S2中所构建的模型进行修正,设计坐标表征方法和热图解码方法;S3, modifying the model constructed in S2, designing a coordinate representation method and a heat map decoding method;
S4、综合S1~S3所述操作,采用轻量化的姿态估计网络解码架构,完成真实场景下野生动物姿态的估计。S4. Combining the operations described in S1 to S3, a lightweight posture estimation network decoding architecture is used to complete the estimation of the posture of wild animals in real scenes.
优选地,S1中所述对数据集进行处理,具体包括如下内容:Preferably, the processing of the data set in S1 specifically includes the following contents:
将公共数据集划分为测试集和训练集,对测试集中增添真实场景下常见的野外环境因素和风格迁移因素,模拟复杂环境场景,以观察不同场景下测试集与训练集的分布差异。The public dataset is divided into a test set and a training set. Common wild environmental factors and style transfer factors in real scenarios are added to the test set to simulate complex environmental scenarios in order to observe the distribution differences between the test set and the training set in different scenarios.
优选地,所述S2具体包括如下内容:Preferably, S2 specifically includes the following contents:
利用非批次归一化和组白化方法,并加入挤压和提取模块,构建基于热图生成的自由简单基线野生动物姿态估计模型,以自适应学习特征权重;所述模型遵循单架构-编码器解码器设计,对于包含关键点的动物输入图像,通过骨干网络进行特征提取得到特征图,解码器中的反卷积模块通过对特征图上采样获取预测热图,最后通过Argmax函数获得热图最大值作为最终的关键点坐标,并通过公式生成目标热图作为监督对象,利用目标热图和预测热图之间应用均方误差损失进行训练,使得所述模型输出的预测热图逼近目标热图。Using non-batch normalization and group whitening methods, and adding squeezing and extraction modules, a free simple baseline wildlife posture estimation model based on heat map generation is constructed to adaptively learn feature weights; the model follows a single architecture-encoder-decoder design. For animal input images containing key points, feature extraction is performed through the backbone network to obtain feature maps. The deconvolution module in the decoder obtains the predicted heat map by upsampling the feature map. Finally, the maximum value of the heat map is obtained as the final key point coordinates through the Argmax function, and the formula is used to calculate the maximum value of the heat map. A target heat map is generated as a supervision object, and training is performed by applying a mean square error loss between the target heat map and the predicted heat map, so that the predicted heat map output by the model is close to the target heat map.
优选地,所述非批次归一化和组白化方法具体包括以下内容:Preferably, the non-batch normalization and group whitening method specifically includes the following contents:
针对所构建模型网络内部估计偏移累积的问题,将白化操作引入到组归一化方法中,提出组白化方法(即非批次维度归一化方法—组归一化方法),将神经元分成组,对每个样本的每组神经元独立应用标准化操作,以避免估计偏移,缓解批量归一化层在小批次下误差迅速变大的缺陷。To address the problem of estimation offset accumulation within the constructed model network, a whitening operation is introduced into the group normalization method, and a group whitening method (i.e., non-batch dimension normalization method—group normalization method) is proposed. Neurons are divided into groups, and normalization operations are applied independently to each group of neurons for each sample to avoid estimation offset and alleviate the defect that the error of the batch normalization layer increases rapidly under small batches.
优选地,所述组归一化方法具体包括如下内容:Preferably, the group normalization method specifically includes the following contents:
将白化操作引入组归一化方法中,采用组白化将样本的神经元分,以便对每组中的神经元进行标准化,然后对所得的组进行去相关,改善自由简单基线姿态估计模型。The whitening operation is introduced into the group normalization method. The neurons of the sample are divided into groups by group whitening so as to standardize the neurons in each group. Then the obtained groups are decorrelated to improve the free simple baseline pose estimation model.
优选地,在特征提取的过程中,不同通道对输出的结果贡献不同,为获得更好的特征表示,引入挤压和提取模块,通过全连接网络根据损失来自适应学习特征权重,所述挤压和提取模块通过全局平均池化操作将特征图通道的特征值压缩为一个实数,压缩后的特征图按照公式e=Fex(fc)=σ·G给每个通道生成不同的权重,利用逐通道乘以权重系数,对特征图的通道维度进行缩放,以此实现对不同特征通道进行自适应地调整。Preferably, in the process of feature extraction, different channels contribute differently to the output results. In order to obtain better feature representation, a squeeze and extract module is introduced. The feature weights are adaptively learned according to the loss through a fully connected network. The squeeze and extract module compresses the feature value of the feature map channel into a real number through a global average pooling operation. The compressed feature map generates different weights for each channel according to the formula e=F ex (f c )=σ·G. The channel dimension of the feature map is scaled by multiplying the weight coefficient channel by channel, so as to achieve adaptive adjustment of different feature channels.
优选地,所述S3包括如下内容:Preferably, S3 includes the following contents:
将高斯热图的高斯内核以亚像素位置为中心进行坐标编码,遵循泰勒定理,通过在预测热图的最大激活处评估的泰勒二阶展开公式来近似激活,以亚像素精度实现联合预测;采用与训练数据具有相似结构的高斯核进行卷积操作,平滑预测热图中的多个峰值,以得到一个峰值;通过取输入特征矩阵的最大值和最小值对平滑后的热图进行调整,以保证处理后的原始热图大小;The Gaussian kernel of the Gaussian heat map is coordinate-encoded with the sub-pixel position as the center. Following Taylor's theorem, the activation is approximated by the Taylor second-order expansion formula evaluated at the maximum activation of the predicted heat map to achieve joint prediction with sub-pixel accuracy. A convolution operation is performed using a Gaussian kernel with a similar structure to the training data to smooth multiple peaks in the predicted heat map to obtain a single peak. The smoothed heat map is adjusted by taking the maximum and minimum values of the input feature matrix to ensure the size of the original heat map after processing.
更具体地包括如下内容:More specifically, it includes the following:
S3.1、基于热图的姿态估计方法是以高斯热图作为监督信息,人工标注的关键点坐标需要经过网络生成高斯热图完成这一目标;按照公式以量化后的坐标为中心生成最终的高斯热图。量化后的坐标存在偏差,导致训练性能下降,本设计允许高斯核以亚像素位置为中心,将其放在非量化位置上,选择亚像素作为中心进行坐标编码,避免量化误差;S3.1. The heat map-based posture estimation method uses Gaussian heat map as supervision information. The coordinates of the manually annotated key points need to be generated by the network to achieve this goal. According to the formula The final Gaussian heat map is generated with the quantized coordinates as the center. The quantized coordinates have deviations, which leads to a decrease in training performance. This design allows the Gaussian kernel to be centered on the sub-pixel position, put it on the non-quantized position, and select the sub-pixel as the center for coordinate encoding to avoid quantization errors;
S3.2、关键点坐标在获取的过程中因存在量化误差,所以只能对应粗略的位置坐标,现引入一种原则性的分布感知坐标解码方法,以亚像素精度实现更准确的联合定位;S3.2, the key point coordinates can only correspond to rough position coordinates due to quantization errors in the acquisition process. Now we introduce a principled distribution-aware coordinate decoding method to achieve more accurate joint positioning with sub-pixel accuracy;
为得到准确的亚像素级别的关键点坐标,使用公式表示待预测关键点坐标相对应的高斯中心的预测热图;其中协方差是一个对角矩阵;借助对数似然优化原则,利用公式对预测热图进行对数变换来促进推理,并保持最大激活的原始位置不变;表明高斯热图最大响应值点μ是函数的极值,该处的一阶导数为0;公式遵循泰勒定理,通过在预测热图的最大激活处评估的泰勒二阶展开公式来近似激活:最后利用公式μ=m-(S″(m))-1S′(m)选择m来近似μ,其代表一个接近最大激活数值的良好的粗略联合预测;To obtain accurate sub-pixel keypoint coordinates, use the formula Represents the predicted heat map of the Gaussian center corresponding to the coordinates of the key points to be predicted; the covariance is a diagonal matrix; with the help of the log-likelihood optimization principle, using the formula Logarithmic transformation of prediction heatmaps is performed to facilitate inference, while keeping the original location of maximum activation unchanged; It shows that the maximum response value point μ of the Gaussian heat map is the extreme value of the function, and the first-order derivative there is 0; the formula follows Taylor's theorem and approximates the activation by evaluating Taylor's second-order expansion formula at the maximum activation of the predicted heat map: Finally, we select m to approximate μ using the formula μ = m-(S″(m)) -1 S′(m), which represents a good rough joint prediction close to the maximum activation value;
S3.3、在动物姿态估计模型中,基于泰勒展开的坐标解码方法不具备良好的高斯结构,预测热图中通常呈现处多个峰值;现采用与训练数据具有相似结构的高斯核K按照公式进行卷积操作来平滑预测热图中的多个峰值。S3.3. In the animal posture estimation model, the coordinate decoding method based on Taylor expansion does not have a good Gaussian structure, and the predicted heat map usually presents multiple peaks. Now we use a Gaussian kernel K with a similar structure to the training data according to the formula A convolution operation is performed to smooth out multiple peaks in the prediction heatmap.
优选地,所述S4具体包括如下内容:Preferably, the S4 specifically includes the following contents:
S4.1、在编码结构部分采用轻量化策略,轻量化架构将反卷积拆分为逐点卷积和深度反卷积;反卷积模块用于重构高分辨率特征图,将反卷积模块的反卷积层拆分为深度反卷积和逐点卷积;S4.1. A lightweight strategy is adopted in the coding structure. The lightweight architecture splits the deconvolution into point-by-point convolution and depth-wise deconvolution. The deconvolution module is used to reconstruct the high-resolution feature map, and the deconvolution layer of the deconvolution module is split into depth-wise deconvolution and point-by-point convolution.
S4.2、轻量化模型先使用深度卷积对输入张量进行特征提取,在使用逐点卷积将特征映射压缩到所需的通道数;S4.2, the lightweight model first uses deep convolution to extract features from the input tensor, and then uses point-by-point convolution to compress the feature map to the required number of channels;
S4.3、为弥补轻量化反卷积操作可能带来的性能下降,引入三分支融合注意力机制,搭建通道注意力、空间注意力和恒等映射三个支路的结构。S4.3. In order to compensate for the performance degradation that may be caused by lightweight deconvolution operations, a three-branch fusion attention mechanism is introduced to build a structure with three branches: channel attention, spatial attention, and identity mapping.
与现有技术相比,本发明提供了一种真实场景下野生动物姿态估计方法,具备以下有益效果:Compared with the prior art, the present invention provides a method for estimating the posture of wild animals in real scenes, which has the following beneficial effects:
本发明基于野生动物姿态估计方法为主体,可以实现真实场景下对野生动物姿态的估计,并可以部署在边缘设备上对野生动物进行实时监测分析,进一步强化了野生动物保护的智能化发展,对生态文明建设、物种多样性保护等方面具有重要意义。The present invention is based on a wild animal posture estimation method, which can realize the estimation of wild animal posture in real scenarios, and can be deployed on edge devices to conduct real-time monitoring and analysis of wild animals, further strengthening the intelligent development of wildlife protection, and is of great significance to the construction of ecological civilization and species diversity protection.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1为本发明提出的一种真实场景下野生动物姿态估计方法的流程示意图;FIG1 is a schematic flow chart of a method for estimating the posture of wild animals in a real scene proposed by the present invention;
图2为本发明实施例3中提到的独立同分布条件下效果对比图;FIG2 is a comparison diagram of effects under independent and identically distributed conditions mentioned in Example 3 of the present invention;
图3为本发明实施例3中提到的Animal-Pose数据集上性能表现(ResNet101)示意图;FIG3 is a schematic diagram of the performance (ResNet101) on the Animal-Pose dataset mentioned in Example 3 of the present invention;
图4为本发明实施例3中提到的Animal-Pose数据集上性能表现(ResNet50)示意图;FIG4 is a schematic diagram of the performance (ResNet50) on the Animal-Pose dataset mentioned in Example 3 of the present invention;
图5为本发明实施例3中提到的小批次下两种数据集的表现示意图。FIG5 is a schematic diagram showing the performance of two data sets under small batches mentioned in Example 3 of the present invention.
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。The technical solutions in the embodiments of the present invention will be described clearly and completely below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are only part of the embodiments of the present invention, rather than all the embodiments.
本发明提供真实场景下野生动物姿态估计方法,图1为本发明的野生动物姿态估计方法的流程图。如图1所示,其公开了包括:构建真实场景下野生动物姿态估计数据集,并基于风格迁移对数据集进行处理;采用组白化操作,构建自由简单基线方法;通过无偏亚像素中心编码、泰勒展开坐标解码、热图分布调制,得到基于分布感知坐标表征方法,构建出基于修正量化误差的野生动物姿态估计方法;构建轻量化野生动物姿态估计方法模型。The present invention provides a method for estimating the posture of wild animals in real scenes, and FIG1 is a flow chart of the method for estimating the posture of wild animals in the present invention. As shown in FIG1, it discloses the following steps: constructing a data set for estimating the posture of wild animals in real scenes, and processing the data set based on style migration; using a group whitening operation to construct a free simple baseline method; obtaining a distribution-aware coordinate representation method through unbiased sub-pixel center encoding, Taylor expansion coordinate decoding, and heat map distribution modulation, and constructing a method for estimating the posture of wild animals based on correcting quantization errors; and constructing a lightweight model for the method for estimating the posture of wild animals.
本发明基于野生动物姿态估计方法模型,利用自由简单基线方法提升模型识别的泛化能力,又考虑到该方法模型处理流程中引入的量化误差问题,设计亚像素编码方式和基于泰勒二阶展开式的解码方式,以减小误差。最后,通过构建轻量化的模型结构,增强边缘部署能力。最终可以实现在复杂环境下对于野生动物姿态的准确估计,以加强野生动物保护工作的有效进行。The present invention is based on a model of a method for estimating the posture of wild animals. It uses a free simple baseline method to improve the generalization ability of model recognition. It also takes into account the quantization error problem introduced in the model processing flow of this method, and designs a sub-pixel encoding method and a decoding method based on Taylor's second-order expansion to reduce the error. Finally, by constructing a lightweight model structure, the edge deployment capability is enhanced. Ultimately, accurate estimation of the posture of wild animals in complex environments can be achieved to strengthen the effective implementation of wildlife protection work.
为了更好的理解上述技术方案,下面将参照附图更详细地描述本发明的示例性实施例。虽然附图中显示了本发明的示例性实施例,然而应当理解,可以以各种形式实现本发明而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更清楚、透彻地理解本发明,并且能够将本发明的范围完整的传达给本领域的技术人员。In order to better understand the above technical solution, exemplary embodiments of the present invention will be described in more detail below with reference to the accompanying drawings. Although exemplary embodiments of the present invention are shown in the accompanying drawings, it should be understood that the present invention can be implemented in various forms and should not be limited by the embodiments described herein. On the contrary, these embodiments are provided to enable a clearer and more thorough understanding of the present invention and to fully convey the scope of the present invention to those skilled in the art.
实施例1:Embodiment 1:
请参阅图1,本发明提出一种真实场景下野生动物姿态估计方法,包括:Referring to FIG. 1 , the present invention proposes a method for estimating the posture of wild animals in a real scene, comprising:
步骤1:step 1:
构建动物姿态估计数据集,对公共数据集进行处理,增添真实场景下常见的野外环境因素和风格迁移因素,并引入数据集的测试部分。Construct an animal pose estimation dataset, process the public dataset, add common wild environment factors and style transfer factors in real scenes, and introduce the test part of the dataset.
步骤2:Step 2:
提出自由简单基线姿态估计模型,该模型遵循简单的架构-编码器解码器设计,对于包含关键点的动物输入图像,通过骨干网络进行特征提取得到特征图,解码器中的反卷积模块通过对特征图上采样获取预测热图,最后通过Argmax函数获得热图最大值作为最终的关键点坐标。A free simple baseline pose estimation model is proposed. This model follows a simple architecture - encoder-decoder design. For animal input images containing key points, feature extraction is performed through the backbone network to obtain feature maps. The deconvolution module in the decoder obtains the predicted heat map by upsampling the feature map. Finally, the maximum value of the heat map is obtained as the final key point coordinates through the Argmax function.
自由简单基线姿态估计模型以标注关键点坐标为核心,按照公式生成的目标热图作为监督对象,然后在预测热图和目标热图之间应用均方误差损失来进行训练,使得模型训练输出的预测热图逼近目标热图。The free simple baseline pose estimation model is centered on the coordinates of the key points. According to the formula The generated target heatmap is used as the supervision object, and then the mean square error loss is applied between the predicted heatmap and the target heatmap for training, so that the predicted heatmap output by the model training is close to the target heatmap.
针对所构建模型网络内部估计偏移累积的问题,将白化操作引入到组归一化方法中,提出组白化方法,把样本的神经元分成组,以便对每组中的神经元进行标准化,然后对这些组进行去相关,以改善自由简单基线姿态估计模型。To solve the problem of estimation offset accumulation within the constructed model network, a whitening operation is introduced into the group normalization method. A group whitening method is proposed to divide the neurons of the sample into groups so as to standardize the neurons in each group, and then decorrelate these groups to improve the free simple baseline pose estimation model.
在特征提取的过程中,不同通道对输出的结果贡献不同,为获得更好的特征表示,引入挤压和提取模块,通过全连接网络根据损失来自适应学习特征权重。In the process of feature extraction, different channels contribute differently to the output results. In order to obtain better feature representation, a squeezing and extraction module is introduced, and the feature weights are adaptively learned according to the loss through a fully connected network.
全局平均池化操作将特征图通道的特征值压缩为一个实数,压缩后的特征图按照公式e=Fex(fc)=σ·G给每个通道生成不同的权重,利用逐通道乘以权重系数,对特征图的通道维度进行缩放,以此实现对不同特征通道进行自适应地调整。The global average pooling operation compresses the eigenvalues of the feature map channels into a real number. The compressed feature map generates different weights for each channel according to the formula e = F ex (f c ) = σ · G. The channel dimension of the feature map is scaled by multiplying the weight coefficient channel by channel, so as to achieve adaptive adjustment of different feature channels.
步骤3:Step 3:
基于热图的姿态估计方法是以高斯热图作为监督信息,人工标注的关键点坐标需要经过网络生成高斯热图完成这一目标。按照公式以量化后的坐标为中心生成最终的高斯热图。量化后的坐标存在偏差,导致训练性能下降,本设计允许高斯核以亚像素位置为中心,将其放在非量化位置上,选择亚像素作为中心进行坐标编码,避免量化误差。The heat map-based pose estimation method uses Gaussian heat maps as supervision information. The manually annotated key point coordinates need to be generated by the network to achieve this goal. According to the formula The final Gaussian heat map is generated with the quantized coordinates as the center. The quantized coordinates are biased, resulting in a decrease in training performance. This design allows the Gaussian kernel to be centered on the sub-pixel position, placing it on the non-quantized position, and selecting the sub-pixel as the center for coordinate encoding to avoid quantization errors.
关键点坐标在获取的过程中因存在量化误差,所以只能对应粗略的位置坐标,现引入一种原则性的分布感知坐标解码方法,以亚像素精度实现更准确的联合定位。Since the key point coordinates can only correspond to rough position coordinates due to quantization errors in the acquisition process, a principled distribution-aware coordinate decoding method is introduced to achieve more accurate joint positioning with sub-pixel accuracy.
为得到准确的亚像素级别的关键点坐标,使用公式表示待预测关键点坐标相对应的高斯中心的预测热图。其中协方差是一个对角矩阵。借助对数似然优化原则,利用公式对预测热图进行对数变换来促进推理,并保持最大激活的原始位置不变。表明高斯热图最大响应值点μ是函数的极值,该处的一阶导数为0。公式遵循泰勒定理,通过在预测热图的最大激活处评估的泰勒二阶展开公式来近似激活:最后利用公式μ=m-(S″(m))-1S′(m)选择m来近似μ,其代表一个接近最大激活数值的良好的粗略联合预测。To obtain accurate sub-pixel keypoint coordinates, use the formula Represents the predicted heat map of the Gaussian center corresponding to the coordinates of the key points to be predicted. is a diagonal matrix. With the help of log-likelihood optimization principle, using the formula The prediction heatmaps are log-transformed to facilitate inference and keep the original location of the maximum activation unchanged. It shows that the point μ of the maximum response value of the Gaussian heat map is the extreme value of the function, and the first-order derivative there is 0. The formula follows Taylor's theorem and approximates the activation by the Taylor second-order expansion formula evaluated at the maximum activation of the predicted heat map: Finally, we select m to approximate μ using the formula μ=m-(S″(m)) - 1S′(m), which represents a good rough joint prediction close to the maximum activation value.
在动物姿态估计模型中,基于泰勒展开的坐标解码方法不具备良好的高斯结构,预测热图中通常呈现处多个峰值。现采用与训练数据具有相似结构的高斯核K按照公式进行卷积操作来平滑预测热图中的多个峰值。In the animal posture estimation model, the coordinate decoding method based on Taylor expansion does not have a good Gaussian structure, and the predicted heat map usually shows multiple peaks. Now we use a Gaussian kernel K with a similar structure to the training data according to the formula A convolution operation is performed to smooth out multiple peaks in the prediction heatmap.
步骤4:Step 4:
在编码结构部分采用轻量化策略,轻量化架构将反卷积拆分为逐点卷积和深度反卷积。反卷积模块用于重构高分辨率特征图,将反卷积模块的反卷积层拆分为深度反卷积和逐点卷积。A lightweight strategy is adopted in the encoding structure. The lightweight architecture splits the deconvolution into point-by-point convolution and depth-wise deconvolution. The deconvolution module is used to reconstruct the high-resolution feature map, and the deconvolution layer of the deconvolution module is split into depth-wise deconvolution and point-by-point convolution.
轻量化模型先使用深度卷积对输入张量进行特征提取,在使用逐点卷积将特征映射压缩到所需的通道数。The lightweight model first uses deep convolution to extract features from the input tensor, and then uses point-by-point convolution to compress the feature map to the required number of channels.
为弥补轻量化反卷积操作可能带来的性能下降,引入三分支融合注意力机制,搭建通道注意力、空间注意力和恒等映射三个支路的结构。In order to compensate for the performance degradation that may be caused by lightweight deconvolution operations, a three-branch fusion attention mechanism is introduced to build a structure with three branches: channel attention, spatial attention and identity mapping.
实施例2:Embodiment 2:
基于实施例1但有所不同之处在于,Based on Example 1, but different in that,
本实施例提出一种真实场景下野生动物姿态估计方法,主要包括:This embodiment proposes a method for estimating the posture of wild animals in a real scene, which mainly includes:
基于公共数据集构建符合本设计的野生动物姿态估计训练集,并对训练集进行处理。A wildlife posture estimation training set that meets the design is constructed based on a public dataset and then processed.
该训练集基于AP-10K数据集和Animal-Pose数据集,在公共数据集的基础上,将真实世界中常见的野外环境因素和风格迁移因素引入数据集的测试部分,以此构建出真实场景,使得模型面对的训练集和测试集处于非独立同分布状态。This training set is based on the AP-10K dataset and the Animal-Pose dataset. On the basis of the public dataset, common field environment factors and style transfer factors in the real world are introduced into the test part of the dataset to construct a real scene, so that the training set and test set faced by the model are in a non-independent and identically distributed state.
其中,采用基于自适应实例归一化的实时风格迁移方法,选取一张经典的野外图像作为风格图像,数据集中的图像作为内容图像,使用VGG网络对内容图像和风格图像进行编码。通过自适应实例归一化层对数据集中的图片进行风格迁移操作,最后经过解码器将特征空间的输出转换到图像空间。Among them, a real-time style transfer method based on adaptive instance normalization is adopted. A classic wild image is selected as the style image, and the image in the dataset is used as the content image. The content image and style image are encoded using the VGG network. The style transfer operation is performed on the images in the dataset through the adaptive instance normalization layer, and finally the output of the feature space is converted to the image space through the decoder.
采用非批次归一化、白化操作、挤压和提取模块,融入到自由简单基线姿态估计模型中,得到基于编码结构的热图生成姿态估计方法,通过上述训练集对模型进行训练,验证真实场景下模型的性能。Non-batch normalization, whitening operation, squeezing and extraction modules are integrated into the free simple baseline pose estimation model to obtain a pose estimation method based on a heat map generation structure based on the encoding structure. The model is trained using the above training set to verify the performance of the model in real scenarios.
其中非批次归一化将神经元分成组,对每个样本的每组神经元独立地应用标准化操作。组白化操作将样本的神经元分成组,以便对每组中的神经元进行标准化,然后对这些组进行去相关。挤压和提取模块通过全连接网络根据损失来自适应学习特征权重。Among them, non-batch normalization divides neurons into groups, and the normalization operation is applied independently to each group of neurons for each sample. The group whitening operation divides the neurons of the sample into groups so that the neurons in each group are normalized and then the groups are decorrelated. The squeezing and extraction module adaptively learns feature weights according to the loss through a fully connected network.
采用无偏亚像素中心编码、泰勒展开坐标编码、热图分布调制方法,得到基于分布感知坐标表征方法。Unbiased sub-pixel center coding, Taylor expansion coordinate coding and heat map distribution modulation method are used to obtain a distribution-aware coordinate representation method.
其中无偏亚像素中心编码允许高斯核以亚像素位置为中心,将其放置在非量化位置上,选择亚像素作为中心进行坐标编码,避免热图编码过程中的量化误差问题。基于泰勒展开的坐标解码,通过一种原则性的分布感知坐标解码方法,以亚像素精度实现更准确的联合定位。热图分布调制采用与训练数据具有相似结构的高斯核K按照公式进行卷积操作来平滑预测热图中的多个峰值。通过取输入特征矩阵的最大值和最小值对平滑后的热图进行调整,使得平滑后的高斯热图的最大激活值与预测热图的最大激活值相对应。Unbiased sub-pixel center coding allows the Gaussian kernel to be centered at the sub-pixel position and placed at an unquantized position. The sub-pixel is selected as the center for coordinate coding to avoid the quantization error problem in the heat map coding process. Coordinate decoding based on Taylor expansion achieves more accurate joint positioning with sub-pixel accuracy through a principled distribution-aware coordinate decoding method. Heat map distribution modulation uses a Gaussian kernel K with a similar structure to the training data according to the formula A convolution operation is performed to smooth the multiple peaks in the predicted heat map. The smoothed heat map is adjusted by taking the maximum and minimum values of the input feature matrix so that the maximum activation value of the smoothed Gaussian heat map corresponds to the maximum activation value of the predicted heat map.
采用轻量化的反卷积头模块和三分支融合注意力机制,得到了基于轻量化反卷积设计的野生动物姿态估计方法模型。By adopting a lightweight deconvolution head module and a three-branch fusion attention mechanism, a wildlife posture estimation method model based on lightweight deconvolution design was obtained.
其中反卷积头模块用于重建高分辨率特征图,在反卷积网络中,其在训练期间可以从输入数据中学习上采样的参数,并将上采样和卷积参数结合到一步。深度反卷积在每个通道上应用单个反卷积核进行上采样获得高分辨率的特征表示,逐点卷积层通过1×1的内核,在高分辨率特征表示的所有通道上进行卷积,为跨通道的信息融合提供了桥梁。三分支融合注意力机制由通道注意力、空间注意力和恒等映射构成。The deconvolution head module is used to reconstruct high-resolution feature maps. In the deconvolution network, it can learn upsampling parameters from the input data during training and combine upsampling and convolution parameters into one step. Deep deconvolution applies a single deconvolution kernel on each channel to upsample and obtain high-resolution feature representation. The point-by-point convolution layer convolves on all channels of the high-resolution feature representation through a 1×1 kernel, providing a bridge for cross-channel information fusion. The three-branch fusion attention mechanism consists of channel attention, spatial attention, and identity mapping.
实施例3:Embodiment 3:
基于实施例1-2但有所不同之处在于,下面结合具体实例,并设计对比实验来对本发明所提出的一种真实场景下野生动物姿态估计方法进行说明,具体包括如下内容。Based on Example 1-2, but with the difference that, the method for estimating the posture of wild animals in real scenarios proposed by the present invention is described below in combination with specific examples and designed comparative experiments, which specifically include the following contents.
一、自由简单基线姿态估计方法模型1. Free Simple Baseline Pose Estimation Method Model
本发明构建了一种考虑真实世界场景偏移的野生动物姿态估计模型。该模型引入非批次维度归一化策略,阻断了模型本身存在的估计偏移累积现象,解决了连续的批次归一化层堆叠对复杂野外场景造成的模型性能下降问题。同时加入白化操作进一步增强了模型在真实场景下的泛化能力。通过在AP-10K和Animal-Pose数据集上验证,相较于SimpleBaseline方法,模型在天气影响、运动模糊和设备噪声等真实野外场景下获得了3.82%~8.18%性能的提高,同时在面对不同的迁移场景时准确率提升了2.07%~7.39%。具体内容如下列表1-4及图2-5所示。The present invention constructs a wild animal posture estimation model that takes into account the real-world scene offset. The model introduces a non-batch dimensional normalization strategy, which blocks the estimation offset accumulation phenomenon in the model itself and solves the problem of model performance degradation caused by continuous batch normalization layer stacking for complex field scenes. At the same time, the addition of whitening operation further enhances the generalization ability of the model in real scenes. Verified on the AP-10K and Animal-Pose datasets, compared with the SimpleBaseline method, the model has achieved a 3.82% to 8.18% performance improvement in real field scenes such as weather effects, motion blur, and equipment noise, and the accuracy rate has increased by 2.07% to 7.39% when facing different migration scenarios. The specific contents are shown in the following Tables 1-4 and Figures 2-5.
(1)非批次维度归一化改进效果(1) Improvement effect of non-batch dimension normalization
1.非批次归一化方法在网络中的位置(如表1所示)1. The location of non-batch normalization methods in the network (as shown in Table 1)
表1不同替代位置的实验效果Table 1 Experimental results of different replacement positions
2.不同非批次归一化方法的表现(如表2所示)2. Performance of different non-batch normalization methods (as shown in Table 2)
表2不同的归一化方法实验效果Table 2 Experimental results of different normalization methods
(2)真实场景下模型实验结果(2) Model experimental results in real scenarios
1.独立同分布下模型实验结果(如图2所示)1. Experimental results of the model under independent and identical distribution (as shown in Figure 2)
2.真实场景下模型实验结果(如表3-4及图3-4所示)2. Model experimental results in real scenarios (as shown in Table 3-4 and Figure 3-4)
表3真实场景下模型效果对比Table 3 Comparison of model effects in real scenarios
表4风格迁移场景下模型效果对比Table 4 Comparison of model effects in style transfer scenarios
(3)小批次下模型性能实验结果(如图5所示)。(3) Experimental results of model performance under small batches (as shown in Figure 5).
二、分布感知坐标表征的姿态估计方法2. Posture Estimation Method Based on Distribution-aware Coordinate Representation
本发明提出了修正姿态估计模型固有量化误差的方法。本发明通过亚像素编码生成无偏热图进行训练监督,同时基于泰勒展开公式和热图分布调制方法进行坐标解码来修正图像分辨率下降带来的关键点坐标误差问题。此外,提出数据增强的无偏处理方法来重新定义坐标轴,通过以采样单位距离作为坐标轴刻度的方法消除坐标变换前后结果不一致的问题。针对两种不同的量化误差提出的改进方案,在没有增加网络复杂度的基础上在AP-10K数据集上准确率指标分别提升了1.31%和2.34%。具体内容如下列表5-6所示。The present invention proposes a method for correcting the inherent quantization error of the posture estimation model. The present invention generates an unbiased heat map through sub-pixel coding for training supervision, and corrects the key point coordinate error problem caused by the reduction of image resolution by coordinate decoding based on the Taylor expansion formula and the heat map distribution modulation method. In addition, an unbiased processing method for data enhancement is proposed to redefine the coordinate axis, and the problem of inconsistent results before and after coordinate transformation is eliminated by using the sampling unit distance as the coordinate axis scale. The improvement schemes proposed for two different quantization errors have improved the accuracy indicators by 1.31% and 2.34% on the AP-10K dataset respectively without increasing the complexity of the network. The specific contents are shown in Tables 5-6 below.
(1)关键点坐标表征算法实验对比(如表5-6所示)(1) Experimental comparison of key point coordinate representation algorithms (as shown in Table 5-6)
表5关键点坐标表征算法实验结果对比Table 5 Comparison of experimental results of key point coordinate representation algorithm
表6不同输入图像分辨率实验对比Table 6 Experimental comparison of different input image resolutions
三、轻量化的姿态估计模型3. Lightweight posture estimation model
本发明设计了动物姿态估计模型的轻量化改进策略。为了便于模型的部署推广,对模型做了轻量化改进。本发明借鉴深度可分离卷积概念,设计了轻量化的深度可分离反卷积模块。同时为了补偿轻量化带来的性能下降,提出了通道、空间和特征恒等映射的三分支融合注意力机制。最终,模型的参数量和计算量在分别减少了18%和12%的基础上性能仅仅下降了0.1%。具体内容如下列表7-8所示。The present invention designs a lightweight improvement strategy for the animal posture estimation model. In order to facilitate the deployment and promotion of the model, the model is lightweighted. Drawing on the concept of deep separable convolution, the present invention designs a lightweight deep separable deconvolution module. At the same time, in order to compensate for the performance degradation caused by lightweighting, a three-branch fusion attention mechanism of channel, space and feature identity mapping is proposed. In the end, the performance of the model only decreased by 0.1% on the basis of reducing the number of parameters and the amount of calculation by 18% and 12% respectively. The specific contents are shown in the following tables 7-8.
(1)轻量化姿态估计模型性能分析(如表7所示)(1) Performance analysis of lightweight posture estimation model (as shown in Table 7)
表7轻量化编解码结构实验对比Table 7 Experimental comparison of lightweight codec structure
(2)不同注意力机制对比实验结果(如表8所示)(2) Comparative experimental results of different attention mechanisms (as shown in Table 8)
表8不同注意力机制实验对比Table 8 Experimental comparison of different attention mechanisms
本领域内的技术人员应明白,本发明的实施例可提供为方法、系统或计算机程序产品。因此,本发明可采用完全硬件实施例、完全软件实施例,或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。It will be appreciated by those skilled in the art that embodiments of the present invention may be provided as methods, systems or computer program products. Therefore, the present invention may take the form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Furthermore, the present invention may take the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program code.
以上,仅为本发明较佳的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,根据本发明的技术方案及其发明构思加以等同替换或改变,都应涵盖在本发明的保护范围之内。The above are only preferred specific implementation modes of the present invention, but the protection scope of the present invention is not limited thereto. Any technician familiar with the technical field can make equivalent replacements or changes according to the technical solutions and inventive concepts of the present invention within the technical scope disclosed by the present invention, which should be covered by the protection scope of the present invention.
Claims (3)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202410101315.9A CN117912116B (en) | 2024-01-24 | 2024-01-24 | Wild animal posture estimation method in real scene |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202410101315.9A CN117912116B (en) | 2024-01-24 | 2024-01-24 | Wild animal posture estimation method in real scene |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN117912116A CN117912116A (en) | 2024-04-19 |
| CN117912116B true CN117912116B (en) | 2024-07-02 |
Family
ID=90685226
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202410101315.9A Active CN117912116B (en) | 2024-01-24 | 2024-01-24 | Wild animal posture estimation method in real scene |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN117912116B (en) |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106778479A (en) * | 2016-11-18 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of action identification method based on section body object |
| CN109891347A (en) * | 2017-02-23 | 2019-06-14 | 深圳市大疆创新科技有限公司 | For simulating the method and system of loose impediment state |
Family Cites Families (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103177468A (en) * | 2013-03-29 | 2013-06-26 | 渤海大学 | Three-dimensional motion object augmented reality registration method based on no marks |
| CN111225558B (en) * | 2017-08-07 | 2023-08-11 | 杰克逊实验室 | Long-term and continuous animal behavior monitoring |
| CN108399617B (en) * | 2018-02-14 | 2020-08-14 | 中国农业大学 | Method and device for detecting animal health condition |
| US11308314B2 (en) * | 2019-03-21 | 2022-04-19 | Toyota Research Institute, Inc. | Systems, devices, and methods for generating a pose estimate of an object |
| CN110689056A (en) * | 2019-09-10 | 2020-01-14 | Oppo广东移动通信有限公司 | A classification method and device, equipment and storage medium |
-
2024
- 2024-01-24 CN CN202410101315.9A patent/CN117912116B/en active Active
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106778479A (en) * | 2016-11-18 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of action identification method based on section body object |
| CN109891347A (en) * | 2017-02-23 | 2019-06-14 | 深圳市大疆创新科技有限公司 | For simulating the method and system of loose impediment state |
Also Published As
| Publication number | Publication date |
|---|---|
| CN117912116A (en) | 2024-04-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN113673307B (en) | A lightweight video action recognition method | |
| CN115841596B (en) | Multi-label image classification method and model training method and device | |
| CN112184577B (en) | Single image defogging method based on multiscale self-attention generation countermeasure network | |
| CN111340738B (en) | An image rain removal method based on multi-scale progressive fusion | |
| CN109671070B (en) | Target detection method based on feature weighting and feature correlation fusion | |
| CN112862792B (en) | Wheat powdery mildew spore segmentation method for small sample image dataset | |
| CN111476219A (en) | Image object detection method in smart home environment | |
| CN105205449B (en) | Sign Language Recognition Method based on deep learning | |
| CN111310666A (en) | High-resolution image ground feature identification and segmentation method based on texture features | |
| CN106529570B (en) | Image Classification Method Based on Deep Ridgelet Neural Network | |
| CN111967537B (en) | A SAR target classification method based on two-way capsule network | |
| CN109840560A (en) | Based on the image classification method for incorporating cluster in capsule network | |
| CN113449691A (en) | Human shape recognition system and method based on non-local attention mechanism | |
| WO2021169160A1 (en) | Image normalization processing method and device, and storage medium | |
| CN108062559A (en) | A kind of image classification method based on multiple receptive field, system and device | |
| CN116519106A (en) | A method, device, storage medium and equipment for measuring the body weight of live pigs | |
| CN119672340A (en) | Semantic detail fusion and context enhancement remote sensing image segmentation method based on DeepLabv3+ | |
| CN111667495A (en) | Image scene analysis method and device | |
| CN119205541A (en) | Weak-light image enhancement method and system based on illumination decomposition | |
| CN117710486A (en) | Camera self-calibration method and system based on face three-dimensional shape estimation | |
| CN117237858A (en) | Loop detection method | |
| CN119131701B (en) | A method, system, device and medium for detecting a building change area | |
| CN117912116B (en) | Wild animal posture estimation method in real scene | |
| CN111027350A (en) | An Improved PCA Algorithm Based on 3D Face Reconstruction | |
| CN119888231A (en) | Lane line semantic segmentation method based on self-attention |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant |