[go: up one dir, main page]

CN111126179A - Information acquisition method and device, storage medium and electronic device - Google Patents

Information acquisition method and device, storage medium and electronic device Download PDF

Info

Publication number
CN111126179A
CN111126179A CN201911239754.1A CN201911239754A CN111126179A CN 111126179 A CN111126179 A CN 111126179A CN 201911239754 A CN201911239754 A CN 201911239754A CN 111126179 A CN111126179 A CN 111126179A
Authority
CN
China
Prior art keywords
clothing
information
video
training
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911239754.1A
Other languages
Chinese (zh)
Inventor
李冠楠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing QIYI Century Science and Technology Co Ltd
Original Assignee
Beijing QIYI Century Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing QIYI Century Science and Technology Co Ltd filed Critical Beijing QIYI Century Science and Technology Co Ltd
Priority to CN201911239754.1A priority Critical patent/CN111126179A/en
Publication of CN111126179A publication Critical patent/CN111126179A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

本申请提供了一种信息的获取方法和装置、存储介质和电子装置,其中,该方法包括:从待检测视频中提取关键帧;检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。

Figure 201911239754

The present application provides an information acquisition method and device, a storage medium and an electronic device, wherein the method includes: extracting key frames from a video to be detected; detecting a first costume of a first costume object in a frame image of the detected key frame region; extract the clothing features of the first clothing object from the first clothing region, wherein the clothing features of the first clothing object include at least one of the following: color information of the first clothing object, posture information of the first clothing object; In the case where a second clothing region in which the clothing features of the included clothing objects and the clothing features of the first clothing object are determined in the multiple reference clothing regions, acquire the target clothing corresponding to the second clothing objects contained in the second clothing region information, wherein each reference clothing area contains at least one clothing object; the region tracking is performed on the first clothing region in the video frame sequence of the video to be detected, and the appearance information of the first clothing object in the video to be detected is determined.

Figure 201911239754

Description

信息的获取方法和装置、存储介质和电子装置Information acquisition method and device, storage medium and electronic device

技术领域technical field

本申请涉及计算机领域,尤其涉及一种信息的获取方法和装置、存储介质和电子装置。The present application relates to the field of computers, and in particular, to a method and device for acquiring information, a storage medium and an electronic device.

背景技术Background technique

目前,在视频节目(例如,综艺节目)中出现服饰时,可以通过添加同款服饰商品的商品信息,为用户提供便捷的服饰商品的购买入口,满足用户购买同款或者同类服饰的需求。At present, when clothing appears in a video program (for example, a variety show), the product information of the same clothing product can be added to provide users with a convenient entrance for purchasing clothing products to meet users' needs for purchasing the same or similar clothing products.

为了得到视频节目中包含的服饰商品的信息,需要识别视频中出现的服饰款式。相关技术中的服饰识别方式,通常利用服饰特征点信息,如袖口、领口、下摆位置信息,进行服装款式的识别。然而,上述服饰识别方式,对于远离特征点区域的服饰特征表达能力有限,在进行训练时容易出现过拟合现象,在实际场景算法性能有限,导致存在服饰识别准确率较低的问题。In order to obtain the information of the clothing products included in the video program, it is necessary to identify the clothing styles appearing in the video. The clothing identification method in the related art usually uses clothing feature point information, such as cuff, neckline, and hem position information, to identify clothing styles. However, the above-mentioned clothing recognition methods have limited ability to express clothing features far away from feature points, and are prone to overfitting during training. The algorithm performance in actual scenarios is limited, resulting in the problem of low clothing recognition accuracy.

发明内容SUMMARY OF THE INVENTION

本申请实施例提供了一种信息的获取方法及装置、存储介质和电子装置,以至少解决相关技术中的服饰识别方式存在由于服饰表达能力弱导致的服饰识别准确率低的问题。The embodiments of the present application provide an information acquisition method and device, a storage medium, and an electronic device, to at least solve the problem of low clothing recognition accuracy due to weak clothing expression ability in clothing identification methods in the related art.

根据本申请实施例的一个方面,提供了一种信息的获取方法,包括:从待检测视频中提取关键帧;检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。According to an aspect of the embodiments of the present application, a method for acquiring information is provided, which includes: extracting key frames from a video to be detected; detecting a first clothing area including a first clothing object in a frame image of the key frame; The clothing features of the first clothing object are extracted from the clothing area, wherein the clothing features of the first clothing object include at least one of the following: color information of the first clothing object, posture information of the first clothing object; In the case where it is determined that the clothing feature of the clothing object contained in the second clothing area matches the clothing feature of the first clothing object, the target clothing information corresponding to the second clothing object contained in the second clothing area is obtained, wherein each Each reference clothing region contains at least one clothing object; the region tracking is performed on the first clothing region in the video frame sequence of the video to be detected, and the appearance information of the first clothing object in the video to be detected is determined.

根据本申请实施例的另一个方面,提供了一种信息的获取装置,包括:第一提取单元,用于从待检测视频中提取关键帧;检测单元,用于检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;第二提取单元,用于从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;第一获取单元,用于在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;第一确定单元,用于对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。According to another aspect of the embodiments of the present application, an apparatus for acquiring information is provided, including: a first extraction unit, configured to extract key frames from a video to be detected; the first clothing area of the first clothing object; the second extraction unit is configured to extract the clothing features of the first clothing object from the first clothing area, wherein the clothing features of the first clothing object include at least one of the following: the first clothing The color information of the object, the posture information of the first clothing object; the first acquisition unit is used to determine the second clothing whose clothing features of the included clothing objects match the clothing features of the first clothing object from a plurality of reference clothing regions In the case of an Area tracking is performed in the video frame sequence of the video to be detected, and the appearance information of the first clothing object in the video to be detected is determined.

可选地,第一提取单元包括:第一提取模块,用于按照目标间隔从待检测视频中提取关键帧;或者,第二提取模块,用于从待检测视频所包含的镜头中抽取与镜头对应的关键帧。Optionally, the first extraction unit includes: a first extraction module, used to extract key frames from the video to be detected according to target intervals; or, a second extraction module, used to extract and shot from the shots included in the video to be detected. corresponding keyframes.

可选地,第二提取单元包括:输入模块,用于将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征,其中,第一特征提取模型是使用第一训练样本对第一初始模型进行训练得到的,第一训练样本是标注出包含的第一训练服饰对象的第一服饰特征的图像。Optionally, the second extraction unit includes: an input module for inputting the first clothing region into the first feature extraction model to obtain clothing features of the first clothing object output by the first feature extraction model, wherein the first feature The extraction model is obtained by training the first initial model by using the first training sample, and the first training sample is an image marked with the first clothing feature of the first training clothing object.

可选地,上述装置还包括:第二获取单元,用于在将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征之前,获取第一训练服饰对象的第一服饰特征,其中,第一服饰特征包括以下至少之一:第一训练服饰对象的第一颜色信息和第一训练服饰对象的第一姿态信息;第一训练单元,用于使用第一训练样本对第一初始模型进行训练,得到第一特征提取模型,其中,第一特征提取模型从第一训练样本中提取出的第二服饰特征与第一服饰特征的相似度大于或者等于第一阈值,其中,第二服饰特征包括以下至少之一:第一训练服饰对象的第二颜色信息和第一训练服饰对象的第二姿态信息。Optionally, the above-mentioned device further includes: a second obtaining unit, configured to obtain the first clothing object before inputting the first clothing region into the first feature extraction model to obtain the clothing features of the first clothing object output by the first feature extraction model. A first clothing feature of a training clothing object, wherein the first clothing feature includes at least one of the following: first color information of the first training clothing object and first posture information of the first training clothing object; a first training unit, using Using the first training sample to train the first initial model to obtain a first feature extraction model, wherein the similarity between the second clothing feature and the first clothing feature extracted by the first feature extraction model from the first training sample is greater than or equal to the first threshold, wherein the second clothing feature includes at least one of the following: second color information of the first training clothing object and second posture information of the first training clothing object.

可选地,第二获取单元包括以下至少之一:获取模块,用于对第一训练样本进行直方图计算和聚类计算,获取第一颜色信息;第一确定模块,用于根据标注的第一训练服饰对象的特征点的第一位置信息和特征点的可见性信息,确定出第一姿态信息。Optionally, the second obtaining unit includes at least one of the following: an obtaining module, configured to perform histogram calculation and clustering calculation on the first training sample, and obtain first color information; The first position information of the feature points of the training clothing object and the visibility information of the feature points are determined to determine the first posture information.

可选地,上述装置还包括:第三获取单元,用于在从第一服饰区域中提取第一服饰对象的服饰特征之后,将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征,其中,第二特征提取模型是使用第二训练样本对第二初始模型进行训练得到的,第二训练样本为标注出包含的第二训练服饰对象的服饰特征和第二训练服饰对象的同款标识的图像,第二训练对象的同款标识用于标识同款的第二训练服饰对象,第二特征提取模型提取出的、同款的第二训练服饰对象的目标特征之间的相似度大于或者等于第二阈值,第二特征提取模型提取出的、不同款的第二训练服饰对象的目标特征之间的相似度小于第二阈值;第四获取单元,用于从多个参考服饰区域中获取目标特征与第一服饰对象的目标特征匹配的候选服饰区域;第二确定单元,用于从候选服饰区域中确定出服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域。Optionally, the above-mentioned device further includes: a third acquisition unit, configured to input the clothing features of the first clothing object into the second feature extraction model after extracting the clothing features of the first clothing object from the first clothing area, and obtain the clothing features of the first clothing object. The target feature of the first clothing object output by the second feature extraction model, wherein the second feature extraction model is obtained by using the second training sample to train the second initial model, and the second training sample is marked with the second The clothing feature of the training clothing object and the image of the same item of the second training clothing object, the same item of the second training object is used to identify the second training clothing object of the same style, and the same item extracted by the second feature extraction model The similarity between the target features of the second training clothing objects is greater than or equal to the second threshold, and the similarity between the target features of the second training clothing objects of different styles extracted by the second feature extraction model is less than the second threshold. The 4th acquisition unit is used to obtain from a plurality of reference clothing areas the candidate clothing area that the target feature matches the target feature of the first clothing object; The second determination unit is used to determine the clothing feature from the candidate clothing area and the first A second clothing region to which the clothing features of a clothing object are matched.

可选地,上述装置还包括:输入单元,用于在将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征之前,将第二训练样本输入到第一特征提取模型,得到第一特征提取模型输出的、第二训练服饰对象的服饰特征;第二训练单元,用于使用第二训练服饰对象的服饰特征和第二训练对象的同款信息对第二初始模型进行训练,得到第二特征提取模型。Optionally, the above-mentioned device further comprises: an input unit, used for inputting the clothing feature of the first clothing object into the second feature extraction model, and before obtaining the target feature of the first clothing object output by the second feature extraction model. The second training sample is input into the first feature extraction model to obtain the clothing features of the second training clothing object output by the first feature extraction model; the second training unit is used for using the clothing features of the second training clothing object and the second training The same information of the object is used to train the second initial model to obtain the second feature extraction model.

可选地,上述装置还包括:添加单元,第一确定单元包括第二确定模块,其中,第二确定模块,用于按照第一服饰区域分别对待检测视频中位于关键帧之前的视频帧和位于的关键帧之后的视频帧进行区域检测,确定第一服饰对象在待检测视频中出现的时间段信息和位置信息,其中,出现信息包括时间段信息和位置信息;添加单元,用于在对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息之后,在待检测视频中添加控制信息,其中,控制信息用于控制在待检测视频被播放到时间段信息对应的时间段时,通过弹窗的方式在待检测视频中与位置信息对应的位置上显示目标服饰信息。Optionally, the above-mentioned device further includes: an adding unit, and the first determining unit includes a second determining module, wherein the second determining module is used for the video frame before the key frame and the video frame located in the video to be detected according to the first clothing area respectively. The video frame after the key frame is subjected to regional detection, and the time period information and position information of the first clothing object appearing in the video to be detected are determined, wherein the appearance information includes the time period information and the position information; the adding unit is used for A clothing area is tracked in the video frame sequence of the video to be detected, and after determining the appearance information of the first clothing object in the video to be detected, control information is added to the video to be detected, wherein the control information is used to control the video to be detected. When the video is played to the time period corresponding to the time period information, the target clothing information is displayed on the position corresponding to the position information in the video to be detected by means of a pop-up window.

根据本发明的又一个实施例,还提供了一种计算机可读的存储介质,存储介质中存储有计算机程序,其中,计算机程序被设置为运行时执行上述任一项方法实施例中的步骤。According to yet another embodiment of the present invention, a computer-readable storage medium is also provided, and a computer program is stored in the storage medium, wherein the computer program is configured to execute the steps in any one of the above method embodiments when running.

根据本发明的又一个实施例,还提供了一种电子装置,包括存储器和处理器,存储器中存储有计算机程序,处理器被设置为运行计算机程序以执行上述任一项方法实施例中的步骤。According to yet another embodiment of the present invention, an electronic device is also provided, comprising a memory and a processor, a computer program is stored in the memory, and the processor is configured to run the computer program to execute the steps in any of the above method embodiments .

通过本发明,采用结合色彩信息和/或姿态信息进行服饰识别的方式,从待检测视频中提取关键帧;检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息,由于结合服饰对象的色彩信息和/或姿态信息进行服饰识别,可以实现增加服饰特征对服饰对象的表达能力的目的,达到提高服饰识别准确性的技术效果,从而解决相关技术中的服饰识别方式存在由于服饰表达能力弱导致的服饰识别准确率低的问题。Through the present invention, the method of clothing recognition combined with color information and/or posture information is used to extract key frames from the video to be detected; the frame image of the detection key frame includes the first clothing area of the first clothing object; The clothing features of the first clothing object are extracted from the region, wherein the clothing features of the first clothing object include at least one of the following: color information of the first clothing object, posture information of the first clothing object; In the case of determining a second clothing area in which the clothing features of the included clothing objects match those of the first clothing object, obtain target clothing information corresponding to the second clothing objects contained in the second clothing area, wherein each At least one clothing object is included in the reference clothing area; regional tracking is performed on the first clothing area in the video frame sequence of the video to be detected, and the appearance information of the first clothing object in the video to be detected is determined. Clothes recognition by gesture information can achieve the purpose of increasing the expressive ability of clothing features to clothing objects, and achieve the technical effect of improving the accuracy of clothing identification, thereby solving the problem of clothing identification in related technologies due to weak clothing expression ability. Identify problems with low accuracy.

附图说明Description of drawings

此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本发明的实施例,并与说明书一起用于解释本发明的原理。The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description serve to explain the principles of the invention.

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,对于本领域普通技术人员而言,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the following briefly introduces the accompanying drawings that need to be used in the description of the embodiments or the prior art. In other words, on the premise of no creative labor, other drawings can also be obtained from these drawings.

图1是根据本申请实施例的一种可选的服务器的硬件结构框图;1 is a block diagram of a hardware structure of an optional server according to an embodiment of the present application;

图2是根据本申请实施例的一种可选的信息的获取方法的流程示意图;2 is a schematic flowchart of an optional information acquisition method according to an embodiment of the present application;

图3是根据本申请实施例的另一种可选的信息的获取方法的流程示意图;3 is a schematic flowchart of another optional information acquisition method according to an embodiment of the present application;

图4是根据本申请实施例的一种可选的模型训练的流程示意图;4 is a schematic flowchart of an optional model training according to an embodiment of the present application;

图5是根据本申请实施例的一种可选的服饰特征数据库的构建的流程示意图;5 is a schematic flowchart of an optional clothing feature database construction according to an embodiment of the present application;

图6是根据本申请实施例的一种可选的视频中服饰识别的流程示意图;以及,FIG. 6 is a schematic flowchart of an optional clothing identification in a video according to an embodiment of the present application; and,

图7是根据本申请实施例的一种可选的信息的获取装置的结构框图。FIG. 7 is a structural block diagram of an optional information acquisition apparatus according to an embodiment of the present application.

具体实施方式Detailed ways

下文中将参考附图并结合实施例来详细说明本发明。需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互组合。Hereinafter, the present invention will be described in detail with reference to the accompanying drawings and in conjunction with embodiments. It should be noted that the embodiments in the present application and the features of the embodiments may be combined with each other in the case of no conflict.

需要说明的是,本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。It should be noted that the terms "first", "second" and the like in the description and claims of the present invention and the above drawings are used to distinguish similar objects, and are not necessarily used to describe a specific sequence or sequence.

根据本申请实施例的一个方面,提供了一种信息的获取方法。可选地,该方法可以在服务器(视频内容播放平台的服务器)、用户终端或者类似的运算装置中执行。以运行在服务器上为例,图1是根据本申请实施例的一种可选的服务器的硬件结构框图。如图1所示,服务器10可以包括一个或多个(图1中仅示出一个)处理器102(处理器102可以包括但不限于微处理器MCU或可编程逻辑器件FPGA等的处理装置)和用于存储数据的存储器104,可选地,上述服务器还可以包括用于通信功能的传输设备106以及输入输出设备108。本领域普通技术人员可以理解,图1所示的结构仅为示意,其并不对上述服务器的结构造成限定。例如,服务器10还可包括比图1中所示更多或者更少的组件,或者具有与图1所示不同的配置。According to an aspect of the embodiments of the present application, a method for acquiring information is provided. Optionally, the method can be executed in a server (a server of a video content playing platform), a user terminal or a similar computing device. Taking running on a server as an example, FIG. 1 is a hardware structural block diagram of an optional server according to an embodiment of the present application. As shown in FIG. 1 , the server 10 may include one or more (only one is shown in FIG. 1 ) processors 102 (the processors 102 may include but are not limited to processing devices such as a microprocessor MCU or a programmable logic device FPGA) and a memory 104 for storing data, optionally, the above server may further include a transmission device 106 and an input and output device 108 for communication functions. Those skilled in the art can understand that the structure shown in FIG. 1 is only for illustration, and does not limit the structure of the above server. For example, server 10 may also include more or fewer components than shown in FIG. 1 , or have a different configuration than that shown in FIG. 1 .

存储器104可用于存储计算机程序,例如,应用软件的软件程序以及模块,如本申请实施例中的信息的获取方法对应的计算机程序,处理器102通过运行存储在存储器104内的计算机程序,从而执行各种功能应用以及数据处理,即实现上述的方法。存储器104可包括高速随机存储器,还可包括非易失性存储器,如一个或者多个磁性存储装置、闪存、或者其他非易失性固态存储器。在一些实例中,存储器104可进一步包括相对于处理器102远程设置的存储器,这些远程存储器可以通过网络连接至服务器10。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory 104 can be used to store computer programs, for example, software programs and modules of application software, such as computer programs corresponding to the information acquisition methods in the embodiments of the present application. The processor 102 executes the computer programs stored in the memory 104 by running the computer programs. Various functional applications and data processing implement the above method. Memory 104 may include high-speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some instances, memory 104 may further include memory located remotely from processor 102, which may be connected to server 10 through a network. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.

传输装置106用于经由一个网络接收或者发送数据。上述的网络具体实例可包括服务器10的通信供应商提供的无线网络。在一个实例中,传输装置106包括一个NIC(Network Interface Controller,网络适配器),其可通过基站与其他网络设备相连从而可与互联网进行通讯。在一个实例中,传输装置106可以为RF(Radio Frequency,射频)模块,其用于通过无线方式与互联网进行通讯。Transmission means 106 are used to receive or transmit data via a network. The specific example of the above-mentioned network may include a wireless network provided by the communication provider of the server 10 . In one example, the transmission device 106 includes a NIC (Network Interface Controller, network adapter), which can be connected to other network devices through a base station so as to communicate with the Internet. In one example, the transmission device 106 may be an RF (Radio Frequency, radio frequency) module, which is used to communicate with the Internet in a wireless manner.

在本实施例中提供了一种运行于上述服务器的信息的获取方法,图2是根据本申请实施例的一种可选的信息的获取方法的流程图,如图2所示,该流程包括如下步骤:This embodiment provides a method for acquiring information running on the above server. FIG. 2 is a flowchart of an optional method for acquiring information according to an embodiment of the present application. As shown in FIG. 2 , the process includes: Follow the steps below:

步骤S202,从待检测视频中提取关键帧;Step S202, extracting key frames from the video to be detected;

步骤S204,检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;Step S204, the frame image of the detection key frame includes the first clothing area of the first clothing object;

步骤S206,从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;Step S206, extracting the clothing feature of the first clothing object from the first clothing area, wherein the clothing feature of the first clothing object includes at least one of the following: color information of the first clothing object, and posture information of the first clothing object;

步骤S208,在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;Step S208, in the case of determining from a plurality of reference apparel regions a second apparel region in which the apparel features of the included apparel objects match the apparel features of the first apparel object, obtain a second apparel region that matches the second apparel region. target clothing information corresponding to the object, wherein each reference clothing area contains at least one clothing object;

步骤S210,对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。Step S210 , performing regional tracking of the first clothing region in the video frame sequence of the video to be detected, and determining the appearance information of the first clothing object in the video to be detected.

可选地,上述步骤的执行主体可以为服务器、用户终端等,但不限于此。Optionally, the execution subject of the above steps may be a server, a user terminal, etc., but is not limited thereto.

通过本实施例,采用结合色彩信息和/或姿态信息进行服饰识别的方式,由于结合服饰对象的色彩信息和/或姿态信息进行服饰识别,可以实现增加服饰特征对服饰对象的表达能力的目的,解决了相关技术中的服饰识别方式存在由于服饰表达能力弱导致的服饰识别准确率低的问题,提高了服饰识别准确性。Through the present embodiment, the method of performing clothing identification in combination with color information and/or posture information, because the clothing identification is performed in combination with the color information and/or posture information of the clothing object, the purpose of increasing the expressive ability of clothing features to the clothing object can be achieved, The problem of low clothing recognition accuracy caused by weak clothing expression ability in the clothing recognition method in the related art is solved, and the clothing recognition accuracy is improved.

下面结合图2对本申请实施例中的信息的获取方法进行说明。The method for acquiring information in the embodiment of the present application will be described below with reference to FIG. 2 .

在步骤S202中,从待检测视频中提取关键帧。In step S202, key frames are extracted from the video to be detected.

对于视频节目(例如,综艺节目)的节目视频,用户可以通过如客户端、网页等方式进行节目视频的观看。For the program video of a video program (for example, a variety show), the user can watch the program video through a client, a web page, or the like.

为了在该节目视频中显示与该节目视频中的服饰对象对应的服饰信息(例如,服饰商品信息),可以在播放该节目视频之间预先添加与该节目视频中的服饰对象对应的服饰信息,或者,在播放节目视频的过程中,实时确定并获取与该节目视频中包含的服饰对象对应的服饰信息。In order to display the clothing information (for example, clothing commodity information) corresponding to the clothing objects in the program video in the program video, the clothing information corresponding to the clothing objects in the program video can be pre-added before playing the program video, Or, in the process of playing the program video, the clothing information corresponding to the clothing object included in the program video is determined and acquired in real time.

在进行服饰信息获取时,首先提取待检测视频(例如,节目视频)的关键帧,以便对提取的关键帧的帧图像进行处理。提取待检测视频的关键帧的方式可以有多种,可以包括但不限于以下之一:等间隔抽取,按照镜头抽取。When acquiring clothing information, firstly extract the key frames of the video to be detected (eg, program video), so as to process the frame images of the extracted key frames. There are many ways to extract the key frames of the video to be detected, which may include but are not limited to one of the following: extraction at equal intervals, extraction by shot.

作为一种可选的实施例,从待检测视频中提取关键帧包括:按照目标间隔从待检测视频中提取关键帧;或者,从待检测视频所包含的镜头中抽取与镜头对应的关键帧。As an optional embodiment, extracting key frames from the video to be detected includes: extracting key frames from the video to be detected according to target intervals; or, extracting key frames corresponding to shots from shots included in the video to be detected.

作为一种可选的实施方式,可以按照目标间隔从待检测视频中提取出关键帧。上述目标间隔可以是目标时间间隔,例如,每隔2s(可以根据需要进行设置或修改)抽取一个视频帧作为当前关键帧,也可以是预定数量的视频帧间隔,例如,每隔50个视频帧(可以根据需要进行设置或修改)抽取一个视频帧作为当前关键帧。具体等间隔抽取关键帧的方式可以根据需要设定,本实施例中对此不作具体限定。As an optional implementation manner, key frames may be extracted from the video to be detected according to target intervals. The above target interval can be a target time interval, for example, every 2s (can be set or modified as needed) to extract a video frame as the current key frame, or it can be a predetermined number of video frame intervals, for example, every 50 video frames (Can be set or modified as needed) Extract a video frame as the current key frame. The specific manner of extracting key frames at equal intervals may be set as required, which is not specifically limited in this embodiment.

作为另一种可选的实施方式,可以从待检测视频所包含的镜头中抽取与镜头对应的关键帧。待检测视频可以包括多个镜头,其中,同一镜头内的视频帧中,相邻视频帧的相似度可以大于或者等于第一阈值。对于每个镜头,可以抽取一个或多个视频帧作为关键帧。具体确定待检测视频中包含的镜头的方式以及从各个镜头中抽取关键帧的方式可以根据需要设定(例如,根据相邻视频帧的相似度),本实施例中对此不作具体限定。As another optional implementation manner, key frames corresponding to the shots may be extracted from shots included in the video to be detected. The video to be detected may include multiple shots, wherein, in the video frames within the same shot, the similarity of adjacent video frames may be greater than or equal to the first threshold. For each shot, one or more video frames can be extracted as keyframes. The specific way of determining the shots included in the video to be detected and the way of extracting key frames from each shot can be set as required (for example, based on the similarity of adjacent video frames), which is not specifically limited in this embodiment.

通过本实施例,通过按照特定间隔抽取待检测视频的关键帧,或者,按照镜头抽取待检测视频的关键帧,可以保证关键帧抽取的合理性,提高关键帧的抽取效率。Through this embodiment, by extracting the key frames of the video to be detected according to a specific interval, or extracting the key frames of the video to be detected according to the shot, the rationality of the key frame extraction can be ensured, and the key frame extraction efficiency can be improved.

在步骤S204中,检测关键帧的帧图像中包含第一服饰对象的第一服饰区域。In step S204, the frame image of the detected key frame includes the first clothing area of the first clothing object.

对于提取出的关键帧,可以依次对各个关键帧的帧图像进行处理,确定各个关键帧的帧图像中包含第一服饰对象的第一服饰区域。其中,第一服饰区域可以为包含第一服饰对象的四边形区域。检测第一服饰区域的方式可以结合相关技术,例如,可以是基于fashion detect算法的服饰区域检测方式,在此不做具体限定。For the extracted key frames, the frame images of each key frame may be processed in sequence to determine a first clothing area including the first clothing object in the frame images of each key frame. The first clothing area may be a quadrilateral area containing the first clothing object. The method of detecting the first clothing area may be combined with related technologies, for example, it may be a clothing area detection method based on the fashion detect algorithm, which is not specifically limited herein.

例如,可以对待处理视频进行关键帧计算,并对每一关键帧进行服饰检测,得到各个关键帧的服饰区域。For example, key frame calculation can be performed on the video to be processed, and clothing detection can be performed on each key frame to obtain the clothing area of each key frame.

在步骤S206中,从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息。In step S206, the clothing features of the first clothing object are extracted from the first clothing area, wherein the clothing features of the first clothing object include at least one of the following: color information of the first clothing object, posture information of the first clothing object .

对于检测出的第一服饰区域,可以从第一服饰区域中提取出第一服饰对象的服饰特征。服饰特征用于表示服饰对象,服饰特征可以包括一种或多种特征,不同服饰对象的服饰特征可以相同,也可以不同。For the detected first clothing region, clothing features of the first clothing object may be extracted from the first clothing region. The clothing features are used to represent clothing objects, and the clothing features may include one or more features, and the clothing features of different clothing objects may be the same or different.

上述服饰特征可以包括以下至少之一:颜色信息,姿态信息,服饰对象的颜色信息可以是整个服饰对象的色彩信息,用于表示服饰对象的整体颜色(主色调),服饰对象的姿态信息可以用于表示服饰对象的当前姿态。除了颜色信息和/或姿态信息以外,服饰特征还可以包括以下至少之一:服饰类别信息,款式信息和特征点信息,其中,The above-mentioned clothing features can include at least one of the following: color information, posture information, and the color information of the clothing object can be the color information of the entire clothing object, which is used to represent the overall color (main tone) of the clothing object, and the attitude information of the clothing object can be used Used to represent the current pose of the clothing object. In addition to color information and/or posture information, clothing features may also include at least one of the following: clothing category information, style information and feature point information, wherein,

(1)服饰类别信息用于服饰对象的类别,例如,短袖T,长袖帽衫等;(1) The clothing category information is used for the category of clothing objects, for example, short-sleeved T, long-sleeved hoodie, etc.;

(2)款式信息用于表示服饰对象的款式,例如,纹理、材质、形状等更细的特征;(2) The style information is used to represent the style of the clothing object, for example, finer features such as texture, material, and shape;

(3)特征点信息用于表示服饰对象的特征点的位置信息,上述特征点可以包括:袖口、领口、下摆、腰侧各两个特征点。(3) The feature point information is used to represent the position information of the feature points of the clothing object, and the above-mentioned feature points may include: two feature points each of the cuff, the neckline, the hem, and the waist.

对于姿态信息,服饰对象的姿态信息可以用于表示服饰对象的当前姿态,例如,正身(服饰对象正面面向镜头),背身(服饰对象背面面向镜头),侧身(服饰对象侧面面向镜头)等。如果服饰对象的当前姿态为侧面,姿态信息还可以表示服饰对象相对于正身所侧转的角度。例如,服饰对象的姿态信息可以表示:服饰对象侧转30°。For the posture information, the posture information of the clothing object can be used to represent the current posture of the clothing object, for example, body (front of the clothing object faces the camera), back (the back of the clothing object faces the camera), sideways (the side of the clothing object faces the camera) and so on. If the current posture of the clothing object is a side view, the posture information may also indicate the angle at which the clothing object is turned with respect to the body. For example, the posture information of the clothing object may indicate that the clothing object is turned sideways by 30°.

作为一种可选的实施例,从第一服饰区域中提取出第一服饰对象的服饰特征包括:将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征,其中,第一特征提取模型是使用第一训练样本对第一初始模型进行训练得到的,第一训练样本是标注出包含的第一训练服饰对象的第一服饰特征的图像。As an optional embodiment, extracting the clothing feature of the first clothing object from the first clothing region includes: inputting the first clothing region into the first feature extraction model, and obtaining the first Clothing features of the clothing object, wherein the first feature extraction model is obtained by training the first initial model using the first training sample, and the first training sample is an image marked with the first clothing feature of the first training clothing object included .

对于第一服饰区域,可以使用第一特征提取模型提取第一服饰对象的服饰特征。可以将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型的输出结果,该输出结果可以作为第一服饰区域包含的第一服饰对象的服饰特征。For the first clothing region, a first feature extraction model may be used to extract clothing features of the first clothing object. The first clothing region can be input into the first feature extraction model to obtain an output result of the first feature extraction model, and the output result can be used as clothing features of the first clothing object included in the first clothing region.

第一特征提取模型可以是使用第一训练样本对第一初始模型(第一初始特征提取模型)进行训练得到的,第一训练样本是标注出包含的第一训练服饰对象的第一服饰特征的图像。The first feature extraction model may be obtained by using the first training sample to train the first initial model (the first initial feature extraction model), and the first training sample is marked with the first clothing feature of the first training clothing object included. image.

通过本实施例,通过使用训练好的第一特征提取模型进行第一服饰对象的服饰特征的提取,可以提高服饰特征提取的效率,提高服饰特征提取的准确性。Through this embodiment, by using the trained first feature extraction model to extract clothing features of the first clothing object, the efficiency of clothing feature extraction and the accuracy of clothing feature extraction can be improved.

在使用第一特征提取模型之前,可以使用第一训练样本对第一初始模型进行训练,以得到该第一特征提取模型。Before using the first feature extraction model, the first initial model may be trained by using the first training sample to obtain the first feature extraction model.

作为一种可选的实施例,在将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征之前,可以获取第一训练服饰对象的第一服饰特征,其中,第一服饰特征包括以下至少之一:第一训练服饰对象的第一颜色信息和第一训练服饰对象的第一姿态信息;使用第一训练样本对第一初始模型进行训练,得到第一特征提取模型,其中,第一特征提取模型从第一训练样本中提取出的第二服饰特征与第一服饰特征的相似度大于或者等于第一阈值,其中,第二服饰特征包括以下至少之一:第一训练服饰对象的第二颜色信息和第一训练服饰对象的第二姿态信息。As an optional embodiment, before inputting the first clothing region into the first feature extraction model and obtaining the clothing features of the first clothing object output by the first feature extraction model, the first training clothing object's first clothing feature may be obtained. A clothing feature, wherein the first clothing feature includes at least one of the following: first color information of the first training clothing object and first posture information of the first training clothing object; using the first training sample to train the first initial model , obtain a first feature extraction model, wherein the similarity between the second clothing feature extracted from the first training sample and the first clothing feature by the first feature extraction model is greater than or equal to the first threshold, wherein the second clothing feature includes At least one of the following: second color information of the first training clothing object and second posture information of the first training clothing object.

对于第一训练样本,可以获取第一训练样本中包含的第一训练服饰对象的第一服饰特征,该第一服饰特征可以是包括以下至少之一:第一训练服饰对象的第一颜色信息;第一训练服饰对象的第一姿态信息。第一服饰特征还可以包括以下至少之一:第一训练服饰对象的第一服饰类别信息,第一训练服饰对象的第一款式信息和第一训练服饰对象的特征点的第一位置信息。For the first training sample, the first clothing feature of the first training clothing object included in the first training sample may be obtained, and the first clothing feature may include at least one of the following: first color information of the first training clothing object; The first pose information of the first training clothing object. The first clothing feature may further include at least one of the following: first clothing category information of the first training clothing object, first style information of the first training clothing object, and first position information of feature points of the first training clothing object.

在获取到第一训练服饰对象的第一服饰特征之后,使用第一训练样本对第一初始模型进行训练,得到第一特征提取模型。在进行模型训练时,可以通过迭代的方式将各个第一训练样本依次输入第一初始模型中,得到该第一初始模型输出的检测结果(第一训练服饰对象的服饰特征),根据第一初始模型输出的、第一训练服饰对象的服饰特征与第一服饰特征的相似度程度,调整第一初始模型的模型参数,以调整第一初始模型输出的、第一训练服饰对象的服饰特征与第一服饰特征的相似度,并在收敛条件时(满足模型训练的目标函数)确定训练完成,得到第一特征提取模型。After acquiring the first clothing feature of the first training clothing object, use the first training sample to train the first initial model to obtain a first feature extraction model. During model training, each first training sample can be sequentially input into the first initial model in an iterative manner to obtain the detection result output by the first initial model (the clothing feature of the first training clothing object). The degree of similarity between the clothing features of the first training clothing object and the first clothing features output by the model, adjust the model parameters of the first initial model to adjust the clothing features of the first training clothing object output by the first initial model and the first training clothing features. A similarity of clothing features, and when the convergence conditions are met (the objective function of the model training is satisfied), it is determined that the training is completed, and a first feature extraction model is obtained.

第一特征提取模型从第一训练样本中提取出的第二服饰特征与第一服饰特征的相似度大于或者等于第一阈值,第二服饰特征可以包括:第一训练服饰对象的第二颜色信息和第一训练服饰对象的第二姿态信息。例如,第一阈值可以是在满足目标函数要求时,调整后的第一初始模型从第一训练样本中提取出的、第一训练服饰对象的服饰特征与第一服饰特征的相似度。The similarity between the second clothing feature extracted from the first training sample and the first clothing feature by the first feature extraction model is greater than or equal to the first threshold, and the second clothing feature may include: second color information of the first training clothing object and the second pose information of the first training apparel object. For example, the first threshold may be the similarity between the clothing features of the first training clothing object and the first clothing features extracted from the first training sample by the adjusted first initial model when the objective function requirements are met.

需要说明的是,使用第一特征提取模型的服务器和训练第一初始模型的服务器可以是相同的服务器,也可以是不同的服务器。It should be noted that the server that uses the first feature extraction model and the server that trains the first initial model may be the same server, or may be different servers.

通过本实施例,通过使用第一训练样本对第一初始模型进行训练,得到第一特征提取模型,可以提高训练得到的第一特征提取模型对服饰对象的服饰特征的检测能力。Through this embodiment, by using the first training sample to train the first initial model to obtain the first feature extraction model, the ability of the trained first feature extraction model to detect clothing features of the clothing object can be improved.

获取第一训练服饰对象的第一服饰特征的方式可以有多种,可以人工标注的方式进行第一服饰特征的获取,也可以通过机器标注的方式进行第一服饰特征的获取。There are many ways to obtain the first clothing feature of the first training clothing object. The first clothing feature can be obtained by manual labeling, or the first clothing feature can be obtained by machine labeling.

作为一种可选的实施例,获取第一训练服饰对象的第一服饰特征包括以下至少之一:对第一训练样本进行直方图计算和聚类计算,获取第一颜色信息;根据标注的第一训练服饰对象的特征点的第一位置信息和特征点的可见性信息,确定出第一姿态信息。As an optional embodiment, obtaining the first clothing feature of the first training clothing object includes at least one of the following: performing histogram calculation and clustering calculation on the first training sample to obtain first color information; The first position information of the feature points of the training clothing object and the visibility information of the feature points are determined to determine the first posture information.

对于第一训练服饰对象的颜色信息,可以首先对第一训练样本进行直方图计算,得到第一训练样本的直方图。然后根据直方图,对第一训练样本的像素点进行聚类,将聚类后得到的多个类中包含像素点最多的目标类的颜色信息,作为第一训练服饰对象的颜色信息。For the color information of the first training clothing object, a histogram calculation may be performed on the first training sample to obtain a histogram of the first training sample. Then, according to the histogram, the pixels of the first training sample are clustered, and the color information of the target class with the most pixels in the multiple classes obtained after the clustering is used as the color information of the first training clothing object.

对于第一训练服饰对象的姿态信息,根据标注的第一训练服饰对象的特征点的第一位置信息和特征点的可见性信息,确定出第一训练服饰对象的第一姿态信息。特征点的可见性信息用于表示特征点是否可见。For the posture information of the first training clothing object, the first posture information of the first training clothing object is determined according to the marked first position information of the feature points of the first training clothing object and the visibility information of the feature points. The visibility information of feature points is used to indicate whether the feature points are visible.

根据标注的第一训练服饰对象的特征点的第一位置信息和特征点的可见性信息,可以对第一训练服饰对象的目标位置区域(例如,胸前、腿前侧等纹服饰纹理可能较为丰富区域)的图像空间注意力掩模进行加权,得到处理后的第一训练样本,并使用处理后的第一训练样本对第一初始模型进行训练。According to the marked first location information of the feature points of the first training clothing object and the visibility information of the feature points, the target location area of the first training clothing object (for example, the pattern clothing textures such as the chest and the front of the legs may be relatively The image space attention mask of the rich region) is weighted to obtain the processed first training sample, and the first initial model is trained using the processed first training sample.

例如,第一特征提取模型可以是服饰图像特征表达模型,可以对较大规模的、包含服饰类别、款式集、特征点信息的数据集,进行直方图计算及聚类计算,采用半自动的数据收集方式挖掘服饰色彩信息(第一颜色信息);结合特征点位置及可见性信息,判断当前服饰的姿态信息,并对胸前、腿前侧等纹服饰纹理可能较为丰富区域的图像空间注意力掩模进行加强;结合服饰的类别、款式、色彩、特征点和姿态信息,训练服饰图像特征表达模型。For example, the first feature extraction model can be a clothing image feature expression model, which can perform histogram calculation and clustering calculation on a relatively large-scale data set including clothing category, style set, and feature point information, and use semi-automatic data collection. The method mines clothing color information (first color information); combines the position and visibility information of feature points to judge the posture information of the current clothing, and pays attention to the image space in the areas where the patterned clothing textures such as the chest and the front of the legs may be richer. The model is strengthened; combined with the clothing category, style, color, feature points and posture information, the clothing image feature expression model is trained.

通过本实施例,通过对训练图像的直方图进行聚类,得到第一训练服饰对象的颜色信息,可以提高第一颜色信息表征服饰对象的色彩的能力;根据服饰对象的特征点位置及可见性信息,确定服饰对象的姿态信息,可以提高服饰对象姿态信息确定的准确性。Through this embodiment, the color information of the first training clothing object is obtained by clustering the histogram of the training image, and the ability of the first color information to represent the color of the clothing object can be improved; according to the position and visibility of the feature points of the clothing object information, and determine the posture information of the clothing object, which can improve the accuracy of determining the posture information of the clothing object.

在得到第一特征提取模型,可以将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型的输出结果,将该输出结果作为第一服饰对象的服饰特征。After the first feature extraction model is obtained, the first clothing region may be input into the first feature extraction model to obtain an output result of the first feature extraction model, and the output result is used as the clothing feature of the first clothing object.

在得到第一服饰对象的服饰特征之后,可以直接根据第一服饰对象的服饰特征进行区域匹配,也可以首先根据第一服饰对象的服饰特征得到第一服饰对象的同款判别特征(目标特征),然后根据同款判断特征进行区域筛选,最后根据第一服饰对象的服饰特征对第一服饰区域与筛选出的区域进行匹配。After the clothing features of the first clothing object are obtained, region matching can be performed directly according to the clothing features of the first clothing object, or the same-style discriminating features (target features) of the first clothing object can be obtained first according to the clothing features of the first clothing object. , and then perform regional screening according to the same judging feature, and finally match the first clothing region with the screened region according to the clothing features of the first clothing object.

作为一种可选的实施例,可以在从第一服饰区域中提取第一服饰对象的服饰特征之后,将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征;从多个参考服饰区域中获取目标特征与第一服饰对象的目标特征匹配的候选服饰区域;从候选服饰区域中确定出服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域。As an optional embodiment, after the clothing features of the first clothing object are extracted from the first clothing area, the clothing features of the first clothing object may be input into the second feature extraction model, and the output of the second feature extraction model may be obtained. the target features of the first clothing object; obtain candidate clothing regions whose target features match the target features of the first clothing object from a plurality of reference clothing regions; determine the clothing features and the clothing of the first clothing object from the candidate clothing regions The second clothing area for which the feature is matched.

可以使用第二特征提取模型提取第一服饰对象的目标特征(同款判别特征)。该第二特征提取模型可以使用第二训练样本对第二初始模型进行训练得到的,第二训练样本为标注出包含的第二训练服饰对象的服饰特征和第二训练服饰对象的同款标识的图像。第二训练对象的同款标识用于标识同款的第二训练服饰对象(标识两个服饰对象是否是同款服饰)。第二特征提取模型提取出的、同款的第二训练服饰对象的目标特征之间的相似度大于或者等于第二阈值,第二特征提取模型提取出的、不同款的第二训练服饰对象的目标特征之间的相似度小于第二阈值。上述第二阈值可以根据经验值进行设定,并可以根据需要进行调整。The second feature extraction model can be used to extract the target feature (discriminative feature of the same item) of the first clothing object. The second feature extraction model can be obtained by training the second initial model by using the second training sample, and the second training sample is marked with the clothing features of the second training clothing object and the same type of identification of the second training clothing object. image. The same style identification of the second training object is used to identify the second training clothing object of the same style (to identify whether the two clothing objects are the same style of clothing). The similarity between the target features of the second training clothing objects of the same style extracted by the second feature extraction model is greater than or equal to the second threshold, and the similarity between the target features of the second training clothing objects of different styles extracted by the second feature extraction model The similarity between the target features is less than the second threshold. The above-mentioned second threshold can be set according to an empirical value, and can be adjusted as required.

在得到第一服饰对象的目标特征之后,从多个参考服饰区域中获取目标特征与第一服饰对象的目标特征匹配的候选服饰区域。例如,可以计算第一服饰对象的目标特征与各个参考服饰区域中包含的参考服饰对象的目标特征之间的相似度,并将包含的参考服饰对象的目标特征与第一服饰对象的目标特征之间的相似度大于或者等于第三阈值的参考服饰区域,作为候选服饰区域。After the target feature of the first apparel object is obtained, a candidate apparel region whose target feature matches the target feature of the first apparel object is acquired from a plurality of reference apparel regions. For example, the similarity between the target feature of the first clothing object and the target feature of the reference clothing object contained in each reference clothing area may be calculated, and the difference between the target feature of the reference clothing object contained and the target feature of the first clothing object may be calculated. The reference clothing region whose similarity is greater than or equal to the third threshold is used as a candidate clothing region.

在得到候选服饰区域之后,可以从候选服饰区域中确定出服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域。例如,可以计算候选服饰区域中包含的服饰对象的服饰特征与第一服饰对象的服饰特征之间的相似度,并将包含的参考服饰对象的服饰特征与第一服饰对象的服饰特征之间的相似度大于或者等于第四阈值的候选服饰区域,作为第二服饰区域。After the candidate clothing regions are obtained, a second clothing region whose clothing features match those of the first clothing object may be determined from the candidate clothing regions. For example, the similarity between the clothing features of the clothing objects contained in the candidate clothing area and the clothing features of the first clothing object can be calculated, and the similarity between the clothing features of the reference clothing objects contained and the clothing features of the first clothing object can be calculated. The candidate clothing area whose similarity is greater than or equal to the fourth threshold is used as the second clothing area.

例如,可以利用训练得到的服饰图像特征表达模型(第一特征提取模型),对检测到的服饰区域进行图像特征计算,得到服饰区域的图像表达特征(第一服饰对象的服饰特征),利用训练得到的同款服饰判别模型(第二特征提取模型),结合上述生成的图像表达特征,计算该服饰区域的同款判别特征(目标特征)。For example, the image feature expression model (the first feature extraction model) of the clothing image obtained by training can be used to perform image feature calculation on the detected clothing area to obtain the image expression feature of the clothing area (the clothing feature of the first clothing object). The obtained same-style clothing discrimination model (second feature extraction model) is combined with the above-generated image expression features to calculate the same-style discrimination feature (target feature) of the clothing region.

在进行区域检索时,可以使用计算得到的同款判别特征在服饰特征数据库(包含多个参考服饰区域)中进行检索查询,将库中符合相似性要求的图像作为同款推荐候选(候选服饰区域);利用计算得到的图像表达特征对同款推荐候选进行同款确认,识别该服饰区域与数据库中的候选款饰是否一致,若一致,则将一致的候选服饰区域作为该视频中的同款服饰区域(第二服饰区域)。When performing regional retrieval, you can use the calculated discriminant features of the same style to search in the clothing feature database (including multiple reference clothing regions), and use the images in the library that meet the similarity requirements as the same style recommendation candidate (candidate clothing region). ); use the calculated image expression features to confirm the same item of recommendation candidates for the same item, and identify whether the clothing area is consistent with the candidate items in the database. If they are consistent, the consistent candidate clothing area is used as the same item in the video. Apparel Area (Second Apparel Area).

对于多个参考服饰区域,可以预先确定各个参考服饰区域的参考服饰对象的服饰特征和目标特征,并将确定的各个参考服饰区域的参考服饰对象的服饰特征和目标特征保存到服饰特征数据库。For multiple reference clothing regions, the clothing features and target features of the reference clothing objects in each reference clothing region may be predetermined, and the determined clothing features and target features of the reference clothing objects in each reference clothing region may be stored in the clothing feature database.

例如,可以预先构建服饰特征数据库。对待推广款式的服饰图像进行服装区域检测,利用生成的服饰图像特征表达模型提取服饰区域图像的图像特征(服饰特征),利用生成的同款服饰判别模型,结合生成的图像特征,提取该服饰的同款判别特征;将图像特征和同款判别特征作为服饰特征向量,构建服饰特征数据库。For example, an apparel feature database can be pre-built. Perform clothing region detection on clothing images of styles to be promoted, extract image features (clothing features) of clothing region images using the generated clothing image feature expression model, use the generated clothing discrimination model of the same style, and combine the generated image features to extract the clothing features. Distinguishing features of the same style; use the image features and the distinguishing features of the same style as clothing feature vectors to construct a clothing feature database.

通过本实施例,通过首先根据同款服饰特征进行参考服饰区域的筛选,然后根据服饰对象的服饰特征进行服饰区域的匹配,可以提高服饰区域匹配的效率,减少服饰区域匹配对处理资源的占用。Through this embodiment, by first screening the reference clothing regions according to the clothing features of the same style, and then matching the clothing regions according to the clothing features of the clothing objects, the efficiency of clothing region matching can be improved and the occupation of processing resources by clothing region matching can be reduced.

可以使用第二训练样本对第二初始模型进行训练,得到第二特征识别模型。The second initial model may be trained using the second training sample to obtain a second feature recognition model.

作为一种可选的实施例,在将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征之前,将第二训练样本输入到第一特征提取模型,得到第一特征提取模型输出的、第二训练服饰对象的服饰特征;使用第二训练服饰对象的服饰特征和第二训练对象的同款信息对第二初始模型进行训练,得到第二特征提取模型。As an optional embodiment, before inputting the clothing features of the first clothing object into the second feature extraction model and obtaining the target features of the first clothing object output by the second feature extraction model, inputting the second training sample To the first feature extraction model, obtain the clothing feature of the second training clothing object output by the first feature extraction model; use the clothing feature of the second training clothing object and the same information of the second training object to train the second initial model , to obtain the second feature extraction model.

对于第二训练样本,可以使用第一特征提取模型提取第二训练样本包含的第二训练服饰对象的服饰特征。然后使用第二训练服饰对象的服饰特征和第二训练对象的同款信息对第二初始模型进行训练,得到第二特征提取模型。For the second training sample, the first feature extraction model may be used to extract clothing features of the second training clothing object included in the second training sample. Then, the second initial model is trained by using the clothing features of the second training clothing object and the same item information of the second training object to obtain a second feature extraction model.

在进行第二特征提取模型训练时,可以通过迭代的方式分别将各个第二训练样本的第二训练服饰对象的服饰特征输入到第二初始模型,得到第二初始模型输出的第二训练服饰对象的目标特征。根据同款信息,调整当前第二训练样本包含的第二训练服饰对象的目标特征和当前第二训练样本之前的第二训练样本包含的第二训练服饰对象的目标特征,使得同款的第二训练服饰对象的目标特征的相似度大于或者等于第二阈值,不同款的第二训练服饰对象的目标特征的相似度小于第二阈值。When training the second feature extraction model, the clothing features of the second training clothing objects of each second training sample can be input into the second initial model in an iterative manner, and the second training clothing objects output by the second initial model can be obtained. target features. According to the information of the same paragraph, adjust the target features of the second training clothing object included in the current second training sample and the target features of the second training clothing object included in the second training sample before the current second training sample, so that the second training clothing object of the same paragraph The similarity of the target features of the training clothing objects is greater than or equal to the second threshold, and the similarity of the target features of the second training clothing objects of different styles is less than the second threshold.

需要说明的是,目标特征可以为同款判别特征,可以认为是多个服饰特征的联合判别特征。It should be noted that the target feature can be a discriminative feature of the same item, and can be considered as a joint discriminant feature of multiple clothing features.

例如,可以结合小规模同款服饰数据集,结合生成的服饰图像特征表达模型(第一特征提取模型),训练同款服饰的相似性判别模型,得到的相似性判别模型,同款服饰判别模型(第二特征提取模型)。For example, a small-scale data set of the same style of clothing can be combined with the generated clothing image feature expression model (the first feature extraction model) to train the similarity discrimination model of the same style of clothing. (Second Feature Extraction Model).

通过本实施例,通过结合第一特征提取模型提取的服饰特征进行第二初始模型的训练,可以提高第二初始模型提取的效率,简化模型训练的流程(无需额外进行第二训练样本中的第二训练服饰对象的服饰特征的标注)。Through this embodiment, the training of the second initial model is performed by combining the clothing features extracted by the first feature extraction model, which can improve the extraction efficiency of the second initial model and simplify the model training process (there is no need to additionally perform the first step in the second training sample). 2. Annotation of clothing features of training clothing objects).

在步骤S208中,在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象。In step S208, in the case of determining a second clothing region in which the clothing features of the included clothing objects match the clothing features of the first clothing object from the plurality of reference clothing regions, obtain the first clothing region that matches the clothing features of the second clothing region. Target clothing information corresponding to two clothing objects, wherein each reference clothing area includes at least one clothing object.

在得到与第一服饰区域匹配的第二服饰区域之后,可以获取与该第二服饰区域包含的第二服饰对象对应的目标服饰信息,该目标服饰信息可以包括但不限于:第二服饰对象的描述信息,链接信息或者其他与第二服饰对象关联的信息。After obtaining the second clothing region matching the first clothing region, target clothing information corresponding to the second clothing object included in the second clothing region may be obtained, and the target clothing information may include but is not limited to: Description information, link information or other information associated with the second clothing object.

获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息可以是:根据第二服饰对象对应的物品标识,从保存有各个服饰对象的服饰对象信息的数据库中,提取出第二服饰对象的目标服饰信息。Obtaining the target clothing information corresponding to the second clothing object contained in the second clothing area may be: according to the item identifier corresponding to the second clothing object, from a database storing clothing object information of each clothing object, extracting the second clothing Object's target apparel information.

在步骤S210中,对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。In step S210, area tracking is performed for the first clothing region in the video frame sequence of the video to be detected, and the appearance information of the first clothing object in the video to be detected is determined.

在获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息之后,可以对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息,上述出现信息用于表示第一服饰对象出现在待检测视频中的信息,可以包括但不限于以下至少之一:时间段信息(时间点位信息),位置信息。After acquiring the target clothing information corresponding to the second clothing object contained in the second clothing region, the region tracking of the first clothing region in the video frame sequence of the video to be detected can be performed to determine that the first clothing object is in the video to be detected. The above-mentioned appearance information is used to indicate that the first clothing object appears in the video to be detected, and may include but not limited to at least one of the following: time period information (time point information), location information.

第一服饰对象在待检测视频中出现的时间段(目标时间段)可以是一个时间段,也可以是多个时间段。在目标时间段内的各视频帧中,第一服饰对象出现的位置信息可以是第一服饰对象在该视频帧中的坐标信息(例如,通过x,y坐标表示的坐标信息),也可以是第一服饰对象在该视频帧中的区域信息(例如,左半区域,右半区域,又例如,中间区域,左上区域,左下区域,右上区域,右下区域)。The time period (target time period) in which the first clothing object appears in the video to be detected may be one time period or multiple time periods. In each video frame within the target time period, the position information of the first clothing object in the video frame may be the coordinate information of the first clothing object in the video frame (for example, the coordinate information represented by the x, y coordinates), or it may be Region information of the first clothing object in the video frame (for example, the left half region, the right half region, and for example, the middle region, the upper left region, the lower left region, the upper right region, and the lower right region).

第一服饰对象在待检测视频中出现的位置信息可以是第一服饰区域的特定点(例如,中心点)在待检测视频中出现的位置信息。在进行位置信息保存时,可以保存第一服饰对象在目标时间段内的每个视频帧中的位置信息,也可以仅保存第一服饰对象在目标时间段内的视频帧中的位置信息的变化。The location information of the first clothing object appearing in the video to be detected may be the location information of a specific point (eg, a center point) of the first clothing region appearing in the video to be detected. When storing the position information, the position information of the first clothing object in each video frame within the target time period may be saved, or only the change of the position information of the first clothing object in the video frames within the target time period may be saved .

例如,第一服饰对象出现在待检测视频的第5s到第10s的视频帧内,其中,在第5-7s的视频帧中,第一服饰对象出现的位置坐标为(x1,y1),在第7-9s的视频帧中,第一服饰对象出现的位置坐标为(x2,y2),在第9-10s的视频帧中,第一服饰对象出现的位置坐标为(x3,y3)。则第一服饰对象在待检测视频中的时间段为:第5-10s,第一服饰对象在待检测视频中的位置为:第5-7s,(x1,y1);第7-9s,(x2,y2)(或者,(x2-x1,y2-y1));第9-10s,(x3,y3)(或者,(x3-x1,y3-y1),或者,(x3-x2,y3-y2))。For example, the first clothing object appears in the video frames from the 5th to the 10th s of the video to be detected, wherein, in the 5th to 7th video frames, the coordinates of the position where the first clothing object appears are (x 1 , y 1 ) , in the video frame of the 7th-9th s, the position coordinates of the first clothing object appearing are (x 2 , y 2 ), and in the video frame of the 9-10s, the position coordinates of the first clothing object appearing are (x 3 ) , y 3 ). Then the time period of the first clothing object in the video to be detected is: 5-10s, and the position of the first clothing object in the video to be detected is: 5-7s, (x 1 , y 1 ); 7-9s , (x 2 , y 2 ) (or, (x 2 -x 1 , y 2 -y 1 )); 9-10s, (x 3 , y 3 ) (or, (x 3 -x 1 , y 3 -y 1 ), or, (x 3 -x 2 , y 3 -y 2 )).

作为一种可选的实施例,对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息包括:按照第一服饰区域分别对待检测视频中位于关键帧之前的视频帧和位于的关键帧之后的视频帧进行区域检测,确定第一服饰对象在待检测视频中出现的时间段信息和位置信息,其中,出现信息包括时间段信息和位置信息。As an optional embodiment, performing regional tracking of the first clothing region in the video frame sequence of the video to be detected, and determining the appearance information of the first clothing object in the video to be detected includes: according to the first clothing region to be detected separately Region detection is performed on the video frame before the key frame and the video frame after the key frame in the video to determine the time period information and position information of the first clothing object appearing in the video to be detected, wherein the appearance information includes the time period information and the position information. location information.

在获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息之后,可以进一步确定第一服饰对象在待检测视频中出现的时间点位信息。确定时间点位信息的方式可以是:对第一服饰区域在待检测视频的视频帧序列中进行双向跟踪,确定第一服饰对象在待检测视频中出现的时间段信息。After acquiring the target clothing information corresponding to the second clothing object contained in the second clothing area, it is possible to further determine the time point information at which the first clothing object appears in the video to be detected. The method of determining the time point information may be: bidirectionally tracking the first clothing area in the video frame sequence of the video to be detected, and determining the time period information of the first clothing object appearing in the video to be detected.

双向追踪的方式可以有多种,例如,可以对待检测视频中位于关键帧之前的视频帧和位于的关键帧之后的视频帧进行区域检测,确定待检测视频的视频帧序列中,第一服饰区域出现在的视频帧,得到第一服饰区域出现的时间点位信息,进而确定出第一服饰区域出现的时间段信息。又例如,可以对待检测视频中位于当前关键帧之前的关键帧和位于当前关键帧之后的关键帧进行区域检测,确定待检测视频的视频帧序列中,第一服饰区域出现在的关键帧,得到第一服饰区域出现的时间点位信息,进而确定出第一服饰区域出现的时间段信息。There are various ways of bidirectional tracking. For example, the video frame before the key frame and the video frame after the key frame in the video to be detected can be detected, and the first clothing area in the video frame sequence of the video to be detected can be determined. The video frame that appears in the first clothing area is obtained, and the time point information of the appearance of the first clothing area is obtained, and then the time period information of the appearance of the first clothing area is determined. For another example, the key frame located before the current key frame and the key frame located after the current key frame in the video to be detected can be subjected to region detection, and the key frame in which the first clothing region appears in the video frame sequence of the video to be detected can be determined to obtain. The time point information of the appearance of the first clothing area is further determined, and the time period information of the appearance of the first clothing area is further determined.

除了时间点位信息之外,还可以确定第一服饰区域出现在的视频帧中,候选区域出现的位置信息,进而确定第一服饰对象在待检测视频中出现的位置信息。In addition to the time point information, it is also possible to determine the location information of the candidate region in the video frame in which the first clothing region appears, and then determine the location information of the first clothing object to appear in the video to be detected.

在对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息之后,在待检测视频中添加控制信息,其中,控制信息用于控制在待检测视频被播放到时间段信息对应的时间段时,通过弹窗的方式在待检测视频中与位置信息对应的位置上显示目标服饰信息。After the region tracking of the first clothing area in the video frame sequence of the video to be detected is performed, and the appearance information of the first clothing object in the video to be detected is determined, control information is added to the video to be detected, wherein the control information is used to control When the video to be detected is played to the time period corresponding to the time period information, the target clothing information is displayed on the position corresponding to the position information in the video to be detected by means of a pop-up window.

在获取到第一服饰对象在待检测视频中出现的时间点位信息和位置信息之后,可以在待检测视频中添加控制信息,其中,控制信息用于控制在待检测视频被播放到与时间段信息对应的时间段时,通过弹窗的方式(也可以通过其他方式)在待检测视频中与位置信息对应的位置上显示目标服饰信息。After obtaining the time point information and position information of the first clothing object appearing in the video to be detected, control information can be added to the video to be detected, wherein the control information is used to control the video to be detected to be played to the same time period When the information corresponds to the time period, the target clothing information is displayed on the position corresponding to the position information in the video to be detected by means of a pop-up window (or other means).

通过本实施例,通过在待检测视频中对第一服饰区域进行区域追踪,可以确定出第一服饰对象在待检测视频中出现的时间段信息和位置信息,从而方便进行服饰信息的添加,提高服饰信息添加的准确性。Through this embodiment, by performing regional tracking on the first clothing area in the video to be detected, the time period information and location information of the first clothing object appearing in the video to be detected can be determined, thereby facilitating the addition of clothing information and improving The accuracy of adding clothing information.

下面结合可选示例对上述的信息的获取方法进行说明。该方法可以运行在视频服务器中。本示例中的信息的获取方法结合色彩信息及人体姿态约束,并采用分阶段模型调优策略,提升同款服饰的识别准确率。The method for obtaining the above information will be described below with reference to optional examples. This method can be run in the video server. The information acquisition method in this example combines color information and human posture constraints, and adopts a phased model tuning strategy to improve the recognition accuracy of the same style of clothing.

如图3所示,本示例中的信息的获取方法可以包括以下步骤:As shown in Figure 3, the method for acquiring information in this example may include the following steps:

步骤S302,训练服饰图像特征表达模型。Step S302, training a clothing image feature expression model.

结合服饰的类别、款式、色彩、特征点及姿态信息,训练识别服饰图像特征的表达模型,即,服饰图像特征表达模型(如图4所示)。Combined with the category, style, color, feature points and posture information of the clothing, an expression model for identifying the characteristics of clothing images, that is, a clothing image feature expression model (as shown in FIG. 4 ) is trained.

步骤S304,训练同款服饰判别模型。Step S304, training a discriminant model for the same style of clothing.

结合小规模的同款服饰数据集,训练同款服饰的相似性判别模型即,判别模型,同款服饰判别模型(如图4所示)。Combined with the small-scale data set of the same style of clothing, the similarity discrimination model of the same style of clothing is trained, that is, the discriminant model, the same style of clothing discriminant model (as shown in Figure 4).

步骤S306,构建服饰特征数据库。Step S306, constructing a clothing feature database.

对服饰数据库中的各图像进行服饰区域检测,利用表达模型提取服饰区域图像的图像特征(服饰特征),进而利用判别模型提取该服饰的同款判别特征(目标特征),构建服饰特征数据库(如图5所示)。Perform clothing area detection on each image in the clothing database, use the expression model to extract the image features (clothing features) of the clothing area images, and then use the discriminant model to extract the same type of clothing discriminant feature (target feature), and build a clothing feature database (such as shown in Figure 5).

步骤S308,视频中的同款服饰区域识别。In step S308, the region of the same style of clothing in the video is identified.

对待处理视频进行关键帧计算,对每一关键帧进行服饰检测,并对检测到的服饰区域进行图像特征及同款判别特征计算,利用同款判别特征在服饰特征数据库中进行检索查询,将符合相似性要求的库中图像作为同款推荐候选,利用图像表达特征进行同款确认,识别该服饰区域与数据库中的候选款饰是否一致,若一致则作为该视频中的同款服饰区域(如图6所示)。Perform key frame calculation on the video to be processed, perform clothing detection on each key frame, and calculate the image features and the same type discriminant feature for the detected clothing area, and use the same type discriminant feature to search and query in the clothing feature database. The images in the library that are required by the similarity are used as the recommendation candidates of the same style, and the same style is confirmed by using the image expression features to identify whether the clothing area is consistent with the candidate accessories in the database. shown in Figure 6).

步骤S310,对视频中出现的同款服饰,生成帧级精度的的时间点位信息。Step S310, generating time point information with frame-level precision for the same style of clothing appearing in the video.

对于识别到的服饰区域在视频序列中进行双向跟踪,获得该款服饰在视频中出现的时间点位信息,在该时间段内,支持以视频弹窗的方式对服饰进行提示及推广。Two-way tracking is performed on the identified clothing area in the video sequence, and the time point information of the clothing appearing in the video is obtained. During this time period, the clothing can be prompted and promoted in the form of a video pop-up window.

通过本示例,结合服饰类别、款式、色彩、特征点及姿态信息进行图像特征提取,能够提升服饰特征的表达能力;采用分阶段的模型训练方式,提升小数据集上的模型调优性能;采用多层级特征结合的识别方法,提升同款识别准确率。Through this example, image feature extraction combined with clothing category, style, color, feature points and posture information can improve the expression ability of clothing features; the staged model training method is adopted to improve the model tuning performance on small data sets; The recognition method of multi-level feature combination improves the recognition accuracy of the same item.

通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到根据上述实施例的方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,或者网络设备等)执行本发明各个实施例所述的方法。From the description of the above embodiments, those skilled in the art can clearly understand that the method according to the above embodiment can be implemented by means of software plus a necessary general hardware platform, and of course can also be implemented by hardware, but in many cases the former is better implementation. Based on this understanding, the technical solutions of the present invention can be embodied in the form of software products in essence or the parts that make contributions to the prior art, and the computer software products are stored in a storage medium (such as ROM/RAM, magnetic disk, CD-ROM), including several instructions to make a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) execute the methods described in the various embodiments of the present invention.

根据本申请实施例的另一个方面,提供了一种用于实施上述信息的获取方法的信息的获取装置。可选地,该装置用于实现上述实施例及优选实施方式,已经进行过说明的不再赘述。如以下所使用的,术语“模块”可以实现预定功能的软件和/或硬件的组合。尽管以下实施例所描述的装置较佳地以软件来实现,但是硬件,或者软件和硬件的组合的实现也是可能并被构想的。According to another aspect of the embodiments of the present application, an apparatus for acquiring information for implementing the above method for acquiring information is provided. Optionally, the apparatus is used to implement the above-mentioned embodiments and preferred implementations, and the descriptions that have already been described will not be repeated. As used below, the term "module" may be a combination of software and/or hardware that implements a predetermined function. Although the apparatus described in the following embodiments is preferably implemented in software, implementations in hardware, or a combination of software and hardware, are also possible and contemplated.

图7是根据本申请实施例的一种可选的信息的获取装置的结构框图,如图7所示,该装置包括:FIG. 7 is a structural block diagram of an optional information acquisition apparatus according to an embodiment of the present application. As shown in FIG. 7 , the apparatus includes:

(1)第一提取单元702,用于从待检测视频中提取关键帧;(1) The first extraction unit 702 is used to extract key frames from the video to be detected;

(2)检测单元704,与第一提取单元702相连,用于检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;(2) the detection unit 704, connected with the first extraction unit 702, is used for detecting the first clothing area of the first clothing object in the frame image of the key frame;

(3)第二提取单元706,与检测单元704相连,用于从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;(3) The second extraction unit 706, connected to the detection unit 704, is used for extracting the clothing features of the first clothing object from the first clothing area, wherein the clothing features of the first clothing object include at least one of the following: the first clothing The color information of the object, the posture information of the first clothing object;

(4)第一获取单元708,与第二提取单元706相连,用于在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;(4) The first acquiring unit 708, connected with the second extracting unit 706, is used to determine a second clothing region in which the clothing features of the included clothing objects match the clothing features of the first clothing object from a plurality of reference clothing regions In the case of , obtain the target clothing information corresponding to the second clothing object contained in the second clothing area, wherein each reference clothing area contains at least one clothing object;

(5)第一确定单元710,与第一获取单元708相连,用于对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。(5) The first determination unit 710, connected with the first acquisition unit 708, is used to perform regional tracking of the first clothing area in the video frame sequence of the video to be detected, and to determine the appearance information of the first clothing object in the video to be detected .

可选地,第一提取单元702可以用于上述实施例中的步骤S202,检测单元704可以用于上述实施例中的步骤S204,第二提取单元706可以用于上述实施例中的步骤S206,第一获取单元708可以用于执行上述实施例中的步骤S208,第一确定单元710可以用于执行上述实施例中的步骤S210。Optionally, the first extracting unit 702 may be used in step S202 in the foregoing embodiment, the detecting unit 704 may be used in step S204 in the foregoing embodiment, and the second extraction unit 706 may be used in step S206 in the foregoing embodiment, The first obtaining unit 708 may be configured to perform step S208 in the foregoing embodiments, and the first determining unit 710 may be configured to execute step S210 in the foregoing embodiments.

通过本实施例,采用结合色彩信息和/或姿态信息进行服饰识别的方式,由于结合服饰对象的色彩信息和/或姿态信息进行服饰识别,可以实现增加服饰特征对服饰对象的表达能力的目的,解决了相关技术中的服饰识别方式存在由于服饰表达能力弱导致的服饰识别准确率低的问题,提高了服饰识别准确性。Through the present embodiment, the method of performing clothing identification in combination with color information and/or posture information, because the clothing identification is performed in combination with the color information and/or posture information of the clothing object, the purpose of increasing the expressive ability of clothing features to the clothing object can be achieved, The problem of low clothing recognition accuracy caused by weak clothing expression ability in the clothing recognition method in the related art is solved, and the clothing recognition accuracy is improved.

作为一种可选的实施例,第一提取单元702包括:As an optional embodiment, the first extraction unit 702 includes:

(1)第一提取模块,用于按照目标间隔从待检测视频中提取关键帧;或者,(1) The first extraction module is used to extract key frames from the video to be detected according to the target interval; or,

(2)第二提取模块,用于从待检测视频所包含的镜头中抽取与镜头对应的关键帧。(2) The second extraction module is used for extracting key frames corresponding to the shots from the shots included in the video to be detected.

作为一种可选的实施例,第二提取单元706包括:As an optional embodiment, the second extraction unit 706 includes:

(1)输入模块,用于将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征,其中,第一特征提取模型是使用第一训练样本对第一初始模型进行训练得到的,第一训练样本是标注出包含的第一训练服饰对象的第一服饰特征的图像。(1) an input module, used for inputting the first clothing area into the first feature extraction model to obtain the clothing features of the first clothing object output by the first feature extraction model, wherein the first feature extraction model uses the first training The sample is obtained by training the first initial model, and the first training sample is an image marked with the first clothing feature of the included first training clothing object.

作为一种可选的实施例,上述装置还包括:As an optional embodiment, the above device also includes:

(1)第二获取单元,用于在将第一服饰区域输入到第一特征提取模型,得到第一特征提取模型输出的、第一服饰对象的服饰特征之前,获取第一训练服饰对象的第一服饰特征,其中,第一服饰特征包括以下至少之一:第一训练服饰对象的第一颜色信息和第一训练服饰对象的第一姿态信息;(1) The second acquisition unit is used to obtain the first clothing object of the first training clothing before inputting the first clothing area into the first feature extraction model to obtain the clothing features of the first clothing object output by the first feature extraction model. A clothing feature, wherein the first clothing feature includes at least one of the following: first color information of the first training clothing object and first posture information of the first training clothing object;

(2)第一训练单元,用于使用第一训练样本对第一初始模型进行训练,得到第一特征提取模型,其中,第一特征提取模型从第一训练样本中提取出的第二服饰特征与第一服饰特征的相似度大于或者等于第一阈值,其中,第二服饰特征包括以下至少之一:第一训练服饰对象的第二颜色信息和第一训练服饰对象的第二姿态信息。(2) A first training unit, used for training the first initial model by using the first training sample to obtain a first feature extraction model, wherein the first feature extraction model extracts the second clothing feature from the first training sample The similarity with the first clothing feature is greater than or equal to the first threshold, wherein the second clothing feature includes at least one of the following: second color information of the first training clothing object and second posture information of the first training clothing object.

作为一种可选的实施例,第二获取单元包括以下至少之一:As an optional embodiment, the second obtaining unit includes at least one of the following:

(1)获取模块,用于对第一训练样本进行直方图计算和聚类计算,获取第一颜色信息;(1) an acquisition module, for performing histogram calculation and clustering calculation on the first training sample to obtain the first color information;

(2)第一确定模块,用于根据标注的第一训练服饰对象的特征点的第一位置信息和特征点的可见性信息,确定出第一姿态信息。(2) A first determination module, configured to determine the first posture information according to the first position information of the feature points of the marked first training clothing object and the visibility information of the feature points.

作为一种可选的实施例,上述装置还包括:As an optional embodiment, the above device also includes:

(1)第三获取单元,用于在从第一服饰区域中提取第一服饰对象的服饰特征之后,将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征,其中,第二特征提取模型是使用第二训练样本对第二初始模型进行训练得到的,第二训练样本为标注出包含的第二训练服饰对象的服饰特征和第二训练服饰对象的同款标识的图像,第二训练对象的同款标识用于标识同款的第二训练服饰对象,第二特征提取模型提取出的、同款的第二训练服饰对象的目标特征之间的相似度大于或者等于第二阈值,第二特征提取模型提取出的、不同款的第二训练服饰对象的目标特征之间的相似度小于第二阈值;(1) The third acquisition unit is used to input the clothing features of the first clothing object into the second feature extraction model after extracting the clothing features of the first clothing object from the first clothing area, and obtain the output of the second feature extraction model , the target feature of the first clothing object, wherein the second feature extraction model is obtained by using the second training sample to train the second initial model, and the second training sample is the clothing feature that marks the included second training clothing object and the image of the same paragraph mark of the second training clothing object, the same paragraph mark of the second training object is used to identify the second training clothing object of the same paragraph, the second training clothing object of the same paragraph extracted by the second feature extraction model The similarity between the target features is greater than or equal to the second threshold, and the similarity between the target features of the second training apparel objects of different models extracted by the second feature extraction model is less than the second threshold;

(2)第四获取单元,用于从多个参考服饰区域中获取目标特征与第一服饰对象的目标特征匹配的候选服饰区域;(2) the 4th acquisition unit, is used to obtain the candidate clothing area that target feature matches with the target feature of the first clothing object from a plurality of reference clothing areas;

(3)第二确定单元,用于从候选服饰区域中确定出服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域。(3) A second determining unit, configured to determine, from the candidate clothing regions, a second clothing region whose clothing features match those of the first clothing object.

作为一种可选的实施例,上述装置还包括:As an optional embodiment, the above device also includes:

(1)输入单元,用于在将第一服饰对象的服饰特征输入到第二特征提取模型,获取第二特征提取模型输出的、第一服饰对象的目标特征之前,将第二训练样本输入到第一特征提取模型,得到第一特征提取模型输出的、第二训练服饰对象的服饰特征;(1) an input unit, used for inputting the clothing features of the first clothing object into the second feature extraction model, and before acquiring the target features of the first clothing object output by the second feature extraction model, inputting the second training sample into The first feature extraction model obtains the clothing features of the second training clothing object output by the first feature extraction model;

(2)第二训练单元,用于使用第二训练服饰对象的服饰特征和第二训练对象的同款信息对第二初始模型进行训练,得到第二特征提取模型。(2) a second training unit, used for training the second initial model by using the clothing features of the second training clothing object and the same item information of the second training object to obtain a second feature extraction model.

作为一种可选的实施例,上述装置还包括:添加单元,第一确定单元包括第二确定模块,其中,As an optional embodiment, the above-mentioned apparatus further includes: an adding unit, and the first determining unit includes a second determining module, wherein,

(1)第二确定模块,用于按照第一服饰区域分别对待检测视频中位于关键帧之前的视频帧和位于的关键帧之后的视频帧进行区域检测,确定第一服饰对象在待检测视频中出现的时间段信息和位置信息,其中,出现信息包括时间段信息和位置信息;(1) The second determination module is used to perform regional detection on the video frame before the key frame and the video frame after the key frame in the video to be detected according to the first clothing area, and determine that the first clothing object is in the video to be detected. Time period information and location information of occurrence, wherein the occurrence information includes time period information and location information;

(2)添加单元,用于在对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息之后,在待检测视频中添加控制信息,其中,控制信息用于控制在待检测视频被播放到时间段信息对应的时间段时,通过弹窗的方式在待检测视频中与位置信息对应的位置上显示目标服饰信息。(2) an adding unit for performing regional tracking of the first clothing area in the video frame sequence of the video to be detected, and after determining the appearance information of the first clothing object in the video to be detected, adding control information in the video to be detected , wherein the control information is used to control when the video to be detected is played to the time period corresponding to the time period information, the target clothing information is displayed on the position corresponding to the position information in the video to be detected by means of a pop-up window.

需要说明的是,上述各个模块是可以通过软件或硬件来实现的,对于后者,可以通过以下方式实现,但不限于此:上述模块均位于同一处理器中;或者,上述各个模块以任意组合的形式分别位于不同的处理器中。It should be noted that the above modules can be implemented by software or hardware, and the latter can be implemented in the following ways, but not limited to this: the above modules are all located in the same processor; or, the above modules can be combined in any combination The forms are located in different processors.

根据本申请实施例的又一个方面,提供了一种计算机可读的存储介质。可选地,该存储介质中存储有计算机程序,其中,该计算机程序被设置为运行时执行本申请实施例中所提供的上述任一项方法中的步骤。According to yet another aspect of the embodiments of the present application, a computer-readable storage medium is provided. Optionally, a computer program is stored in the storage medium, wherein the computer program is configured to execute the steps in any one of the above methods provided in the embodiments of the present application when the computer program is run.

可选地,在本实施例中,上述存储介质可以被设置为存储用于执行以下步骤的计算机程序:Optionally, in this embodiment, the above-mentioned storage medium may be configured to store a computer program for executing the following steps:

S1,从待检测视频中提取关键帧;S1, extract key frames from the video to be detected;

S2,检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;S2, the frame image of the detection key frame includes the first clothing area of the first clothing object;

S3,从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;S3, extract the clothing feature of the first clothing object from the first clothing area, wherein the clothing feature of the first clothing object includes at least one of the following: color information of the first clothing object, and posture information of the first clothing object;

S4,在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;S4, in the case of determining from a plurality of reference clothing regions a second clothing region in which the clothing features of the included clothing objects match the clothing features of the first clothing object, obtain the second clothing object that is included in the second clothing region Corresponding target clothing information, wherein each reference clothing area contains at least one clothing object;

S5,对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。S5, perform regional tracking on the first clothing region in the video frame sequence of the video to be detected, and determine the appearance information of the first clothing object in the video to be detected.

可选地,在本实施例中,上述存储介质可以包括但不限于:U盘、ROM(Read-OnlyMemory,只读存储器)、RAM(Random Access Memory,随机存取存储器)、移动硬盘、磁碟或者光盘等各种可以存储计算机程序的介质。Optionally, in this embodiment, the above-mentioned storage medium may include but is not limited to: U disk, ROM (Read-Only Memory, read-only memory), RAM (Random Access Memory, random access memory), mobile hard disk, magnetic disk Or various media such as optical discs that can store computer programs.

根据本申请实施例的又一个方面,提供了一种电子装置,包括:处理器(该存储器可以是图1中的处理器102)和存储器(该存储器可以是图1中的存储器104),该存储器中存储有计算机程序,该处理器被设置为运行计算机程序以执行本申请实施例中所提供的上述任一项方法中的步骤。According to yet another aspect of the embodiments of the present application, an electronic device is provided, including: a processor (the memory may be the processor 102 in FIG. 1 ) and a memory (the memory may be the memory 104 in FIG. 1 ), the memory A computer program is stored in the memory, and the processor is configured to run the computer program to execute the steps in any one of the above methods provided in the embodiments of the present application.

可选地,上述电子装置还可以包括传输装置(该传输装置可以是图1中的传输装置106)以及输入输出设备(该输入输出设备可以是图1中的输入输出设备108),其中,该传输设备和上述处理器连接,该输入输出设备和上述处理器连接。Optionally, the above electronic device may further include a transmission device (the transmission device may be the transmission device 106 in FIG. 1 ) and an input/output device (the input/output device may be the input/output device 108 in FIG. 1 ), wherein the The transmission device is connected to the above-mentioned processor, and the input and output device is connected to the above-mentioned processor.

可选地,在本实施例中,上述处理器可以被设置为通过计算机程序执行以下步骤:Optionally, in this embodiment, the above-mentioned processor may be configured to execute the following steps through a computer program:

S1,从待检测视频中提取关键帧;S1, extract key frames from the video to be detected;

S2,检测关键帧的帧图像中包含第一服饰对象的第一服饰区域;S2, the frame image of the detection key frame includes the first clothing area of the first clothing object;

S3,从第一服饰区域中提取第一服饰对象的服饰特征,其中,第一服饰对象的服饰特征包括以下至少之一:第一服饰对象的颜色信息,第一服饰对象的姿态信息;S3, extract the clothing feature of the first clothing object from the first clothing area, wherein the clothing feature of the first clothing object includes at least one of the following: color information of the first clothing object, and posture information of the first clothing object;

S4,在从多个参考服饰区域中确定出包含的服饰对象的服饰特征与第一服饰对象的服饰特征匹配的第二服饰区域的情况下,获取与第二服饰区域中包含的第二服饰对象对应的目标服饰信息,其中,每个参考服饰区域中包含至少一个服饰对象;S4, in the case of determining from a plurality of reference clothing regions a second clothing region in which the clothing features of the included clothing objects match the clothing features of the first clothing object, obtain the second clothing object that is included in the second clothing region Corresponding target clothing information, wherein each reference clothing area contains at least one clothing object;

S5,对第一服饰区域在待检测视频的视频帧序列中进行区域追踪,确定第一服饰对象在待检测视频中的出现信息。S5, perform regional tracking on the first clothing region in the video frame sequence of the video to be detected, and determine the appearance information of the first clothing object in the video to be detected.

可选地,本实施例中的可选示例可以参考上述实施例及可选实施方式中所描述的示例,本实施例在此不再赘述。Optionally, for optional examples in this embodiment, reference may be made to the examples described in the foregoing embodiments and optional implementation manners, and details are not described herein again in this embodiment.

显然,本领域的技术人员应该明白,上述的本发明的各模块或各步骤可以用通用的计算装置来实现,它们可以集中在单个的计算装置上,或者分布在多个计算装置所组成的网络上,可选地,它们可以用计算装置可执行的程序代码来实现,从而,可以将它们存储在存储装置中由计算装置来执行,并且在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤,或者将它们分别制作成各个集成电路模块,或者将它们中的多个模块或步骤制作成单个集成电路模块来实现。这样,本发明不限制于任何特定的硬件和软件结合。Obviously, those skilled in the art should understand that the above-mentioned modules or steps of the present invention can be implemented by a general-purpose computing device, and they can be centralized on a single computing device or distributed in a network composed of multiple computing devices Alternatively, they may be implemented in program code executable by a computing device, such that they may be stored in a storage device and executed by the computing device, and in some cases, in a different order than here The steps shown or described are performed either by fabricating them separately into individual integrated circuit modules, or by fabricating multiple modules or steps of them into a single integrated circuit module. As such, the present invention is not limited to any particular combination of hardware and software.

以上所述仅为本发明的优选实施例而已,并不用于限制本发明,对于本领域的技术人员来说,本发明可以有各种更改和变化。凡在本发明的原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above descriptions are only preferred embodiments of the present invention, and are not intended to limit the present invention. For those skilled in the art, the present invention may have various modifications and changes. Any modification, equivalent replacement, improvement, etc. made within the principle of the present invention shall be included within the protection scope of the present invention.

Claims (11)

1. An information acquisition method, comprising:
extracting key frames from a video to be detected;
detecting a first decoration area containing a first decoration object in a frame image of the key frame;
extracting a clothing feature of the first clothing object from the first clothing region, wherein the clothing feature of the first clothing object comprises at least one of: color information of the first garment object, and posture information of the first garment object;
under the condition that a second clothing area, which is matched with clothing characteristics of the first clothing object, is determined from a plurality of reference clothing areas, target clothing information corresponding to the second clothing object contained in the second clothing area is obtained, wherein each reference clothing area contains at least one clothing object;
and carrying out region tracking on the first decoration region in the video frame sequence of the video to be detected, and determining the appearance information of the first decoration object in the video to be detected.
2. The method of claim 1, wherein extracting the key frames from the video to be detected comprises:
extracting the key frames from the video to be detected according to the target interval; or,
and extracting the key frames corresponding to the shots from the shots contained in the video to be detected.
3. The method of claim 1, wherein extracting the clothing feature of the first clothing object from the first clothing region comprises:
inputting the first clothing region into a first feature extraction model to obtain clothing features of the first clothing object output by the first feature extraction model, wherein the first feature extraction model is obtained by training a first initial model by using a first training sample, and the first training sample is an image marked with the first clothing features of the first training clothing object.
4. The method of claim 3, wherein prior to inputting the first clothing region to the first feature extraction model, resulting in clothing features of the first clothing object output by the first feature extraction model, the method further comprises:
obtaining the first apparel feature of the first training apparel object, wherein the first apparel feature comprises at least one of: first color information of the first training clothing object and first pose information of the first training clothing object;
training the first initial model by using the first training sample to obtain the first feature extraction model, wherein the similarity between a second clothing feature extracted from the first training sample by the first feature extraction model and the first clothing feature is greater than or equal to a first threshold, and the second clothing feature comprises at least one of the following: second color information of the first training clothing object and second pose information of the first training clothing object.
5. The method of claim 4, wherein obtaining the first garment characteristic of the first training garment object comprises at least one of:
performing histogram calculation and clustering calculation on the first training sample to obtain the first color information;
and determining the first posture information according to the marked first position information of the characteristic points of the first training clothing object and the marked visibility information of the characteristic points.
6. The method of claim 3, wherein after extracting the apparel feature of the first apparel object from the first apparel area, the method further comprises:
inputting the clothing characteristics of the first clothing object into a second characteristic extraction model, acquiring the target characteristics of the first clothing object output by the second characteristic extraction model, wherein the second feature extraction model is obtained by training a second initial model by using a second training sample, the second training sample is an image marked with a clothing feature of a second training clothing object and a homogenous identification of the second training clothing object, the homogeneous identification of the second training object is used to identify the second training apparel object in homogeneous, the similarity between the target features of the second training clothes object extracted by the second feature extraction model and in the same style is larger than or equal to a second threshold value, the similarity between the target features of the second training clothes object extracted by the second feature extraction model and in different styles is smaller than the second threshold value;
obtaining candidate clothing regions with target characteristics matched with the target characteristics of the first clothing object from the plurality of reference clothing regions;
determining the second clothing region with clothing characteristics matched with the clothing characteristics of the first clothing object from the candidate clothing regions.
7. The method of claim 6, wherein before inputting the clothing feature of the first clothing object into the second feature extraction model and obtaining the target feature of the first clothing object output by the second feature extraction model, the method further comprises:
inputting the second training sample into the first feature extraction model to obtain clothing features of the second training clothing object output by the first feature extraction model;
and training the second initial model by using the clothing features of the second training clothing object and the same-style information of the second training object to obtain the second feature extraction model.
8. The method according to any one of claims 1 to 7,
performing region tracking on the first service region in a video frame sequence of the video to be detected, and determining the occurrence information of the first service object in the video to be detected includes: respectively performing area detection on a video frame before the key frame and a video frame after the key frame in the video to be detected according to the first service area, and determining time period information and position information of the first service object appearing in the video to be detected, wherein the appearance information comprises the time period information and the position information;
after performing region tracking on the first service region in the video frame sequence of the video to be detected and determining the occurrence information of the first service object in the video to be detected, the method further includes: adding control information into the video to be detected, wherein the control information is used for controlling the target clothing information to be displayed on the position corresponding to the position information in the video to be detected in a pop-up window mode when the video to be detected is played to the time period corresponding to the time period information.
9. An apparatus for acquiring information, comprising:
the first extraction unit is used for extracting key frames from a video to be detected;
a detecting unit, configured to detect a first decoration area containing a first decoration object in a frame image of the key frame;
a second extraction unit, configured to extract a clothing feature of the first clothing object from the first clothing region, where the clothing feature of the first clothing object includes at least one of: color information of the first garment object, and posture information of the first garment object;
a first obtaining unit, configured to, when a second clothing region in which clothing features of clothing objects included in the plurality of reference clothing regions are matched with clothing features of the first clothing object is determined, obtain target clothing information corresponding to the second clothing object included in the second clothing region, where each of the reference clothing regions includes at least one clothing object;
the first determining unit is configured to perform region tracking on the first service region in the video frame sequence of the video to be detected, and determine occurrence information of the first service object in the video to be detected.
10. A computer-readable storage medium, in which a computer program is stored, wherein the computer program is configured to carry out the method of any one of claims 1 to 8 when executed.
11. An electronic device comprising a memory and a processor, characterized in that the memory has stored therein a computer program, the processor being arranged to execute the method of any of claims 1 to 8 by means of the computer program.
CN201911239754.1A 2019-12-05 2019-12-05 Information acquisition method and device, storage medium and electronic device Pending CN111126179A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911239754.1A CN111126179A (en) 2019-12-05 2019-12-05 Information acquisition method and device, storage medium and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911239754.1A CN111126179A (en) 2019-12-05 2019-12-05 Information acquisition method and device, storage medium and electronic device

Publications (1)

Publication Number Publication Date
CN111126179A true CN111126179A (en) 2020-05-08

Family

ID=70497544

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911239754.1A Pending CN111126179A (en) 2019-12-05 2019-12-05 Information acquisition method and device, storage medium and electronic device

Country Status (1)

Country Link
CN (1) CN111126179A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112569596A (en) * 2020-12-11 2021-03-30 腾讯科技(深圳)有限公司 Video picture display method and device, computer equipment and storage medium
CN113486855A (en) * 2021-07-30 2021-10-08 浙江大华技术股份有限公司 Clothing identification method, device, equipment and medium
CN113869247A (en) * 2021-09-30 2021-12-31 北京爱奇艺科技有限公司 Apparel identification method, device, equipment and storage medium
CN113887426A (en) * 2021-09-30 2022-01-04 北京爱奇艺科技有限公司 Clothing identification method, device, equipment and storage medium
CN114255375A (en) * 2020-09-10 2022-03-29 北京搜狗科技发展有限公司 Image data processing method and device and electronic equipment
CN113869247B (en) * 2021-09-30 2026-02-06 北京爱奇艺科技有限公司 Garment identification method, device, equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107291825A (en) * 2017-05-26 2017-10-24 北京奇艺世纪科技有限公司 With the search method and system of money commodity in a kind of video
CN107609193A (en) * 2017-10-16 2018-01-19 杭州时间线信息科技有限公司 The intelligent automatic processing method and system of picture in a kind of suitable commodity details page
CN110035314A (en) * 2019-03-08 2019-07-19 腾讯科技(深圳)有限公司 Methods of exhibiting and device, storage medium, the electronic device of information
CN110188701A (en) * 2019-05-31 2019-08-30 上海媒智科技有限公司 Dress ornament recognition methods, system and terminal based on the prediction of human body key node

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107291825A (en) * 2017-05-26 2017-10-24 北京奇艺世纪科技有限公司 With the search method and system of money commodity in a kind of video
CN107609193A (en) * 2017-10-16 2018-01-19 杭州时间线信息科技有限公司 The intelligent automatic processing method and system of picture in a kind of suitable commodity details page
CN110035314A (en) * 2019-03-08 2019-07-19 腾讯科技(深圳)有限公司 Methods of exhibiting and device, storage medium, the electronic device of information
CN110188701A (en) * 2019-05-31 2019-08-30 上海媒智科技有限公司 Dress ornament recognition methods, system and terminal based on the prediction of human body key node

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114255375A (en) * 2020-09-10 2022-03-29 北京搜狗科技发展有限公司 Image data processing method and device and electronic equipment
CN112569596A (en) * 2020-12-11 2021-03-30 腾讯科技(深圳)有限公司 Video picture display method and device, computer equipment and storage medium
CN113486855A (en) * 2021-07-30 2021-10-08 浙江大华技术股份有限公司 Clothing identification method, device, equipment and medium
CN113869247A (en) * 2021-09-30 2021-12-31 北京爱奇艺科技有限公司 Apparel identification method, device, equipment and storage medium
CN113887426A (en) * 2021-09-30 2022-01-04 北京爱奇艺科技有限公司 Clothing identification method, device, equipment and storage medium
CN113887426B (en) * 2021-09-30 2025-07-29 北京爱奇艺科技有限公司 Garment identification method, device, equipment and storage medium
CN113869247B (en) * 2021-09-30 2026-02-06 北京爱奇艺科技有限公司 Garment identification method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
TWI760650B (en) Method and image matching method for neural network training and device thereof
CN106815566B (en) Face retrieval method based on multitask convolutional neural network
CN107944431B (en) A kind of intelligent identification Method based on motion change
CN106933867B (en) An image query method and device
US20130071816A1 (en) Methods and systems for building a universal dress style learner
Chen et al. Poseshop: Human image database construction and personalized content synthesis
CN112905889B (en) Clothing searching method and device, electronic equipment and medium
CN111126179A (en) Information acquisition method and device, storage medium and electronic device
CN111126457A (en) Information acquisition method and device, storage medium and electronic device
US20210035187A1 (en) Device and method for item recommendation based on visual elements
CN109409994A (en) The methods, devices and systems of analog subscriber garments worn ornaments
CN106055710A (en) Video-based commodity recommendation method and device
WO2017005014A1 (en) Method and device for searching matched commodities
Zhao et al. Clothing cosegmentation for shopping images with cluttered background
Barman et al. Shape: A novel graph theoretic algorithm for making consensus-based decisions in person re-identification systems
CN110598017A (en) Self-learning-based commodity detail page generation method
Jiang et al. Online robust action recognition based on a hierarchical model
CN109829418A (en) A kind of punch card method based on figure viewed from behind feature, device and system
CN108052952A (en) A kind of the clothes similarity determination method and its system of feature based extraction
CN112150349A (en) Image processing method and device, computer equipment and storage medium
Li et al. Personrank: Detecting important people in images
CN114513694A (en) Scoring determination method and device, electronic equipment and storage medium
CN117095053A (en) Gesture recognition method, posture recognition model training method and related equipment
CN114332940B (en) Model training method, clothing recognition processing method, related device and terminal
CN115565247A (en) Method, device, equipment and medium for determining action similarity score

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200508