[go: up one dir, main page]

TWM629362U - Identification system - Google Patents

Identification system Download PDF

Info

Publication number
TWM629362U
TWM629362U TW110215752U TW110215752U TWM629362U TW M629362 U TWM629362 U TW M629362U TW 110215752 U TW110215752 U TW 110215752U TW 110215752 U TW110215752 U TW 110215752U TW M629362 U TWM629362 U TW M629362U
Authority
TW
Taiwan
Prior art keywords
person
model
recognition system
identity recognition
data
Prior art date
Application number
TW110215752U
Other languages
Chinese (zh)
Inventor
吳忠倫
賴辰瑜
Original Assignee
關貿網路股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 關貿網路股份有限公司 filed Critical 關貿網路股份有限公司
Priority to TW110215752U priority Critical patent/TWM629362U/en
Publication of TWM629362U publication Critical patent/TWM629362U/en

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The utility model is an identification system, including a data processing module, a feature extraction module, and a classification module. A person’ s image of a person form a received image data is cut to form an input data by the data processing module. After a feature vector of the input data is captured by the feature extraction module, whether the person is a clerk with uniform or a consumer without uniform is determined based on the feature vector by the classification module and the flow of people can be counted. In the utility model, the problem that the conventional face recognition technology is not suitable for the crowd counting of retail stores can be solved.

Description

身分辨識系統 ID system

本創作係關於影像辨識之技術,尤指一種用以對人物之身分別進行識別之身分辨識系統。 This work is about the technology of image recognition, especially an identity recognition system used to identify the bodies of people.

對於商店而言,來客數代表著店家受歡迎的程度,也可增加消費的機會,因此,如何正確統計商店一天或一段時間內之來客人數,對商店之自我評估相當重要。進言之,在智慧零售的應用中,通常需要蒐集消費者行為,作為提升零售之服務體驗的依據,傳統對於消費者行為之蒐集,須透過蒐集問卷或由店員配戴例如計數器之識別裝置進行人流計數。時至今日,蒐集消費者行為中目前較常見的方法,大多利用影像辨識技術進行人流計數、熱區識別以及消費者性別年齡層辨識,以在不影響消費者的情況下,達到取得消費者到店之相關資訊的目的。 For a store, the number of visitors represents the popularity of the store, and it can also increase the chance of consumption. Therefore, how to correctly count the number of visitors to the store in a day or a period of time is very important for the store's self-assessment. In addition, in the application of smart retail, it is usually necessary to collect consumer behavior as a basis for improving the service experience of retail. Traditionally, the collection of consumer behavior must be carried out through the collection of questionnaires or identification devices such as counters worn by store clerks to conduct flow of people. count. Today, the most common methods for collecting consumer behaviors are mostly using image recognition technology to count people, identify hot spots, and identify consumers’ gender and age groups. The purpose of store-related information.

在分析消費者相關資訊時,來店之消費者的人數係最基礎之指標(Metric),然商店內通常同時存在消費者以及為消費者提供服務之店員,因而如何在人流計數上區分出消費者和店員,以便得到真正的消費者數,實為目前智慧零售之重要主題之一。對此,為提供智慧零售對真正消費者之計數,常見之方法係利用人臉辨識(Face Recognition)技術,將已預先註冊面容識別(Face ID)之店員辨 識出來並進行扣除,惟近期由於新冠肺炎(COVID-19)疫情之影響,民衆在公共場合皆會戴上口罩,將影響人臉辨識之辨識效果。再者,依美國國家標準與技術研究所(NIST)之實驗指出,臉部辨識在辨識戴著口罩之人臉時,錯誤率在5%到50%之間,換言之,即使此問題隨著臉部辨識技術的提升已有改善,但人臉辨識仍然會有受臉部角度影響的問題,因而,習知的人臉辨識技術對於身分識別仍存有缺失。 When analyzing consumer-related information, the number of consumers who come to the store is the most basic indicator (Metric). However, there are usually both consumers and shop assistants serving consumers in the store, so how to distinguish consumption in the flow of people count In order to obtain the real number of consumers, it is one of the important themes of smart retail at present. In this regard, in order to provide smart retail counts of real consumers, a common method is to use Face Recognition technology to identify shop assistants who have pre-registered Face ID. However, due to the recent impact of the new crown pneumonia (COVID-19) epidemic, people will wear masks in public places, which will affect the recognition effect of face recognition. Furthermore, according to the experiments of the National Institute of Standards and Technology (NIST), the error rate of facial recognition in recognizing the face of a person wearing a mask is between 5% and 50%. Improvements in facial recognition technology have been improved, but face recognition still has the problem of being affected by the angle of the face. Therefore, the conventional face recognition technology is still lacking in identity recognition.

另外,由於零售店之店通常會穿著制服,是以,另有利用服裝辨識(Clothes Recognition)之方法針對制服進行辨識以區別之,然用於辨識制服之模型對於輸入之服裝角度以及解析度之要求甚高。另外,於訓練模型時,通常會輸入一位模特兒或是一套服裝所構成之靜態影像,惟消費者或店員於零售店內通常呈現動態之情況,故習知之服裝辨識方法實不適用於零售店中人物密集、角度不定或是遮蔽的情況下由閉路電視(Closed-Circuit Television,CCTV)所錄製的串流影片,又或者,利用標誌辨識(Logo Recognition)和光學字元辨識(Optical Character Recognition)辨識制服上標誌(Logo)或字樣之方法,亦存在與前述服裝辨識技術相同之問題,亦即,標誌辨識或光學字元辨識之方法仍須要輸入清楚且相較完整的影像,故同樣不適用於零售店的場域。 In addition, because the stores of retail stores usually wear uniforms, the clothing recognition (Clothes Recognition) method is used to identify the uniforms to distinguish them. Very demanding. In addition, when training the model, a static image composed of a model or a set of clothing is usually input, but consumers or shop assistants usually present a dynamic situation in the retail store, so the conventional clothing identification method is not suitable for Streaming video recorded by Closed-Circuit Television (CCTV) when people in retail stores are crowded, angled or obscured, or, using Logo Recognition and Optical Character Recognition Recognition) The method of recognizing the logo or words on the uniform also has the same problem as the aforementioned clothing recognition technology, that is, the method of logo recognition or optical character recognition still needs to input a clear and relatively complete image, so The same does not apply to the field of retail stores.

鑑於上述問題,如何辨識較大之尺度的串流影片中之人物,特別是,即使人物於串流影片畫面上所顯示之尺寸相對較小或是遭到部分遮蔽,仍能取得相較豐富之相關資訊而進行識別,此將成為目前本技術領域人員急欲追求之目標。 In view of the above problems, how to identify characters in larger-scale streaming videos, especially, even if the characters displayed on the streaming video are relatively small in size or partially obscured, they can still obtain relatively rich Identifying relevant information will become the goal that those skilled in the art are eager to pursue.

為解決上述現有技術之問題,本創作係揭露一種身分辨識系統,係包括:資料處理模組,係用以接收影像資料,且利用多物件追蹤演算法追蹤該影像資料中之人物,以自該影像資料中對應該人物剪裁出多個人物影像而形成輸入資料;特徵擷取模組,係與該資料處理模組耦接,利用人物重識別模型對該輸入資料中該人物進行特徵向量之擷取;以及分類模組,係與該特徵擷取模組耦接,用以將該特徵向量輸入分類模型,判斷該人物之身分別。 In order to solve the above-mentioned problems of the prior art, the present invention discloses an identity recognition system, which includes: a data processing module, which is used for receiving image data, and uses a multi-object tracking algorithm to track the characters in the image data, so as to obtain the information from the image data. In the image data, a plurality of person images are cut out corresponding to the person to form input data; the feature extraction module is coupled with the data processing module, and uses the person re-identification model to extract the feature vector of the person in the input data and a classification module, which is coupled with the feature extraction module, and is used for inputting the feature vector into a classification model to determine the identity of the character.

於一實施例中,該資料處理模組係包括用於對該多個人物影像進行多方向取樣之姿勢方向估計模型,其中,該姿勢方向估計模型係對該多個人物影像進行方向辨識,以篩選出不同方向之人物影像而形成該輸入資料。 In one embodiment, the data processing module includes a posture direction estimation model for sampling the plurality of person images in multiple directions, wherein the posture direction estimation model performs direction recognition on the plurality of person images to The input data is formed by filtering out person images in different directions.

於另一實施例中,該不同方向之人物影像係包括該人物之前、後、左及/或右之方位。 In another embodiment, the image of the person in different directions includes the orientation of the person in front, behind, left and/or right.

於另一實施例中,該資料處理模組係對該多個人物影像進行人臉匿名模糊化,以形成該輸入資料。 In another embodiment, the data processing module anonymously fuzzes the faces of the plurality of human images to form the input data.

於另一實施例中,該特徵向量之擷取係包括取得該人物身上之制服、帽子及/或識別證的特徵向量。 In another embodiment, the extraction of the feature vector includes obtaining the feature vector of the uniform, hat and/or identification card on the person.

於又一實施例中,於模型訓練過程中,該資料處理模組係預先將該多個人物影像依據該人物穿著制服與否進行標記,以形成訓練樣本,俾供該人物重識別模型利用該訓練樣本進行訓練。 In another embodiment, during the model training process, the data processing module pre-marks the plurality of character images according to whether the character is wearing a uniform or not, so as to form a training sample for the character re-identification model to use the character image. training samples for training.

由上可知,本創作之身分辨識系統,係藉由資料處理模組對所接收之影像資料進行人物剪裁,以形成輸入資料,經特徵擷取模組對輸入資料進行人物之特徵向量擷取,使分類模組能依據特徵向量對人物進行分類及計數,據以透過制服辨識而判斷人物之身分別以達到人流計數之目的。 As can be seen from the above, the identity recognition system of this creation uses the data processing module to cut the characters of the received image data to form the input data, and the feature extraction module extracts the character feature vectors of the characters from the input data. The classification module can classify and count the characters according to the feature vector, so as to judge the body of the characters through uniform identification, so as to achieve the purpose of counting the flow of people.

1:身分辨識系統 1: Identity recognition system

11:資料處理模組 11: Data processing module

12:特徵擷取模組 12: Feature extraction module

13:分類模組 13: Classification module

S201~S207:步驟 S201~S207: Steps

S501~S507:流程 S501~S507: Process

U、NU:空間 U, NU: space

圖1係本創作之身分辨識系統之系統架構圖。 Figure 1 is a system architecture diagram of the identity recognition system of this creation.

圖2係本創作之身分辨識系統之人物重識別模型進行訓練之流程圖。 Figure 2 is a flowchart of the training of the character re-identification model of the identity recognition system of the present creation.

圖3係本創作之身分辨識系統進行分類訓練之示意圖。 Figure 3 is a schematic diagram of classification training performed by the identity recognition system of the present creation.

圖4A-4B係本創作之身分辨識系統進行分類推論之示意圖。 4A-4B are schematic diagrams of classification inference performed by the identity recognition system of the present creation.

圖5執行本創作之身分辨識系統的身分辨識方法之流程圖。 FIG. 5 is a flow chart of the identity recognition method for implementing the identity recognition system of the present invention.

以下藉由特定的具體實施形態說明本創作之技術內容,熟悉此技藝之人士可由本說明書所揭示之內容輕易地瞭解本創作之優點與功效。然本創作亦可藉由其他不同的具體實施形態加以施行或應用。 The technical content of the present creation is described below by means of specific embodiments, and those skilled in the art can easily understand the advantages and effects of the present creation from the content disclosed in this specification. However, the present invention can also be implemented or applied by other different specific implementation forms.

圖1為本創作之身分辨識系統之系統架構圖。如圖所示,本創作之身分辨識系統1係包括資料處理模組11、特徵擷取模組12以及分類模組13,其中,資料處理模組11用以自接收之影像資料中剪裁出人物影像,據以形成輸入資料,經特徵擷取模組12對輸入資料進行特徵向量擷取後,由分類模組13依據特徵向量判斷人物影像中人物之身分別,例如為消費者或店員,甚或零售店經理、活動宣傳模特、外送員,並進行人流計數,簡言之,透過特徵向量擷取來進行身分別之判斷,如何提取和辨識影像中人物之特徵成為關鍵,以下以消費者與店員之辨識為例進行說明。 Figure 1 is a system architecture diagram of the identity recognition system created by the author. As shown in the figure, the identity recognition system 1 of the present creation includes a data processing module 11, a feature extraction module 12 and a classification module 13, wherein the data processing module 11 is used to cut out characters from the received image data The image is used to form the input data. After the feature extraction module 12 extracts the feature vector of the input data, the classification module 13 determines the identity of the person in the person image according to the feature vector, such as a consumer or a store clerk, or even a Retail store managers, event promotion models, delivery staff, and people count. In short, through feature vector extraction to make identification judgments, how to extract and identify the characteristics of the characters in the image becomes the key. The identification of a store clerk is described as an example.

關於本創作之身分辨識系統1之說明,詳述如下。 The description of the identity recognition system 1 of the present creation is detailed as follows.

資料處理模組11係用以接收影像資料。在一實施例中,資料處理模組11係接收來自外部之閉路電視(CCTV)所錄製之串流影片作為影像資料,或是由本創作之身分辨識系統1所建置之資料庫中所儲存或攝影機所拍攝之影像資料。進言之,資料處理模組11係利用多物件追蹤(Multiple Object Tracking,MOT)技術,其中,多物件追蹤技術具體為多物件追蹤演算法,能針對影像資料中之人物(例如消費者或店員)進行追蹤,具體而言,資料處理模組11可藉由多物件追蹤演算法以框格框選人物之方式顯示對人物之追蹤,經多物件追蹤技術對人物進行追蹤後,將影像資料中對應所追蹤之人物進行裁剪,即依據框選人物之框格裁剪,由於影像資料係連續之串流影片,因而人物經裁剪後,將產生多個人物影像而形成輸入資料。 The data processing module 11 is used for receiving image data. In one embodiment, the data processing module 11 receives the streaming video recorded by the external closed-circuit television (CCTV) as the image data, or is stored in the database established by the identity recognition system 1 of the present creation or The image data captured by the camera. In other words, the data processing module 11 utilizes the Multiple Object Tracking (MOT) technology, wherein the multiple object tracking technology is specifically a multiple object tracking algorithm, which can target people in the image data (such as consumers or shop assistants) To perform tracking, specifically, the data processing module 11 can display the tracking of the person in the form of frame selection of the person by the multi-object tracking algorithm. After tracking the person through the multi-object tracking technology, the corresponding image data The tracked person is cropped, that is, according to the frame of the selected person. Since the image data is a continuous streaming video, after the person is cropped, multiple person images will be generated to form the input data.

於一實施例中,資料處理模組11係包括用以自多個人物影像中針對人物進行多方取樣之姿勢方向估計模型,由於單一人物經自影像資料中進行剪裁後,通常具有多種型態人物影像作為樣本,亦即多種型態人物影像會包含各類方向或角度的人物影像,因此,即便在連續或隨機取樣後仍可能取得在外觀上幾乎重複的樣本,是以,為求盡可能得到人物面對不同方向的樣本,故本創作採用姿勢方向估計(Pose Orientation Estimation,POE)模型進一步分析,使該資料處理模組11藉由姿勢方向估計模型先辨識出人物之方向,再對該多個人物影像進行多方向取樣,最後僅保留不同方向的人物影像,進而使經多方向取樣後的該些人物影像形成該輸入資料,亦即,本創作之多方向取樣係針對該人物之前、後、左及/或右之方位進行取樣,具體地,形成例如人物面向前方、後方、左方及/或右方之方向的樣本。據此,透過提供單一人物之多方位之人物影 像,藉以提升本創作之身分辨識系統1之辨識結果的正確性,同時排除多餘相似資料,能有助於減少數據運算量。 In one embodiment, the data processing module 11 includes a pose and orientation estimation model for multi-sampling a person from a plurality of person images, since a single person usually has multiple types of people after being cropped from the image data. Images are used as samples, that is, various types of human images will contain human images in various directions or angles. Therefore, even after continuous or random sampling, it is possible to obtain samples that are almost repeated in appearance. Therefore, in order to obtain as much as possible The characters face samples in different directions, so this creation adopts the pose orientation estimation (POE) model for further analysis, so that the data processing module 11 first recognizes the orientation of the character through the pose orientation estimation model, and then analyzes the pose orientation estimation model. Multi-directional sampling is carried out on the personal image, and finally only the personal images in different directions are retained, so that the input data is formed by the multi-directional sampling of these personal images, that is, the multi-directional sampling in this creation is for the before and after the character. , left and/or right orientation are sampled, in particular, samples are formed such as the orientation of the character facing forward, rear, left and/or right. Accordingly, by providing a multi-faceted person shadow of a single person image, so as to improve the accuracy of the identification result of the identity identification system 1 of the present creation, and at the same time exclude redundant similar data, which can help reduce the amount of data calculation.

於一實施例中,該資料處理模組11係對該多個人物影像進行人臉匿名模糊化,以形成該輸入資料,亦即,資料處理模組11對人物影像進行處理步驟中還包含對人物影像中之人臉部位進行模糊化,以使身分辨識系統1無需辨識臉部特徵,本創作藉由將人物影像中之人物的人臉部位模糊化,使得特徵擷取模組12於執行特徵向量擷取時,能專注於人臉以外之部份,例如衣服、帽子或識別證等,甚至是衣服上之標誌(LOGO)部分,使本創作之身分辨識系統1之辨識效果更加精確;另外,透過人物之人臉的模糊化,更能同時達到保護消費者隱私之目的。 In an embodiment, the data processing module 11 anonymously blurs the faces of the plurality of human images to form the input data, that is, the data processing module 11 processes the human images further comprising: The facial parts in the human image are blurred, so that the identity recognition system 1 does not need to recognize facial features. In this creation, the facial parts of the people in the human image are blurred, so that the feature extraction module 12 is When performing feature vector extraction, it can focus on parts other than the face, such as clothes, hats, identification cards, etc., and even the part of the logo (LOGO) on the clothes, so that the identification effect of the identity recognition system 1 of this creation is more accurate. ; In addition, through the blurring of the faces of the characters, the purpose of protecting the privacy of consumers can be achieved at the same time.

特徵擷取模組12係利用人物重識別(Person Re-identification)模型對該輸入資料中該人物進行特徵向量擷取,其中,人物之特徵向量之擷取係包括取得該人物身上之制服、LOGO(例如零售店之商標、店徽等)、帽子及/或識別證等可供辨識身分別(例如為店員或消費者之人物)的人臉以外之特徵向量。 The feature extraction module 12 uses a person re-identification (Person Re-identification) model to extract the feature vector of the person in the input data, wherein the extraction of the feature vector of the person includes obtaining the uniform and LOGO on the person. (For example, the trademark of a retail store, shop emblem, etc.), hats and/or identification cards, etc., feature vectors other than faces that can be used to identify a person (for example, a person of a store clerk or a consumer).

分類模組13係將經由人物重識別模型所擷取之特徵向量輸入至分類模型中,使分類模型依據特徵向量判斷該輸入資料所對應之人物是否穿著制服及其身分別,以進行分類,特別是本創作係分類出店員和非店員(消費者),並能於分類後進行人流計數,據以達到統計消費者之人數的目的。 The classification module 13 inputs the feature vector retrieved by the person re-identification model into the classification model, so that the classification model judges whether the person corresponding to the input data is wearing a uniform and their identity according to the feature vector, so as to classify, especially It is the creation system that classifies shop assistants and non-shop assistants (consumers), and can count the flow of people after the classification, so as to achieve the purpose of counting the number of consumers.

圖2為本創作之身分辨識系統之人物重識別模型進行訓練之流程圖。如圖所示,於身分辨識系統運作前,可預先對人物重識別模型進行訓練,除了建立人物重識別模型外,亦能對人物重識別模型作調整改良,俾以達到對影像資料進行正確且精準之消費者計數之效果。其訓練步驟如下所示。 Figure 2 is a flowchart of the training of the character re-identification model of the identity recognition system created by the author. As shown in the figure, before the operation of the identity recognition system, the person re-identification model can be trained in advance. In addition to establishing the person re-identification model, the person re-identification model can also be adjusted and improved, so as to achieve the correct and accurate image data. The effect of accurate consumer counting. Its training steps are as follows.

於步驟S201中,準備資料。本步驟除了能自本創作之身分辨識系統所建置之資料庫中取得預先儲存之影像資料外,亦可接收自本系統建置之攝影機或外部之CCTV裝置所擷取之影像資料,經由多物件追蹤演算法將該影像資料中的人物裁剪出來做為樣本。 In step S201, data is prepared. This step can not only obtain the pre-stored image data from the database built by the identity recognition system of this creation, but also receive the image data captured by the camera built in the system or the external CCTV device. The object tracking algorithm crops the people in the image data as samples.

於步驟S202中,多方向取樣。本步驟係將一位人物所對應之多個人物影像,經姿勢方向估計模型將人物方向辨識出來,以得到人物面對不同方向之人物影像以作為樣本。於一實施例中,多方向取樣之結果係可保留前、後、左、右之方向的樣本。 In step S202, multi-directional sampling is performed. In this step, a plurality of person images corresponding to a person are identified through the pose direction estimation model to identify the person's direction, so as to obtain the person images with the person facing different directions as samples. In one embodiment, the result of multi-direction sampling can retain samples in front, back, left, and right directions.

於步驟S203中,進行人臉匿名模糊。由於本創作欲使人物重識別模型著重於對人物之制服、LOGO、帽子或識別證等店員專有之特徵進行特徵擷取,因而將人臉進行匿名模糊化,其中,人臉匿名模糊化後對人物重識別的辨識能力幾乎沒有影響;再者,由於身分辨識不需要辨識人臉特徵,是以,為提升本創作之人物重識別模型於實際場域(例如於零售商店中)辨識的泛化性,故在處理輸入之資料(即樣本)時,可將人臉進行匿名模糊,一方面能保護消費者的隱私,另一方面可以讓人物重識別模型專注於服裝之特徵進行辨識,減少對人臉特徵的依賴。 In step S203, face anonymous blurring is performed. Since this creation intends to make the character re-identification model focus on the feature extraction of the unique features of the clerk, such as the uniform, LOGO, hat or identification card of the character, the face is anonymized and blurred. It has almost no effect on the recognition ability of person re-identification; furthermore, since identity recognition does not need to recognize facial features, therefore, in order to improve the general recognition of the character re-identification model of this creation in actual fields (such as in retail stores) Therefore, when processing the input data (ie samples), the face can be anonymous and blurred, on the one hand, it can protect the privacy of consumers; Dependence on facial features.

於步驟S204中,標記資料。本步驟係針對樣本中之人物有穿制服或沒穿制服作為條件進行標記,簡言之,為了區分樣本是否有穿制服或沒穿制服,進而推估該人物為店員與否,本步驟可先利用標記之方式,對樣本中之人物進行有穿制服或沒穿制服之標記。 In step S204, the data is marked. This step is to mark the character in the sample with or without uniform as a condition. In short, in order to distinguish whether the sample is wearing a uniform or not, and then estimate whether the character is a clerk or not, this step can be done first. Using the method of marking, the characters in the sample are marked with or without uniforms.

於步驟S205中,特徵擷取。本步驟係將樣本輸入至人物重識別模型中以擷取特徵向量,據之使人物重識別模型分別針對有穿制服或沒穿制服之人物進行特徵向量擷取,即可得到有穿制服之人物之特徵向量。 In step S205, feature extraction is performed. In this step, the sample is input into the character re-identification model to extract the feature vector, and the character re-identification model is used to extract the feature vector for the person with or without the uniform, and then the person with the uniform can be obtained. eigenvector of .

於步驟S206中,分類。本步驟將特徵向量輸入如支援向量機(Support Vector Machine,SVM)之分類模型中,以將樣本分成店員或消費者兩類別。另外,由於零售店中通常店員數遠比消費者數少,所取得之資料為不平衡資料(Imbalanced Data),故在訓練支援向量機時,採用類別權重設定(例如機器學習中sklearn.svm.SVC的平衡(balanced)模式的類別權重設定),以自動增加樣本較少的店員類別的權重,藉以避免人物重識別模型傾向將人物辨識成樣本較多的消費者類別。 In step S206, classify. In this step, the feature vector is input into a classification model such as a Support Vector Machine (SVM), so as to classify the samples into two categories of shop assistants or consumers. In addition, because the number of employees in a retail store is usually far less than the number of consumers, the obtained data is Imbalanced Data, so when training the support vector machine, the class weight setting is used (for example, sklearn.svm. The category weight setting of the balanced mode of SVC) to automatically increase the weight of the clerk category with fewer samples, so as to avoid the tendency of the person re-identification model to identify the person as the consumer category with more samples.

舉例而言,若用以訓練之影像資料中,具有1,000筆關於消費者之樣本以及100筆關於店員之樣本,此時,若將全部店員皆誤判成消費者,其錯誤率也僅為100/(1,000+100),約為9.1%,其錯誤率可能被忽略,對此,本創作增加店員之權重,亦即令消費者之權重與店員之權重的比例為1:10,使得1,000位消費者與100店員之重要性相同,據此,若遇所有店員皆遭誤判時,其錯誤率為100*10/(1,000*1+100*10)=50%,換言之,若所有店員皆遭誤判時,等同於誤判半數資料,增加了誤判之嚴重性,故本創作透過調整、設定權重以增加模型對於樣本數量相對較少之店員的樣本誤判之嚴重性,以避免誤判情況衍伸之問題。於一具體實施例中,本創作之設定權重能透過平衡(balanced)模式進行自動調整,亦即,平衡模式利用公式:總樣本數/(類別數*該類別的樣本數),以達到自動依據樣本數調整前述之權重之目的,舉例言之,承前述示例,則消費者權重=(1,000+100)/(2*1,000)=0.55,而店員權重=(1,000+100)/(2*100)=5.5,因此,消費 者權重與店員權重之比例為0.55:5.5即為1:10,同於前述示例,如此即能避免誤判所衍伸之模型失衡。 For example, if there are 1,000 samples of consumers and 100 samples of shop assistants in the image data used for training, if all shop assistants are misidentified as consumers, the error rate is only 100/ (1,000+100), about 9.1%, and its error rate may be ignored. For this reason, this creation increases the weight of shop assistants, that is, the ratio of the weight of consumers to the weight of shop assistants is 1:10, so that 1,000 consumers The importance of the 100 clerks is the same. According to this, if all the clerks are misjudged, the error rate is 100*10/(1,000*1+100*10)=50%, in other words, if all the clerks are misjudged , which is equivalent to misjudging half of the data, which increases the severity of the misjudgment. Therefore, this creation adjusts and sets the weight to increase the severity of the model's misjudgment for the sample of shop assistants with a relatively small number of samples, so as to avoid the problem of the extension of the misjudgment situation. In a specific embodiment, the set weight of this creation can be adjusted automatically through the balanced mode, that is, the balanced mode uses the formula: total number of samples/(number of categories * number of samples of this category) to achieve automatic basis The purpose of adjusting the aforementioned weight by the number of samples, for example, following the aforementioned example, the consumer weight=(1,000+100)/(2*1,000)=0.55, and the shop assistant weight=(1,000+100)/(2*100 )=5.5, therefore, consumption The ratio of the operator weight to the clerk weight is 0.55:5.5, which is 1:10, which is the same as the previous example, so that the model imbalance caused by misjudgment can be avoided.

於步驟S207中,權重優化。本步驟依據分類結果與真實類別(Ground Truth)計算損耗值(Loss),更新人物重識別模型和分類模型的權重(Weight),讓Loss變小,進而反覆訓練直到達到終止條件,其中,於真實類別計算上,t為樣本之真實類別,其可為1、-1,以分別代表店員(1)以及消費者(-1),當分類模型預測之結果為y=0.7,且該樣本的真實類別為t=1,則loss為:max(0,1-t*y)=max(0,1-1*0.7)=0.3,此表示若y愈接近t,則loss愈小。 In step S207, the weight is optimized. In this step, the loss value (Loss) is calculated according to the classification result and the ground truth, and the weight (Weight) of the person re-identification model and the classification model is updated to make the Loss smaller, and then the training is repeated until the termination condition is reached. In the category calculation, t is the real category of the sample, which can be 1, -1, to represent the clerk (1) and the consumer (-1) respectively, when the classification model predicts y=0.7, and the real category of the sample The category is t=1, then the loss is: max(0,1-t*y)=max(0,1-1*0.7)=0.3, which means that the closer y is to t, the smaller the loss.

接著,為使本創作之身分辨識系統之人物重辨識模型於訓練後能夠正確的分類,進行訓練如下。 Next, in order to make the character re-identification model of the identity recognition system of this creation can correctly classify after training, the training is carried out as follows.

如圖3所示,係本創作之身分辨識系統進行分類訓練之示意圖。為了讓執行身分辨識之分類模型訓練後能夠正確的分類,訓練的目標設定為訓練人物重識別模型的全身特徵擷取能力並保證以下條件為真:(a)穿著相似且為同一個人時,特徵向量應該要幾乎一樣;(b)穿著相似時,即使是不同個人,特徵向量應該要相近;(c)穿著不同時,特徵向量應該要不相近。 As shown in Figure 3, it is a schematic diagram of classification training performed by the identity recognition system of this creation. In order to enable the classification model for performing identity recognition to classify correctly after training, the training goal is set to train the full-body feature extraction capability of the person re-identification model and ensure that the following conditions are true: (a) When the clothes are similar and the same person, the characteristics The vectors should be almost the same; (b) when the clothes are similar, the eigenvectors should be similar, even for different individuals; (c) when the clothes are different, the eigenvectors should be dissimilar.

首先,將標記穿制服樣本(樣本1-3)以及沒穿制服樣本(樣本4-6)輸入人物重識別模型,以訓練人物重識別模型對人物之全身特徵的擷取能力,使得人物重識別模型對於穿著相似且為同一個人所擷取之特徵向量相同或近乎相同、穿著相似之不同個人所擷取之特徵向量相似以及穿著不同者所擷取之特徵向量不相近。 First, the marked samples in uniform (samples 1-3) and samples without uniforms (samples 4-6) are input into the person re-identification model to train the ability of the person re-identification model to capture the whole body features of the person, so that the person can be re-identified The model has the same or nearly the same eigenvectors captured by the same person in similar clothes, similar eigenvectors captured by different individuals in similar clothes, and dissimilar eigenvectors captured by different individuals.

之後,訓練分類模型對特徵向量進行分類,在前述條件分類下,讓有穿制服與沒穿制服對應的特徵向量會分別落在決策邊界所分割出來的空間U(Uniform,即有穿制服)與空間NU(Non-Uniform,即沒穿制服)。 Afterwards, the classification model is trained to classify the feature vectors. Under the aforementioned conditional classification, the feature vectors corresponding to those with uniforms and those without uniforms will fall into the space U (Uniform, that is, with uniforms) and Space NU (Non-Uniform, ie no uniform).

接著,訓練完分類模型後,即可進行推論,如圖4A和4B所示,係本創作之身分辨識系統進行分類推論之示意圖。本系統之身分辨識系統將輸入樣本進行多方向取樣及人臉匿名模糊,經人物重識別模型對輸入樣本推論出特徵向量,接著,利用分類模型對特徵向量進行分類,針對有穿制服和沒穿制服的樣本,如圖4A所示,若特徵向量落在空間U(Uniform),即人物為有穿制服,反之,若是落在空間NU(Non-Uniform),即人物為沒穿制服,如圖4B所示。 Next, after the classification model is trained, inference can be made, as shown in Figures 4A and 4B, which are schematic diagrams of classification inference by the identity recognition system of the present creation. The identity recognition system of this system samples the input samples in multiple directions and blurs the faces anonymously. The character re-identification model deduces the feature vectors of the input samples. Then, the classification model is used to classify the feature vectors. The uniform sample, as shown in Figure 4A, if the feature vector falls in the space U (Uniform), that is, the character is wearing a uniform, on the contrary, if it falls in the space NU (Non-Uniform), that is, the character is not wearing a uniform, as shown in the figure shown in 4B.

圖5為執行本創作之身分辨識系統的身分辨識方法之流程圖。如圖所示,本創作之身分辨識方法係於電腦、伺服器或雲端系統等電子設備中執行,具體而言,本創作之身分辨識方法係包括以下流程。 FIG. 5 is a flow chart of an identity recognition method for implementing the identity recognition system of the present invention. As shown in the figure, the identification method of this creation is executed in electronic devices such as computers, servers or cloud systems. Specifically, the identification method of this creation includes the following processes.

於流程S501中,接收影像資料。本流程係可自資料庫中取得其所儲存之影像資料,亦可接收自所設置之攝影機或CCTV裝置中所擷取之影像資料。 In the process S501, image data is received. This process can obtain the stored image data from the database, or receive the image data captured from the set camera or CCTV device.

於流程S502中,人物追蹤。本流程係透過多物件追蹤演算法追蹤該影像資料中之人物,以自將該影像資料中對應該人物剪裁出多個人物影像。 In the process S502, a person is tracked. This process uses a multi-object tracking algorithm to track a person in the image data, so as to cut out a plurality of person images from the image data corresponding to the person.

於流程S503中,多方向取樣。本流程係透過使用姿勢方向估計模型對所剪裁出來之多個人物影像進行多方向取樣,具體而言,多方向取樣係針對該人物之前、後、左及/或右之方位進行取樣,以依據經多方向取樣之該多個人物影像形成該輸入資料。 In the process S503, multi-directional sampling is performed. In this process, multi-directional sampling is performed on a plurality of cropped person images by using a pose direction estimation model. The plurality of person images sampled in multiple directions form the input data.

於流程S504中,人臉匿名模糊。本流程係將該多個人物影像進行人臉匿名模糊化,使人臉部位模糊化而無法進行辨識,據以形成該輸入資料,其目的除了保護個資外,亦能降低特徵擷取時的影響。 In the process S504, the face is anonymously blurred. This process is to anonymously blur the faces of the multiple human images, so that the face parts are blurred and cannot be identified, and the input data is formed accordingly. The purpose is not only to protect personal information, but also to reduce the time for feature extraction Impact.

於流程S505中,形成輸入資料。本流程係將上述經剪裁、多方向取樣以及人臉匿名模糊後之多個人物影像形成輸入資料。 In the process S505, input data is formed. In this process, the above-mentioned cropped, multi-directional sampling, and anonymously blurred face images are formed into input data.

於流程S506中,透過人物重識別模型對該輸入資料中該人物之特徵進行特徵向量擷取。前述之特徵向量擷取係包括對人物之制服、LOGO、帽子及/或識別證等進行向量擷取。 In the process S506, the feature vector of the character in the input data is extracted by the character re-identification model. The aforementioned feature vector extraction includes vector extraction of the characters' uniforms, logos, hats, and/or identification cards.

於流程S507中,透過分類模型依據該特徵向量,判斷該人物是否穿著制服及其身分別。本流程是利用分類模型來判斷人物影像其身分別為何,於本創作中,主要判斷人物為消費者或店員,接著還能進行人流計數。 In the process S507, according to the feature vector, it is determined whether the character is wearing a uniform and his identity through the classification model. This process is to use the classification model to determine the identity of the person's image. In this creation, it is mainly determined that the person is a consumer or a clerk, and then the flow of people can be counted.

於一實施例中,於步驟S506中,於利用人物重識別模型前,可先對人物重識別模型進行訓練,以獲得人物重識別模型或後續人物重識別模型之優化。具體而言,訓練步驟包括準備資料、多方向取樣、進行人臉匿名模糊、標記資料、特徵擷取、分類以及權重優化,其詳細內容已於圖2及對應段落說明,故不再贅述。 In one embodiment, in step S506, before using the person re-identification model, the person re-identification model may be trained first to obtain the person re-identification model or the subsequent optimization of the person re-identification model. Specifically, the training steps include data preparation, multi-directional sampling, face anonymous blurring, labeling data, feature extraction, classification, and weight optimization, the details of which have been described in Figure 2 and the corresponding paragraphs, so they will not be repeated here.

在一實施例中,本創作之身分辨識系統及方法係於實際應用場域進行應用,簡言之,於9間大型零售店進行訓練,其訓練樣本包括5,077個沒穿制服樣本以及755個有穿制服樣本,本創作使用經此訓練後之人物重識別模型,以於另外之零售店進行1,074個測試樣本,其中,平均精準率(Precision)、召回率(Recall)皆達到93%以上,詳如下面表1所示。 In one embodiment, the identity recognition system and method of the present creation is applied in a practical application field. In short, training is performed in 9 large-scale retail stores, and the training samples include 5,077 samples without uniforms and 755 samples with Uniform sample, this creation uses the character re-identification model after this training to conduct 1,074 test samples in other retail stores, among which the average precision and recall rate are over 93%. As shown in Table 1 below.

Figure 110215752-A0101-12-0012-1
Figure 110215752-A0101-12-0012-1

此外,本創作還揭示一種電腦可讀媒介,係應用於具有處理器(例如,CPU、GPU等)及/或記憶體的計算裝置或電腦中,且儲存有指令,並可利用此計算裝置或電腦透過處理器及/或記憶體執行此電腦可讀媒介,以於執行此電腦可讀媒介時執行上述之方法及各步驟。 In addition, the present invention also discloses a computer-readable medium, which is applied to a computing device or computer with a processor (eg, CPU, GPU, etc.) and/or memory, and stores instructions, and can utilize the computing device or computer. The computer executes the computer-readable medium through a processor and/or a memory, so as to execute the above-mentioned methods and steps when executing the computer-readable medium.

在一實施例中,本創作的模組、單元、裝置等包括微處理器及記憶體,而演算法、資料、程式等係儲存記憶體或晶片內,微處理器可從記憶體載入資料或演算法或程式進行資料分析或計算等處理。易言之,本創作之身分辨識系統可於電子設備上執行,例如一般電腦、平板或是伺服器,在收到影像資料後執行資料分析與運算,故身分辨識系統所進行程序,可透過軟體設計並架構在具有處理器、記憶體等元件之電子設備上,以於各類電子設備上運行;另外,亦可將身分辨識系統之各模組或單元分別以獨立元件組成,例如設計為計算器、記憶體、儲存器或是具有處理單元的韌體,皆可成為實現本創作之組件,而人物重識別模型或分類模型等相關模型,亦可選擇以軟體程式、硬體或韌體架構呈現。 In one embodiment, the modules, units, devices, etc. of the present invention include a microprocessor and a memory, and algorithms, data, programs, etc. are stored in a memory or a chip, and the microprocessor can load data from the memory. Or algorithms or programs for data analysis or calculation processing. In other words, the identity recognition system of this creation can be executed on electronic devices, such as ordinary computers, tablets or servers, and performs data analysis and calculation after receiving the image data. Designed and constructed on electronic equipment with processors, memory and other components to run on various electronic equipment; in addition, each module or unit of the identity recognition system can also be composed of independent components, such as designed to calculate A device, memory, storage or firmware with a processing unit can be used as components to realize this creation, and related models such as character re-identification model or classification model can also be constructed in software programs, hardware or firmware. render.

綜上所述,本創作之身分辨識系統,係於零售店內藉由辨識制服及相關特徵以區分消費者與店員,透過人物重識別針對人物之全身特徵進行特徵向量之擷取,再將特徵向量輸入到用以辨識制服之分類模型,以進行人物之身分別的判斷,例如該人物為消費者與店員;另外,於訓練過程中,針對樣本 進行標記,例如有穿制服和未穿制服,讓模型能學習並辨識店員的共同特徵像是制服、帽子、識別證等。因此,本創作解決了以往用臉部辨識無法辨識未註冊Face ID的店員,以及辨識率受口罩遮蔽、臉部角度影響之問題,易言之,即使有新進店員未在任何系統註冊過,但只要該新進人員穿著店員制服,即便在背對鏡頭的情況下仍可被身分辨識系統辨識出為店員。 To sum up, the identity recognition system of this creation distinguishes consumers and shop assistants by identifying uniforms and related features in retail stores, and extracts feature vectors for the whole body features of the characters through person re-identification, and then uses the features The vector is input into the classification model used to identify uniforms, so as to judge the characters, such as consumers and shop assistants; in addition, in the training process, for the sample Labeling, for example, uniformed and ununiformed, allows the model to learn and recognize common features of shop assistants such as uniforms, hats, badges, etc. Therefore, this creation solves the problems that face recognition cannot be used to identify store staff who have not registered Face ID in the past, and the recognition rate is affected by mask occlusion and face angle. In other words, even if there are new store staff who have not registered in any system, but As long as the new entrant is wearing a clerk uniform, he can still be identified as a clerk by the ID system even when his back is turned away from the camera.

上述實施例僅為例示性說明,而非用於限制本創作。任何熟習此項技藝之人士均可在不違背本創作之精神及範疇下,對上述實施例進行修飾與改變。因此,本創作之權利保護範圍係由本創作所附之申請專利範圍所定義,只要不影響本創作之效果及實施目的,應涵蓋於此公開技術內容中。 The above-mentioned embodiments are only illustrative, and are not intended to limit the present creation. Anyone skilled in the art can modify and change the above embodiments without departing from the spirit and scope of the present invention. Therefore, the scope of protection of the rights of this creation is defined by the scope of the patent application attached to this creation, and shall be included in the disclosed technical content as long as it does not affect the effect and implementation purpose of this creation.

1:身分辨識系統 1: Identity recognition system

11:資料處理模組 11: Data processing module

12:特徵擷取模組 12: Feature extraction module

13:分類模組 13: Classification module

Claims (6)

一種身分辨識系統,係包括: An identity recognition system, comprising: 資料處理模組,係用以接收影像資料,且利用多物件追蹤演算法追蹤該影像資料中之人物,以自該影像資料中對應該人物剪裁出多個人物影像而形成輸入資料; The data processing module is used to receive the image data, and use the multi-object tracking algorithm to track the person in the image data, so as to cut out a plurality of person images corresponding to the person from the image data to form the input data; 特徵擷取模組,係與該資料處理模組耦接,利用人物重識別模型對該輸入資料中該人物進行特徵向量之擷取;以及 a feature extraction module, coupled to the data processing module, to extract feature vectors of the person in the input data by using a person re-identification model; and 分類模組,係與該特徵擷取模組耦接,用以將該特徵向量輸入分類模型,判斷該人物之身分別。 The classification module is coupled to the feature extraction module, and is used for inputting the feature vector into the classification model to determine the identity of the person. 如請求項1所述之身分辨識系統,其中,該資料處理模組係包括用於對該多個人物影像進行多方向取樣之姿勢方向估計模型,其中,該姿勢方向估計模型係對該多個人物影像進行方向辨識,以篩選出不同方向之人物影像而形成該輸入資料。 The identity recognition system of claim 1, wherein the data processing module includes a posture direction estimation model for sampling the plurality of person images in multiple directions, wherein the posture direction estimation model is used for the plurality of The human image is subjected to direction identification, and the input data is formed by filtering out the human images in different directions. 如請求項2所述之身分辨識系統,其中,該不同方向之人物影像係包括該人物之前、後、左及/或右之方位。 The identity recognition system as claimed in claim 2, wherein the image of the person in different directions includes the front, back, left and/or right directions of the person. 如請求項1所述之身分辨識系統,其中,該資料處理模組係對該多個人物影像進行人臉匿名模糊化,以形成該輸入資料。 The identity recognition system as claimed in claim 1, wherein the data processing module anonymously fuzzes the faces of the plurality of person images to form the input data. 如請求項1所述之身分辨識系統,其中,該特徵向量之擷取係包括取得該人物身上之制服、帽子或識別證的特徵向量。 The identity recognition system according to claim 1, wherein the extraction of the feature vector includes obtaining the feature vector of the uniform, hat or identification card on the person. 如請求項1所述之身分辨識系統,其中,於模型訓練過程中,該資料處理模組係預先將該多個人物影像依據該人物穿著制服與否進行標記,以形成訓練樣本,俾供該人物重識別模型利用該訓練樣本進行訓練。 The identity recognition system according to claim 1, wherein, in the model training process, the data processing module pre-marks the plurality of character images according to whether the character is wearing a uniform or not, so as to form a training sample for the training sample. The person re-identification model is trained using this training sample.
TW110215752U 2021-12-30 2021-12-30 Identification system TWM629362U (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW110215752U TWM629362U (en) 2021-12-30 2021-12-30 Identification system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW110215752U TWM629362U (en) 2021-12-30 2021-12-30 Identification system

Publications (1)

Publication Number Publication Date
TWM629362U true TWM629362U (en) 2022-07-11

Family

ID=83437641

Family Applications (1)

Application Number Title Priority Date Filing Date
TW110215752U TWM629362U (en) 2021-12-30 2021-12-30 Identification system

Country Status (1)

Country Link
TW (1) TWM629362U (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115482486A (en) * 2022-09-09 2022-12-16 杭州海康威视数字技术股份有限公司 Passenger flow identification method and device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115482486A (en) * 2022-09-09 2022-12-16 杭州海康威视数字技术股份有限公司 Passenger flow identification method and device

Similar Documents

Publication Publication Date Title
CN110837784B (en) Examination room peeping and cheating detection system based on human head characteristics
CN109657609B (en) Face recognition method and system
CN105160318B (en) Lie detecting method based on facial expression and system
US10965886B2 (en) System and method of generating a composite frame
CN107679448B (en) Eyeball action-analysing method, device and storage medium
Baraldi et al. Gesture recognition in ego-centric videos using dense trajectories and hand segmentation
CN108229330A (en) Face fusion recognition methods and device, electronic equipment and storage medium
CN105809178A (en) Population analyzing method based on human face attribute and device
CN112001886A (en) A temperature detection method, device, terminal and readable storage medium
WO2022062379A1 (en) Image detection method and related apparatus, device, storage medium, and computer program
JP2013058060A (en) Person attribute estimation device, person attribute estimation method and program
CN114049378B (en) A queuing analysis method and apparatus
JP7183089B2 (en) Information processing device, information processing system, information processing method, and program
CN114743026A (en) Target object orientation detection method, device, equipment and computer readable medium
CN115482569A (en) Target passenger flow statistical method, electronic device and computer readable storage medium
WO2013075295A1 (en) Clothing identification method and system for low-resolution video
CN108229493A (en) Object verification method and device and electronic equipment
Budiman et al. Localization of white blood cell images using Haar cascade classifiers
CN113128373B (en) Image processing-based color spot scoring method, color spot scoring device and terminal equipment
Yang et al. Football referee gesture recognition algorithm based on YOLOv8s
CN111738199B (en) Image information verification method, device, computing device and medium
TWM629362U (en) Identification system
Teng et al. Facial expressions recognition based on convolutional neural networks for mobile virtual reality
TWI796072B (en) Identification system, method and computer readable medium thereof
Kishore et al. Spatial Joint features for 3D human skeletal action recognition system using spatial graph kernels