TWI748685B - Method and server for optimizing joint object detection model - Google Patents
Method and server for optimizing joint object detection model Download PDFInfo
- Publication number
- TWI748685B TWI748685B TW109135316A TW109135316A TWI748685B TW I748685 B TWI748685 B TW I748685B TW 109135316 A TW109135316 A TW 109135316A TW 109135316 A TW109135316 A TW 109135316A TW I748685 B TWI748685 B TW I748685B
- Authority
- TW
- Taiwan
- Prior art keywords
- detection
- object detection
- model
- classification
- image data
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 449
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000013145 classification model Methods 0.000 claims abstract description 77
- 238000012549 training Methods 0.000 claims abstract description 44
- 238000002372 labelling Methods 0.000 claims abstract description 29
- 230000004044 response Effects 0.000 claims description 12
- 238000011895 specific detection Methods 0.000 claims 1
- 238000012552 review Methods 0.000 abstract description 8
- 238000013473 artificial intelligence Methods 0.000 abstract description 7
- 238000012216 screening Methods 0.000 abstract description 2
- 230000007246 mechanism Effects 0.000 description 13
- 238000013480 data collection Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000013136 deep learning model Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012423 maintenance Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003137 locomotive effect Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Landscapes
- Image Analysis (AREA)
Abstract
Description
本發明是有關於一種訓練物件偵測模型的技術,且特別是有關於一種種優化聯合物件偵測模型的方法及伺服器。 The present invention relates to a technique for training an object detection model, and particularly relates to a method and server for optimizing a joint object detection model.
在影像監控應用上,結合電腦視覺之深度學習(deep learning)技術已是發展趨勢,且其中應用最為廣泛的是監督學習,即模型訓練過程的所有數據都需要事先正確標註。然而要產生預測準確率高的深度學習模型,需要收集大量的影像標註訓練樣本。目前影像標註的方法大多採取人工標註,標註者需針對影像資料中的目標物件逐一框選物件範圍與標示關聯的物件類別名稱,並於標註後進行審查,以確保標註內容的準確性,如此標註圖資流程相當費時費力。 In the application of image monitoring, deep learning technology combined with computer vision has become a development trend, and the most widely used is supervised learning, that is, all data in the model training process needs to be correctly labeled in advance. However, to produce a deep learning model with high prediction accuracy, a large number of image annotation training samples need to be collected. At present, most of the methods of image labeling are manual labeling. The labeler needs to select the object range and the object category name associated with the label one by one for the target object in the image data, and review it after labeling to ensure the accuracy of the labeling content. Such labeling The drawing and funding process is time-consuming and labor-intensive.
在影像監控應用領域日漸增加的各種人工智慧影像辨識場景與應用需求,若只依據固定的標註圖資集所訓練的深度學習模型,缺乏數據多樣化,且往往不易適用現實場景與應用的需求。若需要根據現實場景的需求構建專屬的圖資集,實現對特定類別 的分類,需耗費相當的標註與校驗人力來維護。若建置後必須依需求新增一至多項新的物件特徵類別到既有的應用模型中,對原訓練圖資集的新增標註成本亦相當可觀。 In the field of image surveillance applications, various artificial intelligence image recognition scenarios and application requirements are increasing. If only a deep learning model trained based on a fixed set of annotation maps, there is a lack of data diversification, and it is often difficult to adapt to the needs of real scenarios and applications. If you need to build an exclusive set of graphics resources according to the needs of the real scene, realize the specific category The classification requires considerable labeling and verification manpower to maintain. If one or more new object feature categories must be added to the existing application model after construction, the cost of adding new annotations to the original training image collection is also considerable.
此外,習知之自動標註方法多半是為了保證標註結果的客觀性和準確性,但該自動標註的模型在其訓練過程中,需要對標註結果進行篩檢,仍存在大量人工對標註資料審查的成本。另外,對於找出最佳模型的訓練過程,大多依靠既有的訓練資料集進行,但在實務應用中,往往需要針對不同的環境和需求,對不同的類別做偵測或分類,而且類別多樣且繁複,故需構建或擴充包含現場實際資料的訓練資料集,以實現適用之模型對特定類別的分類。 In addition, most of the conventional automatic labeling methods are to ensure the objectivity and accuracy of the labeling results, but the automatic labeling model needs to screen the labeling results during its training process, and there is still a large amount of manual review of the cost of labeling data. . In addition, the training process for finding the best model mostly relies on existing training data sets. However, in practical applications, it is often necessary to detect or classify different categories according to different environments and needs, and the categories are diverse And it is complicated, so it is necessary to construct or expand the training data set containing the actual data on the spot to realize the classification of specific categories by the applicable model.
另外,習知技術對於已達預測精準度目標之模型,遇到變更偵測物件的要求,在時程與成本有限的條件下,不易再將龐大訓練圖資全部檢視,逐一標註出新物件類別後重新精練模型,來為新變更的應用需求提供新物件特徵組合的偵測模型。因此,實務上變通方式會從既有的物件偵測模型中以串連或並接的方式組合出所需要偵測的物件類別,然而,此舉在初期雖然可以省略重新標註圖資與訓練模型的成本,並快速上線應用,但缺點是系統資源需求高且運作效率差,因上線系統的資源有限,加上圖像處理單元(GPU)推論設備的擴充成本高,後續擴增的維運成本與頻繁客訴的壓力,非長久上線應用之良策。 In addition, the conventional technology meets the requirement of changing detected objects for models that have reached the target of prediction accuracy. Under the conditions of limited time and cost, it is not easy to inspect all the huge training images and mark new object categories one by one. Then refine the model again to provide a detection model of new object feature combinations for newly changed application requirements. Therefore, in practice, the workaround is to combine the types of objects that need to be detected from the existing object detection models in a serial or parallel manner. However, this can omit the re-labeling and training of the model in the initial stage. Cost, and quickly launch the application, but the disadvantage is that the system resource requirements are high and the operation efficiency is poor. Because the resources of the online system are limited, and the expansion cost of the image processing unit (GPU) inference equipment is high, the maintenance cost of subsequent expansion is also The pressure of frequent customer complaints is not a good strategy for long-term online applications.
有鑑於此,本發明提供一種優化聯合物件偵測模型的方法及伺服器,其可用於解決上述技術問題。 In view of this, the present invention provides a method and server for optimizing a joint object detection model, which can be used to solve the above technical problems.
本發明提供一種優化聯合物件偵測模型的方法,包括:取得一圖資庫,其中圖資庫包括一第一未標註圖資及多個標註圖資集,各標註圖資集對應於單一類物件;以所述多個標註圖資集訓練多個單物件偵測模型及多個單物件分類模型,其中所述多個標註圖資集一對一地對應於所述多個單物件偵測模型,且所述多個標註圖資集一對一地對應於所述多個單物件分類模型;取得多個指定偵測物件類別,並據以取得多個外部物件偵測模型,其中各外部物件偵測模型用於偵測屬於所述多個指定偵測物件類別的至少其中之一的物件;反應於判定各偵測物件類別對應於所述多個標註圖資集的其中之一,從所述多個單物件偵測模型找出對應於所述多個指定偵測物件類別的多個特定單物件偵測模型;在所述多個標註圖資集中找出對應於所述多個指定偵測物件類別的多個特定標註圖資集,並據以訓練一聯合物件偵測模型;使用所述多個特定單物件偵測模型、所述多個外部物件偵測模型及聯合物件偵測模型對第一未標註圖資進行偵測,以產生多個第一物件偵測結果,其中所述多個第一物件偵測結果對應於所述多個指定偵測物件類別中的一第一指定偵測物件類別;使用所述多個單物件分類模型中對應於第一指定偵測物件類別的一第一單物件分類模型對各第一物件偵測結果進行預測分類以得到各第一物件偵測結 果的一第一物件分類結果;基於各第一物件偵測結果及對應的第一物件分類結果適應性地修正所述多個第一物件偵測結果,並將修正後的所述多個第一物件偵測結果加入所述多個標註圖資集中對應於第一指定偵測物件類別的一第一標註圖資集;基於第一標註圖資集重新訓練所述多個單物件偵測模型中對應於第一指定偵測物件類別的一第一單物件偵測模型、第一單物件分類模型及聯合物件偵測模型。 The present invention provides a method for optimizing a joint object detection model, including: obtaining a map database, wherein the map database includes a first unlabeled map asset and a plurality of annotated map asset sets, and each annotated map asset set corresponds to a single category Objects; training multiple single-object detection models and multiple single-object classification models with the multiple annotated image data sets, wherein the multiple annotated image data sets correspond to the multiple single-object detections one-to-one Model, and the plurality of annotated image data sets correspond to the plurality of single object classification models one-to-one; obtain a plurality of designated detection object categories, and obtain a plurality of external object detection models accordingly, wherein each external object The object detection model is used to detect objects belonging to at least one of the plurality of designated detection object categories; in response to determining that each detection object category corresponds to one of the plurality of annotated image data sets, from The plurality of single-object detection models find out a plurality of specific single-object detection models corresponding to the plurality of designated detection object categories; and find out the plurality of specific single-object detection models corresponding to the plurality of designated detection objects in the plurality of annotated image data sets Detect a plurality of specific annotation map data sets of object types, and train a joint object detection model accordingly; use the plurality of specific single object detection models, the plurality of external object detection models, and joint object detection The model detects the first unlabeled image data to generate a plurality of first object detection results, wherein the plurality of first object detection results correspond to a first of the plurality of specified detection object categories Specify the detection object category; use a first single object classification model corresponding to the first specified detection object category among the plurality of single object classification models to predict and classify the detection results of each first object to obtain each first object Detection knot A first object classification result of the result; based on each first object detection result and the corresponding first object classification result, the plurality of first object detection results are adaptively corrected, and the corrected plurality of first object An object detection result is added to a first annotated image data set corresponding to the first specified detection object category in the plurality of annotated image data sets; the multiple single-object detection models are retrained based on the first annotated image data set A first single-object detection model, a first single-object classification model, and a joint object detection model corresponding to the first designated detection object category.
本發明提供一種優化聯合物件偵測模型的伺服器,包括儲存電路及處理器。儲存電路儲存多個模組。處理器耦接儲存電路,存取所述多個模組以執行下列步驟:取得一圖資庫,其中圖資庫包括一第一未標註圖資及多個標註圖資集,各標註圖資集對應於單一類物件;以所述多個標註圖資集訓練多個單物件偵測模型及多個單物件分類模型,其中所述多個標註圖資集一對一地對應於所述多個單物件偵測模型,且所述多個標註圖資集一對一地對應於所述多個單物件分類模型;取得多個指定偵測物件類別,並據以取得多個外部物件偵測模型,其中各外部物件偵測模型用於偵測屬於所述多個指定偵測物件類別的至少其中之一的物件;反應於判定各偵測物件類別對應於所述多個標註圖資集的其中之一,從所述多個單物件偵測模型找出對應於所述多個指定偵測物件類別的多個特定單物件偵測模型;在所述多個標註圖資集中找出對應於所述多個指定偵測物件類別的多個特定標註圖資集,並據以訓練一聯合物件偵測模型;使用所述多個特定單物件偵測模 型、所述多個外部物件偵測模型及聯合物件偵測模型對第一未標註圖資進行偵測,以產生多個第一物件偵測結果,其中所述多個第一物件偵測結果對應於所述多個指定偵測物件類別中的一第一指定偵測物件類別;使用所述多個單物件分類模型中對應於第一指定偵測物件類別的一第一單物件分類模型對各第一物件偵測結果進行預測分類以得到各第一物件偵測結果的一第一物件分類結果;基於各第一物件偵測結果及對應的第一物件分類結果適應性地修正所述多個第一物件偵測結果,並將修正後的所述多個第一物件偵測結果加入所述多個標註圖資集中對應於第一指定偵測物件類別的一第一標註圖資集;基於第一標註圖資集重新訓練所述多個單物件偵測模型中對應於第一指定偵測物件類別的一第一單物件偵測模型、第一單物件分類模型及聯合物件偵測模型。 The invention provides a server for optimizing a joint object detection model, which includes a storage circuit and a processor. The storage circuit stores multiple modules. The processor is coupled to the storage circuit and accesses the plurality of modules to perform the following steps: obtain a map database, where the map database includes a first unlabeled map resource and a plurality of labeled map resource sets, each of which is labeled A set corresponds to a single type of object; multiple single-object detection models and multiple single-object classification models are trained with the multiple annotated image data sets, wherein the multiple annotated image data sets correspond to the multiple one-to-one A single-object detection model, and the plurality of label map data sets correspond to the plurality of single-object classification models one-to-one; obtain a plurality of designated detection object categories, and obtain a plurality of external object detections accordingly Model, wherein each external object detection model is used to detect objects belonging to at least one of the plurality of specified detection object categories; in response to determining that each detection object category corresponds to the plurality of annotated image data sets One of them is to find a plurality of specific single-object detection models corresponding to the plurality of designated detection object categories from the plurality of single-object detection models; The plurality of specific annotation map data sets of the plurality of designated detection object categories are used to train a joint object detection model; the plurality of specific single object detection models are used Type, the plurality of external object detection models and the joint object detection model detect the first unlabeled image to generate a plurality of first object detection results, wherein the plurality of first object detection results Corresponding to a first specified detection object category in the plurality of specified detection object categories; using a first single object classification model pair corresponding to the first specified detection object category among the plurality of single object classification models Predictive classification is performed on each first object detection result to obtain a first object classification result of each first object detection result; based on each first object detection result and the corresponding first object classification result, the multiple A first object detection result, and adding the corrected detection results of the plurality of first objects to a first annotation image data set corresponding to the first designated detection object category in the plurality of annotated image data sets; Retraining a first single-object detection model, a first single-object classification model, and a joint object detection model corresponding to the first designated detection object category among the plurality of single-object detection models based on the first annotation image data set .
100:伺服器 100: server
102:儲存電路 102: storage circuit
104:處理器 104: processor
310:圖資集 310: Picture Collection
311:整合標註圖資集 311: Integrate annotated map data collection
GA~GH,GN:標註圖資集 GA ~ GH, GN: Annotated image collection
A 0~E 0,N 0:單物件偵測模型 A 0 ~ E 0 , N 0 : Single object detection model
A’~E’,N’:單物件分類模型 A’~E’,N’: Single object classification model
324:聯合偵測模型 324: Joint Detection Model
341~343:偵測結果 341~343: Detection result
S210~S290:步驟 S210~S290: steps
圖1是依據本發明之一實施例繪示的優化聯合物件偵測模型的伺服器示意圖。 FIG. 1 is a schematic diagram of a server for optimizing a joint object detection model according to an embodiment of the present invention.
圖2是依據本發明之一實施例繪示的優化聯合物件偵測模型的方法流程圖。 FIG. 2 is a flowchart of a method for optimizing a joint object detection model according to an embodiment of the present invention.
圖3是依據本發明之一實施例繪示的應用情境圖。 Fig. 3 is an application scenario diagram drawn according to an embodiment of the present invention.
概略而言,為解決上述人工智慧影像辨識應用需求所面臨的問題,本發明包含了建立單物件標註圖資集與單物件偵測模型與單物件分類模型機制、影像預測標註機制、多模型預測結果比對機制等方法組合來達成較佳之應用目標。 Generally speaking, in order to solve the above-mentioned problems faced by the application requirements of artificial intelligence image recognition, the present invention includes the establishment of single-object annotation map resource collection and single-object detection model and single-object classification model mechanism, image prediction and annotation mechanism, and multi-model prediction The result comparison mechanism and other methods are combined to achieve better application goals.
在擴充標註圖資集特徵上,本發明可統一儲存拆解與融合後之外部匯入之標註圖資集與收集外部物件偵測模型之可偵測物件特徵類別之特徵標註內容。 In terms of expanding the features of annotated image collections, the present invention can uniformly store the externally imported annotated image collections after disassembly and fusion and collect the feature annotation content of the detectable object feature categories of the external object detection model.
在多模型比對與訓練效率特徵上,建立單物件特徵的物件偵測模型進行預測標註與修正再訓練的效果,相較於一次標註多種物件特徵圖資的標註和訓練效果,前者效率更好,且人工複審與修正的預測標註資料的品質更佳。 In terms of multi-model comparison and training efficiency characteristics, the effect of establishing a single-object feature object detection model for predictive labeling and correction retraining is better than the labeling and training effects of labeling multiple object feature maps at once. , And the quality of the artificially reviewed and revised forecast annotation data is better.
在臨場應用與組合目標偵測模型特徵上,因單物件偵測模型可較快建立並產生新案場圖資的預測標註資料,可彈性搭配適合案場所需的多物件特徵組合,並可進一步建立與合併新案場之訓練圖資與物件偵測模型。 In the field application and the combination of target detection model features, the single-object detection model can be established quickly and generate the prediction annotation data of the new case map, which can be flexibly matched with the combination of multiple object features required by the case, and can be further established Combine the training graphics and object detection model of the new case.
在模型預測物件類別名稱之自動審查特徵上,可利用標註物件邊界框內之區域影像為訓練圖資,透過不同之演算法與參數,訓練該單物件特徵分類模型以用來即時對預測標註資料之物件類別進行審查,具有減少人力複審成本的效果。 In the automatic review feature of the model predicting the object category name, the regional image within the bounding box of the labeled object can be used as the training image. Through different algorithms and parameters, the single-object feature classification model can be trained for real-time labeling of the predicted data. It has the effect of reducing the cost of manpower review.
在精進人工智慧物件偵測模型迭代週期的工作方法與工作流程特徵上,以上述特徵與利用多模型預測結果比對機制,有著訓練模型更快收斂、標註圖資快速且易多樣化、準確度較高、 預測速度較快、節省標註與審查人力的優勢,讓模型精練更有效率,亦可從案場影像中快速挑出特定的物件影像,可有效改善資料偏頗問題。以下將針對本發明提供詳細說明。 In the work method and workflow characteristics of the iterative cycle of the refined artificial intelligence object detection model, the above-mentioned characteristics and the use of the multi-model prediction result comparison mechanism have the training model to converge faster, the annotation map data is fast and easy to diversify, and the accuracy is Higher, The advantages of faster prediction speed, saving annotation and review manpower, make the model more efficient, and can also quickly pick out specific object images from the scene images, which can effectively improve the problem of data bias. A detailed description of the present invention will be provided below.
請參照圖1,其是依據本發明之一實施例繪示的優化聯合物件偵測模型的伺服器示意圖。在不同的實施例中,優化聯合物件偵測模型的伺服器100例如是各式電腦裝置及智慧型裝置,但可不限於此。如圖1所示,伺服器100包括儲存電路102及處理器104。
Please refer to FIG. 1, which is a schematic diagram of a server that optimizes a joint object detection model according to an embodiment of the present invention. In different embodiments, the
儲存電路102例如是任意型式的固定式或可移動式隨機存取記憶體(Random Access Memory,RAM)、唯讀記憶體(Read-Only Memory,ROM)、快閃記憶體(Flash memory)、硬碟或其他類似裝置或這些裝置的組合,而可用以記錄多個程式碼或模組。
The
處理器104耦接於儲存電路102,並可為一般用途處理器、特殊用途處理器、傳統的處理器、數位訊號處理器、多個微處理器(microprocessor)、一個或多個結合數位訊號處理器核心的微處理器、控制器、微控制器、特殊應用積體電路(Application Specific Integrated Circuit,ASIC)、現場可程式閘陣列電路(Field Programmable Gate Array,FPGA)、任何其他種類的積體電路、狀態機、基於進階精簡指令集機器(Advanced RISC Machine,ARM)的處理器以及類似品。
The
在本發明的實施例中,處理器104可存取儲存電路102
中記錄的模組、程式碼來實現本發明提出的優化聯合物件偵測模型的方法,其細節詳述如下。
In the embodiment of the present invention, the
請參照圖2,其是依據本發明之一實施例繪示的優化聯合物件偵測模型的方法流程圖。本實施例的方法可由圖1的伺服器100執行,以下即搭配圖1所示的元件說明圖2各步驟的細節。此外,為便於理解本發明的概念,以下將另以圖3為例進行說明,其中圖3是依據本發明之一實施例繪示的應用情境圖。
Please refer to FIG. 2, which is a flowchart of a method for optimizing a joint object detection model according to an embodiment of the present invention. The method of this embodiment can be executed by the
首先,在步驟S210中,處理器104可取得圖資庫310,其中此圖資庫例如可包括多個未標註圖資及多個標註圖資集GA~GH及GN,且標註圖資集GA~GH及GN個別可對應於單一類物件。舉例而言,標註圖資集GA例如可僅包括以邊界框(bounding box)或其他標註方式標註有「物件A」(例如人)的各式圖片,標註圖資集GB例如可僅包括標註有「物件B」(例如機車)的各式圖片,標註圖資集GC例如可僅包括標註有「物件C」(例如自行車)的各式圖片,而其餘標註圖資集的內容應可依此類推,於此不另贅述。
First, in step S210, the
之後,在步驟S220中,處理器104可以所述多個標註圖資集訓練多個單物件偵測模型及多個單物件分類模型,其中所述多個標註圖資集GA~GH及GN一對一地對應於所述多個單物件偵測模型,且所述多個標註圖資集一對一地對應於所述多個單物件分類模型。
After that, in step S220, the
舉例而言,處理器104可以標註圖資集GA來訓練用於
對「物件A」進行偵測的單物件偵測模型A 0,以及訓練用於對「物件A」進行分類的單物件分類模型A’。舉另一例而言,處理器104可以標註圖資集B來訓練用於對「物件B」進行偵測的單物件偵測模型B 0,以及訓練用於對「物件B」進行分類的單物件分類模型B’。舉又一例而言,處理器104可以標註圖資集C來訓練用於對「物件C」進行偵測的單物件偵測模型C 0,以及訓練用於對「物件C」進行分類的單物件分類模型C’。換言之,對於圖資庫310中的各個標註圖資集而言,處理器104皆可用以訓練對應的單物件偵測模型及單物件分類模型。
For example, the
應了解的是,本發明所提及的各式模型皆為人工智慧模型,且可採用各式設計者偏好的模型(例如各式神經網路等)實現,但可不限於此。 It should be understood that the various models mentioned in the present invention are all artificial intelligence models, and can be implemented using various models preferred by the designer (for example, various neural networks, etc.), but may not be limited thereto.
在步驟S230中,處理器104可取得多個指定偵測物件類別,並據以取得多個外部物件偵測模型ext1~ext3,其中外部物件偵測模型ext1~ext3個別可用於偵測屬於所述多個指定偵測物件類別的至少其中之一的物件。
In step S230, the
在一實施例中,假設使用者欲使用經本發明的方法訓練而得的聯合物件偵測模型來對某個特定場域(例如某道路、路口等)進行一或多種物件的辨識的話,使用者可依據所欲偵測的物件類別來設定上述指定偵測物件類別。例如,若使用者欲讓聯合物件偵測模型具有偵測屬於「物件A」、「物件B」、「物件C」、「物件D」(例如狗)、「物件E」(例如小客車)及「物件N」(例如騎 士)等物件類別之物件的能力,則使用者可將「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」設定為上述指定偵測物件類別,但可不限於此。 In one embodiment, if the user wants to use the joint object detection model trained by the method of the present invention to identify one or more objects in a specific field (such as a road, intersection, etc.), the user The above-mentioned designated detection object type can be set according to the object type to be detected. For example, if the user wants the joint object detection model to detect objects belonging to "Object A", "Object B", "Object C", "Object D" (such as a dog), "Object E" (such as a car), and "Object N" (e.g. ride 士) and other object types, the user can set "Object A", "Object B", "Object C", "Object D", "Object E" and "Object N" to the above specified detection Object category, but not limited to this.
另外,在一實施例中,上述外部物件偵測模型個別可具有一次偵測多種物件類別的能力。舉例而言,外部物件偵測模型ext1例如可具有偵測「物件A」、「物件B」、「物件K」及「物件S」的能力。舉另一例而言,外部物件偵測模型ext2例如可具有偵測「物件C」、「物件G」及「物件M」的能力。此外,外部物件偵測模型ext3例如可具有偵測「物件A」、「物件D」、「物件E」及「物件X」的能力。由上可知,外部物件偵測模型ext1~ext3個別皆具有偵測屬於所述多個指定偵測物件類別的至少其中之一的物件的能力。 In addition, in one embodiment, the above-mentioned external object detection model may individually have the ability to detect multiple object types at one time. For example, the external object detection model ext1 can detect "object A", "object B", "object K" and "object S". For another example, the external object detection model ext2 can detect "object C", "object G" and "object M", for example. In addition, the external object detection model ext3 can detect "object A", "object D", "object E" and "object X", for example. It can be seen from the above that the external object detection models ext1 to ext3 each have the ability to detect objects belonging to at least one of the plurality of specified detection object categories.
另外,在一些實施例中,處理器104於步驟S230中挑選的外部物件偵測模型也可被要求符合一或多個條件,例如「物件預測區域邊框與原標註區域邊框交疊率(Intersection over Union)>交疊率目標閾值」、「單一物件特徵類別平均精度(Average Precision)>單物件平均精度目標閾值」、「模型準確度(accuracy)>模型準確度目標閾值」、「模型損失函數收斂結果值<模型損失函數目標閾值」等,但可不限於此。
In addition, in some embodiments, the external object detection model selected by the
之後,在一實施例中,處理器104可判斷各指定偵測物件類別是否對應於所述多個標註圖資集的其中之一。
After that, in an embodiment, the
在圖3情境中,由於「物件A」、「物件B」、「物件C」、「物
件D」、「物件E」及「物件N」分別對應於標註圖資集GA~GH及GN,故處理器104可判定各指定偵測物件類別皆對應於所述多個標註圖資集的其中之一,並可接續在步驟S240中從所述多個單物件偵測模型找出對應於所述多個指定偵測物件類別的多個特定單物件偵測模型。
In the scenario in Figure 3, due to "object A", "object B", "object C", "object
"Item D", "Object E", and "Object N" respectively correspond to the annotated image data sets GA~GH and GN, so the
在圖3中,處理器104可將對應於「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」的單物件偵測模型A 0~E 0及N 0作為上述特定單物件偵測模型,但可不限於此。
In FIG. 3, the
然而,在其他實施例中,由於使用者可能會設定未對應於任何標註圖資集的指定偵測物件類別(下稱參考指定偵測物件類別),因此本發明可透過所提出的圖資新增機制來新增對應於參考指定偵測物件類別的標註圖資集(下稱參考標註圖資集),並據以訓練對應的單物件偵測模型(下稱參考單物件偵測模型)及單物件分類模型(下稱參考單物件偵測模型)。 However, in other embodiments, since the user may set a designated detection object category that does not correspond to any labeled map data set (hereinafter referred to as the designated detection object category), the present invention can use the proposed map data update Added mechanism to add annotated image data set corresponding to the reference specified detection object category (hereinafter referred to as reference annotated image data set), and train the corresponding single object detection model (hereinafter referred to as reference single object detection model) and Single object classification model (hereinafter referred to as reference single object detection model).
在一實施例中,在處理器104判定所述多個指定偵測物件類別中的參考指定偵測物件類別(例如「大客車」)未對應於標註圖資集GA~GH及GN的任一時,處理器104可在圖資庫310中新增對應於參考指定偵測物件類別的參考標註圖資集。之後,處理器104可從圖資庫310取得多筆第一圖資,並在各第一圖資中針對對應於參考指定偵測物件類別的特定類物件(例如大客車)進行標註,以產生多筆第一參考標註圖資。在一些實施例中,上述標註行為可由人工對所述第一圖資執行,但可不限於此。之後,
處理器104可基於上述第一參考標註圖資訓練參考單物件偵測模型及參考單物件分類模型。
In one embodiment, when the
在一實施例中,反應於判定所述參考單物件偵測模型及所述參考單物件分類模型個別滿足上述目標條件,處理器104可將所述參考單物件分類模型新增至單物件偵測模型中,並將參考單物件分類模型新增至上述單物件分類模型及上述特定單物件偵測模型中。
In one embodiment, in response to determining that the reference single object detection model and the reference single object classification model meet the aforementioned target conditions, the
在另一實施例中,反應於判定參考單物件偵測模型及參考單物件分類模型皆未滿足上述目標條件,且圖資庫310仍有未針對上述特定類物件(例如大客車)進行標註的多筆第二圖資,處理器104可經配置以:以參考單物件偵測模型對各第二圖資進行標註,以得到多個第二參考標註圖資;修正上述第二參考標註圖資,並將修正後的第二參考標註圖資新增至參考標註圖資集;基於參考標註圖資集再次訓練參考單物件偵測模型及參考單物件分類模型。處理器104可持續重複上述步驟,直至所述參考單物件偵測模型及所述參考單物件分類模型個別滿足上述目標條件,但可不限於此。在一些實施例中,修正各第二參考標註圖資的方式可包括但不限於:調整各第二參考標註圖資中用於標註特定類物件的邊界框位置/大小、刪除非必要之重複邊界框、調整特定類物件所在區域邊框或調整物件邊緣輪廓。
In another embodiment, it is reflected in the determination that the reference single object detection model and the reference single object classification model do not meet the above-mentioned target conditions, and the
舉例而言,在一實施例中,假設圖資庫310中原本不存在標註圖資集GN,則處理器104可執行上述圖資集新增機制以在
圖資庫310中新增對應於「物件N」的標註圖資集GN。之後,處理器104可基於標註圖資集GN訓練單物件偵測模型N 0及單物件分類模型N’。並且,在單物件偵測模型N 0及單物件分類模型N’個別滿足上述目標條件之後,處理器104可將單物件偵測模型N 0亦視為特定單物件偵測模型,但可不限於此。
For example, in one embodiment, assuming that there is no annotated map resource set GN in the
在步驟S250中,處理器104可在所述多個標註圖資集GA~GH、GN中找出對應於所述多個指定偵測物件類別的多個特定標註圖資集,並據以訓練聯合物件偵測模型324。
In step S250, the
在圖3情境中,對應於所述多個指定偵測物件類別的所述多個特定標註圖資集例如包括標註圖資集GA~GE、GN,而其可形成整合標註圖資集311。之後,處理器104可基於整合標註圖資集311來訓練訓練聯合物件偵測模型324。在此情況下,聯合物件偵測模型324應可具有偵測「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」的能力。
In the scenario of FIG. 3, the plurality of specific annotation map resource sets corresponding to the plurality of designated detection object categories include, for example, annotation map resource sets GA to GE, GN, and they can form an integrated annotation map resource set 311. After that, the
在本發明的實施例中,處理器104可對圖資庫310中的一或多個未標註圖資執行以下步驟。為便於說明,以下將針對未標註圖資中的一第一未標註圖資進行說明,但本發明可不限於此。
In an embodiment of the present invention, the
之後,在步驟S260中,處理器104可使用所述多個特定單物件偵測模型(即,單物件偵測模型A 0~E 0及N 0)、所述多個外部物件偵測模型ext1~ext3及聯合物件偵測模型324對第一未標註圖資進行偵測,以產生多個第一物件偵測結果,其中所述多個第
一物件偵測結果對應於所述多個指定偵測物件類別中的第一指定偵測物件類別。
Then, in step S260, the
在圖3中,偵測結果341例如可包括由所述多個特定單物件偵測模型(即,單物件偵測模型A 0~E 0及N 0)對第一未標註圖資進行偵測的結果。如圖3所示,偵測結果341可包括以方框框起的代號A 0~E 0及N 0,此即代表單物件偵測模型A 0~E 0及N 0分別在第一未標註圖資中偵測到屬於「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」的物件,而每個被偵測到的物件例如可由對應的單物件偵測模型以邊界框或其他類似方式予以標註,且可具有對應的偵測信心值。
In FIG. 3, the
此外,偵測結果342例如可包括由所述多個外部物件偵測模型ext1~ext3對第一未標註圖資進行偵測的結果。如圖3所示,偵測結果342可包括以方框框起的代號A 1、B 1、C 2、A 3、D 3及E 3,其中A 1、B 1代表外部物件偵測模型ext1在第一未標註圖資中偵測到屬於「物件A」、「物件B」的物件;C 2代表外部物件偵測模型ext2在第一未標註圖資中偵測到屬於「物件C」的物件;A 3、D 3及E 3代表外部物件偵測模型ext3在第一未標註圖資中偵測到屬於「物件A」、「物件D」、「物件E」的物件。相似地,在第一未標註圖資中,每個被外部物件偵測模型ext1~ext3偵測到的物件可由對應的外部物件偵測模型以邊界框或其他類似方式予以標註,且可具有對應的偵測信心值。
In addition, the
另外,偵測結果343例如可包括由聯合物件偵測模型324
對第一未標註圖資進行偵測的結果。如圖3所示,偵測結果343可包括以方框框起的代號A 4~E 4及N 4,此即代表聯合物件偵測模型324在第一未標註圖資中偵測到屬於「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」的物件。相似地,在第一未標註圖資中,每個被聯合物件偵測模型324偵測到的物件可由聯合物件偵測模型324以邊界框或其他類似方式予以標註,且可具有對應的偵測信心值。
In addition, the
為便於說明,以下假設第一指定偵測物件類別為「物件A」。在此情況下,處理器104於步驟S260中所取得所述多個第一物件偵測結果例如包括偵測結果341中的代號A 0、偵測結果342中的代號A 1、A 3及偵測結果343中的代號A 4,且其可個別具有對應的第一偵測信心值。
For ease of description, the following assumes that the first designated detection object category is "Object A". In this case, the detection results of the plurality of first objects obtained by the
之後,在步驟S270中,處理器104可使用所述多個單物件分類模型中對應於第一指定偵測物件類別的第一單物件分類模型對各第一物件偵測結果進行預測分類以得到各第一物件偵測結果的第一物件分類結果。
After that, in step S270, the
承上例,在第一指定偵測物件類別經假設為「物件A」的情況下,處理器104可將單物件分類模型A’作為對應於第一指定偵測物件類別的第一單物件分類模型。之後,處理器104可使用單物件分類模型A’對上述第一物件偵測結果進行預測分類以得到各第一物件偵測結果的第一物件分類結果。
Continuing from the above example, if the first designated detection object category is assumed to be "Object A", the
在一實施例中,處理器104可使用單物件分類模型A’對
偵測結果341中的代號A 0進行預測分類。例如,處理器104可使用單物件分類模型A’對由單物件偵測模型A 0於第一未標註圖資中框選的影像區域進行預測分類,以得到相應的一第一分類信心值作為對於此影像區域的第一物件分類結果,其中所述第一分類信心值可表徵單物件偵測模型A 0偵測的正確性。
In an embodiment, the
在另一實施例中,處理器104可使用單物件分類模型A’對偵測結果343中的代號A 4進行預測分類。例如,處理器104可使用單物件分類模型A’對由聯合物件偵測模型324於第一未標註圖資中框選的影像區域進行預測分類,以得到相應的一第一分類信心值作為對於此影像區域的第一物件分類結果,其中所述第一分類信心值可表徵聯合物件偵測模型324偵測的正確性。
In another embodiment, the
針對偵測結果342中的代號A 1、A 3,處理器104可使用單物件分類模型A’執行與上述類似的機制,以得到其個別對應的第一物件分類結果,其細節於此不另贅述。
Regarding the codes A 1 and A 3 in the
之後,在步驟S280中,處理器104可基於各第一物件偵測結果及對應的第一物件分類結果適應性地修正所述多個第一物件偵測結果,並將修正後的所述多個第一物件偵測結果加入所述多個標註圖資集中對應於第一指定偵測物件類別(例如「物件A」)的第一標註圖資集(例如,標註圖資集GA)。
Afterwards, in step S280, the
在第一實施例中,反應於判定各第一物件偵測結果的第一偵測信心值皆高於第一偵測信心閾值,所述多個第一物件偵測結果彼此的相似度皆滿足相似度條件,且各第一物件偵測結果對
應的第一分類信心值皆高於第一分類信心閾值,處理器104可相應地判定所述多個第一物件偵測結果不需修正,並將所述多個第一物件偵測結果加入第一標註圖資集(例如,標註圖資集GA)。
In the first embodiment, the first detection confidence value reflecting the determination that the detection result of each first object is higher than the first detection confidence threshold, and the similarity of the detection results of the plurality of first objects to each other is satisfied Similarity conditions, and the detection results of each first object are right
The corresponding first classification confidence values are all higher than the first classification confidence threshold, the
在一些實施例中,上述相似度條件可包括「邊框交疊率(Intersection over Union)大於交疊率差異基準閾值」,但可不限於此。 In some embodiments, the above-mentioned similarity condition may include "the frame overlap rate (Intersection over Union) is greater than the overlap rate difference reference threshold", but it may not be limited thereto.
在第二實施例中,反應於判定任一第一物件偵測結果的第一偵測信心值不高於第一偵測信心閾值,所述多個第一物件偵測結果的其中之二的相似度未滿足上述相似度條件,或任一第一物件偵測結果對應的第一分類信心值未高於第一分類信心閾值,處理器104可相應地判定所述多個第一物件偵測結果需修正,並將修正後的所述多個第一物件偵測結果加入第一標註圖資集(例如,標註圖資集GA)。
In the second embodiment, the first detection confidence value reflecting the determination that any one of the first object detection results is not higher than the first detection confidence threshold value is that two of the plurality of first object detection results If the similarity does not satisfy the above-mentioned similarity condition, or the first classification confidence value corresponding to any of the first object detection results is not higher than the first classification confidence threshold, the
之後,在步驟S290中,處理器104可基於第一標註圖資集重新訓練所述多個單物件偵測模型中對應於第一指定偵測物件類別的第一單物件偵測模型(例如單物件偵測模型A 0)、第一單物件分類模型(例如單物件分類模型A’)及聯合物件偵測模型324。
After that, in step S290, the
在一些實施例中,處理器104可針對圖資庫310中的其他未標註圖資重複地執行步驟S260~S290,直至聯合物件偵測模型324滿足設計者所設定的一或多個目標條件。
In some embodiments, the
例如,對於圖資庫310中的一第二未標註圖資而言,處理器104可經配置以:使用所述多個特定單物件偵測模型、所述 多個外部物件偵測模型及聯合物件偵測模型對第二未標註圖資進行偵測,以產生多個第二物件偵測結果(例如偵測結果341中的代號B 0、偵測結果342中的代號B 1及偵測結果343中的代號B 4),其中所述多個第二物件偵測結果對應於所述多個指定偵測物件類別中的一第二指定偵測物件類別(例如「物件B」);使用所述多個單物件分類模型中對應於第二指定偵測物件類別的第二單物件分類模型(例如單物件分類模型B’)對各第二物件偵測結果進行預測分類以得到各第二物件偵測結果的第二物件分類結果;基於各第二物件偵測結果及對應的第二物件分類結果適應性地修正所述多個第二物件偵測結果,並將修正後的所述多個第二物件偵測結果加入所述多個標註圖資集中對應於第二指定偵測物件類別的第二標註圖資集(例如標註圖資集GB);基於第二標註圖資集重新訓練所述多個單物件偵測模型中對應於第二指定偵測物件類別的第二單物件偵測模型(例如單物件偵測模型B 0)、第二單物件分類模型及聯合物件偵測模型324。 For example, for a second unlabeled image in the image library 310, the processor 104 may be configured to: use the plurality of specific single object detection models, the plurality of external object detection models, and the joint detecting a second object model for detecting unlabeled map data to generate a plurality of second object detection result (e.g. code detection result B 0 341, 342 of the detection result and the detection result of the code B 1 Code B 4 in 343), wherein the detection results of the plurality of second objects correspond to a second specified detection object category (for example, "object B") among the plurality of specified detection object categories; A second single-object classification model (for example, single-object classification model B') corresponding to the second designated detection object category among the plurality of single-object classification models predicts and classifies the detection results of each second object to obtain each second object The second object classification result of the detection result; based on each second object detection result and the corresponding second object classification result, the plurality of second object detection results are adaptively corrected, and the plurality of corrected The second object detection result is added to the second annotated image resource set (for example, annotated image resource set GB) corresponding to the second specified detection object category in the multiple annotated image resource sets; the training institute is retrained based on the second annotated image resource set A second single-object detection model (for example, a single-object detection model B 0 ), a second single-object classification model, and a joint object detection model corresponding to the second designated detection object category among the plurality of single-object detection models 324 .
在不同的實施例中,聯合物件偵測模型324需滿足的目標條件可包括「訓練圖資中的新增案場圖資大於規劃新增圖資數量」、「平均物件預測區域邊框與原標註區域邊框交疊率(Intersection over Union)大於交疊率目標閾值」、「各個單一物件特徵類別平均精度(Average Precision)皆大於單物件平均精度目標閾值」、「聯合物件偵測模型324之所有可偵測之物件特徵類別之平均精度均值(Mean Average Precision)大於模型平均精度目標閾
值」、「模型準確度(accuracy)大於模型準確度目標閾值」、「模型損失函數小於模型損失函數目標閾值」的至少其中之一,但可不限於此。
In different embodiments, the target conditions to be met by the joint
在一些實施例中,在聯合物件模型324訓練的初期時,處理器104在執行步驟S280時可採用第一實施例中記載的機制來將未修正的所述多個第一物件偵測結果加入第一標註圖資集(例如,標註圖資集GA),以較為快速地提升聯合物件偵測模型324的偵測能力。
In some embodiments, at the initial stage of the training of the
另外,在聯合物件模型324訓練的中期時,處理器104在執行步驟S280時可採用第二實施例中記載的機制來將修正後的所述多個第一物件偵測結果加入第一標註圖資集(例如,標註圖資集GA),以藉由較為多樣的訓練資料進一步提升聯合物件偵測模型324的偵測能力。
In addition, in the middle of the training of the
此外,在聯合物件偵測模型324訓練的後期時,由於可能有某些標註圖資集(例如標註圖資集GA)的內容較為缺少,使得聯合物件偵測模型324對於屬於對應指定偵測物件類別(例如「物件A」)的物件的偵測能力較為不佳。因此,在一些實施例中,處理器104在執行步驟S260時可僅要求所述多個特定單物件偵測模型、所述多個外部物件偵測模型及聯合物件偵測模型僅偵測屬於上述指定偵測物件類別的物件,以增加其對應的訓練資料量。
In addition, in the later stage of the training of the joint
例如,假設標註圖資集GA的內容較為缺少,則處理器104在執行步驟S260時可要求對應的第一單物件偵測模型(即,
單物件偵測模型A 0)、外部物件偵測模型ext1~ext3及聯合物件偵測模型324對其他的未標註圖資進行偵測,並依據先前的教示將相關的偵測結果修正後加入標註圖資集GA。之後,處理器104可再依據標註圖資集GA訓練聯合物件偵測模型324(及單物件偵測模型A 0),以強化聯合物件偵測模型324偵測「物件A」的能力,但可不限於此。
For example, assuming that the content of the annotated image collection GA is relatively lacking, the
在一些實施例中,若處理器104判定聯合物件偵測模型324已滿足上述目標條件,則處理器104可判定已完成聯合物件偵測模型324的訓練。在此情況下,訓練後的聯合物件偵測模型324即可用於對上述特定場域進行「物件A」、「物件B」、「物件C」、「物件D」、「物件E」及「物件N」的偵測,但可不限於此。
In some embodiments, if the
綜上所述,本發明至少具備以下特點。 In summary, the present invention has at least the following features.
(1)累積標註成果,彈性組合應用:可將既有的多個訓練標註圖資集,以各個物件特徵類別進行抽離、分類、再彙集建立各個單物件特徵類別的標註資料集,並訓練產生對應的各個單物件特徵類別的單物件偵測模型。以此方式可累積影像標註成果、可簡化後續更新維護成本、增加建立新應用的模型效率與組合彈性。 (1) Cumulative labeling results, flexible combination application: multiple existing training labeling image data sets can be separated, classified, and re-assembled for each object feature category to create a labeling data set for each single object feature category, and train Generate single-object detection models corresponding to each single-object feature category. In this way, image annotation results can be accumulated, subsequent update and maintenance costs can be simplified, and model efficiency and combination flexibility for establishing new applications can be increased.
(2)可融合既有模型成果:對於未提供訓練資料集的模型,透過預測標註模組,對既有圖資庫進行預測標註,除可得到該模型之各個物件特徵類別之預測準確率,亦可進行反向驗證該既有對應之標註資料集的資料是否準確,並修正標註資料。當該 模型具有標註圖資集中未標註的一物件特徵類別,根據本發明的特點,可將該一物件特徵類別之修正後標註資料儲存於該一物件特徵類別標註圖資集,實現融合既有模型成果。 (2) Existing model results can be integrated: For models that do not provide training data sets, predict and label the existing map database through the predictive labeling module. In addition to obtaining the prediction accuracy of each object feature category of the model, It is also possible to reversely verify whether the data in the existing corresponding labeled data set is accurate, and correct the labeled data. When should The model has a feature category of an object that is not labeled in the annotated map data set. According to the characteristics of the present invention, the corrected annotation data of the feature category of the object can be stored in the annotated map data set of the feature category of an object to realize the integration of existing model results .
(3)利用既有模型比對篩選資料:可利用組織內或外部的既有機器學習物件偵測模型幫助篩選未標註圖資,亦可作為目標模型預測標註結果的比對資訊、幫助檢測與建立各個單物件特徵類別的標註圖資集。 (3) Use existing models to compare and filter data: Existing machine learning object detection models within or outside the organization can be used to help filter unlabeled images, and can also be used as comparison information for target model prediction and labeling results to help detect and Establish annotated map data set of each single object feature category.
(4)以單物件偵測模型為預測比對標的:利用單物件特徵類別之AI偵測模型對比多種物件特徵類別之AI偵測模型的優勢,有訓練更快收斂、圖資標註快且易多樣化、準確度較高、預測速度較快的優勢,可為主要AI偵測模型預測的比對標的,取代初期圖資之篩選、標註、審查人力。 (4) The single-object detection model is used as the prediction comparison target: the advantage of using the AI detection model of single object feature category to compare with the AI detection model of multiple object feature categories, has faster training convergence, quick and easy map data labeling The advantages of diversification, higher accuracy, and faster prediction speed can be the comparison target for the main AI detection model prediction, replacing the initial map data selection, labeling, and review manpower.
(5)可快速且彈性擴充訓練標註圖資:當有擴增新物件特徵類別的需求,可從既有圖資庫中標註部分少量的圖資,經由建立單物件特徵類別的物件偵測模型,再透過預測標註模組篩選與擴充標註訓練圖資的迭代方式完成所有新進標註圖資與精煉模型。 (5) Can quickly and flexibly expand the training and labeling image data: when there is a need to expand the feature category of new objects, a small amount of image data can be labeled from the existing image database, and the object detection model of a single object feature category can be created , And then through the iterative method of predictive annotation module screening and expansion of annotation training image resources to complete all new annotation image resources and refined models.
(6)可提供物件特徵類別的菜單自動產生訓練圖資與應用模型:依應用需求挑選欲偵測的物件特徵類別,初期可從標註圖資集快速組合出符合新應用所需的多項物件特徵類別的訓練標註圖資,訓練後得到所需求的聯合物件偵測模型。後續以需求組合內的各個單物件特徵類別的物件偵測模型,對新投入未標註之 圖資進行預測,再以多模型預測結果比對機制進行標註圖資擴充與目標聯合模型的精練。 (6) A menu of object feature categories can be provided to automatically generate training images and application models: select the object feature categories to be detected according to the application needs, and initially can quickly combine multiple object features that meet the requirements of the new application from the set of labeled images The training annotated image data of the category, and the required joint object detection model is obtained after training. Subsequent to the object detection model of each single object feature category in the demand combination, the new investment is not marked The map data is predicted, and then the multi-model prediction result comparison mechanism is used to mark the map data expansion and refine the target joint model.
(7)多模型預測結果比對機制:本發明的多模型預測結果比對機制可更精準挑選出所需要的圖資進行標註;經由對比挑出AI模型預測不準確的新圖資再投入訓練,讓訓練模型更有效率;可快速篩選所需特定物件影像資料,解決訓練圖資偏頗問題;幫助降低AI物件偵測模型的精練成本,減少目標模型的迭代訓練週期,加快AI模型商用時程。 (7) Multi-model prediction result comparison mechanism: The multi-model prediction result comparison mechanism of the present invention can more accurately select the required image data for labeling; through comparison, new images with inaccurate AI model predictions are selected and then put into training. Make training models more efficient; quickly filter the image data of specific objects needed to solve the problem of biased training maps; help reduce the cost of refining the AI object detection model, reduce the iterative training cycle of the target model, and accelerate the commercialization of AI models.
(8)提供對預測結果之物件類別的審查機制:透過建立單物件特徵類別的物件分類模型,可複檢該相同物件特徵類別的物件偵測模型之偵測結果之物件邊界框內之區域影像,是否為該物件特徵類別之正確性。 (8) Provide a review mechanism for the object category of the prediction result: by creating an object classification model of a single object feature category, the area image within the object bounding box of the detection result of the object detection model of the same object feature category can be rechecked , Whether it is the correctness of the feature category of the object.
雖然本發明已以實施例揭露如上,然其並非用以限定本發明,任何所屬技術領域中具有通常知識者,在不脫離本發明的精神和範圍內,當可作些許的更動與潤飾,故本發明的保護範圍當視後附的申請專利範圍所界定者為準。 Although the present invention has been disclosed in the above embodiments, it is not intended to limit the present invention. Anyone with ordinary knowledge in the technical field can make some changes and modifications without departing from the spirit and scope of the present invention. The scope of protection of the present invention shall be subject to those defined by the attached patent scope.
S210~S290:步驟 S210~S290: steps
Claims (14)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW109135316A TWI748685B (en) | 2020-10-13 | 2020-10-13 | Method and server for optimizing joint object detection model |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW109135316A TWI748685B (en) | 2020-10-13 | 2020-10-13 | Method and server for optimizing joint object detection model |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| TWI748685B true TWI748685B (en) | 2021-12-01 |
| TW202215307A TW202215307A (en) | 2022-04-16 |
Family
ID=80680943
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW109135316A TWI748685B (en) | 2020-10-13 | 2020-10-13 | Method and server for optimizing joint object detection model |
Country Status (1)
| Country | Link |
|---|---|
| TW (1) | TWI748685B (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180082152A1 (en) * | 2016-09-21 | 2018-03-22 | GumGum, Inc. | Training machine learning models to detect objects in video data |
| TW202004513A (en) * | 2018-06-05 | 2020-01-16 | 美商克萊譚克公司 | Active learning for defect classifier training |
| TWI702536B (en) * | 2019-12-31 | 2020-08-21 | 財團法人工業技術研究院 | Training method and system of object detection model based on adaptive annotation design |
| TWI707137B (en) * | 2020-01-13 | 2020-10-11 | 憶象有限公司 | Intelligent production line monitoring system and implementation method thereof |
-
2020
- 2020-10-13 TW TW109135316A patent/TWI748685B/en active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180082152A1 (en) * | 2016-09-21 | 2018-03-22 | GumGum, Inc. | Training machine learning models to detect objects in video data |
| TW202004513A (en) * | 2018-06-05 | 2020-01-16 | 美商克萊譚克公司 | Active learning for defect classifier training |
| TWI702536B (en) * | 2019-12-31 | 2020-08-21 | 財團法人工業技術研究院 | Training method and system of object detection model based on adaptive annotation design |
| TWI707137B (en) * | 2020-01-13 | 2020-10-11 | 憶象有限公司 | Intelligent production line monitoring system and implementation method thereof |
Also Published As
| Publication number | Publication date |
|---|---|
| TW202215307A (en) | 2022-04-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN113505261B (en) | Data labeling method and device and data labeling model training method and device | |
| CN114245910B (en) | An AutoML system, method, and device for automated machine learning | |
| JP2018200685A (en) | Forming of data set for fully supervised learning | |
| CN111860494A (en) | Optimal method, device, electronic device and storage medium for image target detection | |
| CN110852347A (en) | Fire detection method using improved YOLO v3 | |
| CN114387499A (en) | Island coastal wetland waterfowl identification method, distribution query system and medium | |
| CN110533086A (en) | The semi-automatic mask method of image data | |
| TWI845797B (en) | Object detection device and object detection method | |
| CN114663687A (en) | Model training method, target recognition method, device, equipment and storage medium | |
| CN114708518A (en) | Bolt defect detection method based on semi-supervised learning and priori knowledge embedding strategy | |
| CN115859108A (en) | Sample expansion method, device, equipment, storage medium and product | |
| CN115859122A (en) | Data identification method, automatic continuous learning model, device and equipment | |
| CN118887382A (en) | A zero-shot target detection method based on embedding localization knowledge of pre-trained models | |
| CN117372424B (en) | A defect detection method, device, equipment and storage medium | |
| CN118628846A (en) | Target detection method, device, electronic device and medium based on multimodal prompting | |
| TWI748685B (en) | Method and server for optimizing joint object detection model | |
| CN114359670A (en) | Unstructured data labeling method and device, computer equipment and storage medium | |
| TWI860629B (en) | Image Annotation Quality Optimization Method | |
| CN118840541A (en) | Open world target detection method based on self-adaptive semantic degradation learning | |
| CN117131222A (en) | Semi-automatic labeling method and device based on open world large model | |
| EP4469916A1 (en) | A system and method for quality check of labelled images | |
| CN116453043A (en) | Kitchen mouse detection method based on background modeling and target detection | |
| CN120031883B (en) | Wafer defect detection method, device, equipment and medium | |
| CN120853209B (en) | Image recognition method, device, equipment and medium based on recognition model | |
| CN115565201B (en) | Taboo picture identification method, apparatus and storage medium |