TWI882800B - Ai-based method, computer program, and computer readable medium for identifying genera and species of mold - Google Patents
Ai-based method, computer program, and computer readable medium for identifying genera and species of mold Download PDFInfo
- Publication number
- TWI882800B TWI882800B TW113117580A TW113117580A TWI882800B TW I882800 B TWI882800 B TW I882800B TW 113117580 A TW113117580 A TW 113117580A TW 113117580 A TW113117580 A TW 113117580A TW I882800 B TWI882800 B TW I882800B
- Authority
- TW
- Taiwan
- Prior art keywords
- mold
- species
- genus
- identified
- identifying
- Prior art date
Links
Landscapes
- Image Analysis (AREA)
Abstract
Description
本發明係一種以人工智慧辨識黴菌之菌屬及菌種的方法、電腦程式、電腦可讀取媒體,特別是指先使用卷積神經網路模型(CNN model)鑑定分類待辨識黴菌之黴菌菌屬,再使用物件偵測模型(Object Detection)在所分類的黴菌菌屬下鑑定分類待辨識黴菌之黴菌菌種的二階段黴菌菌屬及菌種的鑑定分類發明。 The present invention is a method, computer program, and computer-readable medium for identifying the genus and species of mold using artificial intelligence, and particularly refers to a two-stage invention for identifying and classifying the genus and species of mold using a convolutional neural network model (CNN model) to identify and classify the mold genus of the mold to be identified, and then using an object detection model (Object Detection) to identify and classify the mold species of the mold to be identified under the classified mold genus.
台灣氣候高溫潮濕炎熱,對於黴菌生長提供有利環境,除了日常生活環境包括食物、衣物、日用品、家具、住宅等可能發霉之外,人體的黴菌感染有時也常造成致命後果。 Taiwan's hot, humid and warm climate provides a favorable environment for the growth of mold. In addition to the daily environment including food, clothing, daily necessities, furniture, and housing that may be moldy, mold infection in the human body can sometimes cause fatal consequences.
對黴菌感染的治療而言,目前醫院臨床上辨識黴菌種類的方式主要是由檢驗人員以顯微鏡下的形態學進行鑑定,而待辨識黴菌的形態則是建立在玻片培養獲取的圖像,並以待辨識黴菌的圖像上的形態特徵作為黴菌菌屬、菌種、菌名的鑑定分類。 For the treatment of fungal infection, the current clinical identification method of fungal species in hospitals is mainly for inspectors to identify them by morphology under a microscope. The morphology of the fungus to be identified is based on the image obtained by slide culture, and the morphological characteristics of the fungus to be identified are used as the identification and classification of the fungus genus, species, and name.
但是黴菌生長時間長,導致利用玻片培養以獲取圖像的過程相當耗時,且習知黴菌的形態特徵鑑定需依賴具有黴菌檢驗專業且有經驗的檢驗人 員,時常因為檢驗人員的短缺以及經驗不足帶來誤判問題,影響鑑定分類的準確性。 However, mold growth takes a long time, which makes the process of using slide culture to obtain images very time-consuming. In addition, the identification of mold morphological characteristics requires the use of experienced inspectors with mold testing expertise. The shortage of inspectors and lack of experience often lead to misjudgments, affecting the accuracy of identification and classification.
參閱第五圖所示,臨床應用上會使用簡易的膠帶黏貼法直接獲取黴菌菌株,並取得膠帶上的黴菌圖像,藉此取代耗時的玻片培養技術以提升時效。但是採用膠帶黏貼法可能使黴菌菌株的外形破損,影響其辨識率。 As shown in Figure 5, in clinical applications, a simple tape sticking method is used to directly obtain the mold strain and obtain the mold image on the tape, thereby replacing the time-consuming slide culture technology to improve the timeliness. However, the use of the tape sticking method may damage the appearance of the mold strain, affecting its identification rate.
臨床應用上也會進一步配合質譜鑑定來輔助分類黴菌種類,但質譜鑑定的成功率不高,發明人實測在鑑定麴黴菌屬(Aspergillus)上成功率約86%,在鑑定枝孢菌屬(Cladosporium)上成功率約40.5%,在鑑定毛癬菌屬(Trichophyton)上成功率約83.3%,在鑑定青黴菌屬(Penicillium)上成功率約44.3%。因此質譜鑑定對於準確鑑定分類黴菌種類的幫助有限。 In clinical applications, mass spectrometry will be further used to assist in the classification of fungal species, but the success rate of mass spectrometry identification is not high. The inventors have measured that the success rate in identifying Aspergillus is about 86%, the success rate in identifying Cladosporium is about 40.5%, the success rate in identifying Trichophyton is about 83.3%, and the success rate in identifying Penicillium is about 44.3%. Therefore, mass spectrometry identification is of limited help in accurately identifying and classifying fungal species.
由於人工智慧的發展,使用人工智慧來輔助檢驗人員鑑定黴菌種類提高了準確性。例如Md Arafatur Rahman等人提出的「Classification of fungal genera from microscopic images using artificial intelligence」,該案揭露使用卷積神經網路(CNN)模型來辨識89種菌屬(Genus);Lukáš Picek等人提出的「Automatic Fungi Recognition:Deep Learning Meets Mycology」,該案經由產生細緻分類之資料集Danish Fungi 2020(DF20)進行CNN模型訓練,產生低辨識錯誤之菌種等級標籤(species-level labels)輸出;Fahad Jubayer等人提出的「Detection of mold on the food surface using YOLOv5」,該案使用YOLO物件偵測模型來進行黴菌偵測。 With the development of artificial intelligence, the use of artificial intelligence to assist inspectors in identifying mold species has improved accuracy. For example, the "Classification of fungal genera from microscopic images using artificial intelligence" proposed by Md Arafatur Rahman et al., which used a convolutional neural network (CNN) model to identify 89 fungal genera; the "Automatic Fungi Recognition: Deep Learning Meets Mycology" proposed by Lukáš Picek et al., which trained a CNN model using the Danish Fungi 2020 (DF20) dataset that produces finely classified data, and produced species-level labels with low recognition error output; the "Detection of mold on the food surface using YOLOv5" proposed by Fahad Jubayer et al., which used the YOLO object detection model for mold detection.
參閱第六圖及第七圖所示,由於黴菌的鑑定分類主要採用形態學,雖然CNN可以快速根據形態特徵來辨識出黴菌菌屬,但是相同的黴菌菌屬下的黴菌菌種在形態特徵上差異更小,發明人實測以CNN鑑定分類黴菌菌種的 準確率較低。而物件偵測(YOLO)雖可透過框選物件(例如框選孢子或孢子囊等細節的形態特徵為物件)進行偵測,但是AI辨識需要耗費許多運算資源,且僅使用物件偵測框選物件而冀望能根據形態特徵在許多不同黴菌菌屬下找出正確的黴菌菌種仍猶如大海撈針,容易辨識錯誤。 As shown in the sixth and seventh figures, since the identification and classification of molds mainly adopts morphology, although CNN can quickly identify the mold genus based on morphological features, the mold species under the same mold genus have smaller differences in morphological features. The inventors have tested that the accuracy of using CNN to identify and classify mold species is low. Although object detection (YOLO) can detect objects by selecting them (for example, selecting morphological features of details such as spores or spore sacs as objects), AI recognition requires a lot of computing resources, and only using object detection to select objects in the hope of finding the correct mold species under many different mold genera based on morphological features is like looking for a needle in a haystack, and it is easy to make identification errors.
為了使人工智慧如同人為依據形態學辨識黴菌菌屬及菌種,並提升鑑定分類的準確性,本發明提出一種以人工智慧辨識黴菌之菌屬及菌種的方法,步驟包括:一處理單元取得一待辨識黴菌之一待辨識黴菌影像,並以影像處理取得該待辨識黴菌的至少一形態特徵;將所述形態特徵輸入一卷積神經網路模型(CNN model),該卷積神經網路模型係根據複數樣本黴菌影像的至少一樣本形態特徵經卷積神經網路學習而建立,該卷積神經網路模型根據所述形態特徵鑑定分類該待辨識黴菌之一黴菌菌屬。將該待辨識黴菌影像輸入一物件偵測模型(Object Detection),該物件偵測模型係在所述複數樣本黴菌影像上框選待偵測物件為標籤後經物件偵測學習而建立,所述待偵測物件為黴菌的細部形態特徵,該物件偵測模型在所分類的黴菌菌屬下,根據該待辨識黴菌影像鑑定分類該待辨識黴菌之一黴菌菌種。 In order to make artificial intelligence identify the genus and species of molds based on morphology like humans, and to improve the accuracy of identification and classification, the present invention proposes a method for identifying the genus and species of molds using artificial intelligence, the steps of which include: a processing unit obtains an image of a mold to be identified, and obtains at least one morphological feature of the mold to be identified by image processing; the morphological feature is input into a convolutional neural network model (CNN model), the convolutional neural network model is established based on at least one sample morphological feature of a plurality of sample mold images through convolutional neural network learning, and the convolutional neural network model identifies and classifies a mold genus of the mold to be identified based on the morphological feature. The mold image to be identified is input into an object detection model (Object Detection). The object detection model is established by selecting the object to be detected as a label on the plurality of sample mold images through object detection learning. The object to be detected is the detailed morphological features of the mold. The object detection model identifies and classifies a mold species of the mold to be identified according to the mold image to be identified under the classified mold genus.
進一步,該待辨識黴菌係以膠帶黏貼法獲得。 Furthermore, the fungus to be identified was obtained by the tape sticking method.
進一步,係取得該待辨識黴菌在不同視野下之多個待辨識黴菌影像進行鑑定分類。更進一步,在該處理單元上設定一第一辨識閾值,該處理單元計算該多個待辨識黴菌影像分類在同一個黴菌菌屬的比例,並找出比例最高的黴菌菌屬,當該比例最高的黴菌菌屬的比例高於該第一辨識閾值時,所述黴 菌菌屬作為該待辨識黴菌的分類結果,當該比例低於該第一辨識閾值時,輸出一無法辨識訊息;較佳的,所述第一辨識閾值設定在60%或70%。更進一步,在該處理單元上設定一第二辨識閾值,該處理單元計算該多個待辨識黴菌影像分類在同一個黴菌菌種的比例,並找出比例最高的黴菌菌種,當該比例最高的黴菌菌種的比例高於該第二辨識閾值時,所述黴菌菌種作為該待辨識黴菌的分類結果,當該比例低於該第二辨識閾值時,輸出一無法辨識訊息;較佳的,所述第二辨識閾值設定在60%或70%。 Furthermore, a plurality of images of the mold to be identified under different viewing fields are obtained for identification and classification. Furthermore, a first identification threshold is set on the processing unit, and the processing unit calculates the proportion of the plurality of images of the mold to be identified that are classified into the same mold genus, and finds the mold genus with the highest proportion. When the proportion of the mold genus with the highest proportion is higher than the first identification threshold, the mold genus is used as the classification result of the mold to be identified. When the proportion is lower than the first identification threshold, an unidentifiable message is output; preferably, the first identification threshold is set at 60% or 70%. Furthermore, a second identification threshold is set on the processing unit, and the processing unit calculates the ratio of the multiple mold images to be identified that are classified into the same mold species, and finds the mold species with the highest ratio. When the ratio of the mold species with the highest ratio is higher than the second identification threshold, the mold species is used as the classification result of the mold to be identified. When the ratio is lower than the second identification threshold, an unidentifiable message is output; preferably, the second identification threshold is set at 60% or 70%.
進一步,所述形態特徵包括顏色、質地、形狀、大小及生長方式之一或組合。 Furthermore, the morphological characteristics include one or a combination of color, texture, shape, size and growth pattern.
進一步,該卷積神經網路模型係經由卷積神經網路合併轉移學習(transfer learning)並加上注意力模型(attention mechanism)學習而建立。 Furthermore, the convolutional neural network model is established by combining transfer learning with the convolutional neural network and adding attention mechanism learning.
進一步,所述物件偵測模型係選自下列之一:YOLO模型、Faster-RCNN模型、及SSD模型(Single Shot Multi-box Detector)。 Furthermore, the object detection model is selected from one of the following: YOLO model, Faster-RCNN model, and SSD model (Single Shot Multi-box Detector).
本發明再提出一種電腦程式,供安裝於一電腦,而由該電腦程式執行前述以人工智慧辨識黴菌之菌屬及菌種的方法。 The present invention further proposes a computer program for installation in a computer, and the computer program executes the aforementioned method of identifying the genus and species of mold using artificial intelligence.
本發明再提出一種電腦可讀取媒體,係儲存有前述電腦程式。 The present invention further proposes a computer-readable medium that stores the aforementioned computer program.
根據上述技術特徵可達成以下功效: Based on the above technical features, the following effects can be achieved:
1.解決了專業檢驗人員短缺和經驗不足帶來的誤判問題,提高了黴菌菌屬及黴菌菌種鑑定分類的準確性。 1. It solves the problem of misjudgment caused by the shortage of professional inspectors and lack of experience, and improves the accuracy of identification and classification of mold genera and mold species.
2.該待辨識黴菌使用簡易的膠帶黏貼法取得,代替耗時的玻片培養方法,對於醫學的臨床應用,可提升檢驗的時效性。 2. The fungus to be identified is obtained using a simple tape sticking method, replacing the time-consuming slide culture method. For clinical applications in medicine, it can improve the timeliness of testing.
3.取得該待辨識黴菌不同視野之多個待辨識黴菌影像進行鑑定分類,可避免使用單一影像而誤判分類,提高準確性。 3. Obtaining multiple images of the fungus to be identified from different fields of view for identification and classification can avoid misclassification caused by using a single image and improve accuracy.
4.採用「先辨識菌屬、再辨識菌種」的二階段辨識程序,在醫學臨床上應用時,所辨識的黴菌菌屬分類結果若為無法辨識時則逕歸類為其它(MOLD),可不再進行黴菌菌種的辨識,若辨識出黴菌菌屬,則可利用物件偵測進一步辨識黴菌菌種。 4. The two-stage identification procedure of "first identify the genus, then identify the species" is adopted. When applied in clinical medicine, if the classification result of the identified fungus genus is unidentifiable, it will be directly classified as others (MOLD) and the fungus species identification will no longer be performed. If the fungus genus is identified, the fungus species can be further identified by object detection.
5.合併CNN與物件偵測,可以達到快速及準確分類黴菌菌屬及黴菌菌種的效果。其中,CNN先辨識黴菌菌屬猶如將搜尋範圍由大海縮小到小池子的範圍,之後由物件偵測在相同黴菌菌屬的小池子下辨識黴菌菌種的細部形態特徵(例如孢子或孢子囊的形態特徵),辨識準確率較高。 5. Combining CNN and object detection can achieve the effect of quickly and accurately classifying mold genera and mold species. Among them, CNN first identifies the mold genus, which is like narrowing the search range from a large ocean to a small pool. Then object detection identifies the detailed morphological features of the mold species (such as the morphological features of spores or sporangia) in the small pool of the same mold genus, and the recognition accuracy is higher.
6.當CNN合併轉移學習(transfer learning)並加上注意力模型(attention mechanism)時,可以提高辨識準確率達3.4%至23.2%。 6. When CNN incorporates transfer learning and adds an attention mechanism, the recognition accuracy can be improved by 3.4% to 23.2%.
1:卷積神經網路模型 1: Convolutional neural network model
2:物件偵測模型 2: Object detection model
3:待辨識黴菌影像 3: Image of mold to be identified
4:使用介面 4: User interface
5:處理單元 5: Processing unit
6:鑑定分類結果 6: Identification and classification results
[第一圖]係為本發明實施例的功能方塊圖。 [Figure 1] is a functional block diagram of an embodiment of the present invention.
[第二圖]係為本發明實施例的流程圖。 [Figure 2] is a flow chart of an embodiment of the present invention.
[第三圖]係為本發明實施例中,待辨識黴菌在不同視野下之多個待辨識黴菌影像的顯微圖。 [Figure 3] is a microscopic image of multiple mold images to be identified under different viewing fields in an embodiment of the present invention.
[第四圖]係為本發明實施例的使用介面示意圖。 [Figure 4] is a schematic diagram of the user interface of an embodiment of the present invention.
[第五圖]係為本發明實施例中,待辨識黴菌以玻片培養法及膠帶黏貼法在外觀形態比對的顯微圖。 [Fifth Figure] is a microscopic image of the mold to be identified in the embodiment of the present invention, comparing its appearance using the slide culture method and the tape sticking method.
[第六圖]係為不同黴菌菌屬的形態特徵差異比較的顯微圖。 [Figure 6] is a micrograph comparing the morphological characteristics of different mold genera.
[第七圖]係為相同黴菌菌屬下之不同黴菌菌種的形態特徵差異比較的顯微圖。 [Figure 7] is a micrograph comparing the morphological characteristics of different mold species under the same mold genus.
下列所述的實施例,只是輔助說明本發明之以人工智慧辨識黴菌之菌屬及菌種的方法、電腦程式、電腦可讀取媒體,並非用以限制本發明。 The following embodiments are only used to assist in explaining the method, computer program, and computer-readable medium of the present invention for identifying the genus and species of mold using artificial intelligence, and are not intended to limit the present invention.
參閱第一圖及第二圖所示,本實施例包括下列步驟:根據複數樣本黴菌影像的多個樣本形態特徵,使用卷積神經網路學習而建立一卷積神經網路模型1(CNN model),所述樣本形態特徵包括顏色、質地、形狀、大小等等,這些樣本形態特徵的特徵值可以由習知的影像處理獲得,例如可根據像素值(例如RGB值)來獲得顏色、質地的特徵值,可使用幾何輪廓的識別來獲得形狀、大小的特徵值等等。本實施例中,該卷積神經網路模型1進一步由卷積神經網路合併轉移學習(transfer learning)並加上注意力模型(attention mechanism)學習而建立,而初始卷積神經網路學習所根據之複數樣本黴菌影像來自Google的圖像資料庫,共以4108張圖像進行訓練,包括麴菌屬(Aspergillus)1488張(佔36.2%)、枝孢菌屬(Cladosporium)779張(佔19%)、青黴菌屬(Penicillium)522張(佔12.7%)、毛癬菌屬(Trichophyton)487張(佔11.9%)、其它黴菌圖像832張(佔20.3%);其中,由臨床病人檢體常見的麴黴菌屬(Aspergillus)中,黑麴菌(Aspergillus niger)有302張(佔Aspergillus中20.3%)、煙麴菌(Aspergillus fumigatus)有238張(佔Aspergillus中16%)、土麴菌(Aspergillus terreus)有227張(佔Aspergillus中15.3%)、雜色麴菌(Aspergillus versicolor)有117張(佔Aspergillus中7.9%)、黃麴菌(Aspergillus flavus)有111張(佔Aspergillus中7.5%)、其它麴黴菌 (Aspergillus sp.)有493張(佔Aspergillus中31.9%);卷積神經網路的學習中,訓練集佔64%、驗證集佔16%、測試集佔20%。 Referring to the first and second figures, the present embodiment includes the following steps: based on multiple sample morphological features of multiple sample mold images, a convolutional neural network model 1 (CNN model) is established using convolutional neural network learning, wherein the sample morphological features include color, texture, shape, size, etc. The feature values of these sample morphological features can be obtained by known image processing, for example, the feature values of color and texture can be obtained based on pixel values (such as RGB values), and the feature values of shape and size can be obtained by using geometric contour recognition. In the present embodiment, the convolutional neural network model 1 is further combined with transfer learning by the convolutional neural network and an attention model is added. The initial convolutional neural network was built based on a learning mechanism, and the multiple sample mold images based on which the initial convolutional neural network was trained came from the Google image database. A total of 4108 images were used for training, including 1488 Aspergillus images (36.2%), 779 Cladosporium images (19%), 522 Penicillium images (12.7%), 487 Trichophyton images (11.9%), and 832 other mold images (20.3%). Among them, Aspergillus niger, which is common in clinical patient specimens, is the most common. niger) with 302 photos (accounting for 20.3% of Aspergillus), Aspergillus fumigatus with 238 photos (accounting for 16% of Aspergillus), Aspergillus terreus with 227 photos (accounting for 15.3% of Aspergillus), Aspergillus versicolor with 117 photos (accounting for 7.9% of Aspergillus), Aspergillus flavus with 111 photos (accounting for 7.5% of Aspergillus), other Aspergillus (Aspergillus sp.) has 493 images (accounting for 31.9% of Aspergillus); in the learning of convolutional neural networks, the training set accounts for 64%, the validation set accounts for 16%, and the test set accounts for 20%.
在所述複數樣本黴菌影像上框選待偵測物件為標籤後,使用物件偵測學習而建立一物件偵測模型2(Object Detection),例如框選樣本黴菌影像上的孢子或孢子囊作為待偵測物件而作為物件偵測學習的標籤,所述物件偵測模型2可選自下列之一:YOLO模型、Faster-RCNN模型、及SSD模型(Single Shot Multi-box Detector),本實施例使用YOLOv7進行試驗,物件偵測學習的特徵值來自待偵測物件的顏色、質地、形狀、大小及生長方式等等。物件偵測學習中,同樣以訓練集佔64%、驗證集佔16%、測試集佔20%。 After selecting the objects to be detected on the multiple sample mold images as labels, an object detection model 2 (Object Detection) is established using object detection learning. For example, spores or spore sacs on the sample mold images are selected as the objects to be detected and used as labels for object detection learning. The object detection model 2 can be selected from one of the following: YOLO model, Faster-RCNN model, and SSD model (Single Shot Multi-box Detector). This embodiment uses YOLOv7 for testing. The feature values of object detection learning come from the color, texture, shape, size and growth mode of the objects to be detected. In object detection learning, the training set accounts for 64%, the validation set accounts for 16%, and the test set accounts for 20%.
取得所述卷積神經網路模型1及物件偵測模型2後,可用於一待辨識黴菌之黴菌菌屬及黴菌菌種的鑑定分類。 After obtaining the convolutional neural network model 1 and the object detection model 2, they can be used to identify and classify the genus and species of a mold to be identified.
參閱第三圖所示,以膠帶黏貼法取得該待辨識黴菌,並以攝影單元取得該待辨識黴菌在不同視野下的多個待辨識黴菌影像3,第三圖中顯示四個不同視野的待辨識黴菌影像3。 As shown in the third figure, the mold to be identified is obtained by the tape sticking method, and a plurality of mold images 3 of the mold to be identified in different fields of view are obtained by the photographic unit. The third figure shows four mold images 3 of the mold to be identified in different fields of view.
參閱第一圖、第二圖及第四圖所示,由一使用介面4將所述待辨識黴菌影像3輸入一處理單元5,其中,該處理單元5可以是服務端提供的雲端伺服器,該使用介面4為網路介面,或是該處理單元5為單機版電腦,該使用介面4為單機電腦介面皆可。該處理單元5以影像處理取得該待辨識黴菌的形態特徵,包括顏色、質地、形狀、大小等等,再由該卷積神經網路模型1根據所述形態特徵鑑定分類該待辨識黴菌之一黴菌菌屬,所述形態特徵與檢驗人員根據形態學分類黴菌菌屬相當,因而該卷積神經網路模型1可有效的鑑定分類正確的黴菌菌屬。由於本實施例使用不同視野下的多個待辨識黴菌影像3進行鑑定分 類,因此在該處理單元5上設定一第一辨識閾值,所述第一辨識閾值設定在60%或70%,本實施例設定為60%,該處理單元5會計算該卷積神經網路模型1將該多個待辨識黴菌影像3分類在同一個黴菌菌屬的比例,並找出比例最高的黴菌菌屬,當該比例高於該第一辨識閾值時,所述黴菌菌屬作為該待辨識黴菌的鑑定分類結果6,當該比例低於該第一辨識閾值時,輸出一無法辨識訊息,根據試驗,黴菌菌屬整體的鑑定分類正確率約94.9%,其中麴菌屬(Aspergillus)的鑑定分類正確率約98.1%、枝孢菌屬(Cladosporium)的鑑定分類正確率約96.3%、青黴菌屬(Penicillium)的鑑定分類正確率約92.5%、毛癬菌屬(Trichophyton)的鑑定分類正確率約92.6%、其它黴菌菌屬的鑑定分類正確率約90.6%。 Referring to the first, second and fourth figures, the image 3 of the mold to be identified is input into a processing unit 5 through a user interface 4, wherein the processing unit 5 may be a cloud server provided by a server, the user interface 4 may be a network interface, or the processing unit 5 may be a stand-alone computer, or the user interface 4 may be a stand-alone computer interface. The processing unit 5 obtains the morphological features of the mold to be identified by image processing, including color, texture, shape, size, etc., and then the convolution neural network model 1 identifies and classifies a mold genus of the mold to be identified according to the morphological features. The morphological features are equivalent to the mold genus classified by the inspector according to morphology, so the convolution neural network model 1 can effectively identify and classify the correct mold genus. Since the present embodiment uses a plurality of mold images 3 to be identified under different fields of view for identification and classification, a first identification threshold is set on the processing unit 5. The first identification threshold is set at 60% or 70%. In the present embodiment, it is set at 60%. The processing unit 5 calculates the proportion of the plurality of mold images 3 to be identified classified into the same mold genus by the convolutional neural network model 1, and finds the mold genus with the highest proportion. When the proportion is higher than the first identification threshold, the mold genus is used as the identification and classification result 6 of the mold to be identified. When the proportion is lower than the first identification threshold, the mold genus is used as the identification and classification result 6 of the mold to be identified. When the value is set, an unrecognizable message is output. According to the test, the overall identification and classification accuracy of mold genera is about 94.9%, of which the identification and classification accuracy of Aspergillus is about 98.1%, the identification and classification accuracy of Cladosporium is about 96.3%, the identification and classification accuracy of Penicillium is about 92.5%, the identification and classification accuracy of Trichophyton is about 92.6%, and the identification and classification accuracy of other mold genera is about 90.6%.
參閱下表一,本實施例並比較使用卷積神經網路以及合併使用注意力模型下的辨識準確率差異,可以發現當各種卷積神經網路合併使用注意力模型時,可以提高辨識準確率達3.4%至23.2%。 Referring to Table 1 below, this embodiment compares the difference in recognition accuracy when using a convolutional neural network and when using an attention model in combination. It can be found that when various convolutional neural networks are combined with an attention model, the recognition accuracy can be improved by 3.4% to 23.2%.
獲得該待辨識黴菌之黴菌菌屬後,由該物件偵測模型2在該卷積神經網路模型1所分類的黴菌菌屬下,根據該待辨識黴菌影像3分類該待辨識黴菌之一黴菌菌種,其中,由於相同黴菌菌屬的形態特徵差異較小,檢驗人員常根據孢子或孢子囊的型態來分類黴菌菌種,因此物件偵測框選孢子或孢子囊作為待偵測物件可有效鑑定分類正確的黴菌菌種。相同地,由於本實施例使用不同視野下的多個待辨識黴菌影像3進行鑑定分類,因此在該處理單元5上設定一第二辨識閾值,所述第二辨識閾值設定在60%或70%,本實施例設定為60%,該處理單元5會計算該物件偵測模型2將該多個待辨識黴菌影像3分類在同一個黴菌菌種的比例,並找出比例最高的黴菌菌種,當該比例高於該第二辨識閾值時,所述黴菌菌種作為該待辨識黴菌的鑑定分類結果6,當該比例低於該第二辨識閾值時,輸出一無法辨識訊息。 After obtaining the mold genus of the mold to be identified, the object detection model 2 classifies the mold to be identified as a mold species according to the mold image 3 under the mold genus classified by the convolution neural network model 1. Since the morphological features of the same mold genus are relatively small, inspectors often classify mold species according to the morphology of spores or sporangia. Therefore, the object detection frame selects spores or sporangia as the object to be detected, which can effectively identify the correctly classified mold species. Similarly, since the present embodiment uses multiple mold images 3 to be identified under different fields of view for identification and classification, a second identification threshold is set on the processing unit 5. The second identification threshold is set at 60% or 70%. In the present embodiment, it is set at 60%. The processing unit 5 calculates the proportion of the multiple mold images 3 to be identified that are classified into the same mold species by the object detection model 2, and finds the mold species with the highest proportion. When the proportion is higher than the second identification threshold, the mold species is used as the identification and classification result 6 of the mold to be identified. When the proportion is lower than the second identification threshold, an unidentifiable message is output.
參閱下表二及下表三,在物件偵測模型2上,本實施例進一步取得36個待辨識黴菌的不同視野下共135張待辨識黴菌影像3。 Referring to Table 2 and Table 3 below, on the object detection model 2, this embodiment further obtains a total of 135 images 3 of molds to be identified under different views of 36 molds to be identified.
下表二中顯示135張待辨識黴菌影像3中的每一張待辨識黴菌影像3各自以物件偵測模型2進行鑑定分類,並在物件偵測模型2本身設定不同閾值下的辨識準確率,可發現物件偵測模型2本身設定不同閾值下對於待辨識黴菌的黴菌菌種的辨識準確率皆可達90%以上。 Table 2 below shows that each of the 135 mold images 3 to be identified is identified and classified by the object detection model 2, and the recognition accuracy under different thresholds set in the object detection model 2 itself. It can be found that the recognition accuracy of the mold species of the mold to be identified under different thresholds set in the object detection model 2 itself can reach more than 90%.
下表三中顯示將每一個待辨識黴菌輸入不同視野的多張待辨識黴菌影像3進行鑑定分類並設定不同第二辨識閾值下的辨識準確率,可發現當每一個待辨識黴菌以不同視野的多張待辨識黴菌影像3進行鑑定分類且第二辨識閾值設定在60%或70%時,待辨識黴菌的黴菌菌種的辨識準確率皆達100%。 Table 3 below shows the recognition accuracy of each to-be-identified mold when multiple to-be-identified mold images 3 with different fields of view are input for identification and classification and different second identification thresholds are set. It can be found that when each to-be-identified mold is identified and classified with multiple to-be-identified mold images 3 with different fields of view and the second identification threshold is set at 60% or 70%, the recognition accuracy of the mold species of the to-be-identified mold is 100%.
本實施例後續將所述待辨識黴菌影像3儲存於卷積神經網路與物件偵測的學習資料庫做為樣本黴菌影像,使卷積神經網路模型與物件偵測模型的鑑定分類準確度能夠持續提高。 This embodiment subsequently stores the mold image 3 to be identified in the learning database of the convolutional neural network and object detection as a sample mold image, so that the identification and classification accuracy of the convolutional neural network model and the object detection model can be continuously improved.
執行上述以人工智慧辨識黴菌之菌屬及菌種的方法所安裝之電腦程式可儲存於雲端供下載,或可儲存於電腦可讀取媒體中。 The computer program installed to implement the above-mentioned method of using artificial intelligence to identify the genus and species of mold can be stored in the cloud for downloading, or can be stored in a computer-readable medium.
綜合上述實施例之說明,當可充分瞭解本發明之操作、使用及本發明產生之功效,惟以上所述實施例僅係為本發明之較佳實施例,當不能以此限定本發明實施之範圍,即依本發明申請專利範圍及發明說明內容所作簡單的等效變化與修飾,皆屬本發明涵蓋之範圍內。 Combined with the description of the above embodiments, the operation, use and effects of the present invention can be fully understood. However, the above embodiments are only preferred embodiments of the present invention and cannot be used to limit the scope of implementation of the present invention. In other words, simple equivalent changes and modifications made according to the scope of the patent application and the content of the invention description are all within the scope of the present invention.
Claims (12)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW113117580A TWI882800B (en) | 2024-05-13 | 2024-05-13 | Ai-based method, computer program, and computer readable medium for identifying genera and species of mold |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW113117580A TWI882800B (en) | 2024-05-13 | 2024-05-13 | Ai-based method, computer program, and computer readable medium for identifying genera and species of mold |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| TWI882800B true TWI882800B (en) | 2025-05-01 |
| TW202544748A TW202544748A (en) | 2025-11-16 |
Family
ID=96581901
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW113117580A TWI882800B (en) | 2024-05-13 | 2024-05-13 | Ai-based method, computer program, and computer readable medium for identifying genera and species of mold |
Country Status (1)
| Country | Link |
|---|---|
| TW (1) | TWI882800B (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| TW201945732A (en) * | 2018-03-20 | 2019-12-01 | 美商路瑪賽特有限責任公司 | Advanced biophysical and biochemical cellular monitoring and quantification using laser force cytology |
| US20220121884A1 (en) * | 2011-09-24 | 2022-04-21 | Z Advanced Computing, Inc. | System and Method for Extremely Efficient Image and Pattern Recognition and Artificial Intelligence Platform |
| WO2022178095A1 (en) * | 2021-02-19 | 2022-08-25 | Deepcell, Inc. | Systems and methods for cell analysis |
| US20230260273A1 (en) * | 2022-02-14 | 2023-08-17 | Apeel Technology, Inc. | Systems and methods for assessment of produce shelf life using time lapse image data |
-
2024
- 2024-05-13 TW TW113117580A patent/TWI882800B/en active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20220121884A1 (en) * | 2011-09-24 | 2022-04-21 | Z Advanced Computing, Inc. | System and Method for Extremely Efficient Image and Pattern Recognition and Artificial Intelligence Platform |
| TW201945732A (en) * | 2018-03-20 | 2019-12-01 | 美商路瑪賽特有限責任公司 | Advanced biophysical and biochemical cellular monitoring and quantification using laser force cytology |
| WO2022178095A1 (en) * | 2021-02-19 | 2022-08-25 | Deepcell, Inc. | Systems and methods for cell analysis |
| US20230260273A1 (en) * | 2022-02-14 | 2023-08-17 | Apeel Technology, Inc. | Systems and methods for assessment of produce shelf life using time lapse image data |
Also Published As
| Publication number | Publication date |
|---|---|
| TW202544748A (en) | 2025-11-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN106803247B (en) | Microangioma image identification method based on multistage screening convolutional neural network | |
| Li et al. | SAR image change detection using PCANet guided by saliency detection | |
| Quelhas et al. | Cell nuclei and cytoplasm joint segmentation using the sliding band filter | |
| CN109977780A (en) | A kind of detection and recognition methods of the diatom based on deep learning algorithm | |
| CN105320970B (en) | A kind of potato disease diagnostic device, diagnostic system and diagnostic method | |
| Azizi et al. | Identifying potato varieties using machine vision and artificial neural networks | |
| Gan et al. | Automated leather defect inspection using statistical approach on image intensity | |
| CN111382766A (en) | A device fault detection method based on Faster R-CNN | |
| Mohammadpoor et al. | An intelligent technique for grape fanleaf virus detection | |
| CN118941558B (en) | Visual inspection system and method for new energy vehicle manufacturing | |
| CN116681923A (en) | Automatic ophthalmic disease classification method and system based on artificial intelligence | |
| CN107341514B (en) | Abnormal point and edge point detection method based on joint density and angle | |
| Samaniego et al. | Image processing model for classification of stages of freshness of bangus using YOLOv8 algorithm | |
| Shao et al. | Research on automatic identification system of tobacco diseases | |
| CN114626952A (en) | Novel fish morphological characteristic accurate measurement method | |
| Kashyap et al. | Color histogram based image retrieval technique for diabetic retinopathy detection | |
| TWI882800B (en) | Ai-based method, computer program, and computer readable medium for identifying genera and species of mold | |
| Kajale | Detection & reorganization of plant leaf diseases using image processing and Android OS | |
| Li et al. | Variety identification of delinted cottonseeds based on BP neural network | |
| Zhao et al. | A real‐time classification and detection method for mutton parts based on single shot multi‐box detector | |
| García‐Lamont et al. | Efficient nucleus segmentation of white blood cells mimicking the human perception of color | |
| Dong et al. | Dragon fruit disease image segmentation based on FCM algorithm and two-dimensional OTSU algorithm | |
| Zárate et al. | Fruit detection and classification using computer vision and machine learning techniques | |
| Felicetti et al. | Fish blood cell as biological dosimeter: In between measurements, radiomics, preprocessing, and artificial intelligence | |
| CN110321787A (en) | Disease identification method, system and storage medium based on joint sparse representation |