[go: up one dir, main page]

TWI815057B - Visualization methods for cancer lesions - Google Patents

Visualization methods for cancer lesions Download PDF

Info

Publication number
TWI815057B
TWI815057B TW109139265A TW109139265A TWI815057B TW I815057 B TWI815057 B TW I815057B TW 109139265 A TW109139265 A TW 109139265A TW 109139265 A TW109139265 A TW 109139265A TW I815057 B TWI815057 B TW I815057B
Authority
TW
Taiwan
Prior art keywords
slide
visualized
level image
cancer
layer
Prior art date
Application number
TW109139265A
Other languages
Chinese (zh)
Other versions
TW202219827A (en
Inventor
陳震宇
陳志榮
葉肇元
陳啟中
Original Assignee
臺北醫學大學
雲象科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 臺北醫學大學, 雲象科技股份有限公司 filed Critical 臺北醫學大學
Priority to TW109139265A priority Critical patent/TWI815057B/en
Publication of TW202219827A publication Critical patent/TW202219827A/en
Application granted granted Critical
Publication of TWI815057B publication Critical patent/TWI815057B/en

Links

Landscapes

  • Image Analysis (AREA)
  • Investigating Or Analysing Biological Materials (AREA)
  • Medicines Containing Antibodies Or Antigens For Use As Internal Diagnostic Agents (AREA)

Abstract

A visualization method for cancer lesions is provided. The visualization method can extract a pre-pool layer feature map from a cancer identification deep learning model based on slide-level operations, and use a clustering algorithm to generate multiple groups and mark the groups on a slide-level image to be labeled. A class activation map extracted from the deep learning model is then used to filter the groups that should be labeled, and finally the cancer lesions are visualized.

Description

用於癌症病灶的視覺化方法Visualization methods for cancer lesions

本發明涉及一種用於癌症病灶的視覺化方法,特別是涉及一種可提升分離癌細胞及非癌細胞區域的效果的用於癌症病灶的視覺化方法。The present invention relates to a method for visualizing cancer lesions, and in particular, to a method for visualizing cancer lesions that can improve the effect of separating cancer cells and non-cancer cells.

在肺部腫瘤臨床診斷臨床實務上,會透過低密度電腦斷層掃描(Low density CT)掃描病例肺部,若有可疑的肺結節,再針對肺結節進行粗針切片檢查(Core Needle Biopsy)或開刀切除,檢體經處理後製作成病理切片,再由病理醫師進行診斷,最後決定治療方式。病理醫師診斷過程需先正確判讀切片是否有腫瘤成分,再判讀腫瘤的分類,以發病理報告。這個診斷過程需依賴病理醫師的經驗和時間。In the clinical practice of clinical diagnosis of lung tumors, the patient's lungs will be scanned through low-density computed tomography (Low density CT). If there are suspicious pulmonary nodules, Core Needle Biopsy or surgery will be performed on the pulmonary nodules. After resection, the specimen is processed and made into pathological sections, which are then diagnosed by a pathologist and finally the treatment method is decided. During the diagnosis process, the pathologist must first correctly interpret whether the slices contain tumor components, and then interpret the tumor classification to issue a pathology report. This diagnostic process relies on the experience and time of the pathologist.

使用玻片級圖像訓練進行癌症分類的過程中,能直接利用玻片級標註(Slide-Level Annotation)進行深度學習模型的訓練,無需專家提供細節標註(Patch-Level Annotation),能省下大量的時間及標註資源。In the process of using slide-level image training for cancer classification, slide-level annotation (Slide-Level Annotation) can be directly used to train the deep learning model. There is no need for experts to provide detailed annotations (Patch-Level Annotation), which can save a lot of money. time and labeling resources.

儘管玻片級圖像訓練能提供優異的癌症分辨能力,即以玻片 (slide) 為單位判斷是否包含癌細胞的評分能達到與將全玻片切割為小片段進行判斷癌細胞的方法相同的水準,然而,其顯示病灶區域的能力(Leision Localization)並沒有達到目前主流方法的水準。Although slide-level image training can provide excellent cancer discrimination capabilities, that is, the score of judging whether cancer cells are contained in slide units can be the same as the method of cutting the whole slide into small fragments to judge cancer cells. However, its ability to display lesion areas (Leision Localization) has not reached the level of current mainstream methods.

本發明所要解決的技術問題在於,針對現有技術的不足提供一種可提升分離癌細胞及非癌細胞區域的效果的用於癌症病灶的視覺化方法。The technical problem to be solved by the present invention is to provide a visualization method for cancer lesions that can improve the effect of separating cancer cells and non-cancer cells in response to the shortcomings of the existing technology.

為了解決上述的技術問題,本發明所採用的其中一技術方案是提供一種用於癌症病灶的視覺化方法,其包括下列步驟:取得一玻片級待視覺化影像,其中,該玻片級待視覺化影像經過一經訓練深度學習模型判斷為具有癌症細胞;將該玻片級待視覺化影像輸入該經訓練深度學習模型,其中該經訓練深度學習模型包括一特徵擷取網路、一全局池化層及一全連接層;通過該特徵擷取網路將輸入的該玻片級待視覺化影像進行特徵擷取 (feature extraction)以產生一預池化特徵地圖(pre-pool feature map ),其中,該預池化特徵地圖包括多個元素,各該元素用於表示多個特徵的其中之一是否出現在該玻片級待視覺化影像的多個位置的其中之一;將該預池化特徵地圖對一尺寸維度拆解為多個向量,以產生一向量集合,其中該些向量各具有對應於該些特徵的多個頻道單元;通過一分群演算法將該向量集合依據一分群參數分爲多個分群;將該些分群轉換為多個分群影像並呈現於該玻片級待視覺化影像上; 依據該些分群影像於該玻片級待視覺化影像上的對應關係,篩選出該些分群中,對應於該玻片級待視覺化影像中的癌症細胞的至少一應標註分群;以及將該至少一應標註分群依據一分類激活地圖(Class Activation Map, CAM) 標註該玻片級待視覺化影像中。In order to solve the above technical problems, one of the technical solutions adopted by the present invention is to provide a method for visualizing cancer lesions, which includes the following steps: obtaining a slide-level image to be visualized, wherein the slide-level image to be visualized is The visualized image is judged to contain cancer cells by a trained deep learning model; the slide-level image to be visualized is input into the trained deep learning model, where the trained deep learning model includes a feature extraction network and a global pool layer and a fully connected layer; the input slide-level image to be visualized is subjected to feature extraction (feature extraction) through the feature extraction network to generate a pre-pool feature map (pre-pool feature map), Wherein, the pre-pooled feature map includes a plurality of elements, each element is used to indicate whether one of the multiple features appears in one of the multiple positions of the slide-level image to be visualized; the pre-pooled feature map is The feature map decomposes one size dimension into multiple vectors to generate a vector set, where each of these vectors has multiple channel units corresponding to the features; the vector set is divided into a grouping parameter according to a grouping algorithm through a grouping algorithm. Divide into multiple clusters; convert these clusters into multiple cluster images and present them on the slide-level image to be visualized; filter out the images based on the correspondence between the clustered images on the slide-level image to be visualized Among the clusters, at least one should be labeled that corresponds to the cancer cells in the slide-level image to be visualized; and labeling the slide with the at least one labeled group based on a Class Activation Map (CAM) Awaiting visualization.

本發明的其中一有益效果在於,本發明所提供的用於癌症病灶的視覺化方法,其能從基於玻片級運算的癌症辨識深度學習模型擷取預池化層特徵地圖,並使用分群演算法產生多個分群並標註於玻片級待標註圖像上,進而輔以從深度學習模型擷取的分類激活地圖篩選出應標註分群,最終將癌症病灶進行視覺化。One of the beneficial effects of the present invention is that the visualization method for cancer lesions provided by the present invention can extract the pre-pooling layer feature map from the deep learning model of cancer identification based on slide-level operations and use grouping calculations. The method generates multiple clusters and labels them on the slide-level image to be labeled, and then uses the classification activation map extracted from the deep learning model to select the clusters that should be labeled, and finally visualizes the cancer lesions.

因此,本發明所提供的用於癌症病灶的視覺化方法可提升演算法分離癌細胞及壞死區域的效果,並減少演算法在視覺化癌症病灶時,誤將壞死區域辨識為癌細胞的情形,改良了以玻片級圖像運算所訓練之模型視覺化病灶區域的演算法。Therefore, the visualization method for cancer lesions provided by the present invention can improve the effect of the algorithm in separating cancer cells and necrotic areas, and reduce the situation where the algorithm mistakenly identifies necrotic areas as cancer cells when visualizing cancer lesions. The algorithm for visualizing the lesion area using a model trained with slide-level image operations has been improved.

為使能更進一步瞭解本發明的特徵及技術內容,請參閱以下有關本發明的詳細說明與圖式,然而所提供的圖式僅用於提供參考與說明,並非用來對本發明加以限制。In order to further understand the features and technical content of the present invention, please refer to the following detailed description and drawings of the present invention. However, the drawings provided are only for reference and illustration and are not used to limit the present invention.

以下是通過特定的具體實施例來說明本發明所公開有關“用於癌症病灶的視覺化方法”的實施方式,本領域技術人員可由本說明書所公開的內容瞭解本發明的優點與效果。本發明可通過其他不同的具體實施例加以施行或應用,本說明書中的各項細節也可基於不同觀點與應用,在不背離本發明的構思下進行各種修改與變更。另外,本發明的附圖僅為簡單示意說明,並非依實際尺寸的描繪,事先聲明。以下的實施方式將進一步詳細說明本發明的相關技術內容,但所公開的內容並非用以限制本發明的保護範圍。另外,本文中所使用的術語“或”,應視實際情況可能包括相關聯的列出項目中的任一個或者多個的組合。The following is a specific example to illustrate the implementation of the "visualization method for cancer lesions" disclosed in the present invention. Those skilled in the art can understand the advantages and effects of the present invention from the content disclosed in this specification. The present invention can be implemented or applied through other different specific embodiments, and various details in this specification can also be modified and changed based on different viewpoints and applications without departing from the concept of the present invention. In addition, the drawings of the present invention are only simple schematic illustrations and are not depictions based on actual dimensions, as is stated in advance. The following embodiments will further describe the relevant technical content of the present invention in detail, but the disclosed content is not intended to limit the scope of the present invention. In addition, the term "or" used in this article shall include any one or combination of more of the associated listed items depending on the actual situation.

圖1為根據本發明實施例繪示的用於癌症病灶的視覺化方法的流程圖。參閱圖1所示,本發明實施例提供一種用於癌症病灶的視覺化方法,其主要是以一個在玻片等級下進行運算的肺癌辨識深度學習模型為基礎,且深度學習模型以及本發明的用於癌症病灶的視覺化方法均可通過至少包括記憶體及處理器的一電腦系統所執行。FIG. 1 is a flow chart of a method for visualizing cancer lesions according to an embodiment of the present invention. Referring to Figure 1, an embodiment of the present invention provides a method for visualizing cancer lesions, which is mainly based on a deep learning model for lung cancer identification that operates at the slide level, and the deep learning model and the present invention Any method for visualizing cancer lesions can be executed by a computer system including at least a memory and a processor.

詳細而言,肺癌辨識深度學習模型為一經訓練的深度學習模型,其係利用全玻片訓練方式來訓練卷積神經網路(Convolutional Neural Network, CNN),而訓練完畢的深度學習模型用來在肺部腫塊病理數位切片進行肺腺癌、鱗狀上皮癌、其他癌別與非癌症類別分類。Specifically, the deep learning model for lung cancer identification is a trained deep learning model that uses a whole-slide training method to train a convolutional neural network (CNN), and the trained deep learning model is used in Pathological digital sections of lung masses were used to classify lung adenocarcinoma, squamous cell carcinoma, other cancer types and non-cancer types.

其中,其模型訓練學習使用之資料為臺北醫學大學附設醫院所提供之 7,003 張肺部病理切片,包含肺腺癌 (adenocarcinoma) 3,876 張、鱗狀上皮癌 (squamous cell carcinoma) 1,088 張、非癌症 (non-cancer) 2,039 張,以及病理科醫師對玻片的臨床診斷資訊。Among them, the data used for model training and learning are 7,003 lung pathological slides provided by the Taipei Medical University Hospital, including 3,876 lung adenocarcinoma (adenocarcinoma), 1,088 squamous cell carcinoma, and non-cancer ( non-cancer) 2,039 slides, as well as pathologists' clinical diagnostic information on the slides.

舉例而言,以卷積神經網路為基礎的深度學習模型大多是由數個層堆疊所組成,當一張玻片級病理影像輸入時,第一個層會將影像做轉換得到中間特徵圖(Intermediate Feature Map)。接著,第二個層以先前產生之特徵圖(不限前一層產生之特徵圖)作為輸入(Input)轉換為另一張特徵圖,以此類推將所有層依次做計算後,最後一個特徵圖即為模型預測此張玻片是否包含癌細胞的結果。For example, deep learning models based on convolutional neural networks are mostly composed of several stacked layers. When a slide-level pathological image is input, the first layer will convert the image to obtain an intermediate feature map. (Intermediate Feature Map). Then, the second layer uses the previously generated feature map (not limited to the feature map generated by the previous layer) as input to convert it into another feature map. By analogy, after all layers are calculated in sequence, the last feature map This is the result of the model predicting whether this slide contains cancer cells.

依照每一個層的運算式不同可以分為不同種類的層,常見的種類包含卷積層(Convolutional Layer)、池化層(Pooling Layer)、標準化層(Normalization Layer)、全局池化層(Global Pooling Layer)、全連接層(Fully-Connected Layer)等等。According to the different calculation formulas of each layer, it can be divided into different types of layers. Common types include Convolutional Layer, Pooling Layer, Normalization Layer, and Global Pooling Layer. ), Fully-Connected Layer, etc.

以卷積神經網路為基礎的深度學習模型可例如為ResNet或DenseNet,都採用類似的結構。可參考圖2,圖2為根據本發明實施例繪示的用於癌症病灶的視覺化方法的流程示意圖。如圖2所示,深度學習模型可包括輸入層IN、多個隱含層HID及輸出層OUT,且隱含層HID可包括特徵擷取網路FEN、全局池化層GP及全連接層FC,且特徵擷取網路FEN可包括選自由上述卷積層、池化層及標準化層組成的群組的多個層。Deep learning models based on convolutional neural networks can be, for example, ResNet or DenseNet, both of which adopt similar structures. Reference may be made to FIG. 2 , which is a schematic flowchart of a method for visualizing cancer lesions according to an embodiment of the present invention. As shown in Figure 2, the deep learning model may include an input layer IN, multiple hidden layers HID, and an output layer OUT, and the hidden layer HID may include a feature extraction network FEN, a global pooling layer GP, and a fully connected layer FC. , and the feature extraction network FEN may include multiple layers selected from the group consisting of the above convolution layer, pooling layer and normalization layer.

在功能性上,開頭由多層結構形成的特徵擷取網路FEN將輸入的病理影像做特徵擷取(Feature Extraction),即辨識細胞、組織的型態並將資訊保留在輸出的預池化特徵圖中,不重要的特徵在過程中則會被拋棄,全局池化層GP則是將圖上不同位置擷取到的特徵做整合,即保留各個特徵是否在此張玻片上的任何一個地方出現,拋棄掉此特徵出現位置資訊;而最後的全連接層則是將各個擷取到的特徵做整合,得到最後的預測結果。而本發明即是在上述基礎下,進一步強化深度學習模型對癌症病灶的視覺化能力。In terms of functionality, the feature extraction network FEN formed by a multi-layer structure at the beginning performs feature extraction (Feature Extraction) on the input pathological images, that is, it identifies the types of cells and tissues and retains the information in the output pre-pooled features. In the picture, unimportant features will be discarded in the process. The global pooling layer GP integrates the features captured at different locations on the picture, that is, it retains whether each feature appears anywhere on this slide. , discarding the location information of this feature; and the final fully connected layer integrates the extracted features to obtain the final prediction result. On the basis of the above, the present invention further strengthens the deep learning model's ability to visualize cancer lesions.

一併參閱圖1及圖2所示,用於癌症病灶的視覺化方法至少包括下列幾個步驟:Referring to Figures 1 and 2 together, the visualization method for cancer lesions includes at least the following steps:

步驟S100:取得玻片級待視覺化影像IMG。其中,玻片級待視覺化影像經過上述經訓練深度學習模型判斷為具有癌症細胞或疑似具有癌症細胞。Step S100: Obtain the slide-level image IMG to be visualized. Among them, the slide-level image to be visualized is judged to contain cancer cells or is suspected to contain cancer cells through the above-mentioned trained deep learning model.

步驟S101:將玻片級待視覺化影像IMG輸入經訓練深度學習模型。Step S101: Input the slide-level image IMG to be visualized into the trained deep learning model.

步驟S102:通過特徵擷取網路將輸入的玻片級待視覺化影像進行特徵擷取 (feature extraction)以產生預池化特徵地圖(pre-pool feature map ) PPFM。Step S102: Perform feature extraction on the input slide-level image to be visualized through a feature extraction network to generate a pre-pool feature map (pre-pool feature map) PPFM.

詳細而言,預池化特徵地圖PPFM可以 來表示,為一個 大小的張量(Tensor),其中 為此張量的尺寸維度,亦即,及 維度對應於該玻片級待視覺化影像的高與寬, 為頻道(Channel)數量,其表示擷取特徵的最大數量。 In detail, the pre-pooled feature map PPFM can to represent, as a Tensor of size, where The dimensions of this tensor, that is, and The dimensions correspond to the height and width of the slide-level image to be visualized, is the number of channels, which represents the maximum number of retrieved features.

預池化特徵地圖PPFM可包括多個元素 ,其中,任意一個元素 用於表示多個特徵的其中之一是否出現在玻片級待視覺化影像IMG中的某一個位置(例如,座標 h, w)。而元素 的數值越大,表示其所對應的特徵越明顯。 Pre-pooled feature map PPFM can include multiple elements , where any element Used to indicate whether one of multiple features appears at a certain position (for example, coordinates h, w) in the slide-level image IMG to be visualized. while elements The larger the value, the more obvious the corresponding feature is.

此處,先針對全局池化層GP及全連接層FC在經訓練深度學習模型所執行的功能進行說明。在上文中提到,全局池化層GP將圖上不同位置擷取到的特徵做整合,換言之,全局池化層GP將預池化特徵地圖PPFM中的尺寸維度(亦即H×W)進行降維,以產生全局池化向量。全局池化向量可由下式表示:Here, the functions performed by the global pooling layer GP and the fully connected layer FC in the trained deep learning model will be explained first. As mentioned above, the global pooling layer GP integrates features captured at different locations on the map. In other words, the global pooling layer GP integrates the size dimensions (i.e. H×W) in the pre-pooled feature map PPFM. Dimensionality reduction to produce global pooling vectors. The global pooling vector can be expressed by:

.

其中, 即是全局池化向量,為一個 大小的向量(vector),每一個元素表示某個特徵是否在此玻片級待視覺化影像IMG中出現。 in, That is, the global pooling vector is a A vector of size, each element indicates whether a certain feature appears in the slide-level image to be visualized IMG.

另一方面,全連接層FC則是用於對全局池化向量 進行加權加總,以產生一評估分數。此評估分數即是用於指示玻片級待視覺化影像是否包含癌症細胞,其可由下式表示: On the other hand, the fully connected layer FC is used to pool the global vector A weighted sum is performed to produce an evaluation score. This evaluation score is used to indicate whether the slide-level image to be visualized contains cancer cells, which can be expressed by the following formula:

.

其中, 為評估分數且為純量(Scalar), 為全局池化向量, 為全連接層的第一權重, 為全連接層的第二權重, 為可學習的權重,其係在深度學習模型訓練過程中決定,且用於控制每個特徵的重要程度。 in, is the evaluation score and is a scalar quantity (Scalar), is the global pooling vector, is the first weight of the fully connected layer, is the second weight of the fully connected layer, and It is a learnable weight, which is determined during the training process of the deep learning model and is used to control the importance of each feature.

在上述深度學習模型中,可進一步通過產生分類激活地圖(Class Activation Map, CAM)來表示在玻片級待視覺化影像IMG上,判別為癌症的機率。可進一步參考圖3,其爲根據本發明實施例所繪示用於產生分類激活地圖的流程圖。In the above-mentioned deep learning model, a Class Activation Map (CAM) can be further generated to represent the probability of being diagnosed as cancer on the slide-level image IMG to be visualized. Further reference may be made to FIG. 3 , which is a flow chart for generating a classification activation map according to an embodiment of the present invention.

步驟S300:將預池化特徵地圖PPFM對尺寸維度( )拆解為多個向量Vi,以產生向量集合。向量集合可表示為 ,且多個向量各具有對應於該些特徵的多個頻道單元。 Step S300: Convert the pre-pooled feature map PPFM to the size dimension ( ) into multiple vectors Vi to produce a vector set. The vector set can be expressed as , and each of the multiple vectors has multiple channel units corresponding to the features.

步驟S301:將向量集合的向量以全連接層FC的第一權重及第二權重進行權重加總以產生加總評分向量,由下式表示:Step S301: Sum the vectors of the vector set using the first weight and the second weight of the fully connected layer FC to generate a summed scoring vector, which is expressed by the following formula:

.

其中 為該加總評分向量, 為向量集合,W為第一權重,b為第二權重。 in is the summed rating vector, is a set of vectors, W is the first weight, and b is the second weight.

步驟S302:將加總評分向量進行拼接以產生分類激活地圖。其中,分類激活地圖為二維張量,其大小為尺寸維度( ),且分類激活地圖中,每個位置的值表示在該位置對應的玻片級待視覺化影像IMG上,判別為癌症的機率。而本發明進一步利用可藉由所產生的分類激活地圖中,各個位置上數值的大小來輔助標示出癌症病灶區域。 Step S302: Splice the summed score vectors to generate a classification activation map. Among them, the classification activation map is a two-dimensional tensor, and its size is the size dimension ( ), and in the classification activation map, the value of each position represents the probability of being diagnosed as cancer on the slide-level image to be visualized IMG corresponding to the position. The present invention further utilizes the magnitude of the values at each position in the generated classification activation map to assist in marking the cancer lesion area.

深度學習模型作為一個強分類器 (Strong Classifier) ,會盡可能將訓練資料 (Training Data) 中的所有資訊做擷取。因此實務上,在分辨癌症的工作中,除了癌細胞與癌症直接相關,深度學習模型必須學習擷取癌細胞型態外,經常伴隨癌細胞出現的細胞壞死 (Necrosis)、結締組織增生 (Desmoplasia) 亦會被深度學習模型辨識而做出 「疑似癌症」的判斷,這會使得CAM標示此類區域。As a strong classifier, the deep learning model will try its best to capture all the information in the training data. Therefore, in practice, in the work of distinguishing cancer, in addition to the direct correlation between cancer cells and cancer, the deep learning model must learn to capture the shape of cancer cells, as well as the cell necrosis (Necrosis) and desmoplasia (Desmoplasia) that often accompany cancer cells. It will also be recognized by the deep learning model and make a "suspected cancer" judgment, which will cause the CAM to mark such areas.

事實上,在一個用玻片級訓練集所訓練完成的深度學習模型中,癌細胞、壞死、結締組織增生等特徵會被模型以預池化特徵圖(Pre-Pool Feature Map) 中的不同頻道 (Channel) 所表示。也就是說,預池化特徵圖上的有些頻道識別癌細胞,另外的一些頻道識別壞死。然而,在上述產生 CAM的流程中,這些值經過加權加總而產生最終的預測結果後,並無法藉由單一一個數字來識別是由於辨識到癌細胞還是壞死造成數值偏高的狀況。In fact, in a deep learning model trained with a slide-level training set, features such as cancer cells, necrosis, and connective tissue hyperplasia will be used by the model as different channels in the Pre-Pool Feature Map. (Channel) represents. That is, some channels on the pre-pooled feature map identify cancer cells, and other channels identify necrosis. However, in the above process of generating CAM, after these values are weighted and summed to produce the final prediction result, it is not possible to identify with a single number whether the high value is due to the identification of cancer cells or necrosis.

為此,本發明基於上述前提來將癌細胞特徵與其他伴隨特徵做分離,即藉由分析預池化特徵圖中的每個向量的分佈,來得到哪些頻道(channel)是用來識別癌細胞的。To this end, the present invention is based on the above premise to separate cancer cell features from other accompanying features, that is, by analyzing the distribution of each vector in the pre-pooled feature map, to obtain which channels are used to identify cancer cells of.

癌細胞區域對應到的向量集合中的向量有很低的分類內距離(intra-class dissimilarity),因為癌細胞的特徵會使得擷取癌細胞特徵的頻道有很高的數值,其他頻道則有很低的數值,使用任意的距離評估方式,例如歐式距離 (Euclidian distance)及餘弦相似性(cosine similarity)都能得到較低的數值;相反的,癌細胞區域及壞死區域對應到的向量集合中的向量之間則有很高的分類間距離(inter-class dissimilarity),因為這兩類區域分別對不同的頻道激活(activate)。The vectors in the vector set corresponding to the cancer cell area have a very low intra-class dissimilarity, because the characteristics of cancer cells will cause the channel that captures the characteristics of cancer cells to have a high value, while other channels will have a very high value. For low values, any distance evaluation method, such as Euclidian distance and cosine similarity, can yield lower values; on the contrary, in the vector set corresponding to the cancer cell area and necrosis area There is a high inter-class dissimilarity between vectors because the two types of regions activate different channels respectively.

有這樣的特性,就能藉由使用分群(clustering)演算法將癌細胞、壞死區域的向量分進不同的分群(cluster)中,做到分離癌細胞及壞死區域的效果。With such characteristics, the vectors of cancer cells and necrotic areas can be divided into different clusters by using a clustering algorithm to achieve the effect of separating cancer cells and necrotic areas.

請復參考圖1,方法進入步驟S103:將預池化特徵地圖PPFM對尺寸維度( )拆解為多個向量Vi,以產生向量集合。此步驟與步驟S300相同,故不在此贅述。 Please refer to Figure 1 again. The method proceeds to step S103: compare the pre-pooled feature map PPFM with the size dimension ( ) into multiple vectors Vi to produce a vector set. This step is the same as step S300, so it will not be described again here.

步驟S104:通過分群演算法將向量集合依據分群參數分爲多個分群Gi。分群演算法可例如採用k-means 演算法,且可例如以歐式距離(Euclidean distance)作爲評估距離(dissimilarity)的標準。在本發明的實施例中,分群參數可例如為該些分群的數量,例如k(可設定k=5)。k的值需要人工調整,原則上k只要夠大就能將癌細胞及其他類細胞分離,太大的k值可能會將癌細胞區域分成兩群或兩群以上。Step S104: Divide the vector set into multiple clusters Gi according to the clustering parameters through a clustering algorithm. The clustering algorithm may, for example, adopt the k-means algorithm, and may, for example, use Euclidean distance as a criterion for evaluating distance (dissimilarity). In an embodiment of the present invention, the grouping parameter may be, for example, the number of these groups, such as k (k can be set to 5). The value of k needs to be manually adjusted. In principle, cancer cells and other types of cells can be separated as long as k is large enough. A too large k value may divide the cancer cell area into two or more groups.

步驟S105:將該些分群轉換為多個分群影像並呈現於玻片級待視覺化影像上。如先前描述的,由於太大的k值可能會將癌細胞區域分成兩群以上(含),因此可將每個分群內的區域呈現在原圖上,藉由最終的人工校閱來確認其中的哪幾群為癌細胞的應標註區域,並將辨認為癌細胞之分群對應回原玻片影像對應位置做標記。Step S105: Convert the clusters into multiple cluster images and present them on the slide-level image to be visualized. As described previously, since a too large k value may divide the cancer cell area into two or more groups (inclusive), the areas within each group can be presented on the original image, and the final manual review can be used to confirm which of them Several groups of cancer cells should be marked in the area, and the groups identified as cancer cells should be mapped back to the corresponding positions in the original slide image to mark.

步驟S106:依據分類激活地圖計算該些分群的平均分類激活地圖。可參考圖4,其爲根據本發明實施例繪示的玻片級待視覺化影像、平均分類激活地圖及應標註分群的示意圖。Step S106: Calculate the average classification activation map of the clusters based on the classification activation map. Please refer to FIG. 4 , which is a schematic diagram of a slide-level image to be visualized, an average classification activation map, and groups to be labeled according to an embodiment of the present invention.

步驟S107:依據該些分群影像於該玻片級待視覺化影像IMG上的對應關係及平均CAM,篩選出該些分群中,對應於玻片級待視覺化影像IMG中的癌症細胞的至少一應標註分群。Step S107: Based on the correspondence between the grouped images on the slide-level image IMG to be visualized and the average CAM, select at least one of the groupings corresponding to the cancer cells in the slide-level image IMG to be visualized. Grouping should be marked.

步驟S108:將至少一應標註分群依據分類激活地圖標註玻片級待視覺化影像中。如圖4所示,應標註分群可經過圖像化後疊加於玻片級待視覺化影像上。Step S108: Group at least one annotated group according to the classification activation map annotation in the image to be visualized at the slide level. As shown in Figure 4, the labeled groups can be imaged and superimposed on the slide-level image to be visualized.

可進一步參考圖5,其爲本發明的用於癌症病灶的視覺化方法的視覺化結果與先前技術的比較結果。其中圖5的(a)、(c)部分為現有的CAM視覺化結果,而圖5的(b)、(d)部分為本發明的CAM視覺化結果。如圖所示,相較於現有的CAM演算法,本發明的用於癌症病灶的視覺化方法能減少偽陽性(False Positive)區域面積(如圖5的(a)、(c)部分所示圓框處,即正常組織被錯誤標示為有癌症的區域),並維持真陽性(True Positive)區域(即癌症病灶被正確標示)。Further reference may be made to FIG. 5 , which is a comparison between the visualization results of the method for visualizing cancer lesions of the present invention and the prior art. Parts (a) and (c) of Figure 5 are the existing CAM visualization results, while parts (b) and (d) of Figure 5 are the CAM visualization results of the present invention. As shown in the figure, compared with the existing CAM algorithm, the visualization method for cancer lesions of the present invention can reduce the area of false positive (False Positive) areas (as shown in parts (a) and (c) of Figure 5 The round box, that is, the area where normal tissue is incorrectly marked as having cancer), and maintains the True Positive area (that is, the cancer lesion is correctly marked).

[實施例的有益效果][Beneficial effects of the embodiment]

本發明的其中一有益效果在於,本發明所提供的用於癌症病灶的視覺化方法,其能從基於玻片級運算的癌症辨識深度學習模型擷取預池化層特徵地圖,並使用分群演算法產生多個分群並標註於玻片級待標註圖像上,進而輔以從深度學習模型擷取的分類激活地圖篩選出應標註分群,最終將癌症病灶進行視覺化。One of the beneficial effects of the present invention is that the visualization method for cancer lesions provided by the present invention can extract the pre-pooling layer feature map from the deep learning model of cancer identification based on slide-level operations and use grouping calculations. The method generates multiple clusters and labels them on the slide-level image to be labeled, and then uses the classification activation map extracted from the deep learning model to select the clusters that should be labeled, and finally visualizes the cancer lesions.

因此,本發明所提供的用於癌症病灶的視覺化方法可提升演算法分離癌細胞及壞死區域的效果,並減少演算法在視覺化癌症病灶時,誤將壞死區域辨識為癌細胞的情形,改良了以玻片級圖像運算所訓練之模型視覺化病灶區域的演算法。Therefore, the visualization method for cancer lesions provided by the present invention can improve the effect of the algorithm in separating cancer cells and necrotic areas, and reduce the situation where the algorithm mistakenly identifies necrotic areas as cancer cells when visualizing cancer lesions. The algorithm for visualizing the lesion area using a model trained with slide-level image operations has been improved.

以上所公開的內容僅為本發明的優選可行實施例,並非因此侷限本發明的申請專利範圍,所以凡是運用本發明說明書及圖式內容所做的等效技術變化,均包含於本發明的申請專利範圍內。The contents disclosed above are only preferred and feasible embodiments of the present invention, and do not limit the scope of the patent application of the present invention. Therefore, all equivalent technical changes made by using the description and drawings of the present invention are included in the application of the present invention. within the scope of the patent.

IN:輸入層 HID:隱含層 OUT:輸出層 FEN:特徵擷取網路 GP:全局池化層 FC:全連接層 IMG:玻片級待視覺化影像 PPFM:預池化特徵地圖 Vi:向量 Gi:分群 IN: input layer HID: hidden layer OUT: output layer FEN: Feature Extraction Network GP: global pooling layer FC: fully connected layer IMG: slide-level image to be visualized PPFM: Pre-pooled feature map Vi: vector Gi: grouping

圖1為根據本發明實施例繪示的用於癌症病灶的視覺化方法的流程圖。FIG. 1 is a flow chart of a method for visualizing cancer lesions according to an embodiment of the present invention.

圖2為根據本發明實施例繪示的用於癌症病灶的視覺化方法的流程示意圖。FIG. 2 is a schematic flowchart of a method for visualizing cancer lesions according to an embodiment of the present invention.

圖3爲根據本發明實施例所繪示用於產生分類激活地圖的流程圖。FIG. 3 is a flow chart for generating a classification activation map according to an embodiment of the present invention.

圖4爲根據本發明實施例繪示的玻片級待視覺化影像、平均分類激活地圖及分群影像的示意圖。FIG. 4 is a schematic diagram of a slide-level image to be visualized, an average classification activation map, and a grouped image according to an embodiment of the present invention.

圖5爲本發明的用於癌症病灶的視覺化方法的視覺化結果與先前技術的比較結果。Figure 5 is a comparison between the visualization results of the method for visualizing cancer lesions of the present invention and the prior art.

代表圖為流程圖,故無符號簡單說明 The representative diagram is a flow chart, so there are no symbols for simple explanation.

Claims (10)

一種用於癌症病灶的視覺化方法,其包括下列步驟:取得一玻片級待視覺化影像,其中,該玻片級待視覺化影像經過一經訓練深度學習模型判斷是否具有癌症細胞;在判斷具有癌症細胞時,將該玻片級待視覺化影像輸入該經訓練深度學習模型,其中該經訓練深度學習模型包括一特徵擷取網路、一全局池化層及一全連接層;通過該特徵擷取網路將輸入的該玻片級待視覺化影像進行特徵擷取(feature extraction)以產生一預池化特徵地圖(pre-pool feature map),其中,該預池化特徵地圖包括多個元素,各該元素用於表示多個特徵的其中之一是否出現在該玻片級待視覺化影像的多個位置的其中之一;將該預池化特徵地圖對一尺寸維度拆解為多個向量,以產生一向量集合,其中該些向量各具有對應於該些特徵的多個頻道單元;通過一分群演算法將該向量集合依據一分群參數分為多個分群;將該些分群轉換為多個分群影像並呈現於該玻片級待視覺化影像上;依據該些分群影像於該玻片級待視覺化影像上的對應關係,篩選出該些分群中,對應於該玻片級待視覺化影像中的癌症細胞的至少一應標註分群;以及將該至少一應標註分群依據一分類激活地圖(Class Activation Map,CAM)標註該玻片級待視覺化影像中。 A method for visualizing cancer lesions, which includes the following steps: obtaining a slide-level image to be visualized, wherein the slide-level image to be visualized is judged by a trained deep learning model to determine whether there are cancer cells; When cancer cells are detected, the slide-level image to be visualized is input into the trained deep learning model, wherein the trained deep learning model includes a feature extraction network, a global pooling layer and a fully connected layer; through the feature The acquisition network performs feature extraction on the input slide-level image to be visualized to generate a pre-pool feature map, where the pre-pool feature map includes multiple Elements, each element is used to indicate whether one of the multiple features appears in one of the multiple locations of the slide-level image to be visualized; the pre-pooled feature map is disassembled into multiple dimensions in one dimension. vectors to generate a vector set, wherein each of the vectors has multiple channel units corresponding to the characteristics; use a grouping algorithm to divide the vector set into multiple groups according to a grouping parameter; convert the groupings are multiple grouped images and are presented on the slide-level image to be visualized; based on the correspondence between the grouped images on the slide-level image to be visualized, filter out the groups corresponding to the slide-level images. At least one of the cancer cells in the image to be visualized should be labeled; and the at least one should be labeled in the slide-level image to be visualized based on a Class Activation Map (CAM). 如請求項1所述的用於癌症病灶的視覺化方法,其中該癌症細胞為肺腺癌細胞或鱗狀細胞癌。 The method for visualizing cancer lesions as described in claim 1, wherein the cancer cells are lung adenocarcinoma cells or squamous cell carcinomas. 如請求項1所述的用於癌症病灶的視覺化方法,其中該經訓 練深度學習模型包括一輸入層、多個隱含層及一輸出層,且該些隱含層包括該特徵擷取網路、該全局池化層及該全連接層,且該特徵擷取網路包括選自由卷積層、池化層及標準化層組成的群組的多個層。 The method for visualizing cancer lesions as described in claim 1, wherein the training The deep learning model includes an input layer, a plurality of hidden layers and an output layer, and the hidden layers include the feature extraction network, the global pooling layer and the fully connected layer, and the feature extraction network The path includes a plurality of layers selected from the group consisting of convolutional layers, pooling layers, and normalization layers. 如請求項1所述的用於癌症病灶的視覺化方法,其中該預池化特徵地圖為一H×W×C大小的張量(Tensor),其中H×W為該尺寸維度,且對應於該張量的高與寬,C為一頻道數量,且HW維度對應於該玻片級待視覺化影像的高與寬。 The method for visualizing cancer lesions as described in request 1, wherein the pre-pooled feature map is a tensor (Tensor) of size H × W × C , where H × W is the size dimension, and corresponds to The height and width of the tensor, C is the number of channels, and the H and W dimensions correspond to the height and width of the slide-level image to be visualized. 如請求項4所述的用於癌症病灶的視覺化方法,更包括:通過該全局池化層用於將該預池化特徵地圖中的該尺寸維度進行降維,以產生一全局池化向量;以及通過該全連接層對該全局池化向量進行加權加總,以產生一評估分數,其中該評估分數用於指示該玻片級待視覺化影像是否包含癌症細胞,且由下式表示:Z=W.E+b,其中Z為該評估分數且為一純量,E為該全局池化向量,W為該全連接層的第一權重,b為該全連接層的第二權重。 The method for visualizing cancer lesions as described in claim 4, further comprising: using the global pooling layer to reduce the size dimension in the pre-pooling feature map to generate a global pooling vector. ; and perform a weighted summation of the global pooling vector through the fully connected layer to generate an evaluation score, where the evaluation score is used to indicate whether the slide-level image to be visualized contains cancer cells, and is expressed by the following formula: Z=W. E+b, where Z is the evaluation score and is a scalar, E is the global pooling vector, W is the first weight of the fully connected layer, and b is the second weight of the fully connected layer. 如請求項5所述的用於癌症病灶的視覺化方法,更包括:將該向量集合的該些向量以該全連接層的該第一權重及該第二權重進行權重加總以產生一加總評分向量,由下式表示:Z' hw=W.E' hw+b,其中Z' hw為該加總評分向量,E' hw為該向量集合,W為該全連接層的第一權重,b為該全連接層的第二權重;將該加總評分向量進行拼接以產生該分類激活地圖,其中該分類激活地圖為一二維張量,其大小為該尺寸維度,且該 CAM的每個位置的值表示在該位置對應的該預池化特徵地圖上,判別為癌症細胞的機率。 The method for visualizing cancer lesions as described in claim 5, further comprising: summing the vectors of the vector set with the first weight and the second weight of the fully connected layer to generate a sum The total rating vector is expressed by the following formula: Z ' hw =W. E ' hw + b, where Z ' hw is the total score vector, E ' hw is the vector set, W is the first weight of the fully connected layer, and b is the second weight of the fully connected layer; add the The total score vectors are spliced to generate the classification activation map, where the classification activation map is a two-dimensional tensor whose size is the size dimension, and the value of each position of the CAM represents the pre-pooling corresponding to that position. On the feature map, the probability of identifying cancer cells. 如請求項6所述的用於癌症病灶的視覺化方法,更包括:依據該分類激活地圖計算該些分群的多個平均分類激活地圖;以及依據該些分群影像於該玻片級待視覺化影像上的對應關係及該些平均分類激活地圖,篩選出該些分群中的該至少一應標註分群。 The method for visualizing cancer lesions as described in claim 6, further comprising: calculating a plurality of average classification activation maps of the clusters based on the classification activation map; and based on the clustering images to be visualized at the slide level The corresponding relationship on the image and the average classification activation map are used to filter out at least one of the clusters that should be labeled. 如請求項1所述的用於癌症病灶的視覺化方法,其中該經訓練深度學習模型為ResNet或DenseNet。 The method for visualizing cancer lesions as described in claim 1, wherein the trained deep learning model is ResNet or DenseNet. 如請求項1所述的用於癌症病灶的視覺化方法,其中該分群演算法為k-means演算法。 The method for visualizing cancer lesions as described in claim 1, wherein the grouping algorithm is a k-means algorithm. 如請求項9所述的用於癌症病灶的視覺化方法,其中該k-means演算法以歐式距離作為評估距離的標準,且該分群參數為該些分群的數量。The method for visualizing cancer lesions as described in claim 9, wherein the k-means algorithm uses Euclidean distance as a criterion for evaluating distance, and the clustering parameter is the number of clusters.
TW109139265A 2020-11-11 2020-11-11 Visualization methods for cancer lesions TWI815057B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW109139265A TWI815057B (en) 2020-11-11 2020-11-11 Visualization methods for cancer lesions

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW109139265A TWI815057B (en) 2020-11-11 2020-11-11 Visualization methods for cancer lesions

Publications (2)

Publication Number Publication Date
TW202219827A TW202219827A (en) 2022-05-16
TWI815057B true TWI815057B (en) 2023-09-11

Family

ID=82558699

Family Applications (1)

Application Number Title Priority Date Filing Date
TW109139265A TWI815057B (en) 2020-11-11 2020-11-11 Visualization methods for cancer lesions

Country Status (1)

Country Link
TW (1) TWI815057B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI689944B (en) * 2017-11-30 2020-04-01 美商南坦生物組學有限責任公司 Using deep learning to detect intra-tumor heterogeneity of molecular subtypes in pathological slice images
CN111553462A (en) * 2020-04-08 2020-08-18 哈尔滨工程大学 A Class Activation Mapping Method
US20200272864A1 (en) * 2017-11-06 2020-08-27 University Health Network Platform, device and process for annotation and classification of tissue specimens using convolutional neural network
CN111899212A (en) * 2020-05-28 2020-11-06 江苏君英天达人工智能研究院有限公司 Pulmonary nodule CT image detection method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200272864A1 (en) * 2017-11-06 2020-08-27 University Health Network Platform, device and process for annotation and classification of tissue specimens using convolutional neural network
TWI689944B (en) * 2017-11-30 2020-04-01 美商南坦生物組學有限責任公司 Using deep learning to detect intra-tumor heterogeneity of molecular subtypes in pathological slice images
CN111553462A (en) * 2020-04-08 2020-08-18 哈尔滨工程大学 A Class Activation Mapping Method
CN111899212A (en) * 2020-05-28 2020-11-06 江苏君英天达人工智能研究院有限公司 Pulmonary nodule CT image detection method and system

Also Published As

Publication number Publication date
TW202219827A (en) 2022-05-16

Similar Documents

Publication Publication Date Title
CN109166105B (en) Tumor malignancy risk layered auxiliary diagnosis system based on artificial intelligent medical image
Sainju et al. Automated bleeding detection in capsule endoscopy videos using statistical features and region growing
CN110889853B (en) Tumor segmentation method based on residual error-attention deep neural network
US7620222B2 (en) Method, apparatus, and program for judging images
CN111524170B (en) Pulmonary CT image registration method based on unsupervised deep learning
KR102559616B1 (en) Method and system for breast ultrasonic image diagnosis using weakly-supervised deep learning artificial intelligence
CN115631387B (en) Method and device for predicting pathological high risk factors of lung cancer based on graph convolutional neural network
CN111275699B (en) Methods, devices, equipment, and storage media for medical image processing
Shi et al. Automatic nodule segmentation method for CT images using aggregation-U-Net generative adversarial networks
CN114170133B (en) Intelligent focus positioning method, system, device and medium based on medical image
US11282193B2 (en) Systems and methods for tumor characterization
CN108427963B (en) A classification and recognition method of melanoma skin disease based on deep learning
Huang et al. Benign-malignant classification of pulmonary nodule with deep feature optimization framework
Sun et al. A novel gastric ulcer differentiation system using convolutional neural networks
Zhang et al. A lung cancer detection and recognition method combining convolutional neural network and morphological features
Khattar et al. Deep learning-based lung cancer detection using convolutional neural networks
Bhatt et al. Image retrieval using bag-of-features for lung cancer classification
TWI815057B (en) Visualization methods for cancer lesions
CN115114984A (en) Method and system for assisting in building image omics model based on semi-supervised learning
Samundeeswari et al. An efficient fully automated lung cancer classification model using googlenet classifier
Yang et al. Leveraging auxiliary information from EMR for weakly supervised pulmonary nodule detection
Lin et al. EDSNet: A deep supervision-based classification framework for non-standard medical images of lung cancer
Pradhan An early diagnosis of lung nodule using CT images based on hybrid machine learning techniques
CN109978846B (en) A system and method for pulmonary nodule texture feature extraction based on body local orientation ternary pattern
Liu et al. Rethinking pulmonary nodule detection in multi-view 3D CT point cloud representation