HK40051109B - Multiple instance learner for prognostic tissue pattern identification - Google Patents
Multiple instance learner for prognostic tissue pattern identification Download PDFInfo
- Publication number
- HK40051109B HK40051109B HK62021041323.8A HK62021041323A HK40051109B HK 40051109 B HK40051109 B HK 40051109B HK 62021041323 A HK62021041323 A HK 62021041323A HK 40051109 B HK40051109 B HK 40051109B
- Authority
- HK
- Hong Kong
- Prior art keywords
- block
- blocks
- image
- mil
- patient
- Prior art date
Links
Description
技术领域Technical Field
本发明涉及数字病理学领域,更特别地涉及图像分析领域。This invention relates to the field of digital pathology, and more particularly to the field of image analysis.
背景技术Background Technology
已知若干种图像分析方法,这些图像分析方法可用于辅助诊断过程以及基于对组织样品图像的分析来识别合适的治疗。Several image analysis methods are known that can be used to assist in the diagnostic process and to identify appropriate treatments based on the analysis of tissue sample images.
一些图像分析技术基于使用不同的程序来搜索图像中的结构,已知该结构用作特定疾病的存在和/或用特定药物成功治疗该疾病的可能性的指标。例如,只有在距癌细胞一定距离处存在某些免疫细胞的情况下,在癌症患者免疫疗法过程中使用的一些药物才起作用。在这种情况下,尝试自动识别组织图像中的这些对象,即某些细胞类型或某些亚细胞和超细胞结构,以便能够做出关于疾病的存在情况和/或推荐的治疗的说明。这种方法的缺点是图像分析算法只识别那些为该图像分析算法开发的结构。因此,这种类型的图像分析基于关于某些细胞和组织结构与某些疾病或它们的治疗方案之间关系的现有医学知识。因此,该图像分析方法不适合检测关于某种疾病和/或该疾病治疗的未知预测特征,并且受限于在某个时间可用的医学知识。该图像分析方法不适合扩展医疗关系的知识,即识别迄今为止未知的特征和组织结构,以预测是否存在某种形式的疾病和/或某种药物是否对这种疾病有效。Some image analysis techniques are based on using different procedures to search for structures in images known to serve as indicators of the presence of a specific disease and/or the likelihood of successful treatment with a particular drug. For example, some drugs used in immunotherapy for cancer patients are effective only if certain immune cells are present at a certain distance from cancer cells. In this case, an attempt is made to automatically identify these objects in tissue images—that is, certain cell types or certain subcellular and supercellular structures—in order to make a statement about the presence of a disease and/or recommended treatment. A drawback of this approach is that the image analysis algorithm only identifies structures developed for that algorithm. Therefore, this type of image analysis relies on existing medical knowledge about the relationship between certain cell and tissue structures and certain diseases or their treatments. Consequently, this image analysis method is not suitable for detecting unknown predictive features about a disease and/or its treatment, and is limited by the medical knowledge available at a given time. This image analysis method is also not suitable for extending the knowledge of medical relationships, i.e., identifying features and tissue structures previously unknown to predict the presence of a certain form of disease and/or the effectiveness of a certain drug for that disease.
其他图像分析方法,特别是非监督式机器学习方法,也能够考虑其预测能力为专业界所未知的和/或病理学家在图像分析中无法察觉的组织模式和特征,因为这些特征可以是,例如,由若干其他特征的存在、不存在和/或可表达性产生的衍生特征。这些方法的缺点是它们通常像黑箱一样工作。换句话说,使用这些技术的病理学家必须依赖于这些算法的预测能力,而无法准确指出哪种组织性状对预测具有最终决定性作用。这可能是一个显著缺点,例如在药物批准中,因为出于此目的,必须明确指出受益于某种治疗的患者群体。在决定给某个患者服用潜在有效但副作用强烈的药物是否可行时,不得不完全或部分依赖这个“黑箱”,而不能以语言表述潜在的“决策逻辑”,这对于医生和患者来说都是不尽如人意的。Other image analysis methods, particularly unsupervised machine learning methods, can also consider their predictive power in tissue patterns and features unknown to the professional community and/or imperceptible to pathologists in image analysis. This is because these features can be, for example, derived features resulting from the presence, absence, and/or expressibility of several other features. The drawback of these methods is that they often operate like black boxes. In other words, pathologists using these techniques must rely on the predictive power of these algorithms without being able to pinpoint exactly which tissue characteristics are ultimately decisive in the prediction. This can be a significant drawback, for example, in drug approval, where the patient population that will benefit from a particular treatment must be clearly identified for this purpose. Having to rely entirely or partially on this "black box" when deciding whether to administer a potentially effective but highly contagious drug to a patient, without being able to articulate the underlying "decision logic," is undesirable for both doctors and patients.
发明内容Summary of the Invention
本发明目的是提供一种经改进的识别指示患者相关属性值的组织模式的方法以及一种如独立权利要求中指出的相应图像分析系统。在从属权利要求中给出了本发明的实施例。如果本发明的实施例不是互相排斥的,则可以彼此自由地组合。The object of this invention is to provide an improved method for identifying tissue patterns that indicate patient-related attribute values, and a corresponding image analysis system as pointed out in the independent claims. Embodiments of the invention are given in the dependent claims. Embodiments of the invention may be freely combined with each other if they are not mutually exclusive.
在一个方面,本发明涉及一种识别指示患者相关属性值的组织模式的方法。该方法包括:In one aspect, the present invention relates to a method for identifying organizational patterns that indicate patient-related attribute values. The method includes:
-针对一组患者中的每个患者,通过图像分析系统,接收该患者的组织样品的至少一个数字图像,该至少一个图像已分配至少两个不同的预定义标签中的一个标签,每个标签指示在用标签标记的图像中描绘其组织的患者的患者相关属性值;- For each patient in a group of patients, at least one digital image of a tissue sample of that patient is received via an image analysis system, the at least one image being assigned one of at least two different predefined labels, each label indicating a patient-related attribute value of the patient whose tissue is depicted in the labeled image;
-通过图像分析系统,将每个接收到的图像拆分成图像块集,每个块已分配了分配给用于创建块的图像的标签;- Using an image analysis system, each received image is split into a set of image blocks, and each block has been assigned a label that was used to create the image;
-针对所述块中的每一个块,通过图像分析系统,计算特征向量,该特征向量包含从所述块中描绘的组织模式选择性地提取的图像特征;- For each of the blocks, a feature vector is calculated using an image analysis system. This feature vector contains image features selectively extracted from the organizational patterns depicted in the block.
-基于针对该组中的所有患者接收到的所有图像的所有块和相应特征向量训练多实例学习(MIL)程序,每个块集被MIL程序处理为具有相同标签的块包,该训练包括分析特征向量以便针对所述块中的每一个计算数值,该数值指示与块相关联的特征向量相对于分配给导出块的图像的标签的预测能力;以及- A multi-instance learning (MIL) procedure is trained based on all blocks and corresponding feature vectors from all images received for all patients in the group. Each block set is processed by the MIL procedure into a block bundle with the same label. The training includes analyzing the feature vectors to compute a numerical value for each block, which indicates the predictive power of the feature vector associated with the block relative to the label of the image assigned to the derived block; and
-经由图像分析系统的GUI,输出图像块报告库,该报告库包括块的子集(该块的子集根据它们的相应计算出的数值进行排序),及/或包括它们的相应数值的图形表示。- The image analysis system outputs an image patch report library via its GUI. This report library includes a subset of patches (which are sorted according to their respective calculated values) and/or a graphical representation of their respective values.
这种方法可能是有利的,因为它可将基于明确生物医学专家知识的图像分析方法的优势与机器学习方法的优势结合起来:在机器学习中,多实例学习(MIL)是一种监督式学习。学习器不是接收单独用标签标记的实例集,而是接收用标签标记的包的集合,每个集合包含许多实例。在多实例二元分类的简单情况下,如果包中的所有实例都是负,则该包可用标签标记为负。在另一方面,如果包中至少有一个实例为正,则该包用标签标记为正。从一组用标签标记的包中,学习器尝试(i)引入将正确用标签标记单独示例的概念,或(ii)学习如何在不引入概念的情况下用标签标记包。在Babenko,Boris."Multiple instancelearning:algorithms and applications"(2008)中给出了MIL的方便和简单的示例。然而,根据一些实施例的MIL程序还涵盖基于两个以上不同标签(终点)的训练。This approach can be advantageous because it combines the strengths of image analysis methods based on explicit biomedical expert knowledge with the strengths of machine learning methods: in machine learning, multiple instance learning (MIL) is a form of supervised learning. Instead of receiving individual sets of labeled instances, the learner receives sets of labeled packets, each containing many instances. In the simple case of multi-instance binary classification, a packet is labeled negative if all instances in it are negative. On the other hand, a packet is labeled positive if at least one instance is positive. From a set of labeled packets, the learner attempts to (i) introduce the concept that will correctly label individual instances, or (ii) learn how to label packets without introducing the concept. A convenient and simple example of MIL is given in Babenko and Boris, "Multiple instance learning: algorithms and applications" (2008). However, MIL procedures according to some embodiments also cover training based on more than two different labels (endpoints).
根据本发明的实施例,MIL程序用于计算包中的每个实例(块)(优选地,具有特定标签值的某个患者的组织切片的一个或多个图像的所有块)的预测值并且因此也适用于在块中分别描绘的组织模式。在这一步中,MIL程序可识别新的生物医学知识,因为在训练数据中,图像和相应块的标签作为训练的终点,但不是从块中导出的特征向量的单独特征,块与标签强烈(正或负)相关,因此可预测该标签。此外,针对单独块计算出的预测值也与库中相关块的图形表示一起输出。例如,库中的块可根据数值进行排序。在这种情况下,块在库中的位置允许病理学家或其他人类用户识别经发现对特定标签具有高度预测性的块中描绘的组织模式。此外,或替代性地,数值可显示为空间上接近其相应块,从而使用户能够检查和理解在一个或多个块中描绘的组织的组织模式,所述一个或多个块具有与特定标签相似的数值。According to embodiments of the invention, the MIL program is used to calculate predicted values for each instance (block) in the package (preferably, all blocks of one or more images of a tissue slice from a patient with a specific label value) and is therefore also applicable to tissue patterns depicted separately in the blocks. In this step, the MIL program can identify new biomedical knowledge because, in the training data, the labels of the images and corresponding blocks serve as the endpoints of training, but not as individual features of the feature vectors derived from the blocks; the blocks are strongly (positively or negatively) correlated with the labels, and therefore the labels can be predicted. Furthermore, the predicted values calculated for individual blocks are also output along with graphical representations of related blocks in the library. For example, the blocks in the library may be sorted according to numerical values. In this case, the position of a block in the library allows pathologists or other human users to identify tissue patterns depicted in blocks found to be highly predictive of a specific label. Additionally, or alternatively, numerical values may be displayed as spatially proximate to their corresponding blocks, thereby enabling users to examine and understand tissue patterns depicted in one or more blocks that have numerical values similar to a specific label.
因此,作为训练阶段的输出而生成的图像块库可揭示相对于患者的特定患者相关属性值具有预测性的组织签名。结合图像块呈现数值可能具有以下好处:至少在许多情况下,病理学家可通过比较具有相似数值的库中的若干块与具有高得多或低得多的数值的其他块并且通过比较报告库中块的这些子集中描绘的组织签名来识别和表述预测组织模式(也可称为“组织签名”)。Therefore, the image patch library generated as output of the training phase can reveal predictive tissue signatures relative to specific patient-related attribute values. Combining the numerical representation of image patches may have the following benefits: at least in many cases, pathologists can identify and characterize predictive tissue patterns (also known as "tissue signatures") by comparing several patches in the library with those having much higher or lower values, and by comparing the tissue signatures depicted in these subsets of patches in the report library.
在进一步有益的方面,使用将图像块作为实例处理的MIL程序以及已分配特定标签的同一患者的所有图像的所有块的整体(例如“对药物D有应答=真”、“微卫星状态=MSX”、“HER2表达状态=+”)特别适用于在整个载玻片组织样品图像的情况下预测患者相关特征。这是因为整个载玻片组织样品通常覆盖许多不同的组织区域,只有一些组织区域可具有任何预测值。例如,微转移的直径可能只有几毫米,但载玻片和相应的整个载玻片图像的长度可达许多厘米。尽管整个图像是用标签标记的-根据对样品来源患者的经验观察-使用特定标签,例如“对药物D有应答=真”,包括许多免疫细胞且预测正向应答的微转移周围的组织区域也可能仅覆盖几毫米。因此,大多数块不包括相对于图像方式和通常的患者方式标签可预测的任何组织区域。MIL程序特别适用于基于数据实例包识别预测特征,其中假定大部分实例没有任何预测值。In a further advantageous aspect, the use of the MIL procedure, which treats image patches as instances and the entirety of all patches across all images of the same patient with specific labels assigned (e.g., "Response to drug D = True", "Microsatellite status = MSX", "HER2 expression status = +"), is particularly suitable for predicting patient-related features in the case of whole slide tissue sample images. This is because whole slide tissue samples typically cover many different tissue regions, only some of which may have any predictive value. For example, a micrometastasis may only be a few millimeters in diameter, but the slide and the corresponding whole slide image can be many centimeters long. Even though the entire image is labeled—based on empirical observation of the patient from whom the sample originated—using specific labels such as "Response to drug D = True," the tissue region surrounding a micrometastasis that includes many immune cells and predicts a positive response may only cover a few millimeters. Therefore, most patches do not include any tissue regions that are predictable relative to the image manner and the usual patient manner of labeling. The MIL procedure is particularly suitable for identifying predictive features based on data instance packages, where it is assumed that most instances have no predictive value.
根据实施例,接收数字图像包括组织样品的数字图像,该组织样品的数字图像的像素强度值与非生物标志物特异性染色剂,特别是H&E染色剂的量相关。According to an embodiment, the received digital image includes a digital image of a tissue sample, the pixel intensity value of which is related to the amount of a non-biomarker-specific staining agent, particularly H&E staining agent.
例如,每包块可代表对特定药物的应答已知的相应患者。该患者专用包中包含的实例是从该特定患者的相应组织样品导出的一个或多个图像的块,该组织样品已用非生物标志物特异性染色剂(诸如H&E)染色。该患者的所有组织图像以及由此导出的所有块已分配标签“患者对药物D有应答=真”。For example, each package block may represent a corresponding patient for whom a response to a specific drug is known. The instances contained in this patient-specific package are blocks of one or more images derived from a corresponding tissue sample from that particular patient, stained with a non-biomarker-specific staining agent (such as H&E). All tissue images of that patient, and all blocks derived therefrom, are assigned the label “Patient responds to drug D = True”.
这可能是有利的,因为H&E染色的组织图像代表染色组织图像的最常见形式,并且仅这种类型的染色已揭示大量可用于预测患者相关属性值的数据,例如特定肿瘤的亚型或分期。此外,许多医院包括从过去多年治疗的患者导出的H&E染色组织图像的大型数据库。通常,医院还拥有关于特定患者是否对特定治疗有应答和/或疾病发展的速度或严重程度的数据。因此,可以得到可用相应的结果(例如,通过特定药物治疗是/否成功、无进展存活期超过一年、无进展存活期超过两年等)用标签标记的训练图像的大量语料库。This can be advantageous because H&E-stained tissue images represent the most common form of stained tissue images, and this type of staining alone has revealed a wealth of data that can be used to predict patient-related attribute values, such as the subtype or stage of a specific tumor. Furthermore, many hospitals include large databases of H&E-stained tissue images derived from patients treated over many years. Typically, hospitals also possess data on whether a particular patient responded to a particular treatment and/or the rate or severity of disease progression. Therefore, a large corpus of training images labeled with corresponding outcomes (e.g., whether treatment with a particular drug was successful, progression-free survival exceeding one year, progression-free survival exceeding two years, etc.) can be obtained.
根据实施例,接收数字图像包括组织样品的数字图像,该组织样品的数字图像的像素强度值与生物标志物特异性染色剂的量相关。生物标志物特异性染色剂是适于对组织样品中包含的生物标志物选择性染色的染色剂。例如,生物标志物可为特定蛋白质,诸如HER-2、p53、CD3、CD8等。生物标志物特异性染色剂可为与选择性结合上述生物标志物的抗体偶联的明视野显微镜或荧光显微镜染色剂。According to an embodiment, the received digital image includes a digital image of a tissue sample, the pixel intensity value of which is related to the amount of biomarker-specific staining agent. The biomarker-specific staining agent is a staining agent suitable for selectively staining biomarkers contained in the tissue sample. For example, the biomarker may be a specific protein, such as HER-2, p53, CD3, CD8, etc. The biomarker-specific staining agent may be a bright-field microscopy or fluorescence microscopy staining agent conjugated to an antibody that selectively binds to the aforementioned biomarker.
例如,每包块可代表对特定药物的应答已知的相应患者。该患者专用包中包含的实例是从该特定患者的相应组织样品导出的一个或多个图像的块。一个或多个组织样品已用一种或多种生物标志物特异性染色剂染色。例如,块可从一个、两个或三个组织图像导出,所有组织图像都描绘了同一患者的相邻组织载玻片已用HER2-特异性染色剂染色。根据另一示例,块可从描绘已用HER2-特异性染色剂染色的第一组织样品的第一组织图像导出,并且从描绘已用p53特异性染色剂染色的第二组织样品的第二组织图像导出,以及从描绘已用FAP-特异性染色剂染色的第三组织样品的第三组织图像导出。第一、第二和第三组织样品从同一患者导出。例如,它们可为相邻的组织样品切片。尽管所述三个组织图像描绘了三个不同的生物标志物,但所有组织图像从同一患者导出,且因此由此衍生的所有块已分配标签“患者对药物D有应答=真”。For example, each package block may represent a corresponding patient whose response to a specific drug is known. The instances contained in this patient-specific package are blocks of one or more images derived from a corresponding tissue sample from that particular patient. One or more tissue samples have been stained with one or more biomarker-specific staining agents. For example, a block may be derived from one, two, or three tissue images, all depicting adjacent tissue slides from the same patient stained with a HER2-specific staining agent. According to another example, a block may be derived from a first tissue image depicting a first tissue sample stained with a HER2-specific staining agent, a second tissue image depicting a second tissue sample stained with a p53-specific staining agent, and a third tissue image depicting a third tissue sample stained with a FAP-specific staining agent. The first, second, and third tissue samples are derived from the same patient. For example, they may be sections of adjacent tissue samples. Although the three tissue images depict three different biomarkers, all tissue images are derived from the same patient, and therefore all blocks derived therefrom are assigned the label "Patient responds to drug D = True".
基于像素强度值与生物标志物特异性染色剂的量相关的数字图像的图像块训练MIL程序可能具有以下优势:识别组织中一种或多种特定生物标志物的存在和位置可揭示关于特定疾病和疾病亚型的高度特异性和预后信息。预后信息可包括观察到的两种或更多种生物标志物的存在的正相关和负相关。例如,已观察到某些疾病(诸如肺癌或结肠癌)的推荐治疗方案和预后在很大程度上取决于癌症的突变签名和表达谱。有时,单独的单个标志物的表达不具有预测能力,但多个生物标志物的组合表达和/或特定的另外的生物标志物的缺失可能具有相对于特定患者相关属性值的高预测能力。Training a MIL program using image patches from digital images that correlate pixel intensity values with the amount of biomarker-specific staining agents may have the following advantages: identifying the presence and location of one or more specific biomarkers in tissue can reveal highly specific and prognostic information about specific diseases and disease subtypes. Prognostic information may include positive and negative correlations between the presence of two or more observed biomarkers. For example, recommended treatments and prognoses for certain diseases (such as lung or colon cancer) have been observed to depend heavily on the mutation signature and expression profile of the cancer. Sometimes, the expression of a single biomarker alone is not predictive, but the combined expression of multiple biomarkers and/or the absence of a specific additional biomarker may have high predictive power relative to patient-specific relevant attribute values.
根据实施例,接收数字图像包括该组织样品的像素强度值与第一生物标志物特异性染色剂的量相关的组织样品的数字图像的组合和该组织样品的像素强度值与非生物标志物特异性染色剂的量相关的组织样品的数字图像的组合。生物标志物特异性染色剂是适于对组织样品中包含的生物标志物选择性染色的染色剂。描绘同一组织样品和/或描绘来自同一患者的相邻组织样品的所有数字图像已分配相同的标签。MIL配置为将从所述数字图像导出的所有块处理为同块包的成员。According to an embodiment, the received digital images include a combination of digital images of tissue samples whose pixel intensity values are related to the amount of a first biomarker-specific staining agent and a combination of digital images of tissue samples whose pixel intensity values are related to the amount of a non-biomarker-specific staining agent. The biomarker-specific staining agent is a staining agent suitable for selectively staining biomarkers contained in a tissue sample. All digital images depicting the same tissue sample and/or adjacent tissue samples from the same patient have been assigned the same label. The MIL is configured to process all blocks derived from the digital images as members of the same block package.
这种方法的优势在于,结合由H&E染色揭示的富含信息的组织签名,识别组织中一种或多种特定生物标志物的存在和位置,可提供关于特定疾病和疾病亚型的高度特异性和预后信息。预后信息可包括观察到的两种或更多种生物标志物的存在的正相关和负相关和/或通过H&E染色在视觉上揭示的组织签名。The advantage of this method lies in its ability to identify the presence and location of one or more specific biomarkers in tissue by combining information-rich tissue signatures revealed by H&E staining, providing highly specific and prognostic information about specific diseases and disease subtypes. Prognostic information may include positive and negative correlations in the presence of two or more observed biomarkers and/or tissue signatures visually revealed by H&E staining.
根据实施例,图像块报告库中显示的图像块从接收到的图像中的一个或多个不同图像导出。方法包括,针对报告块库中描绘的一个或多个图像中的每一个:According to an embodiment, the image patches displayed in the image patch report library are derived from one or more different images among the received images. The method includes, for each of the one or more images depicted in the report patch library:
-识别报告库中的块中的一个块,所述一个块已从所述图像导出并且已分配从所述图像的导出的所有块的最高分;根据一个实施例,该得分是由MIL针对每个块计算出的数值;根据替代性实施例,该得分是通过如本文针对本发明的实施例所述的注意力-MLL针对每个块计算出的权重;根据更进一步的实施例,该得分是由MIL计算出的所述数值和由注意力MLL针对所述块计算出的所述权重的组合,由此该组合可以是,例如,数值与权重的乘法;- Identify a block in a report library, said block having been derived from the image and assigned the highest score among all blocks derived from the image; according to one embodiment, the score is a numerical value calculated by the MIL for each block; according to an alternative embodiment, the score is a weight calculated by the attention-MLL for each block as described herein with respect to embodiments of the invention; according to a further embodiment, the score is a combination of the numerical value calculated by the MIL and the weight calculated by the attention-MLL for the block, whereby the combination may be, for example, a multiplication of the numerical value and the weight;
-针对图像的另外一些块中的每一个,通过将另一个块的得分与具有最高得分的块的得分进行比较来计算相关度指标;相关度指标是与另一个块的得分和具有最高得分的块的得分的差异负相关的数值;- For each of the other blocks in the image, a correlation metric is calculated by comparing the score of the other block with the score of the block with the highest score; the correlation metric is a numerical value that is negatively correlated with the difference between the score of the other block and the score of the block with the highest score.
-作为相关度指标针的函数计算图像的相关度热图;因此,相关度热图的像素颜色和/或像素强度指示针对所述图像中的块计算出的相关度指标;以及- The correlation heatmap of the image is calculated as a function of the correlation index; therefore, the pixel color and/or pixel intensity of the correlation heatmap indicate the correlation index calculated for blocks in the image; and
-显示所述相关度热图。例如,相关度热图可显示为在报告块库中,在空间上接近计算相关度热图的整个载玻片图像。- Display the correlation heatmap. For example, the correlation heatmap can be displayed as an image of the entire slide that is spatially close to the calculation of the correlation heatmap in the report block library.
例如,具有与图像的最高分块的得分高度相似的得分的图像区域和相应块可在相关度热图中用第一颜色(例如“红色”)或高强度值来指示,并且其得分与该图像的块的最高分相异的图像区域和相应块可在相关度热图中用不同于第一颜色的第二颜色(例如“蓝色”)或低强度值来表示。For example, image regions and corresponding blocks with scores highly similar to the highest-scoring block of an image can be indicated in the correlation heatmap with a first color (e.g., "red") or a high-intensity value, and image regions and corresponding blocks whose scores differ from the highest-scoring block of the image can be represented in the correlation heatmap with a second color (e.g., "blue") or a low-intensity value, different from the first color.
这可能是有利的,因为GUI自动计算并呈现相关度热图,该相关度热图指示具有高预测能力(或“预后值”)的组织区域和相应图像块的位置和覆盖范围。相关度热图可突出显示具有高相关度指标的组织区域。块通常只是整个载玻片图像的一个小的子区域,且此类报告块库可能无法提供整个组织样品的概览。关于具有高预测相关度的组织模式的位置和覆盖范围的概览信息可由优选以高度直观和智能的方式与整个载玻片组织图像的原始图像组合的相关度热图提供。This can be advantageous because the GUI automatically calculates and presents a correlation heatmap, which indicates the location and coverage of tissue regions with high predictive power (or "prognostic value") and their corresponding image patches. The correlation heatmap highlights tissue regions with high correlation indices. Patches are often just small sub-regions of the entire slide image, and such reportable patch libraries may not provide an overview of the entire tissue sample. An overview of the location and coverage of tissue patterns with high predictive correlation can be provided by a correlation heatmap, preferably combined with the original image of the entire slide tissue image in a highly intuitive and intelligent manner.
基于MIL的数值计算相关度热图可具有以下优势:可能不需要实施和训练注意力MLL。因此,系统架构可更容易实现。Numerical computation of relevance heatmaps based on MIL may have the following advantages: it may not require the implementation and training of attention MIL. Therefore, the system architecture can be more easily implemented.
基于注意力MLL计算出的权重计算相关度热图可能具有以下优势:除了MIL的数值之外,用于块预后相关度的第二个数值度量在相关度热图中评估和表示。The relevance heatmap calculated based on the weights of attention MIL may have the following advantages: in addition to the numerical value of MIL, a second numerical measure for block prognostic relevance is evaluated and represented in the relevance heatmap.
基于从由MIL计算出的数值和由注意力MLL针对特定块计算出的权重导出的组合相关度得分来计算相关度热图可能具有以下优势:两个独立计算出的块预测相关度数值度量集成在组合值和基于组合值的相关度热图中并由其表示。这可增加相关组织切片识别的准确性。Calculating a relevance heatmap based on a combined relevance score derived from numerical values calculated by MIL and weights calculated by Attention MLL for a specific block may have the following advantages: the two independently calculated block-predicted relevance numerical measures are integrated into and represented by the combined value and the relevance heatmap based on the combined value. This can increase the accuracy of relevance tissue slice identification.
根据实施例,GUI使用户能够选择相关度热图是基于MIL的数值或基于注意力MLL的权重或基于组合得分计算出的。这可以允许用户识别关于块的预测能力的MIL的和注意力MLL的输出是否显著不同。According to an embodiment, the GUI allows users to select whether the relevance heatmap is calculated based on the numerical value of the MIL, the weights of the attention MIL, or a combined score. This allows users to identify whether the outputs of the MIL and the attention MIL differ significantly regarding the predictive power of a block.
计算和显示相关度热图可能是有利的,因为该热图指示关于用于训练MIL和/或注意力MLL的端点的块的预测能力。因此,向用户显示相关度热图使用户能够快速识别具有可预测整个载玻片图像内的特定标签的组织模式的块的位置和覆盖范围。Calculating and displaying a correlation heatmap can be advantageous because it indicates the predictive power of blocks with respect to the endpoints used to train the MIL and/or attention MLL. Therefore, displaying a correlation heatmap to the user enables them to quickly identify the location and extent of blocks with tissue patterns that predict specific labels throughout the entire slide image.
根据实施例,报告库中显示的图像块是可选择的。GUI经配置为计算和显示相似性搜索块库,该计算包括:According to an embodiment, the image patches displayed in the report library are selectable. The GUI is configured to calculate and display a similarity search patch library, the calculation including:
-接收用户对报告库图像块中的特定块的选择;- Receive the user's selection of a specific block from the image blocks in the report library;
-通过识别从所有接收到的图像获得的已分配了特征向量的所有块来识别从所有接收到的图像获得的描绘与所选择块相似的组织模式的所有块,所述特征向量与所选择块的特征向量的相似性超过阈值;以及- Identify all blocks from all received images that depict organizational patterns similar to the selected blocks, where the similarity between the feature vectors of the selected blocks and the feature vectors of the selected blocks exceeds a threshold, by identifying all blocks from all received images that have been assigned feature vectors; and
-显示相似性搜索库,相似性搜索库选择性地包括所述识别的块。- Display a similarity search library, which selectively includes the identified blocks.
根据实施例,相似性搜索块库的计算和显示进一步包括:According to an embodiment, the calculation and display of the similarity search block library further includes:
-确定所述块内的块的数量和/或分数,所述块描绘了与所选块相似的组织模式,该所选块已分配与所选块相同的标签;以及- Determine the number and/or score of blocks within the block, the blocks depicting an organizational pattern similar to the selected block, which has been assigned the same labels as the selected block; and
-在相似性搜索库中显示确定的数量和/或分数。- Display the specified quantity and/or score in the similarity search library.
这些特征可能是有利的,因为人类用户能够快速确定特定组织模式在被检查的患者组中以及在具有特定标签的患者的子集中有多常见。因此,人类用户能够快速且直观地验证特定块和其中描绘的组织模式是否真正具有高预测能力。These features can be advantageous because human users can quickly determine how common a particular tissue pattern is in the examined patient group and in a subset of patients with specific labels. Therefore, human users can quickly and intuitively verify whether a particular block and the tissue pattern depicted within it truly possesses high predictive power.
例如,用户可以选择报告库的块中的一个,该报告库已分配最高数值且因此关于图像标签的最高预测能力。在选择了块之后,用户可以发起跨块和许多不同患者的图像的基于块的相似性搜索,这些患者可已分配与当前选择的块不同的标签。相似性搜索基于特征向量和块的比较,用于确定基于相似特征矢量的相似块和相似组织模式。通过评估并显示与所选块(及其组织模式)相似但具有与所选块的标签不同的标签(例如“患者对药物D有应答=假”而不是“患者对药物D有应答=真”)的块(及相应组织模式)的数量和/或分数。For example, a user can select a block from a report library that has been assigned the highest numerical value and therefore the highest predictive power regarding image labels. After selecting a block, the user can initiate a block-based similarity search across blocks and images from many different patients, who may have been assigned labels different from the currently selected block. The similarity search is based on the comparison of feature vectors and blocks to determine similar blocks and similar tissue patterns based on similar feature vectors. This is achieved by evaluating and displaying the number and/or score of blocks (and corresponding tissue patterns) that are similar to the selected block (and its tissue pattern) but have labels different from those of the selected block (e.g., "Patient responded to drug D = false" instead of "Patient responded to drug D = true").
因此,病理学家可以通过选择由MIL程序返回的称为“高度预后”的块来轻松检查由MIL程序识别的组织模式的预测能力,特别是敏感性和特异性,以执行相似性搜索,揭示数据集中多少具有相似特征向量的块已分配与所选块相同的标签。与最新的机器学习应用程序相比,这是一个巨大的优势,机器学习应用程序也可提供组织图像预后特征的指示,但我们不允许用户识别和验证这些特征。基于报告库和相似性搜索库,人类用户可以验证所提出的高预后组织模式,并且还可以用语言表述在所有具有高预测能力的块中显示并与相似特征向量相关联的共同特征和结构。Therefore, pathologists can easily examine the predictive power, particularly sensitivity and specificity, of tissue patterns identified by the MIL program by selecting blocks returned by the MIL program as "high prognostic," to perform a similarity search, revealing how many blocks in the dataset with similar feature vectors have been assigned the same label as the selected block. This is a significant advantage compared to state-of-the-art machine learning applications, which also provide indications of prognostic features in tissue images, but do not allow users to identify and verify these features. Based on a report library and a similarity search library, human users can verify the proposed high-prognostic tissue patterns and can also verbally describe the common features and structures displayed and associated with similar feature vectors in all blocks with high predictive power.
报告库中的块是可选的并且选择触发执行相似性搜索以识别和显示具有与用户选择的块相似的特征向量/组织模式的其他块的特征可使用户能够自由选择他或她感兴趣的报告块库中的任何图像。例如,病理学家可能对如上所述的具有最高预测能力(由MIL计算出的最高数值)的组织模式和相应块感兴趣。替代性地,病理学家可对通常具有特别低的预测能力(特别低的数值)的伪影感兴趣。另外替代性地,病理学家可以出于任何其他原因对特定组织模式感兴趣,例如,因为它揭示了药物的一些副作用或任何其他相关的生物医学信息。病理学家可自由选择相应报告块库中的任何一个块。从而,病理学家触发相似性搜索以及以相似性块库的形式计算和显示结果。完成相似性搜索后,可以自动刷新显示和GUI。The blocks in the report block library are optional, and the selection triggers a similarity search to identify and display features of other blocks with similar feature vectors/organizational patterns to the user-selected block. This allows the user to freely select any image from the report block library that interests them. For example, a pathologist might be interested in an organization pattern and corresponding block with the highest predictive power (the highest value calculated by the MIL), as described above. Alternatively, a pathologist might be interested in artifacts that typically have particularly low predictive power (particularly low values). Furthermore, a pathologist might be interested in a particular organization pattern for any other reason, such as because it reveals some side effects of a drug or any other relevant biomedical information. The pathologist is free to select any block from the corresponding report block library. Thus, the pathologist triggers a similarity search, and the results are calculated and displayed in the form of a similarity block library. The display and GUI can be automatically refreshed after the similarity search is completed.
根据一些实施例,相似性搜索库的计算和显示包括相似性热图的计算和显示。热图以颜色和/或像素强度对相似块和相应特征向量进行编码。具有相似特征向量的图像区域和块在热图中以相似颜色和/或高或低像素强度表示。因此,用户可快速获得整个载玻片图像中特定组织模式签名的分布的概览。只需选择不同的块即可轻松刷新热图,因为该选择会根据新选择的块的特征向量自动诱导特征向量相似性的重新计算。According to some embodiments, the calculation and display of the similarity search library includes the calculation and display of a similarity heatmap. The heatmap encodes similar blocks and their corresponding feature vectors using color and/or pixel intensity. Image regions and blocks with similar feature vectors are represented in the heatmap by similar colors and/or high or low pixel intensities. Therefore, users can quickly obtain an overview of the distribution of a specific tissue pattern signature across the entire slide image. The heatmap can be easily refreshed simply by selecting a different block, as this selection automatically induces a recalculation of feature vector similarity based on the feature vector of the newly selected block.
根据实施例,相似性搜索库包括相似性热图。该方法包括通过子方法创建相似性热图,该子方法包括:According to an embodiment, the similarity search library includes a similarity heatmap. The method includes creating the similarity heatmap via a sub-method, which includes:
-选择报告库中的一个块;- Select a block from the report library;
-针对一些或所有接收到的图像的其他块中的每一个,通过将从同一图像和其他图像导出的其他块的特征向量与所选块的特征向量进行比较来计算关于所选块的相似性得分;- For each of the other blocks in some or all of the received images, a similarity score is calculated about the selected block by comparing the feature vectors of other blocks derived from the same image and other images with the feature vector of the selected block;
-针对块用于计算相应的相似性得分的图像中的每一个,计算作为相似性得分的函数的相应的相似性热图,相似性热图的像素颜色和/或像素强度指示所述图像中的块与所选块的相似性;以及- For each block in the image used to calculate the corresponding similarity score, calculate a corresponding similarity heatmap as a function of the similarity score, whereby the pixel color and/or pixel intensity of the similarity heatmap indicates the similarity between the block in the image and the selected block; and
-显示相似性热图。- Displays a similarity heatmap.
根据实施例,相似性搜索库中显示的图像块也是可选择的。According to an embodiment, the image patches displayed in the similarity search library are also selectable.
相似性热图可以提供有价值的概览信息,该信息允许人类用户轻松感知目标特定组织模式在特定组织中或在具有特定标签的患者亚群的组织样品中出现的广泛程度。用户可以自由选择搜索库中的任意块,从而分别诱导基于分配给当前所选块的特征向量重新计算相似性热图,以及自动刷新包含相似性热图的GUI。Similarity heatmaps provide valuable overview information, allowing human users to easily perceive the prevalence of a specific tissue pattern in a particular tissue or in tissue samples from a patient subgroup with a specific label. Users can freely select any block from the search library, thereby inducing a recalculation of the similarity heatmap based on the feature vector assigned to the currently selected block, and automatically refreshing the GUI containing the similarity heatmap.
根据实施例,报告库和/或相似性搜索块库中的图像块基于从患者组织样品图像导出的块进行分组。根据替代实施例,报告库和/或相似性搜索块库中的图像块基于分配给从中导出块的图像的标签进行分组。According to one embodiment, image blocks in the report library and/or similarity search block library are grouped based on blocks derived from patient tissue sample images. According to an alternative embodiment, image blocks in the report library and/or similarity search block library are grouped based on tags assigned to images from which blocks are derived.
通常,从同一患者导出的所有图像将具有相同的标签,并且来自特定患者的那些图像的所有块将被MIL处理为同一“包”的成员。但是,在某些特殊情况下,可能是同一患者的不同图像分配了不同的标签。例如,如果第一图像描绘了患者的第一转移,且第二图像描绘了同一患者的第二转移,并且观察结果是第一转移对药物D的治疗作出应答而消失,而第二转移继续生长,则患者相关属性值可按图像方式进行分配,而不是按患者方式进行分配。在这种情况下,每个患者可能会有多包块。Typically, all images derived from the same patient will have the same label, and all blocks from those images of a specific patient will be processed by MIL as members of the same "package". However, in certain special cases, different images from the same patient may be assigned different labels. For example, if the first image depicts the patient's first metastasis, and the second image depicts the same patient's second metastasis, and the observation is that the first metastasis responds to treatment with drug D and disappears, while the second metastasis continues to grow, then patient-related attribute values may be assigned image-wise rather than patient-wise. In this case, each patient may have multiple package blocks.
根据另一示例,在用特定药物治疗之前和之后拍摄的患者组织样品的图像以及用于训练MIL和/或应用经训练的MIL的终点(标签)是属性值“组织状态=用药物D治疗后”或属性值“组织状态=用药物D治疗前”。基于所述患者相关属性值训练MIL可具有识别组织模式的优势,该模式指示药物对肿瘤的活性和形态学影响。这种确定的与药物效应相关的组织模式可以验证和探索药物的作用方式以及潜在的药物副作用。According to another example, images of patient tissue samples taken before and after treatment with a specific drug, and the endpoints (labels) used for training the MIL and/or applying the trained MIL, are the attribute value "tissue state = after treatment with drug D" or the attribute value "tissue state = before treatment with drug D". Training the MIL based on these patient-related attribute values has the advantage of identifying tissue patterns that indicate the drug's activity and morphological effects on the tumor. Such identified tissue patterns related to drug effects can validate and explore the drug's mode of action and potential drug side effects.
根据实施例,该方法进一步包括:通过创建额外的块集以计算方式增加块包的数量,每个额外的块集被MIL程序处理为额外的块包,该块包分配了与生成源块的组织图像相同的标签。额外的块集的创建特别地包括:对至少块的子集应用一个或多个伪影生成算法以创建包括伪影的新块。此外,或替代性地,额外的块包的创建可包括提高或降低至少块的子集的分辨率以创建比它们相应的源块粒度更细或粒度更粗的新块。According to an embodiment, the method further includes: computationally increasing the number of block packets by creating additional block sets, each additional block set being processed by a MIL program as an additional block packet, the block packet being assigned the same label as the tissue image from which the source blocks were generated. The creation of the additional block sets specifically includes: applying one or more artifact generation algorithms to at least a subset of the blocks to create new blocks including artifacts. Furthermore, or alternatively, the creation of additional block packets may include increasing or decreasing the resolution of at least a subset of the blocks to create new blocks with finer or coarser granularity than their corresponding source blocks.
例如,可以通过随机选择从所述患者获得的一个或多个组织图像的一些或所有块为患者中的每一个获得子集。伪影生成算法模拟图像伪影。图像伪影可以是,例如,在组织制备、染色和/或图像采集期间产生的伪影类型(例如边缘伪影、过度染色、染色不足、灰尘、斑点伪影(通过高斯模糊等进行模拟)。此外,或替代性地,伪影可以是通用噪声类型(例如通过遮挡、颜色抖动、高斯噪声、椒盐噪声、旋转、翻转、歪斜失真等进行模拟)。For example, a subset can be obtained for each patient by randomly selecting some or all blocks from one or more tissue images obtained from the patient. The artifact generation algorithm simulates image artifacts. Image artifacts can be, for example, types of artifacts generated during tissue preparation, staining, and/or image acquisition (e.g., edge artifacts, over-staining, under-staining, dust, speckle artifacts (simulated by Gaussian blur, etc.)). Furthermore, or alternatively, artifacts can be general noise types (e.g., simulated by occlusion, color jitter, Gaussian noise, salt-and-pepper noise, rotation, flipping, skew distortion, etc.).
额外的块包的创建可能具有从有限的可用训练数据集生成额外的训练数据的优势。额外的训练数据代表图像数据,该图像数据的质量可能会因经常发生在样品制备和图像采集的情况下的常见的失真、伪影和噪声而降低。因此,扩大的训练数据集可确保避免训练期间MIL程序基础模型的过度拟合。Creating additional block packs can have the advantage of generating additional training data from a limited available training dataset. This additional training data represents image data, the quality of which may be degraded by common distortions, artifacts, and noise that frequently occur during sample preparation and image acquisition. Therefore, an expanded training dataset can ensure that overfitting of the MIL procedure's underlying model is avoided during training.
根据实施例,该方法进一步包括计算从一个或多个接收数字图像获得的块群集,其中块基于它们的特征向量的相似性被分组到群集中。优选地,针对患者中的每一个计算群集。这意味着如果块的特征向量足够相似,则来自描绘同一患者的不同组织载玻片的不同图像的块可分组到同一群集中。According to an embodiment, the method further includes calculating a cluster of blocks obtained from one or more received digital images, wherein blocks are grouped into clusters based on the similarity of their feature vectors. Preferably, clusters are calculated for each patient. This means that if the feature vectors of the blocks are sufficiently similar, blocks from different images of different tissue slides depicting the same patient can be grouped into the same cluster.
根据其他实施例,针对源自所有患者的所有块一起计算群集。According to other embodiments, clusters are calculated together for all blocks originating from all patients.
在这两种聚集块的方法中(不同患者的所有块在一起或每个患者的所有块),看起来彼此相似(即具有相似特征向量)的块被聚集到同一群集中。In both of these methods of clustering blocks (all blocks from different patients together or all blocks from each patient), blocks that appear similar to each other (i.e. have similar feature vectors) are clustered into the same cluster.
例如,在“不同患者聚集的所有块”的情况下,聚集的结果可能是生成例如64组(群集)块用于所有患者的所有块。64个群集中的每一个都包含从不同患者导出的相似块。相反,在每个患者聚集的情况下,每个患者将拥有自己的64个群集。For example, in the case of "all blocks from different patients," the result of aggregation might be to generate, for example, 64 groups (clusters) of blocks for all patients. Each of these 64 clusters contains similar blocks derived from different patients. Conversely, in the case of per-patient aggregation, each patient would have its own 64 clusters.
如果为每个患者创建群集,则可能是患者图像没有包含脂肪的块或包含脂肪的块非常少。在这种情况下,可能不会创建“脂肪群集”,因为没有足够的数据来学习围绕“脂肪”特征向量的群集。但是对所有患者的所有块一起执行聚集方法可具有以下优势:可使用最大数量的可用数据识别更多的群集/组织类型:在“所有患者块”聚集中,可能会识别出“脂肪”组织模式的群集,因为至少有些患者的活检中有一些脂肪细胞。因此,数据集中描绘块的脂肪细胞数量足够的概率,将创建脂肪细胞群集(也适用于脂肪细胞含量非常少的患者)。如果为所有患者的所有块在一起创建群集,并且一个群集代表脂肪细胞,则所有含有来自所有患者的脂肪细胞的块都将被分组到该群集中。这意味着针对专用的患者/包,所有带有脂肪细胞的块都将在所述群集中进行分组,并且如果群集采样用于包,则选择属于所述群集的一定数量的块(来自当前患者/包)。If clusters are created for each patient, it's possible that the patient images contain no fat patches or very few fat patches. In this case, "fat clusters" might not be created because there isn't enough data to learn clusters around the "fat" feature vector. However, performing clustering on all patches from all patients together has the following advantages: more clusters/tissue types can be identified using the largest amount of available data: in "all patient patches" clusters, clusters of "fat" tissue patterns are likely to be identified because at least some patients' biopsies contain some fat cells. Therefore, there is a sufficient probability that the number of fat cells depicting patches in the dataset will create fat cell clusters (also applicable to patients with very few fat cells). If clusters are created on all patches from all patients together, and one cluster represents fat cells, then all patches containing fat cells from all patients will be grouped into that cluster. This means that for a specific patient/package, all patches with fat cells will be grouped into said cluster, and if cluster sampling is used for a package, a certain number of patches belonging to said cluster (from the current patient/package) are selected.
块的聚集可能是有利的,因为该操作可揭示在特定患者中可观察到的组织模式的数量和/或类型。根据一些实施例,GUI包括用户可选择的元素,该元素使用户能够在聚集库视图中触发块的聚集以及块群集的呈现。这可以帮助用户直观且快速地理解在患者的特定组织样品中观察到的重要类型的组织模式。Clustering blocks can be advantageous because this operation can reveal the number and/or type of tissue patterns observable in a particular patient. According to some embodiments, the GUI includes user-selectable elements that allow the user to trigger the clustering of blocks and the presentation of block clusters within a cluster library view. This can help the user intuitively and quickly understand the important types of tissue patterns observed in a specific tissue sample from a patient.
根据实施例,MIL程序的训练包括对块集进行重复地采样以便从块集中挑选块的子集,并基于块的子集训练MIL程序。According to an embodiment, training the MIL program includes repeatedly sampling the block set to select a subset of blocks from the block set, and training the MIL program based on the subset of blocks.
本文使用的术语“采样”是在数据分析或训练机器学习算法的情况下使用的技术,该技术包括从数据集(从患者的一个或多个图像中获得的块的总体)中的多个N数据项(实例、块)中挑选特定数量的L样品。根据实施例,“采样”包括根据假定在统计学上表示训练数据集中的N块的总体的概率分布,从N数据项的数量内选择数据项的子集。这可以允许更准确地了解整个人群的特征。概率分布代表了指导机器学习过程并使“从数据中学习”可行的统计假设。As used herein, the term "sampling" is a technique employed in data analysis or training machine learning algorithms that involves selecting a specific number of L samples from a plurality of N data items (instances, blocks) in a dataset (a population of blocks obtained from one or more images of a patient). According to an embodiment, "sampling" involves selecting a subset of data items from the number of N data items based on a probability distribution that is assumed to statistically represent the population of N blocks in the training dataset. This allows for a more accurate understanding of the characteristics of the entire population. The probability distribution represents the statistical assumption that guides the machine learning process and makes "learning from data" feasible.
根据一些实施例,通过随机选择块的子集以提供采样的块包来执行采样。According to some embodiments, sampling is performed by randomly selecting a subset of blocks to provide a block packet for sampling.
根据实施例,聚集和采样组合如下:采样包括从针对患者获得的块群集中的每一个选择块,使得在采样中创建的块的每个子集中的块数量对应于取自所述块的群集的大小。According to an embodiment, the aggregation and sampling are combined as follows: sampling includes selecting each block from a cluster of blocks obtained for the patient, such that the number of blocks in each subset of the blocks created in the sampling corresponds to the size of the cluster taken from said blocks.
例如,可以从特定患者的数字组织图像创建1000个块。聚集创建了显示包括300个块的背景组织载玻片区域的第一群集,显示包括400个块的基质组织区域的第二群集,显示包括200个块的转移性肿瘤组织的第三群集,显示包括40个块的特定染色伪影的第四群集,显示包括60个块的具有微血管的组织的第五群集。For example, 1,000 blocks can be created from digital tissue images of a specific patient. Aggregation creates a first cluster showing a background tissue slide area comprising 300 blocks, a second cluster showing a stromal tissue area comprising 400 blocks, a third cluster showing metastatic tumor tissue comprising 200 blocks, a fourth cluster showing specific staining artifacts comprising 40 blocks, and a fifth cluster showing tissue with microvessels comprising 60 blocks.
根据一个实施例,采样包括从群集中的每一个选择特定部分的块,例如50%。这将意味着来自群集1的150个块、来自群集2的200个块、来自群集3的100个块、来自群集4的20个块和来自群集5的30个块。According to one embodiment, sampling includes selecting a specific portion of blocks from each of the clusters, such as 50%. This would mean 150 blocks from cluster 1, 200 blocks from cluster 2, 100 blocks from cluster 3, 20 blocks from cluster 4, and 30 blocks from cluster 5.
根据优选实施例,采样包括从每个群集中选择相等数量的块。这种采样方法可具有以下优势:从不同类型的聚集中抽取相同数量的块/组织模式示例,从而使训练数据集更加平衡。如果期望的预测特征在训练数据集中很少见,这可增加经训练的MIL和/或经训练的注意力MLL的准确性。According to a preferred embodiment, sampling involves selecting an equal number of blocks from each cluster. This sampling method has the advantage of drawing the same number of block/organization pattern examples from different types of clusters, thus making the training dataset more balanced. This can increase the accuracy of the trained MIL and/or the trained attention MIL if the desired predictive features are rare in the training dataset.
聚集和采样的组合可能是特别有利的,因为可以通过采样增加用于训练的数据基础,而不会无意中“丢失”实际上具有高预测能力的少数块。通常在数字病理学的背景下,组织样品的绝大多数区域不包括由特定疾病或其他患者相关属性修改或预后的组织区域。例如,组织样品的仅一个小的子区域可能实际上包含肿瘤细胞,其余部分可能显示正常组织。通过首先执行块聚集且然后从群集中的每一个选择块可确保显示预后组织模式的至少一些块,例如,确保肿瘤细胞或微血管始终是样品的一部分。The combination of clustering and sampling can be particularly advantageous because sampling can increase the data base for training without inadvertently "losing" a few blocks that actually have high predictive power. Typically, in the context of digital pathology, the vast majority of a tissue sample does not include tissue areas modified or prognostic by specific disease or other patient-related attributes. For example, only a small subregion of a tissue sample may actually contain tumor cells, while the rest may show normal tissue. By first performing block clustering and then selecting blocks from each cluster, it is ensured that at least some blocks show prognostic tissue patterns, for example, ensuring that tumor cells or microvessels are always part of the sample.
特征提取方法Feature extraction methods
根据实施例,针对块中的每一个的特征向量的计算包括接收块中描绘了其组织样品的患者的患者相关数据,以及以特征向量中的一个或多个特征的形式表示患者相关数据,患者相关数据特别选自包括以下项的组:基因组数据、RNA序列数据、患者的已知疾病、年龄、性别、体液中的代谢物浓度、健康参数和当前用药。According to an embodiment, the calculation of the feature vector for each of the blocks includes receiving patient-related data of the patient in the block that depicts a tissue sample therein, and representing the patient-related data in the form of one or more features in the feature vector, the patient-related data being specifically selected from the group consisting of: genomic data, RNA sequence data, known diseases of the patient, age, sex, concentration of metabolites in body fluids, health parameters, and current medications.
根据实施例,特征向量的计算由经训练的机器学习逻辑执行,特别是由包括至少一个瓶颈层的经训练的全卷积神经网络执行。According to an embodiment, the computation of the feature vector is performed by trained machine learning logic, particularly by a trained fully convolutional neural network including at least one bottleneck layer.
根据实施例,用于特征提取(“特征提取MLL”)的经训练的机器学习逻辑通过采用包括瓶颈的全卷积网络类型的MLL(如UNET)在监督方法中接受训练。“Unet”架构由OlafRonneberger、Philipp Fischer和Thomas Brox在“U-Net:用于生物医学图像分割的卷积网络”中描述,德国弗莱堡大学计算机科学系及BIOSS生物信号研究中心(arXiv:1505.04597v1,2015年5月18日)。该文件可通过康奈尔大学图书馆https://arxiv.org/abs/1505.04597下载。According to an embodiment, the trained machine learning logic for feature extraction (“Feature Extraction MLL”) is trained in a supervised manner using an MLL of the type including a bottleneck, such as UNET. The “UNET” architecture is described by Olaf Ronneberger, Philipp Fischer, and Thomas Brox in “U-Net: A Convolutional Network for Biomedical Image Segmentation,” Department of Computer Science and BIOSS Centre for Biosignals, University of Freiburg, Germany (arXiv:1505.04597v1, May 18, 2015). This document is available for download from Cornell University Library at https://arxiv.org/abs/1505.04597.
例如,可训练特征提取MLL来执行组织图像分割任务,由此待识别的片段包括两个或更多个以下组织图像片段类型:肿瘤组织、健康组织、坏死组织、包括特定对象的组织,诸如如肿瘤细胞、血管、基质、淋巴细胞等以及背景区域。根据一些实施例,使用分类网络(诸如Resnet、ImageNet或SegNet)以监督的方式训练特征提取MLL,通过训练它对具有特定预定类别或对象的图像块进行分类。For example, a feature extraction MLL can be trained to perform tissue image segmentation tasks, whereby the fragments to be identified include two or more of the following tissue image fragment types: tumor tissue, healthy tissue, necrotic tissue, tissue containing specific objects such as tumor cells, blood vessels, stroma, lymphocytes, etc., and background regions. According to some embodiments, the feature extraction MLL is trained in a supervised manner using a classification network (such as ResNet, ImageNet, or SegNet) to classify image patches with specific predetermined categories or objects.
在特征提取MLL经训练后,MLL被拆分成“编码器”部分(包括输入层、一个或多个中间层和瓶颈层)和“解码器”,即输出生成部分。根据本发明的实施例,使用“编码器”部分达到经训练的MLL的瓶颈层来提取和计算每个输入块的特征向量。瓶颈层是神经网络的一层,该瓶颈层包含明显少于输入层的神经元。例如,瓶颈层可以是包含小于输入层的60%或甚至小于20%的“神经元”的层。根据不同的网络架构,不同层中神经元的数量和比例可能会有很大差异。瓶颈层是一个隐藏层。After training, the feature extraction MLL is split into an "encoder" part (including an input layer, one or more intermediate layers, and a bottleneck layer) and a "decoder," i.e., the output generation part. According to embodiments of the invention, the "encoder" part is used to reach the bottleneck layer of the trained MLL to extract and compute the feature vector for each input block. The bottleneck layer is a layer of the neural network that contains significantly fewer neurons than the input layer. For example, the bottleneck layer may be a layer containing fewer than 60% or even less than 20% of the "neurons" of the input layer. Depending on the network architecture, the number and proportion of neurons in different layers can vary considerably. The bottleneck layer is a hidden layer.
根据一个示例,特征提取MLL的网络具有基于UNET的网络架构。它有一个具有512*512*3(512x512 RGB)个神经元的输入层和具有9*9*128个神经元的瓶颈层。因此,瓶颈层的神经元数量约为输入层的神经元数量的1.5%。As an example, the feature extraction MLL network has a UNET-based network architecture. It has an input layer with 512*512*3 (512x512 RGB) neurons and a bottleneck layer with 9*9*128 neurons. Therefore, the number of neurons in the bottleneck layer is approximately 1.5% of the number of neurons in the input layer.
根据一个示例,特征提取MLL的网络具有Resnet架构,该架构实现了监督式或无监督式学习算法。输入层包含512x512x3个神经元,瓶颈层和瓶颈层输出的相应特征向量通常包含1024或2048个元素(神经元/数字)。As an example, the network for feature extraction MLL has a ResNet architecture, which implements supervised or unsupervised learning algorithms. The input layer contains 512x512x3 neurons, and the bottleneck layer and the corresponding feature vectors output by the bottleneck layer typically contain 1024 or 2048 elements (neurons/digits).
根据实施例,特征提取由基于ImageNet自然图像数据集训练的ResNet-50(He etal.,2016)架构的特征提取程序模块执行。Pierre Courtiol,EricW中描述了一些基于此架构从图像中提取特征的详细示例。Tramel、Marc Sanselme和Gilles Wainrib:,,仅使用全球标签的组织病理学分类和疾病定位:弱监督方法”,arXiv:1802.02212,于2018年2月1日提交,可通过康奈尔大学图书馆在线获取https://arxiv.org/pdf/1802.02212.pdf。According to the embodiments, feature extraction is performed by a feature extraction procedure module based on the ResNet-50 architecture (He et al., 2016) trained on the ImageNet Natural Image dataset. Pierre Courtiol, Eric W. describes some detailed examples of feature extraction from images based on this architecture. Tramel, Marc Sanselme, and Gilles Wainrib, “Histopathological Classification and Disease Localization Using Global Labels Only: A Weakly Supervised Approach,” arXiv:1802.02212, submitted February 1, 2018, is available online at Cornell University Library at https://arxiv.org/pdf/1802.02212.pdf.
根据实施例,将特定块的经训练的特征提取MLL其中一层生成的输出用作MIL程序从块提取的特征向量。这一层可以是,特别地,瓶颈层。根据实施例,特征提取MLL以无监督或自监督的方式训练,如Mathilde Caron和Piotr Bojanowski以及Armand Joulin和MatthijsDouze中所述:“视觉特征无监督式学习的深度聚集”,可通过https://arxiv.org/abs/1807.05520以电子方式获取CoRR,1807.05520,2018。According to an embodiment, the output generated by one layer of a trained feature extraction MLL for a specific block is used as the feature vector extracted from the block by the MIL procedure. This layer may be, in particular, a bottleneck layer. According to an embodiment, the feature extraction MLL is trained in an unsupervised or self-supervised manner, as described in Mathilde Caron and Piotr Bojanowski and Armand Joulin and Matthijs Douze, “Deep Aggregation for Unsupervised Learning of Visual Features,” available electronically at https://arxiv.org/abs/1807.05520, CoRR, 1807.05520, 2018.
替代性地,可以根据Spyros Gidaris、Praveer Singh、Nikos Komodakis训练特征提取MLL:“通过预测图像旋转进行无监督表示学习”,2018年2月15日,ICLR 2018会议电子版可通过https://openreview.net/forum?id=S1v4N2l0-获取。Alternatively, MLL can be extracted from features trained by Spyros Gidaris, Praveer Singh, and Nikos Komodakis: “Unsupervised Representation Learning by Predicting Image Rotation,” February 15, 2018. The electronic version of the ICLR 2018 conference is available at https://openreview.net/forum?id=S1v4N2l0-.
另外替代性地,根据Elad Hoffer,Nir Ailon可训练特征提取MLL。“通过度量嵌入进行半监督深度学习”,2016年11月4日,ICLR 2017电子版可通过https://openreview.net/forum?id=r1R5Z19le获得。Alternatively, MLLs can be extracted from trainable features, according to Elad Hoffer and Nir Ailon. “Semi-supervised deep learning through metric embeddings,” ICLR 2017, November 4, 2016. The electronic version is available at https://openreview.net/forum?id=r1R5Z19le.
用于训练特征提取MLL的数据集可以是另一个组织图像数据集和/或稍后用于训练MIL程序的组织图像集。在训练阶段,特征提取MLL不会评估或以其他方式使用与训练图像相关联的任何标签,因为特征提取MLL经训练用于识别组织类型和相应的图像片段,而不是识别患者的患者相关属性值(用作MIL程序学习阶段的终点)。The dataset used to train the Feature Extraction MLL can be another dataset of tissue images and/or a set of tissue images later used to train the MIL procedure. During the training phase, the Feature Extraction MLL does not evaluate or otherwise use any labels associated with the training images because the Feature Extraction MLL is trained to identify tissue types and corresponding image fragments, rather than to identify patient-related attribute values (which serve as the endpoint of the MIL procedure's learning phase).
利用基于接近性相似性标签的特征提取方法Feature extraction method based on proximity similarity labels
根据实施例,特征向量由特征提取机器学习逻辑(“特征提取MLL”)计算,该逻辑已经基于包括用标签标记的块对的训练数据集训练,由此每个标签表示由块对描绘的两个组织模式的相似性,且作为块对中两个块的空间距离的函数进行计算。According to an embodiment, the feature vector is computed by a feature extraction machine learning logic (“Feature Extraction MLL”), which has been trained on a training dataset including labeled block pairs, whereby each label represents the similarity of two organizational patterns depicted by the block pair and is computed as a function of the spatial distance between the two blocks in the block pair.
根据优选实施例,每个标签表示由块对描绘的两个组织模式的相似性,并且作为块对中的两个块的空间距离的函数来计算,从而使用空间距离作为两个块的相似性的唯一度量。According to a preferred embodiment, each tag represents the similarity of two organizational patterns depicted by the block pair, and is calculated as a function of the spatial distance between the two blocks in the block pair, thereby using spatial distance as the sole measure of the similarity between the two blocks.
根据优选实施例,标签被完全自动地分配给训练数据集中的块对。According to a preferred embodiment, labels are automatically assigned to block pairs in the training dataset.
由于多种原因,这种方法可能是有益的:两个图像区域的空间接近度是组织样品的每个数字图像中始终且固有可用的特征。问题在于图像和相应组织区域本身的空间接近度通常不会揭示相对于生物医学问题的任何相关信息,诸如组织类型分类、疾病分类、特定疾病持久性的预测或图像分割任务。申请人惊奇地观察到,两个图像区域(“块”)的空间接近度传达的信息是两个图像区域的相似性的准确指标,至少是否在MLL的训练阶段期间分析了大量块及其相应距离。因此,通过利用两个块的“空间接近度”的固有可用信息来为两个进行比较的块自动分配组织模式相似性标签,可自动提供可用于训练MLL的大型注释数据集。经训练的MLL可用于自动确定作为输入接收到的两个图像或图像块是否描绘了相似或相异的组织模式。然而,该数据集还可以用于其他且更复杂的任务,例如图像相似性搜索、图像分割、组织类型检测和组织模式聚集。因此,申请人惊奇地观察到,块的空间接近度传达的信息可用于自动创建带注释的训练数据,允许训练可靠地确定图像相似性的MLL,此外还可以训练输出特征向量的MLL,所述特征向量可由额外的数据处理单元用于数字病理学中的多个复杂图像分析任务。这些方法都不需要领域专家手动注释训练数据。This approach may be beneficial for several reasons: the spatial proximity of two image regions is a consistently and inherently available feature in every digital image of a tissue sample. The problem is that the spatial proximity of the image and the corresponding tissue region itself typically does not reveal any relevant information relative to biomedical problems, such as tissue type classification, disease classification, prediction of the persistence of a specific disease, or image segmentation tasks. The applicant surprisingly observed that the spatial proximity of two image regions (“blocks”) conveys an accurate indicator of the similarity between the two image regions, at least whether a large number of blocks and their corresponding distances were analyzed during the training phase of the MLL. Therefore, by leveraging the inherently available information of the “spatial proximity” of two blocks to automatically assign tissue pattern similarity labels to two compared blocks, a large annotated dataset that can be automatically provided for training an MLL can be provided. The trained MLL can be used to automatically determine whether two images or image blocks received as input depict similar or dissimilar tissue patterns. However, this dataset can also be used for other and more complex tasks, such as image similarity search, image segmentation, tissue type detection, and tissue pattern aggregation. Therefore, the applicant surprisingly observed that the information conveyed by the spatial proximity of blocks can be used to automatically create annotated training data, allowing the training of MLLs that reliably determine image similarity, and further, the training of MLLs that output feature vectors, which can be used by additional data processing units for multiple complex image analysis tasks in digital pathology. None of these methods require manual annotation of training data by domain experts.
当包含许多不同组织模式(例如“非肿瘤”和“肿瘤”)的训练图像被拆分成许多不同的块时,两个块之间的距离越小,两个相比较块描绘相同组织图的概率就越高,例如“非肿瘤”。然而,在描绘不同组织模式的两个不同模式的边界旁边会有一些块对(例如,第一块“肿瘤”,另一块“非肿瘤”)。这些块对产生噪声,因为它们描绘了不同的组织模式,尽管它们在空间上彼此非常接近。申请人惊奇地观察到,由跨越不同组织模式之间的边界的块对结合简化假设(空间接近度指示所描绘的组织模式的相似性)产生的噪声不会显著降低经训练的MLL的准确性。事实上,申请人观察到根据本发明的实施例训练的MLL的准确性能够胜过现有的基准方法。When training images containing many different tissue patterns (e.g., "non-tumor" and "tumor") are split into many different blocks, the smaller the distance between two blocks, the higher the probability that the two compared blocks depict the same tissue pattern, e.g., "non-tumor". However, there are some block pairs (e.g., the first block "tumor", the other "non-tumor") next to the boundary between two different patterns depicting different tissue patterns. These block pairs generate noise because they depict different tissue patterns, even though they are very close to each other spatially. The applicant surprisingly observed that the noise generated by block pairs crossing the boundary between different tissue patterns, combined with the simplifying assumption (spatial proximity indicates the similarity of the depicted tissue patterns), does not significantly reduce the accuracy of the trained MLL. In fact, the applicant observed that the accuracy of the MLL trained according to embodiments of the present invention can outperform existing benchmark methods.
在进一步的有益方面,现在可以快速且完全自动地针对许多不同的图像集创建训练数据。目前,缺乏可用的注释数据集来捕捉组织病理学图像中的自然和实际可变性。例如,即使现有的大型数据集(如Camelyon)也只包含一种染色(苏木精和曙红)和一种癌症(乳腺癌)。在来自不同癌症类型、不同组织染色类型和不同组织类型的图像中,组织病理学图像纹理和对象形状可能会有很大差异。此外,组织病理学图像包含许多具有不同领域特定含义的不同纹理和对象类型(例如基质、肿瘤浸润淋巴细胞、血管、脂肪、健康组织、坏死等)。因此,本发明的实施例可以允许针对多种不同癌症类型、癌症亚型、染色方法和患者组(例如治疗/未治疗、男性/女性、比阈值年龄年长/年幼、生物标志物阳性/生物标志物阴性等)中的每一个自动创建注释数据集。因此,本发明的实施例可允许自动创建注释训练数据并基于训练数据训练相应的MLL,使得经训练的MLL适于以高度特定的方式准确解决多个不同患者组中的每一个的生物医学问题。与基于手动注释的乳腺癌数据集训练的MLL针对结肠癌患者提供次优结果的现有技术方法相反,本发明的实施例可允许分别针对不同患者组中的每一个创建MLL。A further advantage is that training data can now be created rapidly and fully automatically for many different image sets. Currently, there is a lack of available annotated datasets to capture the natural and real-world variability in histopathological images. For example, even existing large datasets (such as Camelyon) contain only one staining (hematoxylin and eosin) and one type of cancer (breast cancer). Histopathological image textures and object shapes can vary considerably across images from different cancer types, different tissue staining methods, and different tissue types. Furthermore, histopathological images contain many different textures and object types (e.g., matrix, tumor-infiltrating lymphocytes, blood vessels, fat, healthy tissue, necrosis, etc.) with different domain-specific meanings. Therefore, embodiments of the present invention can allow the automatic creation of annotated datasets for each of many different cancer types, cancer subtypes, staining methods, and patient groups (e.g., treated/untreated, male/female, older/younger than a threshold age, biomarker positive/biomarker negative, etc.). Thus, embodiments of the present invention can allow the automatic creation of annotated training data and the training of corresponding MLLs based on this training data, such that the trained MLLs are adapted to accurately address biomedical problems in many different patient groups in a highly specific manner. In contrast to existing techniques that train MLLs on manually annotated breast cancer datasets to provide suboptimal results for colon cancer patients, embodiments of the present invention allow for the creation of MLLs separately for each of different patient groups.
根据实施例,指示两个组织模式的相似性程度的标签是二进制数据值,即可以具有两个可能选项中的一个的值。例如,标签可以是“1”或“相似”,并且指示两个块描绘相似的组织模式。替代性地,标签可以是“0”或“相异”,并且指示两个块描绘不同的组织模式。根据其他实施例,标签可为更细粒度,例如,可以是从三个或更多数据值的有限集合中选择的数据值,例如“相异”、“相似”和“高度相似”。根据另一些实施例,标签可为更细粒度并且可为数值,其中数值的量与相似性程度呈正相关。例如,可以将数值计算为将成对的两个块之间的空间距离线性和逆变换为表示组织模式相似性的数值的函数。空间距离越大,指示组织模式相似性的数值越小。存在多种MLL架构,可以处理和使用训练数据集中不同类型的标签(例如序数或数值)。选择MLL的类型,使其能够处理训练数据集的自动创建的标签。According to embodiments, the labels indicating the degree of similarity between two organizational patterns are binary data values, meaning they can have a value that is one of two possible options. For example, the label could be "1" or "Similar," indicating that the two blocks depict similar organizational patterns. Alternatively, the label could be "0" or "Dissimilar," indicating that the two blocks depict different organizational patterns. According to other embodiments, the labels can be more granular, for example, they could be data values selected from a finite set of three or more data values, such as "Dissimilar," "Similar," and "Highly Similar." According to still other embodiments, the labels can be more granular and can be numerical, where the magnitude of the numerical value is positively correlated with the degree of similarity. For example, the numerical value can be computed as a function of the linear and inverse transformation of the spatial distance between a pair of blocks into a numerical value representing the similarity of organizational patterns. The larger the spatial distance, the smaller the numerical value indicating the similarity of organizational patterns. Various MLL architectures exist that can handle and utilize different types of labels (e.g., ordinal or numerical) from the training dataset. The type of MLL is chosen to handle automatically generated labels from the training dataset.
根据实施例,基于自动注释的训练数据集训练并且将用于特征提取的MLL适于根据监督式学习算法进行学习。监督式学习是关于找到将一组输入特征转换为一个或多个输出数据值的映射。输出数据值在训练期间作为标签提供,例如作为二元期权标签“相似”或“相异”或作为相似性定量度量的数值。换句话说,在训练过程中,将要预测的数据值以训练数据标签的形式明确提供给MLL的模型。监督式学习带来的问题是需要用标签标记训练数据,以便为每个样品定义输出空间。According to an embodiment, an MLL trained on an automatically annotated training dataset and adapted for feature extraction is trained using a supervised learning algorithm. Supervised learning is about finding a mapping that transforms a set of input features into one or more output data values. These output data values are provided as labels during training, such as binary option labels "similar" or "dissimilar," or as numerical values that quantify similarity. In other words, during training, the data values to be predicted are explicitly provided to the MLL model in the form of training data labels. A challenge with supervised learning is the need to label the training data to define the output space for each sample.
根据实施例,至少一些或所有块对分别描绘包含在同一组织切片中的两个组织区域。组织切片中的每一个在接收数字图像中的相应一个中描绘。块之间的距离是在2D坐标系内计算的,该坐标系由接收到的从中导出该对中的块数字图像的x和y维度定义。According to an embodiment, at least some or all of the block pairs respectively depict two tissue regions contained in the same tissue slice. Each of the tissue slices is depicted in a corresponding one in the received digital image. The distance between the blocks is calculated in a 2D coordinate system defined by the x and y dimensions of the received digital image from which the blocks of the pair are derived.
根据实施例,通过在多个不同图像的每一个内随机选择块对来生成块对。基于随机选择确保每对中的块之间的空间距离会有所不同。对相似性标签,例如以与两个块之间的距离成反比的数值形式进行计算并分配给每对。According to an embodiment, block pairs are generated by randomly selecting block pairs within each of a plurality of different images. This random selection ensures that the spatial distance between blocks in each pair will be different. Similarity labels, for example, are calculated in a numerical form inversely proportional to the distance between the two blocks and assigned to each pair.
根据其他实施例,通过选择每个接收到的图像的至少一些或所有块作为起始块来生成块对;针对每个起始块,选择所有或预定义数量的“附近块”,其中“附近块”是以起始块为中心的第一圆内的块,由此该圆的半径与第一空间接近度阈值相同;针对每个起始块,选择全部或预定义数量的“远处块”,其中“远处块”是在以起始块为中心的第二圆之外的块,其中所述圆的半径与第二空间度接近阈值相同;可以通过在相应图像区域内随机选择该数量的块来执行预定义数量的选择。第一接近阈值和第二接近阈值可以相同,但优选地,第二接近阈值大于第一接近阈值。例如,第一接近阈值可以是1mm并且第二接近阈值可以是10mm。然后,选择第一块对集,由此每个块对包括起始块和位于第一圆内的附近块。第一集中的每个块对都分配了“相似”组织模式的标签。此外,选择第二块对集,由此所述集中的每一对包括起始块和“远处块”之一。第二集中的每个块对都分配了“相异”组织模式的标签。例如,该实施例可用于创建“相似”或“相异”的“二进制”标签。According to other embodiments, block pairs are generated by selecting at least some or all of the blocks in each received image as starting blocks; for each starting block, all or a predefined number of "nearby blocks" are selected, where "nearby blocks" are blocks within a first circle centered on the starting block, where the radius of the circle is the same as a first spatial proximity threshold; for each starting block, all or a predefined number of "distant blocks" are selected, where "distant blocks" are blocks outside a second circle centered on the starting block, where the radius of the circle is the same as a second spatial proximity threshold; the predefined number of selections can be performed by randomly selecting this number of blocks within the corresponding image region. The first proximity threshold and the second proximity threshold can be the same, but preferably, the second proximity threshold is greater than the first proximity threshold. For example, the first proximity threshold can be 1 mm and the second proximity threshold can be 10 mm. Then, a first set of block pairs is selected, whereby each block pair includes a starting block and a nearby block located within the first circle. Each block pair in the first set is assigned a label for a "similar" organization pattern. Furthermore, a second set of block pairs is selected, whereby each pair in the set includes a starting block and one of the "distant blocks". Each block pair in the second set is assigned a label for a "dissimilar" organization pattern. For example, this embodiment can be used to create “binary” tags that are “similar” or “different”.
根据实施例,在从中导出块的数字图像的x轴和y轴定义的2D坐标系内测量块之间的距离。这些实施例可用于以下情况:其中多个组织样品图像可用,所述多个组织样品图像描绘不同患者和/或同一患者内的不同区域的组织样品,由此所述不同区域彼此远离或由此精确位置所述两个区域相对于彼此是未知的。在这种情况下,块之间的空间接近度仅在由数字图像定义的2D像素平面内测量。基于图像采集设备(例如显微镜的相机或载玻片扫描仪)的已知分辨率因子,原始图像的块之间的距离可用于计算由两个块描绘的组织样品中的组织区域之间的距离。According to embodiments, the distance between blocks is measured within a 2D coordinate system defined by the x-axis and y-axis of a digital image from which the blocks are derived. These embodiments can be used in situations where multiple tissue sample images are available, depicting tissue samples from different patients and/or different regions within the same patient, where these different regions are far apart from each other or where the precise location of the two regions relative to each other is unknown. In this case, the spatial proximity between blocks is measured only within a 2D pixel plane defined by the digital image. Based on the known resolution factor of the image acquisition device (e.g., a camera for a microscope or a slide scanner), the distance between blocks in the original image can be used to calculate the distance between tissue regions in the tissue samples depicted by the two blocks.
根据实施例,至少一些或所有块对描绘一堆相邻组织切片的两个不同组织切片中包含的两个组织区域。组织切片中的每一个在接收数字图像中的相应一个中描绘。接收到的图像(该图像描绘一堆相邻组织切片的组织切片)在3D坐标系中彼此对齐。块之间的距离是在3D坐标系内计算的。According to an embodiment, at least some or all of the block pairs depict two tissue regions contained in two different tissue slices from a set of adjacent tissue slices. Each of the tissue slices is depicted in a corresponding one in the received digital image. The received image (which depicts tissue slices from a set of adjacent tissue slices) is aligned with each other in a 3D coordinate system. The distance between the blocks is calculated in the 3D coordinate system.
例如,一些或所有接收到的数字图像可描绘相邻组织切片的组织块内的切片的组织样品。在这种情况下,数字图像可在公共3D坐标系中彼此对齐,使得数字图像在3D坐标系中的位置再现组织块内分别描绘的组织切片的位置。这可以允许确定3D坐标系中的块距离。“附近”和“远处”块的选择可以如上文所描述的针对2D坐标系情况执行,唯一的区别在于至少一些块对中的块是从接收到的图像中的不同图像导出的。For example, some or all of the received digital images can depict tissue samples of slices within a tissue block of adjacent tissue slices. In this case, the digital images can be aligned with each other in a common 3D coordinate system such that the position of the digital images in the 3D coordinate system reproduces the position of the tissue slices depicted separately within the tissue block. This allows for the determination of block distances in the 3D coordinate system. The selection of “nearby” and “far” blocks can be performed as described above for the 2D coordinate system case, the only difference being that at least some of the blocks in the block pairs are derived from different images in the received images.
根据一些实施例,带注释的训练数据包括从相同数字图像导出的块对以及从已在公共3D坐标系中彼此对齐的不同图像导出的块对。这可能是有益的,因为在只有少量相应组织样品的图像可用的情况下,考虑第三维(代表不同组织样品中组织区域的块的空间接近度)可能会极大地增加训练数据中的块数量,由此组织样品属于同一个细胞块,例如3D活检细胞块。According to some embodiments, the annotated training data includes block pairs derived from the same digital image as well as block pairs derived from different images that have been aligned with each other in a common 3D coordinate system. This can be beneficial because, when only a small number of images of the corresponding tissue samples are available, considering the third dimension (representing the spatial proximity of blocks of tissue regions in different tissue samples) can greatly increase the number of blocks in the training data, where tissue samples belong to the same cell block, such as a 3D biopsy cell block.
根据实施例,每个块描绘具有小于0.5mm,优选地小于0.3mm的最大边缘长度的组织或背景区域。According to an embodiment, each block depicts a tissue or background area having a maximum edge length of less than 0.5 mm, preferably less than 0.3 mm.
小块尺寸可具有以下优点:描述不同组织模式的混合物的块的数量和面积分数减少。这可以帮助减少由描绘两个或更多个不同组织模式的块和由描绘两个不同组织模式的“组织模式边界”旁边的块对产生的噪声。此外,小块尺寸可以允许生成和用标签标记大量块对,从而增加用标签标记的训练数据的量。Smaller patch sizes offer several advantages: a reduction in the number and area fraction of patches describing mixtures of different tissue patterns. This helps reduce noise generated by patches depicting two or more distinct tissue patterns and by patch pairs adjacent to the “tissue pattern boundary” depicting two distinct tissue patterns. Furthermore, smaller patch sizes allow for the generation and labeling of a large number of patch pairs, thereby increasing the amount of labeled training data.
根据实施例,块对的自动生成包括:使用第一空间接近度阈值生成第一块对集;由第一集中的每个块对的两个块描绘的两个组织区域由小于第一空间接近阈值的距离彼此分开;使用第二空间接近度阈值生成第二块对集;由第二集中的每个块对的两个块描绘的两个组织区域由大于第二空间阈值的距离彼此分开。例如,这可以通过选择多个起始块、基于每个起始块周围的第一空间接近度阈值和第二空间接近度阈值计算第一圆和第二圆并选择包括起始块和“附近块”(第一集)或“远处块”(第二集),如上文针对本发明的实施例所述。According to an embodiment, the automatic generation of block pairs includes: generating a first set of block pairs using a first spatial proximity threshold; separating two tissue regions depicted by the two blocks of each block pair in the first set by a distance less than the first spatial proximity threshold; generating a second set of block pairs using a second spatial proximity threshold; and separating two tissue regions depicted by the two blocks of each block pair in the second set by a distance greater than the second spatial proximity threshold. For example, this can be achieved by selecting multiple starting blocks, calculating a first circle and a second circle based on the first and second spatial proximity thresholds around each starting block, and selecting a starting block and either a "nearby block" (first set) or a "distant block" (second set), as described above with respect to embodiments of the invention.
根据实施例,第一空间接近度阈值和第二空间接近度阈值是相同的,例如1mm。According to an embodiment, the first spatial proximity threshold and the second spatial proximity threshold are the same, for example, 1 mm.
根据优选实施例,第二空间接近度阈值比第一空间接近度阈值至少大2mm。这可能是有利的,因为在组织模式从一种模式逐渐变为另一种模式的情况下,“远处块”中描绘的组织模式与“附近”块中描绘的组织模式之间的差异可更清楚并且学习效果可得到提高。According to a preferred embodiment, the second spatial proximity threshold is at least 2 mm larger than the first spatial proximity threshold. This may be advantageous because, in the case of the organizational pattern gradually changing from one pattern to another, the difference between the organizational pattern depicted in the "distant block" and the organizational pattern depicted in the "nearby" block can be clearer and the learning effect can be improved.
根据实施例,第一空间接近度阈值是小于2mm、优选小于1.5mm、特别是1.0mm的距离。According to an embodiment, the first spatial proximity threshold is a distance of less than 2 mm, preferably less than 1.5 mm, and particularly 1.0 mm.
此外或替代性地,第二空间接近度阈值是大于4mm、优选大于8mm、特别是10.0mm的距离。Alternatively, the second spatial proximity threshold is a distance greater than 4 mm, preferably greater than 8 mm, and particularly 10.0 mm.
这些距离阈值指的是数字图像中描绘的组织区域(或切片背景区域)与相应块的距离。基于图像采集设备的已知放大倍数和数字图像的分辨率,该距离可在数字图像的2D或3D坐标系内转换。These distance thresholds refer to the distance between a depicted tissue region (or slice background region) and the corresponding block in a digital image. Based on the known magnification of the image acquisition device and the resolution of the digital image, this distance can be converted within the 2D or 3D coordinate system of the digital image.
例如,可以测量块(以及其中描绘的组织区域)之间的距离,例如2d或3D坐标系中两个块的中心之间。根据替代性实施变型,在2D或3D坐标系中彼此最靠近的两个块边缘(图像区域边缘)之间测量距离。For example, distances between blocks (and the tissue regions depicted within them) can be measured, such as between the centers of two blocks in a 2D or 3D coordinate system. According to alternative implementation variations, distances can be measured between the edges of two blocks (image region edges) that are closest to each other in a 2D or 3D coordinate system.
已经观察到上述阈值可提供用标签标记的训练数据,该数据允许自动生成经训练的MLL,该经训练的MLL能够准确识别乳腺癌患者的相似和相异的组织模式。在一些其他实施示例中,第一空间接近度阈值和第二空间接近度阈值可以具有其他值。特别是在使用显示不同组织类型或癌症类型的不同接收数字图像集的情况下,第一空间接近度阈值和第二空间接近度阈值可具有不同于以上所提供的距离阈值的其他值。It has been observed that the aforementioned thresholds provide labeled training data that allows for the automatic generation of trained MLLs capable of accurately identifying similar and dissimilar tissue patterns in breast cancer patients. In some other implementation examples, the first and second spatial proximity thresholds may have other values. Particularly when using different sets of received digital images displaying different tissue types or cancer types, the first and second spatial proximity thresholds may have values other than the distance thresholds provided above.
根据实施例,该方法进一步包括创建用于训练特征提取MLL的训练数据集。该方法包括接收多个数字训练图像,每个图像描绘一个组织样品;将接收到的训练图像中的每一个拆分成多个块(“特征提取训练块”);自动生成块对,每个块对已分配一个标签,该标签指示在该对的两个块中描绘的两个组织模式的相似性程度,其中相似性程度作为该对中的两个块的空间接近度的函数进行计算,其中距离与相异性正相关;训练机器学习逻辑–MLL–使用用标签标记的块对作为训练数据来生成经训练的MLL,经训练的MLL已学会从数字组织图像中提取特征向量,所述数字组织图像以相似的图像具有相似的特征向量并且相异的图像具有相异的特征向量的方式表示图像;并且使用所述经训练的MLL或其分量作为用于计算块的特征向量的特征提取MLL。According to an embodiment, the method further includes creating a training dataset for training a feature extraction MLL. The method includes receiving a plurality of digital training images, each depicting a tissue sample; splitting each of the received training images into multiple blocks (“feature extraction training blocks”); automatically generating block pairs, each block pair having been assigned a label indicating the degree of similarity between two tissue patterns depicted in the two blocks of the pair, wherein the degree of similarity is calculated as a function of the spatial proximity of the two blocks in the pair, where distance is positively correlated with dissimilarity; training the machine learning logic – MLL – using the labeled block pairs as training data to generate a trained MLL, the trained MLL having learned to extract feature vectors from digital tissue images, the digital tissue images representing images in such a way that similar images have similar feature vectors and dissimilar images have dissimilar feature vectors; and using the trained MLL or components thereof as a feature extraction MLL for computing the feature vectors of the blocks.
这种方法可能是有益的,因为可以根据每个数字病理图像中包含的固有的信息自动创建训练数据集的标签,可以创建带注释的数据集,用于训练特征提取MLL,该特征提取MLL特别适于当前解决的生物医学问题,只需选择相应的训练图像。所有进一步的步骤,如拆分、标记和机器学习步骤,都可以全自动或半自动执行。This approach can be beneficial because labels for the training dataset can be automatically created based on the inherent information contained in each digital pathology image. Annotated datasets can be created to train a feature extraction MLL, which is particularly well-suited to the current biomedical problem being addressed, simply by selecting the appropriate training images. All further steps, such as splitting, labeling, and machine learning steps, can be performed fully or semi-automatically.
根据实施例,经训练的MLL是孪生神经网络,包括通过它们的输出层连接的两个神经元子网络。经训练的孪生神经网络的子网络中的一个单独存储在存储介质上,并用作经训练的MLL的分量,该MLL用于计算特征向量。According to an embodiment, the trained MLL is a Siamese neural network, comprising two neuronal subnetworks connected through their output layers. One of the subnetworks of the trained Siamese neural network is stored separately on a storage medium and used as a component of the trained MLL for computing feature vectors.
由MIL程序处理的标签根据实施例,标签选自包括以下项的组:患者对特定药物有应答的指示;患者已发展出转移或特定形式的转移(例如微转移)的指示;癌症患者对特定治疗显示出病理学完全缓解(pCR)的指示;患者患有具有特定形态学状态或微卫星状态的癌症的指示;患者已对特定药物发展出不良反应的指示;遗传特征,特别是基因签名;和/或RNA表达谱。According to an embodiment, the tags processed by the MIL program are selected from the group consisting of: indications that a patient has responded to a specific drug; indications that a patient has developed metastasis or a specific form of metastasis (e.g., micrometastasis); indications that a cancer patient has shown pathological complete remission (pCR) to a specific treatment; indications that a patient has cancer with a specific morphological state or microsatellite state; indications that a patient has developed an adverse reaction to a specific drug; genetic characteristics, particularly gene signature; and/or RNA expression profile.
这些标签可能有助于诊断以及寻找治疗疾病的合适药物。然而,上述标签仅是示例。其他患者相关属性也可用作标签(即训练MIL程序的端点),如上所述。术语“患者相关”还可以包括治疗相关,因为疾病的特定治疗的有效性也与接受治疗的患者有关。These labels can aid in diagnosis and finding suitable medications to treat diseases. However, the labels described above are merely examples. Other patient-related attributes can also be used as labels (i.e., endpoints for training MIL programs), as mentioned above. The term "patient-related" can also include treatment-related, since the effectiveness of a specific treatment for a disease is also relevant to the patient receiving that treatment.
MIL程序和注意力MLL的组合Combination of MIL program and Attention MLL
根据本发明的实施例,MIL程序与基于注意力的MLL相结合,用于计算指示特定块相对于分配从中导出块的图像的标签的预测能力的数值。例如,可以在训练MIL程序时执行组合,如针对图6中描绘的方法和相应系统的实施例所述。根据另一个示例,可以在训练MIL程序时执行组合,如针对图7中描绘的方法和相应系统的实施例所述。According to embodiments of the invention, the MIL procedure is combined with an attention-based MLL to compute a numerical value indicating the predictive power of a particular block relative to a label assigned to an image from which the block is derived. For example, this combination can be performed during the training of the MIL procedure, as described in embodiments of the method and corresponding system depicted in FIG. 6. According to another example, this combination can be performed during the training of the MIL procedure, as described in embodiments of the method and corresponding system depicted in FIG. 7.
根据实施例,注意力MLL是机器学习逻辑,该机器学习逻辑适于计算权重,该权重指示块的特征向量相对于分配从中导出块的图像的标签的预测能力,并且该权重可作为MIL的输入提供或可与MIL输出的数值组合。According to an embodiment, the attention MLL is a machine learning logic adapted to compute weights that indicate the predictive power of a block’s feature vector relative to the labels of the image from which the block is derived, and these weights may be provided as input to the MLL or may be combined with numerical values from the MLL output.
根据实施例,MIL程序和注意力MLL程序都学习识别特征向量和相应块(因此,其中描绘的组织模式)具有相对于患者相关属性值的预测能力。注意力MLL程序可以作为一部分来实现,例如MIL程序的一个子模块。According to embodiments, both the MIL program and the attention MLL program learn to recognize feature vectors and corresponding blocks (and thus, the depicted tissue patterns) with predictive capabilities relative to patient-related attribute values. The attention MLL program can be implemented as a part, for example, a submodule of the MIL program.
根据一些实施例,注意力MLL程序实现置换不变变换运算,MIL程序使用该运算来聚合相对于在包中的块的所有特征向量中编码的包的标签的预测能力。这种置换不变变换针对基于所有块的包生成单个聚合数值。根据实施例,聚合数值与实际分配给包的标签的差异也被认为是在反向传播期间将被最小化的MIL程序的“损失”形式。置换不变变换运算在训练阶段由MIL使用,而且在测试阶段也由过训练的MIL程序使用。According to some embodiments, the attention MIL program implements a permutation-invariant transformation operation, which the MIL program uses to aggregate the predictive power of the bag's label relative to the labels encoded in all feature vectors of the blocks within the bag. This permutation-invariant transformation generates a single aggregated value for the bag based on all blocks. According to embodiments, the difference between the aggregated value and the actual label assigned to the bag is also considered as a form of "loss" of the MIL program that will be minimized during backpropagation. The permutation-invariant transformation operation is used by the MIL during the training phase and also by the overtrained MIL program during the testing phase.
置换不变变换运算可允许指出在训练阶段如何考虑在包的所有块中编码的信息。Permutation-invariant transformations allow for specifying how information encoded in all blocks of a packet should be considered during the training phase.
根据实施例,置换不变变换运算是最大运算。这可能是有益的,因为在训练MIL时生成的预测模型强烈反映了块中描述的组织模式,该块具有相对于包的标签具有最高预测能力的特征向量。该模型不受与标签无关的组织区域/块的负面影响。但是,最大运算将忽略除最高得分的块之外的所有块中包含的所有信息。因此,可能会错过也可能相关的块/组织模式的预测能力。According to the embodiment, the permutation-invariant transformation operation is a maximization operation. This can be beneficial because the predictive model generated during MIL training strongly reflects the organizational pattern described in the block, which has the feature vector with the highest predictive power relative to the bag label. The model is unaffected by the negative impact of organizational regions/blocks unrelated to the label. However, the maximization operation will ignore all information contained in all blocks except the highest-scoring block. Therefore, the predictive power of potentially relevant block/organizational patterns may be missed.
根据实施例,置换不变变换运算是平均运算,例如数值的算术平均值或中位数表示每个单独特征向量相对于特定标签的预测能力。这可能是有益的,因为在训练MIL时生成的预测模型考虑了所有块中描绘的组织模式。然而,考虑与特定标签的出现实际上无关的组织模式和相应块可能导致经训练的MIL的预测准确性的恶化和降低。According to an embodiment, the permutation-invariant transformation operation is an averaging operation, such as the arithmetic mean or median of the values, representing the predictive power of each individual feature vector relative to a particular label. This can be beneficial because the predictive model generated during the training of the MIL takes into account the organizational patterns depicted in all blocks. However, considering organizational patterns and corresponding blocks that are not actually relevant to the occurrence of a particular label can lead to a deterioration and reduction in the predictive accuracy of the trained MIL.
根据实施例,MIL程序的置换不变变换运算是AVERAGE或MEDIAN运算。According to the embodiments, the permutation-invariant transformation operation of the MIL program is the AVERAGE or MEDIAN operation.
根据一个实施例,置换不变变换运算是平均运算,例如数值的算术平均值或中位数表示每个单独特征向量相对于特定标签的预测能力,并且注意力MLL用于计算块中的每一个的权重。针对特定块和相应特征向量计算出的权重表示MIL将在训练阶段针对该块绘制的“注意力”。According to one embodiment, the permutation-invariant transformation operation is an averaging operation, such as the arithmetic mean or median of the values, representing the predictive power of each individual feature vector relative to a particular label, and the attention MIL is used to compute the weights for each in the block. The weights computed for a particular block and its corresponding feature vectors represent the "attention" that MIL will draw for that block during the training phase.
“平均”置换不变变换运算与针对计算特定块的权重配置的注意力MLL的组合可能具有以下优点:AVERAGE运算所提供的益处(考虑到在所有块中传达的信息)可以在不接受此运算的缺点(不相关的组织模式对MIL程序预测模型训练的影响)的情况下使用。这可以允许提高经训练的MIL程序的预测模型的准确性:通过从分配更高权重的块中选择性地/主要地学习,在学习过程中平衡不重要的块。The combination of the "average" permutation-invariant transformation operation and the attention MIL operation with weight configurations tailored to specific blocks may offer the following advantages: the benefits provided by the AVERAGE operation (considering the information conveyed across all blocks) can be used without accepting its drawbacks (the impact of irrelevant organizational patterns on the training of the MIL program's predictive model). This can allow for improved accuracy of the predictive model trained on the MIL program by selectively/primarily learning from blocks assigned higher weights, balancing less important blocks during the learning process.
结合如本文中针对本发明的实施例所描述的注意力MLL程序和MIL程序可具有以下优点:注意力MLL程序(特别是当实现除MAX运算之外的置换不变变换运算时,例如AVERAGE或MEDIAN运算)允许MIL程序在每次迭代中从多个实例(块)中学习,例如,与MAX运算的示例相反,MAX运算是一种在每次迭代中仅选择所有包的一个实例进行学习的稀疏方法。通常,不优选使用AVERAGE或MEDIAN运算,因为该运算可能会导致由MIL程序学习的模型的劣化,这是由没有预测能力的块的特征向量造成的。然而,如果基于注意力MLL的独立估计,这些块的特征向量已分配低的权重,则MIL程序的训练过程可能会受益于使用AVERAGE或MEDIAN而不是MAXIMUM运算作为置换不变变换。Combining the attention MLL procedure and the MIL procedure as described herein with respect to embodiments of the invention offers the following advantages: The attention MLL procedure (particularly when implementing permutation-invariant transformation operations other than the MAX operation, such as AVERAGE or MEDIAN) allows the MIL procedure to learn from multiple instances (blocks) in each iteration. For example, unlike the example of the MAX operation, which is a sparse method that selects only one instance of all bags for learning in each iteration. Generally, the AVERAGE or MEDIAN operation is not preferred because it can lead to degradation of the model learned by the MIL procedure due to the feature vectors of blocks that lack predictive power. However, if the feature vectors of these blocks have been assigned low weights based on independent estimates from the attention MLL, the training process of the MIL procedure may benefit from using AVERAGE or MEDIAN instead of the MAXIMUM operation as a permutation-invariant transformation.
例如,在训练MIL程序时使用注意力MLL可以按照Maximilian Ilse、JakubM.Tomczak、Max Welling中的描述执行:“基于注意力的深度多实例学习”,2018年2月,可通过https://arxiv.org/abs/1802.04712以电子方式获得。For example, using attention-based MLL when training a MIL program can be performed as described in Maximilian Ilse, Jakub M. Tomczak, and Max Welling, “Attention-Based Deep Multi-Instance Learning,” February 2018, available electronically at https://arxiv.org/abs/1802.04712.
根据实施例,GUI被配置为创建并呈现由注意力MLL程序针对从特定数字图像导出的所有块计算出的权重的热图。权重被归一化,例如到0-1的范围内,然后块的归一化权重进行颜色编码。块的权重越相似,基于注意力MLL的热图的颜色就越相似。According to one embodiment, the GUI is configured to create and render a heatmap of weights computed by an attention MLL procedure for all blocks derived from a specific digital image. The weights are normalized, for example, to a range of 0-1, and then the normalized weights of the blocks are color-coded. The more similar the weights of the blocks, the more similar the colors of the attention MLL-based heatmap.
提供加权数值的注意力MLL程序Attention MLL program that provides weighted values
根据实施例(参见例如图6),该方法包括针对块中的每一个计算以加权数值的形式指示与块相关联的特征向量的预测能力的数值。块的每个加权数值被计算为由注意力MLL针对所述块计算出的权重的和由MIL针对所述块计算出的数值的函数。特别地,加权数值可以通过将由注意力MLL计算出的权重乘以相应块的数值来计算。According to an embodiment (see, for example, Figure 6), the method includes computing, for each block, a numerical value indicating the predictive power of the feature vector associated with the block in the form of a weighted value. Each weighted value for a block is calculated as a function of the weights calculated by the attention MLL for the block and the value calculated by the MIL for the block. In particular, the weighted value can be calculated by multiplying the weights calculated by the attention MLL by the value of the corresponding block.
提供加权特征向量的注意力MLL程序Attention MLL program that provides weighted feature vectors
根据实施例,该方法包括针对块中的每一个计算加权特征向量形式的特征向量。加权特征向量被计算为由注意力MLL针对所述块计算出的权重的和由特征提取程序针对所述块计算出的特征向量的函数。特别是,注意力MLL针对特定块所提供的权重可乘以该块的特征向量。According to an embodiment, the method includes computing a feature vector in the form of a weighted feature vector for each block. The weighted feature vector is computed as a function of weights computed by the attention MLL for the block and a feature vector computed by a feature extraction procedure for the block. In particular, the weights provided by the attention MLL for a particular block can be multiplied by the feature vector of that block.
根据另一实施例,实施MIL的训练使得由MIL针对相对于特定标签的特定块而输出的且指示相对于所述包(图像的)标签的块的预测能力的数值乘以由注意力MLL针对该块计算的权重。在反向传播过程中,权重会影响MIL的预测模型的适应性。特定特征向量对训练期间学习的MIL的预测模型的影响与由注意力MLL针对特定块计算出的权重呈正相关。According to another embodiment, training the MIL is performed such that the numerical value output by the MIL for a specific block relative to a specific label, indicating the predictive power of the block relative to the label of the bag (image), is multiplied by the weights calculated by the attention MLL for that block. During backpropagation, the weights affect the adaptability of the MIL's predictive model. The effect of a specific feature vector on the predictive model of the MIL learned during training is positively correlated with the weights calculated by the attention MLL for that specific block.
根据一个实施例,实施MIL的训练使得注意力MLL所提供的权重与特征向量一起提供作为MIL程序的输入。实施MIL的训练,使得MIL从特征向量具有较高权重的块中学到的比从特征向量具有较低权重的块中学到较多。换句话说,块及其特征向量对训练期间学习的MIL的预测模型的影响与由注意力MLL针对特定块计算出的权重呈正相关。According to one embodiment, training the MIL is performed such that the weights provided by the attention MLL, along with the feature vectors, are used as input to the MIL program. The MIL is trained such that it learns more from blocks with higher weights on their feature vectors than from blocks with lower weights. In other words, the impact of blocks and their feature vectors on the predictive model of the MIL learned during training is positively correlated with the weights computed by the attention MLL for a particular block.
使用注意力MLL来计算每个特征向量的权重可能是有利的,因为MIL将从具有高预测潜力的少数块中学到较多,而从显示不相关组织切片的大多数块中学到的较少。结果,经训练的MIL程序的准确性增加。Using attention-based MIL to compute weights for each feature vector can be advantageous because the MIL will learn more from a few blocks with high predictive potential and less from the majority of blocks showing unrelated tissue slices. As a result, the accuracy of the trained MIL program increases.
进一步的实施例Further embodiments
根据实施例,该方法进一步包括:According to an embodiment, the method further includes:
-针对另外一组患者中的每个患者,通过所述图像分析系统接收所述患者的组织样品的至少一个另外的数字图像,每个另外的图像已分配了所述预定义标签中的一个;- For each patient in another group of patients, at least one additional digital image of the patient's tissue sample is received via the image analysis system, each additional image having been assigned one of the predefined labels;
-通过所述图像分析系统,将每个接收到的另外的图像拆分成另外的图像块集,每个块已分配了分配给用于创建另外的块的图像的标签;-The image analysis system splits each received additional image into additional image block sets, each block being assigned a label to the image used to create the additional block;
-针对所述另外的块中的每一个块,通过所述图像分析系统计算另外的特征向量,所述另外的特征向量包含从所述另外的块并从其中描绘的组织模式选择性地提取的图像特征;- For each of the additional blocks, the image analysis system calculates additional feature vectors, which contain image features selectively extracted from the additional blocks and from the organizational patterns depicted therein;
-在针对所述另外的组中的所有患者接收到的所有另外的图像的所述另外的块和相应的另外的特征向量上应用经训练的多实例学习(MIL)程序,以便针对所述另外的块中的每一个块,计算指示从中导出所述另外的块的图像已分配了特定标签的可能性的数值,所述数值作为学习到的所述另外的块的所述特征向量的非线性变换函数进行计算;以及- Apply a trained Multiple Instance Learning (MIL) procedure to the additional blocks and corresponding additional feature vectors of all additional images received for all patients in the additional group, so as to compute, for each of the additional blocks, a numerical value indicating the probability that an image from which the additional block is derived has been assigned a specific label, the numerical value being computed as a nonlinear transformation function of the learned feature vectors of the additional block; and
-经由图像分析系统的GUI,输出另外的图像块报告库,所述另外的报告库包含多个另外的块,所述块根据它们的相应计算出的数值进行排序和/或包含它们的相应数值的图形表示。- An additional image block report library is output via the GUI of the image analysis system. The additional report library contains multiple additional blocks, which are sorted according to their corresponding calculated values and/or contain graphical representations of their corresponding values.
这可能是有利的,因为经训练的MIL程序可以容易地应用于新的图像数据,从而简化相对于目标患者相关属性的新图像的分析和解释,例如通过自动呈现报告库,选择性地呈现新图像的块中,由经训练的MIL程序识别为相对于这种患者相关属性具有高预测能力的块。This can be advantageous because the trained MIL program can be easily applied to new image data, thereby simplifying the analysis and interpretation of new images relative to target patient-related attributes, for example by automatically presenting blocks of new images that are identified by the trained MIL program as having high predictive power relative to such patient-related attributes through an automated report library.
根据实施例,MIL程序在训练阶段学习以将特征向量转换为可以表示特定标签的概率的值。标签可以代表一个类别(例如对特定药物D治疗有应答的患者)或数字终点值(例如指示应答程度的数字或百分比值)。这种学习可以在数学上描述为非线性变换函数的学习,该函数将特征值转换为训练期间所提供的标签之一。根据一些实施例,在测试时,一些小的结构改变被应用于经训练的MIL程序(诸如禁用Dropout层等)并且不发生测试数据的采样。在测试时应用经训练的MIL程序的主要变化是,MIL程序分析测试数据包中的所有实例(块),以计算最终数值,该数值指示块中的每一个以及在训练阶段所提供的多个标签的预测能力。最后,通过聚合为多个标签的图像的块计算出的数值,针对整个图像或特定患者计算最终数值。将经训练的MIL程序应用于患者的一张或多张图像的最终结果是具有最高概率的标签之一(例如“患者将对药物D的治疗有应答!”)。此外,可以在报告图像块库中呈现相对于该标签具有最高预测能力的块之一,该库在结构上等同于上文针对训练阶段描述的报告图像块库。According to embodiments, the MIL program learns during the training phase to transform feature vectors into values that can represent probabilities of a specific label. A label can represent a category (e.g., a patient who responds to treatment with a specific drug D) or a numerical endpoint value (e.g., a number or percentage value indicating the degree of response). This learning can be mathematically described as learning a non-linear transformation function that transforms the feature values into one of the labels provided during training. According to some embodiments, at test time, some minor structural changes are applied to the trained MIL program (such as disabling Dropout layers, etc.) and no sampling of test data occurs. The main change in applying the trained MIL program at test time is that the MIL program analyzes all instances (blocks) in the test data packet to calculate a final value indicating the predictive power of each block and the multiple labels provided during the training phase. Finally, the final value is calculated for the entire image or a specific patient by aggregating the values calculated from the blocks of images into multiple labels. The final result of applying the trained MIL program to one or more images of a patient is one of the labels with the highest probability (e.g., "The patient will respond to treatment with drug D!"). In addition, one of the blocks with the highest predictive power relative to the label can be presented in the report image block library, which is structurally equivalent to the report image block library described above for the training phase.
根据实施例,该方法还包括自动选择或使用户能够选择一个或多个“高预测力块”。“高预测能力块”是其数值指示其特征向量相对于特定标签的预测能力超过高预测能力阈值的块;和/或According to an embodiment, the method further includes automatically selecting or enabling a user to select one or more "high predictive power blocks". A "high predictive power block" is a block whose numerical value indicates that the predictive power of its feature vector relative to a specific label exceeds a high predictive power threshold; and/or
此外或替代性地,该方法还包括自动选择或使用户能够选择一个或多个“伪影块”。伪影块是一种块,其数值指示其特征向量相对于特定标签的预测能力低于最小预测能力阈值或描绘一个或多个伪影。In addition, or alternatively, the method also includes automatically selecting or enabling the user to select one or more "artifact blocks". An artifact block is a block whose value indicates that its feature vector has a predictive power of less than a minimum predictive power threshold relative to a particular label or depicts one or more artifacts.
响应于对一个或多个高预测能力块和/或伪影块的选择,自动重新训练所述MIL程序,从而从训练集中排除所述高预测能力块和所述伪影块。In response to the selection of one or more high-predictability blocks and/or artifact blocks, the MIL program is automatically retrained, thereby excluding the high-predictability blocks and the artifact blocks from the training set.
这些特征可能具有重新训练的MIL程序可能更准确的优点,因为在重新训练期间将不再考虑排除的伪影块。因此,通过基于不包含伪影块的训练数据集的简化版本重新训练MIL程序,可以避免和消除由训练数据集中描述伪影的块引起的学习转换中的任何偏差。These features may have the advantage that a retrained MIL program might be more accurate because excluded artifact patches are no longer considered during retraining. Therefore, by retraining the MIL program based on a simplified version of the training dataset that does not contain artifact patches, any bias in the learning transformation caused by patches describing artifacts in the training dataset can be avoided and eliminated.
使用户能够从训练数据集中移除高预测性的块可能违反直觉,但仍然提供了重要的益处:有时,某些组织模式相对于某些标签的预测能力是不言而喻的。Allowing users to remove highly predictive blocks from the training dataset may be counterintuitive, but it still offers important benefits: sometimes, the predictive power of certain organizational patterns relative to certain labels is self-evident.
例如,包含许多表达肺癌特异性生物标志物的肿瘤细胞的组织切片当然是疾病肺癌存在的重要预后标志物。然而,病理学家可能对一些不太明显的组织模式更感兴趣,例如非肿瘤细胞的存在和/或位置,例如FAP+细胞。For example, tissue sections containing tumor cells expressing many lung cancer-specific biomarkers are certainly important prognostic markers for the presence of lung cancer. However, pathologists may be more interested in some less obvious tissue patterns, such as the presence and/or location of non-tumor cells, such as FAP+ cells.
根据另一示例,MIL训练用于识别肺癌中吸烟诱发的组织模式,这可能具有相对于标签“患者对特定药物D的治疗表现出低应答”的预测潜力。MIL可以计算对应于包括吸烟引起的残留物的肺组织的第一组织模式的最高数值/预测能力。去除显示具有吸烟诱发残留物的组织区域的块可能会发现另一种具有中等预测能力的组织模式。在特征向量包括患者的遗传和/或生理属性值的情况下,在具有最高数值的块被“列入黑名单”之后,这些额外的特征的预测能力的影响也可能变得更加相关。这些遗传相关或生理相关的预测特征也可以反映在特定的组织模式中,因此可以允许病理学家通过检查基于不包含列入黑名单的块的训练块集重新训练MIL后生成的结果块库中的相应块来识别和理解遗传相关或生理相关的属性。In another example, MIL training is used to identify smoking-induced tissue patterns in lung cancer, which may have predictive potential relative to the label "patients show low response to treatment with specific drug D". MIL can calculate the highest numerical value/predictive power corresponding to a first tissue pattern that includes smoking-induced residues. Removing blocks showing tissue regions with smoking-induced residues may reveal another tissue pattern with moderate predictive power. When the feature vector includes the patient's genetic and/or physiological attribute values, the predictive power of these additional features may become more relevant after the block with the highest value is "blacklisted". These genetically or physiologically relevant predictive features can also be reflected in specific tissue patterns, thus allowing pathologists to identify and understand genetically or physiologically relevant attributes by examining corresponding blocks in the resulting block library generated after retraining MIL based on a training block set that does not contain blacklisted blocks.
因此,当删除所有显示肿瘤细胞作为最重要预后因素的块,并基于剩余的训练数据集重新训练MIL程序时,重新训练的MIL将能够更可靠地识别不太突出但仍然重要的预后因素和组织模式。Therefore, when all blocks showing tumor cells as the most important prognostic factor are removed, and the MIL program is retrained based on the remaining training dataset, the retrained MIL will be able to more reliably identify less prominent but still important prognostic factors and tissue patterns.
在另一方面,本发明涉及一种用于识别指示患者相关属性值的组织模式的图像分析系统。图像分析系统包括:In another aspect, the present invention relates to an image analysis system for identifying tissue patterns that indicate patient-related attribute values. The image analysis system includes:
-至少一个处理器;- At least one processor;
-易失性或非易失性存储介质,包括一组患者的组织的数字组织图像,其中针对该组患者中的每个患者,该患者的组织样品的至少一个数字图像存储在该存储介质中,至少一个图像已分配至少两个不同的预定义标签中的一个,每个标签指示其组织在用标签标记的图像中描绘的患者的患者相关属性值;- A volatile or non-volatile storage medium comprising digital tissue images of tissues from a group of patients, wherein for each patient in the group, at least one digital image of a tissue sample of that patient is stored in the storage medium, and at least one image is assigned one of at least two different predefined labels, each label indicating a patient-related attribute value of the tissue in the labeled image depicting the patient.
-图像拆分模块,可由至少一个处理器执行并配置为将图像中的每一个拆分成图像块集,每个块已分配了分配给用于创建块的图像的标签;- An image splitting module, which can be executed by at least one processor and configured to split each of an image into a set of image blocks, each block having been assigned a label to the image used to create the block;
-特征提取模块,可由至少一个处理器执行并且被配置成针对块中的每一个块计算特征向量,该特征向量包括从所述块中描绘的组织模式选择性地提取的图像特征;- A feature extraction module, which can be executed by at least one processor and is configured to compute a feature vector for each block in the block, the feature vector including image features selectively extracted from the organizational patterns depicted in the block;
-多实例学习(MIL)程序,可由至少一个处理器执行并且被配置成在MIL程序的训练阶段接收该组中所有患者的所有图像的所有块和各相应特征向量,MIL程序被配置成在训练阶段将每个块集处理为具有相同标签的块包,该训练包括分析特征向量以便针对块中的每一个块计算数值,该数值指示与该块相关联的特征向量相对于分配给从中导出块的图像的标签的预测能力;- A multi-instance learning (MIL) program, which can be executed by at least one processor and is configured to receive all blocks and corresponding feature vectors of all images of all patients in the group during the training phase of the MIL program. The MIL program is configured to process each block set into a block bundle with the same label during the training phase. The training includes analyzing the feature vectors to compute a numerical value for each block in the block, which indicates the predictive power of the feature vector associated with the block relative to the label assigned to the image from which the block is derived.
-GUI生成模块,可由至少一个处理器执行并且被配置成生成并输出包含图像块报告库的GUI,报告库包含块的子集,该块的子集根据它们的相应计算出的数值进行排序和/或包括它们相应数值的图形表示;以及- A GUI generation module, executable by at least one processor and configured to generate and output a GUI containing a report library of image patches, the report library containing a subset of patches sorted according to their respective calculated values and/or including graphical representations of their respective values; and
-适合显示带有图像块报告库的GUI的显示器。-Suitable for displays of GUIs with image block report libraries.
如本文所用,“组织样品”是可通过本发明的方法分析的细胞的3D组件。3D组件可以是离体细胞块组件的切片。例如,样品可以在从患者收集的组织中制备,例如来自癌症患者的肝脏、肺、肾脏或结肠组织样品。样品可以是显微镜载玻片上的全组织或TMA切片。制备载玻片固定组织样品的方法是本领域众所周知的并且适用于本发明。As used herein, a “tissue sample” is a 3D assembly of cells that can be analyzed by the methods of the present invention. The 3D assembly can be a slice of an ex vivo cell block assembly. For example, the sample can be prepared from tissue collected from a patient, such as liver, lung, kidney, or colon tissue samples from a cancer patient. The sample can be a whole tissue or a TMA slice on a microscope slide. Methods for preparing slide-fixed tissue samples are well known in the art and are applicable to the present invention.
可以使用任何试剂或生物标志物标记对组织样品进行染色,诸如直接与特定生物标志物或各种类型的细胞或细胞区室反应的染料或染色剂、组织化学物质或免疫组织化学物质。并非所有染色剂/试剂都兼容。因此,应充分考虑所用染色剂的类型及其应用顺序,但本领域技术人员可以容易地确定。此类组织化学物质可以是透射显微镜可检测的发色团或荧光显微镜可检测的荧光团。通常,可以将含有细胞的样品与包含至少一种组织化学物质的溶液一起孵育,所述组织化学物质将与靶标的化学基团直接反应或结合。一些组织化学物质通常与媒染剂或金属共同孵育以进行染色。可以将含有样品的细胞与至少一种对目标组分染色的组织化学物质和用作复染剂并结合目标组分外的区域的另一种组织化学物质的混合物一起孵育。替代性地,可以在染色中使用多种探针的混合物,并提供一种鉴定特定探针位置的方法。对含有细胞的样品进行染色的程序是本领域公知的。Tissue samples can be stained using any reagent or biomarker, such as dyes or staining agents, histochemicals, or immunohistochemicals that react directly with specific biomarkers or various cell types or cell compartments. Not all staining agents/reagents are compatible. Therefore, the type of staining agent used and the order of its application should be carefully considered, but this can be readily determined by those skilled in the art. Such histochemicals can be chromophores detectable by transmission microscopy or fluorophores detectable by fluorescence microscopy. Typically, samples containing cells can be incubated with a solution containing at least one histochemical that will react directly with or bind to the target chemical groups. Some histochemicals are often co-incubated with a mordant or metal for staining. Cells containing a sample can be incubated with a mixture of at least one histochemical that stains the target component and another histochemical used as a counterstain and binding to regions outside the target component. Alternatively, a mixture of multiple probes can be used in staining, providing a method for identifying specific probe sites. Procedures for staining samples containing cells are well known in the art.
如本文所用的“图像分析系统”是一种系统,例如一种计算机系统,适于评估和处理数字图像,特别是组织样品的图像,以帮助用户评估或解释图像和/或提取隐含或明确包含在图像中的生物医学信息。例如,计算机系统可以是标准的台式计算机系统或分布式计算机系统,例如云系统。通常,计算机化组织病理学图像分析将相机捕获的单通道或多通道图像作为其输入,并试图提供额外的定量信息以帮助诊断或治疗。As used herein, an "image analysis system" is a system, such as a computer system, suitable for evaluating and processing digital images, particularly images of tissue samples, to help users assess or interpret the images and/or extract biomedical information implicit or explicitly contained within them. For example, the computer system could be a standard desktop computer system or a distributed computer system, such as a cloud system. Typically, computerized histopathological image analysis takes single-channel or multi-channel images captured by a camera as its input and attempts to provide additional quantitative information to aid in diagnosis or treatment.
本发明的实施例可用于确定较大患者群中的哪个患者亚群将可能从特定药物中获益。个性化医学(PM)是一个新的医学领域,其目的是根据个人的基因组、表观基因组和蛋白质组学特征提供有效的、量身定制的治疗策略。PM不仅尝试治疗患者,还防止患者受到因无效治疗的副作用。肿瘤发展时经常发生的一些突变会引起对某些治疗的抵抗。因此,可以至少部分地通过生物标志物特异性染色的组织样品的组织图像揭示的患者的突变谱将允许经训练的MIL程序明确决定特定治疗是否对个体患者有效。目前,有必要通过试错法来确定处方药对患者是否有效。试错过程可能会产生许多副作用,诸如不希望的和复杂的药物相互作用、处方药物的频繁更换、有效药物被确定之前的长时间延迟、疾病进展等。PM是基于将个体分成亚群,这些亚群对针对其特定疾病的治疗剂的应答各不相同。例如,一些ALK激酶抑制剂是用于治疗约5%的ALK基因表达升高的NSCLC肺癌患者的有用药物。然而,一段时间后,由于ALK基因或ALK信号级联下游其他基因的突变,激酶抑制剂变得无效。.因此,肺癌患者的智能分子表征允许通过患者分层优化使用某些突变特异性药物。因此,从中获取训练图像或测试图像的“患者组”可以是诸如“100名乳腺癌患者”、100名HER+乳腺癌患者、“200名结肠癌患者”等的组。Embodiments of the present invention can be used to determine which patient subgroups within a larger patient population are likely to benefit from a particular drug. Personalized medicine (PM) is a new field of medicine that aims to provide effective, tailored treatment strategies based on an individual's genomic, epigenomic, and proteomic characteristics. PM not only attempts to treat patients but also prevents them from suffering the side effects of ineffective treatments. Some mutations that frequently occur during tumor development can cause resistance to certain treatments. Therefore, a patient's mutation profile, which can be revealed at least partially through tissue images of tissue samples stained with biomarkers specifically, will allow a trained MIL (Mutual Injection and Treatment) program to definitively determine whether a particular treatment is effective for an individual patient. Currently, it is necessary to determine whether a prescription drug is effective for a patient through trial and error. The trial-and-error process can produce many side effects, such as undesirable and complex drug interactions, frequent changes in prescription drugs, long delays before an effective drug is identified, disease progression, etc. PM is based on dividing individuals into subgroups that respond differently to therapeutic agents targeting their specific diseases. For example, some ALK kinase inhibitors are useful drugs for treating approximately 5% of NSCLC lung cancer patients with elevated ALK gene expression. However, after a period of time, kinase inhibitors become ineffective due to mutations in the ALK gene or other genes downstream of the ALK signaling cascade. Therefore, intelligent molecular characterization of lung cancer patients allows for optimized use of certain mutation-specific drugs through patient stratification. Thus, the "patient group" from which training or test images are obtained could be a group such as "100 breast cancer patients," "100 HER+ breast cancer patients," "200 colon cancer patients," etc.
这里使用的“数字图像”是二维图像的数字表示,通常是二进制的。通常,组织图像是光栅类型的图像,意味着该图像是分别分配了至少一个强度值的像素的光栅(“矩阵”)。一些多通道图像可能具有每个颜色通道具有一个强度值的像素。数字图像包含固定数量的像素行和列。像素是图像中最小的单独元素,保存着代表给定颜色在任何特定点的亮度的过时值。通常,像素作为光栅图像或光栅地图(小整数的二维数组)存储在计算机内存中。这些值通常以压缩形式传输或存储。可以获取数字图像,例如通过数码相机、扫描仪、坐标测量机、显微镜、载玻片扫描装置等。The term "digital image" as used here refers to a digital representation of a two-dimensional image, typically in binary form. Generally, an organized image is a raster-type image, meaning that the image is a raster ("matrix") of pixels, each assigned at least one intensity value. Some multi-channel images may have pixels with one intensity value for each color channel. A digital image contains a fixed number of rows and columns of pixels. A pixel is the smallest individual element in an image, holding an outdated value representing the brightness of a given color at any given point. Pixels are typically stored in computer memory as raster images or raster maps (two-dimensional arrays of small integers). These values are usually transmitted or stored in compressed form. Digital images can be acquired, for example, via digital cameras, scanners, coordinate measuring machines, microscopes, slide scanning devices, etc.
此处使用的“标签”是数据值,例如字符串或数值,表示并指出患者相关属性值。标签的示例可以是“患者对药物D的应答=真”、“患者对药物D的应答=假”、“无进展存活时间=6个月”等。The “labels” used here are data values, such as strings or numbers, that represent and indicate relevant patient attribute values. Examples of labels could be “Patient response to drug D = True”, “Patient response to drug D = False”, “Progression-free survival = 6 months”, etc.
本文使用的“图像块”是数字图像的子区域。通常,从数字图像创建的块可以具有任何形状,例如圆形、椭圆形、多边形、矩形、正方形等,并且可以重叠或不重叠。根据优选实施例,从图像生成的块是矩形的,优选地重叠块。使用重叠块可具有的优势在于,否则将被块生成过程破碎的组织模式也在包中表示。例如,两个重叠块的重叠可以覆盖20-30%,例如单个块面积的25%。The term "image block" as used herein refers to a sub-region of a digital image. Typically, blocks created from a digital image can have any shape, such as circles, ellipses, polygons, rectangles, squares, etc., and can overlap or not. According to a preferred embodiment, the blocks generated from the image are rectangular, and preferably overlapping blocks. The advantage of using overlapping blocks is that organizational patterns that would otherwise be fragmented by the block generation process are also represented in the package. For example, the overlap of two overlapping blocks can cover 20-30%, such as 25% of the area of a single block.
根据实施例,图像块库,例如图像块报告库和/或图像相似性搜索块库是GUI上块的网格样式组织,其中块在图像块库中的空间组织独立于它们在从中导出块的图像的空间排列。According to an embodiment, the image block library, such as the image block report library and/or the image similarity search block library, is a grid-style organization of blocks on a GUI, wherein the spatial organization of blocks in the image block library is independent of their spatial arrangement in the images from which the blocks are derived.
这里使用的“特征向量”是包含描述对象的重要特征的信息的数据结构。数据结构可以是单维或多维数据结构,其中特定类型的数据值存储在该数据结构内的相应位置。例如,数据结构可以是向量、数组、矩阵等。特征向量可以被认为是代表某个对象的数值特征的n维向量。在图像分析中,特征可以有多种形式。图像的简单特征表示是每个像素的原始强度值。然而,更复杂的特征表示也是可能的。例如,从图像或图像块中提取的特征也可以是SIFT描述符特征(规模不变特征变换)。这些特征捕捉了不同线条方向的普遍性。其他特征可以指示图像或图像块的对比度、梯度方向、颜色组成和其他方面。The "feature vector" used here is a data structure containing information describing the important features of an object. The data structure can be one-dimensional or multi-dimensional, where data values of a specific type are stored in corresponding positions within the structure. For example, the data structure can be a vector, array, matrix, etc. A feature vector can be considered an n-dimensional vector representing the numerical features of an object. In image analysis, features can take many forms. A simple feature representation of an image is the raw intensity value of each pixel. However, more complex feature representations are also possible. For example, features extracted from an image or image patch can also be SIFT descriptor features (Scale Invariant Feature Transform). These features capture the generality of different line directions. Other features can indicate the contrast, gradient direction, color composition, and other aspects of an image or image patch.
这里使用的“热图”是数据的图形表示,其中矩阵中包含的单独值以颜色和/或强度值表示。根据一些实施例,热图是不透明的并且包括组织载玻片图像的至少一些结构,热图是基于这些结构创建的。根据其他实施例,热图是半透明的并且显示为用于创建热图的组织图像顶部的覆盖层。根据一些实施例,热图通过相应的颜色或像素强度指示多个相似性得分或相似性得分范围中的每一个。The term "heatmap" as used herein is a graphical representation of data in which individual values contained in a matrix are represented by color and/or intensity values. According to some embodiments, the heatmap is opaque and includes at least some structures of a tissue slide image upon which the heatmap is created. According to other embodiments, the heatmap is semi-transparent and appears as a cover layer on top of the tissue image used to create the heatmap. According to some embodiments, the heatmap indicates each of a plurality of similarity scores or ranges of similarity scores by corresponding color or pixel intensity.
如本文所用,“生物标志物特异性染色剂”是选择性染色特定生物标志物,例如特定的蛋白质如HER,但不是一般的其他生物标志物或组织成分。As used in this article, “biomarker-specific staining agents” are selective staining agents for specific biomarkers, such as specific proteins like HER, but not for other biomarkers or tissue components in general.
如本文所用,“非生物标志物特异性染色剂”是具有更一般的结合行为的染色剂。非生物标志物特异性染色剂不会选择性地染色单独蛋白质或DNA序列,而是染色具有特定物理或化学特性的更大组的物质和亚细胞以及超细胞结构。例如,苏木精和曙红分别是非生物标志物特异性染色剂。苏木精是一种呈碱性/阳性的深蓝色或紫色染色剂。苏木精与嗜碱性物质(诸如DNA和RNA,呈酸性且带负电荷)结合。细胞核中的DNA/RNA和粗面内质网核糖体中的RNA都是酸性的,因为核酸的磷酸骨架带负电荷。这些骨架与含有正电荷的碱性染料形成盐。因此,像苏木精这样的染料会与DNA和RNA结合并将它们染成紫色。曙红是一种呈酸性且呈阴性的红色或粉红色染色剂。曙红与嗜酸物质结合,诸如带正电荷的氨基酸侧链(例如赖氨酸、精氨酸)。某些细胞的细胞质中的大多数蛋白质是碱性的,因为精氨酸和赖氨酸氨基酸残基使它们带正电荷。这些与含有负电荷的酸性染料(如曙红)形成盐。因此,曙红与这些氨基酸/蛋白质结合并将它们染成粉红色。这包括肌肉细胞中的细胞质细丝、细胞内膜和细胞外纤维。As used herein, “non-biomarker-specific staining agents” are staining agents with more general binding behavior. Non-biomarker-specific staining agents do not selectively stain individual protein or DNA sequences, but rather stain larger groups of substances and subcellular and supracellular structures with specific physical or chemical properties. For example, hematoxylin and eosin are non-biomarker-specific staining agents. Hematoxylin is a basic/positive deep blue or purple staining agent. Hematoxylin binds to basophilic substances (such as DNA and RNA, which are acidic and negatively charged). DNA/RNA in the cell nucleus and RNA in rough endoplasmic reticulum ribosomes are acidic because the phosphate backbone of nucleic acids is negatively charged. These backbones form salts with positively charged basic dyes. Therefore, dyes like hematoxylin bind to DNA and RNA and stain them purple. Eosin is an acidic and negative red or pink staining agent. Eosin binds to acidophilic substances, such as positively charged amino acid side chains (e.g., lysine, arginine). In some cells, most proteins in the cytoplasm are basic because the amino acid residues of arginine and lysine give them a positive charge. These form salts with negatively charged acidic dyes such as eosin. Therefore, eosin binds to these amino acids/proteins and stains them pink. This includes cytoplasmic filaments, intracellular membranes, and extracellular fibers in muscle cells.
本文使用的“注意力机器学习逻辑程序”是经训练的以将权重分配给特定参数的MLL,由此权重指示重要性以及其他程序可能在分析这些参数上花费的注意力。注意力MLL背后的想法是模拟人脑选择性地关注与当前上下文特别相关的可用数据子集的能力。使用注意力MLL,例如在文本挖掘领域,有选择地针对特定单词分配权重和计算资源,这些单词对于从句子中获取含义特别重要。并非所有词都同等重要。其中一些比其他更能表征一个句子。通过基于训练数据集训练注意力MLL生成的注意力模型可以指出句子向量可以对“重要”词有更多的注意力。根据一个实施例,经训练的注意力MLL适于计算检查的每个特征向量中的每个特征值的权重,以及计算每个特征向量中所有特征值的加权和。这个加权和体现了块的整个特征向量。The “attention machine learning logic program” used in this paper is an MLL trained to assign weights to specific parameters, whereby the weights indicate importance and the attention that other programs might spend analyzing these parameters. The idea behind attention MLLs is to simulate the human brain’s ability to selectively focus on a subset of available data that is particularly relevant to the current context. Using attention MLLs, for example in the field of text mining, weights and computational resources are selectively assigned to specific words that are particularly important for extracting meaning from a sentence. Not all words are equally important. Some are more representative of a sentence than others. By training an attention MLL on a training dataset, an attention model can be generated that indicates how much attention can be paid to “important” words in a sentence vector. According to one embodiment, the trained attention MLL is adapted to compute the weight of each feature value in each feature vector examined, as well as to compute a weighted sum of all feature values in each feature vector. This weighted sum reflects the entire feature vector of the block.
根据实施例,注意力MLL是包括神经注意力机制的MLL,该机制适于为神经网络配备专注于其输入(或特征)的子集的能力:它选择特定输入。设x∈Rd为输入向量,z∈Rk为特征向量,a∈[0,1]k为注意力向量,g∈Rk注意力一瞥(attention glimpse),以及fφ(x)为带参数φ的注意力网络。According to an embodiment, an attention MLL is an MLL that includes a neural attention mechanism adapted to equip a neural network with the ability to focus on a subset of its inputs (or features): it selects specific inputs. Let x∈Rd be the input vector, z∈Rk be the feature vector, a∈[0,1]k be the attention vector, g∈Rk be the attention glimpse, and fφ(x) be the attention network with parameter φ.
通常,注意力实现为Typically, attention is implemented as
ag=fφ(x),=a⊙z,ag=fφ(x),=a⊙z,
其中⊙是逐元素乘法,而z是另一个具有参数θ的神经网络fθ(x)的输出。我们可以谈论软注意力,它将特征与零到一之间的值的(软)掩码相乘,或者硬注意力,当这些值限制为恰好为零或一时,即a∈{0,1}k。在后一种情况下,我们可以使用hard attention mask直接索引特征向量:g-=z[a](在Matlab符号中),它改变了它的维度,现在g-∈Rm及m≤k。Here, ⊙ represents element-wise multiplication, and z is the output of another neural network fθ(x) with parameter θ. We can talk about soft attention, which multiplies the features by a (soft) mask of values between zero and one, or hard attention, when these values are restricted to exactly zero or one, i.e., a∈{0,1}k. In the latter case, we can directly index the feature vector using a hard attention mask: g-=z[a] (in Matlab notation), which changes its dimension, now g-∈Rm and m≤k.
如本文所用,术语“强度信息”或“像素强度”是在数字图像的像素上捕获的或由数字图像的像素表示的电磁辐射(“光”)的量的量度。如本文所用,术语“强度信息”可包括额外的、相关的信息,例如特定颜色通道的强度。MLL可以使用该信息以计算方式提取诸如数字图像中包含的梯度或纹理的衍生信息,并且可以在训练期间和/或在由经训练的MLL进行特征提取期间从数字图像中隐含地或明确地提取衍生信息。例如,表述“数字图像的像素强度值与一种或多种特异性染色剂的强度相关”可以暗示强度信息(包括颜色信息)允许MLL并且还可能允许用户识别已由所述一种或多种染色剂中的一种特异性染色剂染色的组织样品的区域。例如,描绘用苏木精染色的样品区域的像素在蓝色通道中可能具有高像素强度,描绘用快速红染色的样品区域的像素在红色通道中可以具有高像素强度。As used herein, the term "intensity information" or "pixel intensity" is a measure of the amount of electromagnetic radiation ("light") captured at or represented by pixels of a digital image. As used herein, the term "intensity information" may include additional, related information, such as the intensity of a particular color channel. MLLs can use this information to computationally extract derived information such as gradients or textures contained in a digital image, and may implicitly or explicitly extract derived information from a digital image during training and/or during feature extraction by a trained MLL. For example, the statement "pixel intensity values of a digital image are correlated with the intensity of one or more specific staining agents" may imply that intensity information (including color information) allows the MLL and may also allow the user to identify regions of a tissue sample stained with one of the one or more specific staining agents. For example, pixels depicting a sample region stained with hematoxylin may have high pixel intensity in the blue channel, and pixels depicting a sample region stained with fast red may have high pixel intensity in the red channel.
如本文所用,“全卷积神经网络”是由卷积层组成的神经网络,没有任何通常在网络末端发现的全连接层或多层感知器(MLP)。全卷积网络在每层学习过滤器。甚至网络末端的决策层也学习过滤器。全卷积网络试图学习表示并根据局部空间输入做出决策。As used in this paper, a "fully convolutional neural network" is a neural network composed of convolutional layers, without any fully connected layers or multilayer perceptrons (MLPs) typically found at the ends of the network. Fully convolutional networks learn filters at each layer. Even the decision layers at the ends of the network learn filters. Fully convolutional networks attempt to learn representations and make decisions based on local spatial inputs.
根据实施例,全卷积网络是仅具有以下形式的层的卷积网络,其激活函数在满足以下特性的特定层中的位置(I,j)生成输出数据向量yij:According to an embodiment, a fully convolutional network is a convolutional network with only layers of the following form, whose activation function generates an output data vector y<sub>ij</sub> at position (i, j) in a specific layer that satisfies the following properties:
yij=fks({xsi+δi,sj+δj}0≤δi,δj≤k)y ij =f ks ({x si+δi, sj+δj } 0≤δi, δj≤k )
其中xij为特定层中位置(i;j)的数据向量,yij为以下层所述位置的数据向量,其中yij为网络激活函数产生的输出,其中k称为内核尺寸,s是步幅或子采样因子,fks确定层类型:卷积或平均池化的矩阵乘法,最大池化的空间最大值,或激活函数的元素非线性等其他类型层。这种函数形式在组合下保持不变,内核尺寸和步幅遵循转换规则:Where x <sub>ij </sub> is the data vector at position (i; j) in a specific layer, y <sub>ij</sub> is the data vector at position (i ; j) in the following layer, y<sub>ij</sub> is the output of the network activation function, k is called the kernel size, s is the stride or subsampling factor, and f <sub>ks</sub> determines the layer type: matrix multiplication of convolution or average pooling, spatial maximum of max pooling, or other types of layers such as element-wise nonlinear activation function. This functional form remains invariant under combination, and the kernel size and stride follow transformation rules:
虽然一般深度网络计算一般非线性函数,但只有这种形式的层的网络计算非线性过滤器,也称为深度滤波器或全卷积网络。FCN自然地对任意尺寸的输入进行操作,并产生相应(可能重新采样)空间维度的输出。有关若干全卷积网络的特征的更详细描述,请参阅Jonathan Long、Evan Shelhamer和Trevor Darrell:“用于语义分割的全卷积网络”,CVPR2015。While general deep networks compute nonlinear functions, only networks with this type of layer compute nonlinear filters, also known as deep filters or fully convolutional networks. FCNs naturally operate on inputs of arbitrary size and produce outputs with corresponding (possibly resampled) spatial dimensions. For a more detailed description of the characteristics of several fully convolutional networks, see Jonathan Long, Evan Shelhamer, and Trevor Darrell, “Fully Convolutional Networks for Semantic Segmentation,” CVPR 2015.
如本文所用,“机器学习逻辑(MLL)”是一种程序逻辑,例如一种软件,如经训练的神经元网络或支持向量机等,已经训练或可以在训练过程中进行训练,并且-作为学习阶段的结果-已经学会根据所提供的训练数据执行一些预测和/或数据处理任务。因此,MLL可以是至少部分未由程序员明确指出的程序代码,但在从样品输入构建一个或多个隐含或明确模型的数据驱动学习过程中隐含学习和修改。机器学习可以采用监督式或无监督式学习。有效的机器学习通常很困难,因为找到模式很困难,而且通常没有足够的训练数据可用。As used in this article, "Machine Learning Logic (MLL)" is a program logic, such as software like a trained neural network or support vector machine, that has been trained or can be trained during training and—as a result of the learning phase—has learned to perform some prediction and/or data processing tasks based on the provided training data. Therefore, MLL can be at least partially unspecified program code that is implicitly learned and modified during data-driven learning processes that build one or more implicit or explicit models from sample inputs. Machine learning can be supervised or unsupervised. Effective machine learning is often difficult because finding patterns is challenging and there is often insufficient training data available.
如本文所用,术语“生物标志物”是一种分子,该分子可在生物样品中作为组织类型、正常或致病过程或对治疗干预的应答的指标进行测量。在一个特定实施例中,生物标志物选自:蛋白质、肽、核酸、脂质和碳水化合物。更特别地,生物标志物可以是特定的蛋白质,例如EGRF、HER2、p53、CD3、CD8、Ki67等。某些标志物是特定细胞的特征,而其他标志物已鉴定为与特定疾病或病症相关。As used herein, the term "biomarker" is a molecule that can be measured in a biological sample as an indicator of tissue type, normal or pathogenic process, or response to therapeutic intervention. In one particular embodiment, biomarkers are selected from proteins, peptides, nucleic acids, lipids, and carbohydrates. More specifically, biomarkers can be specific proteins, such as EGRF, HER2, p53, CD3, CD8, Ki67, etc. Some biomarkers are characteristic of specific cells, while others have been identified as being associated with specific diseases or conditions.
基于组织样品图像的图像分析来确定特定肿瘤的阶段,可能需要用多种生物标志物特异性染色剂对样品进行染色。组织样品的生物标志物特异性染色通常涉及使用选择性结合目标生物标志物的一抗。特别是这些一抗,以及染色方案的其他组分,可能很贵,因此在许多应用场景中,特别是高通量筛选,由于成本原因,可能会排除可用的图像分析技术的使用。Image analysis of tissue samples to determine the stage of a specific tumor may require staining the samples with multiple biomarker-specific staining agents. Biomarker-specific staining of tissue samples typically involves the use of primary antibodies that selectively bind to the target biomarker. In particular, these primary antibodies, along with other components of the staining protocol, can be expensive, thus cost may exclude the use of available image analysis techniques in many applications, especially high-throughput screening.
通常,组织样品用背景染色(“反染色”)染色,例如苏木精染色剂或苏木精和曙红染色剂的组合(“H&E”染色),以揭示大规模组织形态以及细胞和细胞核的边界。除了背景染色之外,可以根据要回答的生物医学问题应用多种生物标志物特异性染色剂,例如肿瘤的分类和分期,检测组织中某些细胞类型的数量和相对分布等。Typically, tissue samples are stained with background staining (“reverse staining”), such as hematoxylin staining or a combination of hematoxylin and eosin staining (“H&E” staining), to reveal large-scale tissue morphology and cell and nucleus boundaries. In addition to background staining, a variety of biomarker-specific stains can be applied depending on the biomedical question being answered, such as tumor classification and staging, detecting the number and relative distribution of certain cell types in tissues, etc.
附图说明Attached Figure Description
在以下实施例中,仅通过示例,参考附图更详细地解释本发明,其中:In the following embodiments, the invention is explained in more detail by way of example only, with reference to the accompanying drawings, wherein:
图1描绘了根据本发明的实施例的方法的流程图;Figure 1 depicts a flowchart of a method according to an embodiment of the present invention;
图2描绘了根据本发明实施例的图像分析系统的框图;Figure 2 depicts a block diagram of an image analysis system according to an embodiment of the present invention;
图3描绘了根据本发明的实施例的具有报告图像块库的GUI;Figure 3 depicts a GUI with a report image block library according to an embodiment of the present invention;
图4描绘了根据本发明的实施例的具有相似性搜索图像块库的GUI;Figure 4 depicts a GUI for a similarity-based image patch library according to an embodiment of the present invention;
图5描绘了根据本发明实施例的特征提取MLL程序的网络架构;Figure 5 depicts the network architecture of the feature extraction MLL program according to an embodiment of the present invention;
图6描绘了一种用于结合MIL程序和注意力MLL的可能系统架构;Figure 6 depicts a possible system architecture for combining MIL procedures and attention MLL;
图7描述了另一种用于结合MIL程序和注意力MLL的可能系统架构;Figure 7 illustrates another possible system architecture for combining MIL procedures and attention MLL;
图8说明了2D和3D坐标系中块的空间距离;Figure 8 illustrates the spatial distances of blocks in 2D and 3D coordinate systems;
图9描绘了根据本发明的实施例的孪生神经网络的架构;Figure 9 illustrates the architecture of a twin neural network according to an embodiment of the present invention;
图10描述了作为截短的孪生神经网络实现的特征提取MLL;Figure 10 illustrates the feature extraction MLL implemented as a truncated Siamese neural network;
图11描述了在图像数据库中基于相似性搜索使用特征向量的计算机系统;Figure 11 illustrates a computer system that uses feature vectors for similarity-based search in an image database;
图12显示基于它们的空间接近度用标签标记的“相似”和“相异”块对;以及Figure 12 shows pairs of "similar" and "dissimilar" blocks labeled based on their spatial proximity; and
图13显示了基于相似性搜索结果的特征向量,该特征向量由基于接近度的相似性标签训练的特征提取MLL提取。Figure 13 shows the feature vector based on similarity search results, which is extracted by feature extraction MLL trained on proximity-based similarity labels.
具体实施方式Detailed Implementation
图1示出了根据本发明的实施例的方法流程图。该方法可以用于例如预测患者的患者相关属性值,诸如,例如,生物标志物状态、诊断、治疗结果、特定癌症(诸如结直肠癌或乳腺癌)的微卫星状态(MSS)、淋巴结中的微转移以及诊断活检中的病理完全缓解(pCR)。预测基于使用基于-优选无假设-特征提取的深度学习的组织学载玻片的数字图像。Figure 1 illustrates a flowchart of a method according to an embodiment of the present invention. This method can be used, for example, to predict patient-related attribute values, such as, for example, biomarker status, diagnosis, treatment outcome, microsatellite status (MSS) of a specific cancer (such as colorectal cancer or breast cancer), micrometastases in lymph nodes, and pathological complete remission (pCR) in diagnostic biopsies. The prediction is based on digital images of histological slides using deep learning based on—preferably—assumption-free—feature extraction.
方法100可用于训练弱监督深度学习计算机算法,该算法设计用于识别和提取迄今为止未知的预测性组织学签名。该方法允许识别指示患者相关属性值的组织模式。Method 100 can be used to train a weakly supervised deep learning computer algorithm designed to identify and extract previously unknown predictive histological signatures. This method allows for the identification of tissue patterns that indicate patient-related attribute values.
可以提供来自患者的组织标本,例如以FFPET组织块的形式。需要从具有预先确定和预先知道的终点(例如存活、应答、基因签名等)的患者身上获取组织块,以用作标签。Tissue specimens from patients can be provided, for example, in the form of FFPET tissue blocks. Tissue blocks need to be obtained from patients with predetermined and known endpoints (e.g., survival, response, genetic signature, etc.) for use as labels.
将组织块切片,并将切片放置在显微镜载玻片上。然后,切片用一种或多种组织学相关的染色剂染色,例如H&E和/或各种生物标志物特异性染色剂。图像取自染色的组织切片,例如使用载玻片扫描显微镜。The tissue block is sectioned and placed on a microscope slide. The section is then stained with one or more histology-associated staining agents, such as H&E and/or various biomarker-specific staining agents. Images are taken from the stained tissue sections, for example, using a scanning microscope on a slide.
在第一步骤102中,图像分析系统(例如,参考图2所描述的)针对一组患者中的每个患者接收所述患者的组织样品的至少一个数字图像212。In the first step 102, the image analysis system (e.g., as described with reference to FIG2) receives at least one digital image 212 of a tissue sample from each of the patients in a group of patients.
读取可包括从数据库读取图像。例如,图像可以是多年以前的组织样品图像。旧图像数据集的优势在于许多相关事件的结果,例如治疗成功、疾病进展、副作用同时是已知的并且可用于创建训练数据集,该训练数据集包括将已知事件分配为标签的组织图像。此外或替代性地,可直接从图像采集系统接收到的图像,例如显微镜或载玻片扫描仪。标签可手动或自动分配给接收到的图像。例如,用户可配置载玻片扫描仪的软件,使得所获取的图像在其获取过程中用特定标签自动标记。这在依次获取具有相同患者相关属性值/端点的大量患者的组织样品图像的场景中可能是有帮助的,例如已知显示对特定药物D有应答的第一组100名乳腺癌患者的100个组织图像以及已知未显示出这种应答的第二组120名乳腺癌患者的120个组织图像。用户可能必须在获取第一组图像之前仅一次设置要分配给捕获图像的标签,然后在获取第二组图像之前第二次设置标签。Retrieving images may include reading images from a database. For example, images could be tissue sample images from many years ago. The advantage of older image datasets is that the outcomes of many relevant events, such as treatment success, disease progression, and side effects, are known simultaneously and can be used to create a training dataset that includes tissue images labeled with these known events. Alternatively, images can be received directly from an image acquisition system, such as a microscope or slide scanner. Labels can be assigned manually or automatically to the received images. For example, a user can configure the software of a slide scanner so that the acquired images are automatically labeled with specific tags during their acquisition process. This can be helpful in scenarios where tissue sample images from a large number of patients with the same patient-related attribute values/endpoints are acquired sequentially, such as 100 tissue images from a first group of 100 breast cancer patients known to have responded to a specific drug D, and 120 tissue images from a second group of 120 breast cancer patients known not to have shown such a response. The user may have to set the labels to be assigned to the captured images only once before acquiring the first group of images, and then set the labels a second time before acquiring the second group of images.
针对每个患者,检索一个或多个图像。例如,可以根据不同的染色方案对相同的组织样品进行多次染色,由此针对每个染色方案获取图像。替代性地,若干相邻的组织样品切片可分别用相同或不同的染色方案染色,并且针对组织样品载玻片中的每一个获取图像。接收到的图像中的每一个已分配至少两个不同的预定义标签中的一个。每个标签指示在用标签标记的图像中描绘其组织的患者的患者相关属性值。属性值可以是任何类型,例如布尔值、数字、字符串、有序参数值等。For each patient, one or more images are retrieved. For example, the same tissue sample may be stained multiple times according to different staining schemes, thereby acquiring an image for each staining scheme. Alternatively, several adjacent tissue sample sections may be stained with the same or different staining schemes, and an image may be acquired for each of the tissue sample slides. Each of the received images has been assigned one of at least two different predefined labels. Each label indicates a patient-related attribute value of the patient whose tissue is depicted in the labeled image. Attribute values can be of any type, such as Boolean values, numbers, strings, ordered parameter values, etc.
接下来在步骤104,图像分析系统将每个接收到的图像拆分成图像块集216。因此,每个块已分配了分配给用于创建块的图像的标签。Next, in step 104, the image analysis system splits each received image into a set of image blocks 216. Therefore, each block has been assigned a label that was originally assigned to the image used to create the block.
例如,作为2016年“CAMELYON16”挑战的基础发布的图像数据集可以用作训练数据集。CAMELYON16数据集由乳腺癌患者H&E染色淋巴结组织切片的270个全玻片图像组成,作为训练图像数据集提供(160个正常组织图像,110个肿瘤转移图像)。该数据集可在https://camelyon16.grand-challenge.org/data/下获得。在10倍放大率下,该数据集的图像可用于从尺寸为256x256像素的非背景区域生成1,113,403个RGB块,每个块都没有重叠。For example, the image dataset released as the basis for the 2016 "CAMELYON16" challenge can be used as a training dataset. The CAMELYON16 dataset consists of 270 whole-slide images of H&E-stained lymph node tissue sections from breast cancer patients, provided as the training image dataset (160 normal tissue images and 110 tumor metastasis images). This dataset is available at https://camelyon16.grand-challenge.org/data/ . At 10x magnification, the images in this dataset can be used to generate 1,113,403 RGB blocks from non-background regions of size 256x256 pixels, with no overlap between blocks.
根据一个实施例,接收到的图像以及生成的块是多通道图像。通过创建具有不同尺寸、放大级别和/或包括一些模拟伪影和噪声的现有块的修改副本,可以增加块的数量以丰富训练数据集。在一些情况下,可以通过如本文针对本发明的实施例所描述的对包中的实例重复采样并将所选实例放置在额外的包中来创建多个包。这种“取样”也可能具有丰富训练数据集的积极作用。According to one embodiment, the received image and the generated blocks are multi-channel images. The number of blocks can be increased to enrich the training dataset by creating modified copies of existing blocks with different sizes, magnification levels, and/or including some simulated artifacts and noise. In some cases, multiple packets can be created by resampling instances in a packet as described herein with respect to embodiments of the invention and placing selected instances in additional packets. This “sampling” can also have a positive effect on enriching the training dataset.
在某些情况下,可以将特征向量聚集为N个群集,并且可以从每个群集中随机选择M个实例(块)放入伪包中,以生成包中实例的聚集等变种群。In some cases, feature vectors can be clustered into N clusters, and M instances (blocks) can be randomly selected from each cluster and placed into a pseudo-packet to generate a variant of the clusters of instances in the pack.
接下来在步骤106中,图像分析系统为块中的每一个计算特征向量220。特征向量包括从所述块中描绘的组织模式选择性地提取的图像特征。任选地,特征向量还可以包括遗传特征或其他患者或患者相关数据,这些数据可用于从中导出图像和相应块的患者。根据一些实施例,特征提取由经训练的特征提取MLL执行。特征提取MLL可以针对训练数据集中的每个块生成特征向量,同时保留特征-向量-标签关系。然而,其他实施例可以使用明确编程的特征提取算法来提供多种特征,这些特征描述了计算特征向量的块中描绘的组织区域。Next, in step 106, the image analysis system computes a feature vector 220 for each block. The feature vector includes image features selectively extracted from the tissue patterns depicted in the block. Optionally, the feature vector may also include genetic features or other patient or patient-related data from which the image and the patient of the corresponding block are derived. According to some embodiments, feature extraction is performed by a trained feature extraction MLL. The feature extraction MLL can generate feature vectors for each block in the training dataset while preserving feature-vector-label relationships. However, other embodiments may use an explicitly programmed feature extraction algorithm to provide a variety of features that describe the tissue regions depicted in the block from which the feature vectors are computed.
接下来在步骤108中,基于针对该组中的所有患者接收到的所有图像的所有块和相应特征向量训练多实例学习(MIL)程序226。因此,MIL程序将每个块集处理为具有相同标签的块包。训练包括分析训练数据集中的块的特征向量220,针对块中的每一个计算数值228。该数值指示与块相关联的特征向量相对于分配给从中导出块的图像的标签的预测能力。换句话说,该数值表示特定特征向量的预测能力,即“预测值/能力”,用于分配给块的标签的出现/观察。由于特征向量的特征已经完全地或至少部分地从相应块中包含的图像信息中提取,所以特征向量表示该块中描绘的组织区域的光学特性。因此,特征向量可视为电子组织签名。Next, in step 108, a multi-instance learning (MIL) program 226 is trained based on all blocks and corresponding feature vectors from all images received for all patients in the group. Thus, the MIL program processes each block set as a bundle of blocks with the same label. Training includes analyzing the feature vectors 220 of the blocks in the training dataset and calculating a value 228 for each block. This value indicates the predictive power of the feature vector associated with the block relative to the label assigned to the image from which the block is derived. In other words, this value represents the predictive power of a particular feature vector, i.e., the "predictive value/power," for the occurrence/observation of the label assigned to the block. Since the features of the feature vector have been extracted entirely or at least partially from the image information contained in the corresponding block, the feature vector represents the optical properties of the tissue region depicted in that block. Therefore, the feature vector can be considered an electronic tissue signature.
例如,可以训练MIL程序以预测特定组织区域的一个或多个可能的标签和/或可以训练MIL程序以回归标签如果需要浮点标签预测。在某些情况下,训练额外的注意力MLL以了解哪些特征向量与预测标签最相关。在某些情况下,注意力MLL计算出的权重乘以每个载玻片的特征向量值。作为乘法的结果,针对每个块及其特征向量获得具有加权特征值的特征向量,并在训练时用作MIL程序的输入。在其他实施例中,由注意力MLL计算出的权重乘以由MIL针对每个块的特征向量计算出的数值。这会创建一个加权数值,用作特定块及其相对于标签的特征值的预测能力的指标。加权数值可与训练时的基本事实进行比较,以评估经训练的MIL程序的准确性。在某些情况下,平均、最小、最大最小-最大池化(或它们的组合)可应用于由MIL程序在其置换不变变换运算的训练期间作为块特定结果获得的特征向量。For example, a MIL program can be trained to predict one or more possible labels for a specific tissue region and/or a MIL program can be trained to regress labels if floating-point label prediction is required. In some cases, an additional attention MIL is trained to understand which feature vectors are most relevant to the predicted labels. In other cases, weights computed by the attention MIL are multiplied by the feature vector values for each slide. As a result of the multiplication, a feature vector with weighted feature values is obtained for each block and its feature vector, and is used as input to the MIL program during training. In other embodiments, weights computed by the attention MIL are multiplied by values computed by the MIL for each block's feature vector. This creates a weighted value that serves as an indicator of the predictive power of a particular block and its feature values relative to the label. The weighted value can be compared to the underlying facts during training to evaluate the accuracy of the trained MIL program. In some cases, averaging, min, max-min-max pooling (or combinations thereof) can be applied to the feature vectors obtained by the MIL program as block-specific results during training of its permutation-invariant transformation operations.
接下来在步骤110中,图像分析系统经由由图像分析软件生成的GUI232输出图像块报告库206。GUI示例包括图3中描绘的报告图像块库。报告库包括块的子集,由此根据它们的相应计算出的数值对子集进行排序。此外或替代性地,报告图像块图库包括与相应块相关联的数值的图形表示。Next, in step 110, the image analysis system outputs an image patch report library 206 via a GUI 232 generated by the image analysis software. An example GUI includes the report image patch library depicted in Figure 3. The report library comprises subsets of patches, which are thereby sorted according to their corresponding calculated numerical values. Additionally or alternatively, the report image patch library includes graphical representations of the numerical values associated with the respective patches.
最后,作为训练阶段的结果,获得了经训练的MIL程序。经训练的MIL程序可应用于从其他患者队列导出的图像块。Finally, as a result of the training phase, a trained MIL program was obtained. This trained MIL program can be applied to image patches exported from other patient cohorts.
出于测试目的,还可以将可用数据集拆分为子集(包括例例如约75%的图像)用作训练数据集;以及另一子集(包括例如约25%的图像)用作测试数据集。据观察,经训练的MIL程序对相关使用领域(FOVs)达到很高的预测值。这些包括组织模式,该组织模式直到现在还没有被认为对pCR的预测有影响。For testing purposes, the available dataset can be split into a subset (including, for example, approximately 75% of the images) for use as the training dataset, and another subset (including, for example, approximately 25% of the images) for use as the test dataset. The trained MIL program has been observed to achieve high prediction values for relevant fields of use (FOVs). These include organizational patterns, which have not yet been considered to affect pCR predictions.
因此,本发明的实施例可以允许使用在药物开发过程中可用的大量数据,该数据来自组织学和临床成像、来自基因组学和测序、来自真实世界的数据和来自诊断方法。该方法可以允许提取新颖的见解和开发新技术。Therefore, embodiments of the present invention can allow the use of a wealth of data available in the drug development process, derived from histological and clinical imaging, genomics and sequencing, real-world data, and diagnostic methods. This method can allow for the extraction of novel insights and the development of new technologies.
在病理学和组织学分析的背景下,手动识别预测性潜在组织纹理或组织相关签名的任务可令人生畏,由于多通道、多染色、多模态、高倍率图像中可用的信息剪切量,每个图像都有数十亿像素。因此,这种探索通常基于对人类生成假设的探索,并且因此仅限于关于肿瘤和生物学机制的预先存在的知识的边界,以及手动查看大量高倍率组织学图像的复杂性和劳动要求。本发明的实施例可以允许揭示显微病理组织学图像中的隐藏信息,使得机器学习逻辑和人类都可以解释识别为具有高预测能力的特征。In the context of pathological and histological analysis, the task of manually identifying predictive latent tissue textures or tissue-related signatures can be daunting due to the shearing amount of information available in multi-channel, multi-staining, multi-modal, high-magnification images, each containing billions of pixels. Therefore, such exploration is often based on the exploration of human-generated hypotheses and is thus limited to the boundaries of pre-existing knowledge about tumors and biological mechanisms, as well as the complexity and labor demands of manually examining large numbers of high-magnification histological images. Embodiments of the present invention can allow the revelation of hidden information in microscopic pathological histological images, enabling both machine learning logic and humans to interpret and identify them as highly predictive features.
根据实施例,经训练的MIL可以用于对患者组进行分层。这意味着将患者按给定治疗之外的因素进行划分。可以根据在训练MIL或注意力MLL时不用作标签的患者相关属性执行分层。例如,这种患者相关属性可以是年龄、性别、其他人口统计因素或特定的遗传或生理性状。GUI使用户能够基于未用作标签的所述患者相关属性中的任何一个选择其组织图像用于训练MIL的患者的亚群,并选择性地在该亚群上计算经训练的MLL的预测精度.例如,亚群可以由女性患者或60岁以上的患者组成。针对相应亚群选择性获得的准确性,例如女性/男性或60岁以上/60岁以下的患者可能会在某些亚群中揭示经训练的MIL的特定高或低准确性。这可能允许混淆变量(研究人员正在研究的变量以外的变量),从而使研究人员更容易检测和解释变量之间的关系,并识别将从特定药物中受益最大的患者群体。According to an embodiment, the trained MIL can be used to stratify patient groups. This means segmenting patients based on factors other than a given treatment. Stratification can be performed based on patient-related attributes that are not used as labels when training the MIL or attention MIL. For example, such patient-related attributes could be age, sex, other demographic factors, or specific genetic or physiological traits. The GUI allows users to select a subgroup of patients whose tissue images are used to train the MIL based on any of the said patient-related attributes that are not used as labels, and selectively compute the predictive accuracy of the trained MIL on that subgroup. For example, the subgroup could consist of female patients or patients over 60 years of age. The accuracy selectively obtained for a given subgroup, such as female/male or patients over 60/under 60, may reveal specific high or low accuracy of the trained MIL in certain subgroups. This may allow for confounding variables (variables other than those being studied by researchers), making it easier for researchers to detect and interpret relationships between variables and identify patient groups that will benefit most from a particular drug.
图2描绘了根据本发明实施例的图像分析系统200的框图。Figure 2 depicts a block diagram of an image analysis system 200 according to an embodiment of the present invention.
图像分析系统200包括一个或多个处理器202和易失性或非易失性存储介质210。例如,存储介质可以是硬盘驱动器,例如电磁或闪存驱动器。它可以是磁性的、基于半导体的或光学的数据存储。存储介质可以是易失性介质,例如主存储器,仅临时包含数据。Image analysis system 200 includes one or more processors 202 and volatile or non-volatile storage media 210. For example, the storage media may be a hard disk drive, such as an electromagnetic or flash drive. It may be magnetic, semiconductor-based, or optical data storage. The storage media may be a volatile medium, such as main memory, that only temporarily contains data.
存储介质包括来自具有已知端点的患者的组织样品的多个用标签标记的数字图像212。The storage medium includes multiple labeled digital images 212 of tissue samples from patients with known endpoints.
图像分析系统包括拆分模块214,该模块配置为将图像212中的每一个拆分成多个块。块分组到包216中,由此通常同一包中的所有块从自同一患者导出。包的标签是患者的已知终点,包的所有块都分配了包的标签。The image analysis system includes a splitting module 214 configured to split each image 212 into multiple blocks. The blocks are grouped into packages 216, whereby all blocks within the same package are typically derived from the same patient. The package label is a known endpoint for the patient, and all blocks within the package are assigned a package label.
特征提取模块218被配置为从块216中的每一个中提取多个图像特征。在一些实施例中,特征提取模块218可以是经训练的MLL或经训练的MLL的编码部分。提取的特征作为特征向量220与从中导出特征向量的块相关联存储在存储介质210中。任选地,特征向量可以用从其他来源导出的患者特征来丰富,例如基因组数据,例如微阵列数据。Feature extraction module 218 is configured to extract multiple image features from each of blocks 216. In some embodiments, feature extraction module 218 may be a trained MLL or the encoding portion of a trained MLL. The extracted features are stored as feature vectors 220 associated with the blocks from which the feature vectors are derived in storage medium 210. Optionally, the feature vectors may be enriched with patient features derived from other sources, such as genomic data or microarray data.
任选地,图像分析系统可以包括采样模块215,该模块适于选择用于训练的图像样品(子集)并在剩余的图像块上测试经训练的MIL。采样模块可以在执行采样之前首先基于它们的特征向量对块执行聚集。Optionally, the image analysis system may include a sampling module 215 adapted to select image samples (subsets) for training and test the trained MIL on the remaining image patches. The sampling module may first perform clustering on the patches based on their feature vectors before performing sampling.
任选地,图像分析系统可以包括注意力MLL程序222,该程序配置为计算特征向量中的每一个和相应块的权重。权重可以与特征向量一起用作训练MIL程序226时的输入或用于对作为MIL程序训练结果的MIL针对块中的每一个返回的数值进行加权。Optionally, the image analysis system may include an attention MLL procedure 222 configured to compute weights for each feature vector and the corresponding block. The weights may be used, along with the feature vectors, as input to the MIL procedure 226 during training, or to weight the values returned by the MIL for each block as a result of the MIL procedure training.
图像分析系统包括多实例学习程序(MIL程序226)。在训练期间,MLL程序226接收特征向量220(或由注意力MLL 222生成的加权特征向量224)以及分配给相应块的标签。作为训练的结果,提供了经训练的MIL程序226。此外,针对块中的每一个,计算数值228,该数值指示块的预测能力和本文描绘的针对分配给块的标签的组织模式。这些数值也可以称为“数值块相关度得分”。The image analysis system includes a multi-instance learning procedure (MIL procedure 226). During training, MIL procedure 226 receives feature vectors 220 (or weighted feature vectors 224 generated by attention MIL 222) and labels assigned to the corresponding blocks. As a result of training, the trained MIL procedure 226 is provided. Furthermore, for each block, a numerical value 228 is calculated, which indicates the predictive power of the block and the organization pattern of the labels assigned to the block as described in this paper. These values may also be referred to as "numerical block relevance scores".
图像分析系统进一步包括模块230,该模块配置为生成显示在图像分析系统的屏幕204上的GUI 232。The image analysis system further includes module 230, which is configured to generate a GUI 232 displayed on screen 204 of the image analysis system.
GUI包括报告块库206,该报告块库包括至少一些块和针对这些块计算出的数值228。数值228可以明确显示,例如作为相应块上的覆盖层,和/或隐含地,例如以块的排序顺序的形式根据它们的相应数值228进行排序。当用户选择块中的一个时,显示从中导出块的图像的整个载玻片热图。在其他实施例中,除了默认的报告块库206之外,还可以显示热图。The GUI includes a report block library 206, which comprises at least some blocks and numerical values 228 calculated for these blocks. The numerical values 228 may be explicitly displayed, for example as overlays on the corresponding blocks, and/or implicitly, for example, sorted according to their respective numerical values 228 in the order of the blocks' sorting. When the user selects one of the blocks, a thermal image of the entire slide from which the image of the block is derived is displayed. In other embodiments, a thermal image may be displayed in addition to the default report block library 206.
程序模块214、215、218、222、226、230中的每一个都可以实现为大型MIL训练框架软件应用程序的子模块。替代性地,一个或多个模块可以分别代表与图像分析系统的其他程序和模块可互操作的独立软件应用程序。每个模块和程序可以是例如用Java、Python、C#或任何其他合适的编程语言编写的软件。Each of program modules 214, 215, 218, 222, 226, and 230 can be implemented as a submodule of a large MIL training framework software application. Alternatively, one or more modules can each represent an independent software application interoperable with other programs and modules of the image analysis system. Each module and program can be software written, for example, in Java, Python, C#, or any other suitable programming language.
图3描绘了根据本发明实施例的具有报告图像块库的GUI 300。报告库(行标签302、304、306和308下方的块矩阵)允许用户探索由MIL程序识别的组织模式,以对特定标签具有高预测能力。库包括相对于目标特定标签具有最高数值的块中的一个,例如由MIL计算的“对药物D治疗有反应=真”。块基于从中导出块的组织载玻片图像进行分组,并根据它们的相应数值在它们的组内分类,该数值指示块相对于分配给用于训练MIL的图像的特定标签的预测能力。此外,图库可包括针对图库中的块中的每一个,在训练之后可能已经自动确定的整体预测准确性。此外,或替代性地,报告库可包括分配给相应图像的标签和针对该标签获得的每个包的预测准确性。例如,“基本事实=0”可以代表标签“患者对药物D有应答”,“基本事实=1”可以代表标签“患者对药物D没有应答”。在使用注意力MLL计算权重的情况下,排序也可以基于如本文针对本发明的实施例所描述的针对由注意力MLL生成的块的权重和由MIL计算出的数值的组合(例如乘积)的每个块计算的组合得分值。由MIL计算出的特定图像的所有块的最高数值显示为从所述图像导出的块组顶部的“预测值”。Figure 3 depicts a GUI 300 with a report image block library according to an embodiment of the present invention. The report library (the block matrix below row labels 302, 304, 306, and 308) allows the user to explore tissue patterns identified by the MIL program for high predictive power for specific labels. The library includes one of the blocks with the highest value relative to a target specific label, such as "responsive to drug D treatment = true" calculated by the MIL. Blocks are grouped based on tissue slide images from which the blocks are derived and categorized within their groups according to their corresponding values, which indicate the predictive power of the block relative to a specific label assigned to the image used to train the MIL. Furthermore, the library may include the overall predictive accuracy for each of the blocks in the library, which may have been automatically determined after training. Additionally, or alternatively, the report library may include the label assigned to the corresponding image and the predictive accuracy for each packet obtained for that label. For example, "baseline fact = 0" could represent the label "patient responds to drug D," and "baseline fact = 1" could represent the label "patient does not respond to drug D." When using Attention MLL to calculate weights, ranking can also be based on a combined score value calculated for each block, as described herein with respect to embodiments of the invention, for a combination (e.g., a product) of the weights of the blocks generated by Attention MLL and the values calculated by MIL. The highest value of all blocks for a particular image calculated by MIL is displayed as the "predicted value" at the top of the block group derived from said image.
在所描绘的库中,块行302显示第一患者的六个块。所述块中的第一个已分配最高数值(预后值)指示特定组织载玻片/整个载玻片图像相对于标签的预测能力。每个载玻片组的第一块可以额外地或替代性地分配从特定组织载玻片图像导出的所有块的最高组合值(从由MIL所提供的数值和由注意力MLL计算出的权重导出)。In the depicted library, block row 302 shows six blocks for the first patient. The first assigned highest value (prognostic value) in the block indicates the predictive power of the specific tissue slide/entire slide image relative to the label. The first block of each slide group may additionally or alternatively be assigned the highest combined value of all blocks derived from the specific tissue slide image (derived from the values provided by MIL and the weights calculated by Attention MLL).
如图3中所示的GUI描绘,最高数值可以显示在每位患者的最高评分块顶部。As shown in Figure 3, the highest value can be displayed at the top of the highest score block for each patient.
仅包括具有最高预测能力的块的子集的报告块库可能是有利的,因为病理学家不需要检查整个载玻片。相反,病理学家的注意力自动指向每个整个载玻片图像的少量子区域(块),该图像的组织模式已识别相对于目标标签具有最高预测能力。A report block library that includes only a subset of the blocks with the highest predictive power may be advantageous, as pathologists do not need to examine the entire slide. Instead, the pathologist's attention is automatically directed to a small sub-region (block) of each full slide image whose tissue pattern has been identified as having the highest predictive power relative to the target label.
根据图3中描绘的实施例,报告图像块库显示从H&E染色图像导出的图像块。报告图像块库的组织方式如下:According to the embodiment depicted in Figure 3, the report image block library displays image blocks derived from H&E staining images. The report image block library is organized as follows:
行302包括分配了由MIL程序计算出的最高数值(指示预测能力,即预后值)的六个块,这些块在从第一患者的特定整个载玻片图像312导出的所有块内。根据其他实施例,基于与由MIL计算出的数值相同的得分值或者是由MIL计算出的数值的导数值来执行排序。例如,导数值可以是作为由MIL针对块计算出的数值的和由注意力MLL针对所述块计算出的权重的组合来计算的组合得分。例如,该组合可以是数值和权重的乘积。根据另一些实施例,块仅根据注意力-MLL计算出的权重进行排序,并且MIL计算出的数值以不同的方式显示给用户,例如,以覆盖相应块的数字或呈现在空间上接近相应块的数字的形式。Row 302 includes six blocks assigned the highest values (indicating predictive ability, i.e., prognostic values) calculated by the MIL procedure, these blocks being among all blocks derived from a specific whole slide image 312 of the first patient. According to other embodiments, sorting is performed based on either the same score value calculated by the MIL or a derivative value of the MIL-calculated value. For example, the derivative value could be a combined score calculated as a combination of the value calculated by the MIL for the block and a weight calculated by the Attention-MLL for said block. For example, this combination could be the product of the value and the weight. According to other embodiments, blocks are sorted only according to the weights calculated by the Attention-MLL, and the MIL-calculated values are displayed to the user in different ways, e.g., as numbers covering the corresponding blocks or presented in a form spatially close to the corresponding blocks.
用于生成其中一些显示在行312中的块的第一患者的组织样品的相应整个载玻片图像312在空间上接近于该高度相关的块的所选集312。The corresponding whole slide image 312 of the first patient’s tissue sample, some of which are shown in the blocks in row 312, is spatially close to the selected set 312 of the highly correlated blocks.
此外,显示的任选的相关度热图322突出显示了所有整个载玻片图像区域,由MIL计算出的数值类似于图像312的一个块的数值,其中最高数值指示预测能力已计算。在这种情况下,自动识别和选择计算出的最高数值的块之一(例如,在行312中第一位置的块)并用作计算相关度热图322的基础。根据替代性实施方式,相关度热图322不表示块的数值与针对图像的所有块计算出的最高数值的相似性,而是表示块与针对图像的所有块计算出的最高组合得分的相似度。组合得分可以是由注意力MLL针对块计算出的权重的与由MIL计算出的指示块相对于图像标签的预测能力的数值的组合,例如乘积。根据更进一步的实施例,相关度热图322表示由注意力MLL计算出的块的权重与由注意力MLL针对图像的所有块计算出的最高权重的相似度。Furthermore, the optional relevance heatmap 322 displayed highlights all entire slide image regions where the values calculated by the MIL are similar to the values of a block in image 312, with the highest value indicating the predictive power calculated. In this case, one of the blocks with the highest calculated values (e.g., the block at the first position in row 312) is automatically identified and selected as the basis for calculating the relevance heatmap 322. According to an alternative embodiment, the relevance heatmap 322 does not represent the similarity of a block's value to the highest value calculated for all blocks in the image, but rather the similarity of the block to the highest combined score calculated for all blocks in the image. The combined score can be a combination, for example, a product of the weights calculated by the attention MLL for the block and a value calculated by the MIL indicating the predictive power of the block relative to the image label. According to a further embodiment, the relevance heatmap 322 represents the similarity of the weights of the block calculated by the attention MLL to the highest weights calculated by the attention MLL for all blocks in the image.
列304包括分配了由MIL程序计算出的最高数值的六个块,这些块从第二患者的特定整个载玻片图像314导出。相应的整个载玻片图像314在空间上接近所选的高度相关的块集。此外,显示的相关度热图324突出显示了所有整个载玻片图像区域,该图像区域由MIL计算出的相应数值与由MIL计算出的最高数值的整个载玻片图像314的块高度相似。Column 304 comprises six blocks assigned the highest values calculated by the MIL program, derived from a specific whole slide image 314 of the second patient. The corresponding whole slide image 314 spatially approximates the selected set of highly correlated blocks. Furthermore, the displayed correlation heatmap 324 highlights all whole slide image regions whose corresponding values calculated by the MIL are highly similar to the blocks in the whole slide image 314 with the highest values calculated by the MIL.
列306包括分配了由MIL程序计算出的最高数值的六个块,这些块从第三患者的特定整个载玻片图像316导出。相应的整个载玻片图像316在空间上接近所选的高度相关的块集。此外,显示的相关度热图326突出显示了所有整个载玻片图像区域,该图像区域由MIL计算出的相应数值与由MIL计算出的最高数值的整个载玻片图像316的块高度相似。Column 306 comprises six blocks assigned the highest values calculated by the MIL program, derived from a specific whole slide image 316 of the third patient. The corresponding whole slide image 316 spatially approximates the selected set of highly correlated blocks. Furthermore, the displayed correlation heatmap 326 highlights all whole slide image regions whose corresponding values calculated by the MIL are highly similar to the blocks in the whole slide image 316 with the highest values calculated by the MIL.
列308包括分配了由MIL程序计算出的最高数值的六个块,这些块从患者的特定整个载玻片图像318导出。相应的整个载玻片图像318在空间上接近所选的高度相关的块集。此外,显示的相关度热图328突出显示了所有整个载玻片图像区域,该图像区域由MIL计算出的相应数值与由MIL计算出的最高数值的整个载玻片图像318的块高度相似。Column 308 comprises six blocks assigned the highest values calculated by the MIL program, derived from a specific whole slide image 318 of the patient. The corresponding whole slide image 318 is spatially close to the selected set of highly correlated blocks. Furthermore, the displayed correlation heatmap 328 highlights all whole slide image regions whose corresponding values calculated by the MIL are highly similar to the blocks in the whole slide image 318 with the highest values calculated by the MIL.
根据实施例,在报告块库中呈现的相关度热图指示预测能力、或基于注意力的权重、或它们的组合。在所描绘的示例中,热图中的亮像素描绘了图像中块具有高预测值、高基于注意力的权重或它们的组合的区域。根据实施例,相关度热图的计算包括确定块的得分(例如,数值、权重或组合值)是否高于图像的最高得分块的得分的最小百分比值。如果是,则相关度热图中的相应块由第一颜色或“亮”强度值表示,例如“255”。如果不是,则相关度热图中的各个块由第二种颜色或“暗”强度值表示,例如“0”。According to embodiments, the relevance heatmap presented in the report block library indicates predictive power, attention-based weights, or a combination thereof. In the depicted example, bright pixels in the heatmap depict regions in the image where blocks have high predictive values, high attention-based weights, or a combination thereof. According to embodiments, the calculation of the relevance heatmap includes determining whether a block's score (e.g., a numerical, weighted, or combined value) is higher than a minimum percentage value of the score of the highest-scoring block in the image. If yes, the corresponding block in the relevance heatmap is represented by a first color or "bright" intensity value, such as "255". If not, the individual blocks in the relevance heatmap are represented by a second color or "dark" intensity value, such as "0".
用户可以选择报告块库中的每个块以启动相似性搜索(例如,通过双击块或通过单击选择块,然后选择GUI元素“搜索”),然后将显示一个相似性搜索块库,例如如图4所示。Users can select each block in the report block library to start a similarity search (e.g., by double-clicking the block or by clicking to select the block and then selecting the GUI element "Search"), and a similarity search block library will be displayed, as shown in Figure 4.
可选GUI元素集310中的“黑名单”和“重新训练”元素使用户能够定义块的黑名单并基于除黑名单中的块和与黑名单中的块高度相似的块之外的所有块重新训练MIL程序。例如,黑名单可以包括手动选择的具有特别低数值(预测值)的块集,例如因为它们包含伪影,或具有特别高的数值(排除具有非常高预测能力的块可增加MIL识别额外的、迄今为止未知的组织模式的能力,这些模式相对于目标标签也具有预测能力)。图像分析系统可以配置为响应于用户将特定块添加到黑名单,自动识别其特征向量与添加到黑名单的块的特征向量的相似度超过最小相似度阈值的所有块。识别出的块也自动添加到黑名单中。当用户选择重新训练-GUI元素时,除了黑名单中的块外,MIL将基于训练数据集的所有块重新训练。The "Blacklist" and "Retrain" elements in the optional GUI element set 310 allow users to define a blacklist of blocks and retrain the MIL program based on all blocks except those in the blacklist and those highly similar to them. For example, the blacklist can include a manually selected set of blocks with particularly low numerical values (predicted values), such as because they contain artifacts, or with particularly high numerical values (excluding blocks with very high predictive power increases MIL's ability to identify additional, previously unknown organizational patterns that are also predictive relative to the target label). The image analysis system can be configured to automatically identify all blocks whose feature vectors are more than a minimum similarity threshold to the feature vectors of blocks added to the blacklist in response to a user adding a specific block to the blacklist. The identified blocks are also automatically added to the blacklist. When the user selects the Retrain GUI element, MIL will be retrained based on all blocks in the training dataset, except those in the blacklist.
图4描绘了根据本发明的实施例的具有相似性搜索图像块库的GUI400。相似性搜索由基于用户的对报告库中430块的选择触发。Figure 4 depicts a GUI 400 with a similarity search image patch library according to an embodiment of the present invention. The similarity search is triggered by the user's selection of 430 patches in the report library.
该搜索在从整个载玻片图像412-418中的每一个生成的块内识别例如基于比较特征向量的相似性的六个最相似的块的子集。在相似性搜索中识别的块按每个整个载玻片图像或每个患者分组,并根据它们与选择触发相似性搜索的块430(“查询块”)的相似性以降序进行排序。The search identifies a subset of the six most similar blocks within each block generated from the entire slide images 412-418, for example, based on the similarity of the compared feature vectors. The blocks identified in the similarity search are grouped by each entire slide image or by each patient and sorted in descending order according to their similarity to the block 430 (“query block”) that triggered the similarity search.
整个载玻片图像412-418和相似性热图422-428指示其特征向量(以及因此所描绘的组织模式)与所选块的特征向量最相似的块的位置。The entire slide images 412-418 and similarity heatmaps 422-428 indicate the location of the block whose feature vector (and thus the depicted tissue pattern) is most similar to the feature vector of the selected block.
任选地,相似度搜索块库还包括以下一项或多项数据:Optionally, the similarity search block library also includes one or more of the following data:
-标签分配给从中导出描绘的块的图像;图4中描绘的一个标签是“基本事实:0”;- Labels are assigned to the images from which the depicted blocks are derived; one label depicted in Figure 4 is "Fundamental Fact: 0";
-由MIL程序计算出的每个包(图像)相对于包的标签的预测准确性;- The prediction accuracy of each packet (image) relative to the packet's label, calculated by the MIL program;
-整张载玻片图像中类似块的计数和/或相似块与相异块比较的百分比(分数)(例如,通过阈值处理)- The count of similar blocks in the entire slide image and/or the percentage (fraction) of similar blocks compared to dissimilar blocks (e.g., by thresholding).
-整个载玻片图像中所有块的相似性值的平均值、中值或直方图。- The average, median, or histogram of similarity values for all blocks in the entire slide image.
图5根据本发明的实施例描绘了特征提取MLL程序的网络架构600,该网络架构支持用于特征向量生成的监督式学习方法。由一系列自动编码器604组成的深度神经网络基于以分层方式从图像块中提取的多个特征进行训练。经训练的网络能够稍后执行分类任务,例如基于从图像块中提取的光学特征,将块中描绘的组织分类为“基质组织”、“背景载玻片区域”、“肿瘤细胞”、“转移组织”等类别之一。网络架构包括瓶颈层606,该瓶颈层具有比输入层603少得多的神经元并且随后可以是进一步的隐藏层和分类层。根据一个示例,瓶颈层包括输入层神经元数量的大约1.5%。输入层和瓶颈层之间可能有数百甚至数千个隐藏层,并且瓶颈层提取的特征可以称为“深度瓶颈特征”(DBNF)。Figure 5 depicts a network architecture 600 for a feature extraction MLL procedure according to an embodiment of the present invention, which supports a supervised learning method for feature vector generation. A deep neural network consisting of a series of autoencoders 604 is trained based on multiple features extracted hierarchically from image patches. The trained network is later able to perform classification tasks, such as classifying tissue depicted in an image patch into one of the categories of “stromal tissue,” “background slide region,” “tumor cells,” “metastatic tissue,” etc., based on optical features extracted from the image patch. The network architecture includes a bottleneck layer 606, which has far fewer neurons than the input layer 603 and may be followed by further hidden layers and classification layers. According to one example, the bottleneck layer comprises approximately 1.5% of the number of neurons in the input layer. There may be hundreds or even thousands of hidden layers between the input layer and the bottleneck layer, and the features extracted by the bottleneck layer may be referred to as “deep bottleneck features” (DBNF).
图6描绘了一种可能的系统架构,用于组合MIL程序和注意力MLL。根据所描绘的实施例,MIL程序的训练包括基于所有接收到的图像的所有块的特征向量220、708-714和标签216、702-706训练注意力机器学习逻辑程序222以计算用于块中的每一个的权重。由注意力MLL计算出的权重指示特征向量和相应块相对于由块的标签指示的患者相关属性值的预测能力。然后,图6中描绘的机器学习系统针对从接收到的训练图像获得的块中的每一个计算组合预测值。组合预测值是由MIL针对块计算出的数值的和由注意力MLL针对块计算出的权重的函数。组合数值可以是例如MIL的数值的和注意力MLL的权重的乘积或平均值。组合数值指示特征向量和相应块相对于由块的标签指示的患者相关属性值的预测能力。然后,计算出的损失值指示针对特定标签获得的组合预测值与分配给块的实际标签的差异。然后,基于计算出的损失值,使用反向传播迭代地适应MIL程序的模型。Figure 6 depicts a possible system architecture for combining the MIL procedure and the attention MLL. According to the depicted embodiment, training the MIL procedure involves training an attention machine learning logic program 222 based on feature vectors 220, 708-714 and labels 216, 702-706 for all blocks of all received images to compute weights for each block. The weights computed by the attention MLL indicate the predictive power of the feature vectors and the corresponding blocks relative to the patient-related attribute values indicated by the block's label. The machine learning system depicted in Figure 6 then computes a combined prediction value for each block obtained from the received training images. The combined prediction value is a function of the numerical value computed by the MIL for the block and the weights computed by the attention MLL for the block. The combined numerical value can be, for example, the product or average of the MIL numerical value and the attention MLL weights. The combined numerical value indicates the predictive power of the feature vectors and the corresponding blocks relative to the patient-related attribute values indicated by the block's label. A calculated loss value then indicates the difference between the combined prediction value obtained for a specific label and the actual label assigned to the block. Based on the calculated loss value, a model of the MIL procedure is then iteratively adapted using backpropagation.
图7描绘了另一可能的系统架构,用于组合MIL程序和注意力MLL。MIL程序的训练包括基于所有接收到的图像的所有块的特征向量220和标签216训练注意力机器学习逻辑程序222-注意力MLL程序,以计算块中的每一个的权重。权重指示特征向量和相应块相对于由块的标签表示的患者相关属性值的预测能力。然后,图7中描绘的机器学习系统,针对块中的每一个,计算加权特征向量,作为由注意力MLL针对该块计算出的权重的和从该块提取的特征向量的函数。加权特征向量输入到MIL中,使得MIL能够使用加权特征向量而不是最初从相应块以及任选的其他数据源提取的特征向量来计算块中的每一个的数值。然后,MIL程序计算损失值,该值指示针对特定标签获得的数值与分配给块的实际标签之间的差异。在训练期间,MIL使用基于计算出的损失值的反向传播迭代地适应其模型。Figure 7 illustrates another possible system architecture for combining the MIL procedure and the attention MLL. Training the MIL procedure involves training an attention machine learning logic procedure 222—an attention MLL procedure—based on feature vectors 220 and labels 216 for all blocks of all received images to compute weights for each block. Weights indicate the predictive power of the feature vectors and the corresponding blocks relative to patient-related attribute values represented by the block's label. The machine learning system depicted in Figure 7 then computes a weighted feature vector for each block as a function of the weights computed by the attention MLL for that block and the feature vector extracted from that block. This weighted feature vector is input into the MIL, enabling it to compute values for each block using the weighted feature vectors instead of the feature vectors initially extracted from the corresponding blocks and optional other data sources. The MIL procedure then computes a loss value indicating the difference between the numerical value obtained for a particular label and the actual label assigned to the block. During training, the MIL iteratively adapts its model using backpropagation based on the computed loss value.
图8示出了2D和3D坐标系中块的空间距离,这些坐标系用于基于从块的空间接近度自动导出的相似性标签自动将相似性标签分配给块对。因此,提供了用于训练特征提取MLL的训练数据集,该数据集不需要领域专家手动注释图像或块。Figure 8 illustrates the spatial distances of blocks in 2D and 3D coordinate systems used to automatically assign similarity labels to block pairs based on similarity labels derived automatically from the spatial proximity of the blocks. Therefore, a training dataset for training the Feature Extraction MLL is provided, which does not require manual annotation of images or blocks by domain experts.
图8A示出了由数字组织样品训练图像800的x轴和y轴定义的2D坐标系中块的空间距离。训练图像800描绘了患者的组织样品。从患者获得组织样品后,将样品置于显微镜载玻片上并用一种或多种组织学相关的染色剂染色,例如H&E和/或各种生物标志物特异性染色剂。训练图像800是从染色的组织样品中获取的,例如使用载玻片扫描显微镜。根据一些实施变型,所接收到的训练图像中的至少一些是从不同患者和/或是从同一患者的不同组织区域(活组织检查)导出的并且因此不能在3D坐标系中彼此对齐。在这种情况下,可以在由如下所述的图像的x和y坐标定义的2D空间内计算块距离。Figure 8A illustrates the spatial distances of blocks in a 2D coordinate system defined by the x and y axes of a digital tissue sample training image 800. The training image 800 depicts a tissue sample from a patient. After obtaining the tissue sample from the patient, the sample is placed on a microscope slide and stained with one or more histology-associated staining agents, such as H&E and/or various biomarker-specific staining agents. The training image 800 is obtained from the stained tissue sample, for example, using a slide scanning microscope. According to some implementation variations, at least some of the received training images are derived from different patients and/or from different tissue regions (biopsy) from the same patient and therefore cannot be aligned with each other in a 3D coordinate system. In this case, the block distances can be calculated in a 2D space defined by the x and y coordinates of the image as described below.
训练图像800拆分成多个块。出于说明目的,图8A中的块尺寸大于通常的块尺寸。The training image 800 is divided into multiple blocks. For illustrative purposes, the block size in Figure 8A is larger than the usual block size.
可以通过以下方法自动用标签标记训练数据集:首先,选择起始块802。然后,确定围绕该起始块的第一圆区域。第一圆的半径也称为第一空间接近度阈值808。第一圆内的所有块,例如块806,被认为是起始块802的“附近”块。此外,还确定了围绕该起始块的第二圆区域。第二圆的半径也称为第二空间接近度阈值810。第二圆之外的所有块,例如块804是相对于起始块802的“远处”块。The training dataset can be automatically labeled using the following method: First, a starting block 802 is selected. Then, a first circular region surrounding this starting block is determined. The radius of the first circle is also called the first spatial proximity threshold 808. All blocks within the first circle, such as block 806, are considered "nearby" blocks of the starting block 802. Furthermore, a second circular region surrounding this starting block is determined. The radius of the second circle is also called the second spatial proximity threshold 810. All blocks outside the second circle, such as block 804, are "far away" blocks relative to the starting block 802.
然后,创建第一块对集,其中第一集的每个块对包括起始块和起始块的“附近”块。例如,该步骤可包括创建与第一圆中包含的附近块一样多的块对。替代性地,该步骤可包括随机选择可用附近块的子集并通过将起始块添加到所选附近块来为所选附近块中的每一个创建块对。Then, a first set of block pairs is created, where each block pair in the first set includes a starting block and a “nearby” block of the starting block. For example, this step may include creating as many block pairs as the number of nearby blocks contained in the first circle. Alternatively, this step may include randomly selecting a subset of available nearby blocks and creating a block pair for each of the selected nearby blocks by adding the starting block to the selected nearby blocks.
创建第二块对集。第二集的每个块对包括起始块和相对于起始块的“远处”块。例如,该步骤可包括创建与在第二圆之外的图像800中包含的远处块一样多的块对。替代性地,该步骤可包括随机选择可用的远处块的子集并通过将起始块添加到所选远处块来为所选远处块中的每一个创建块对。Create a second set of block pairs. Each block pair in the second set includes a starting block and a “distant” block relative to the starting block. For example, this step may include creating as many block pairs as there are distant blocks contained in the image 800 outside the second circle. Alternatively, this step may include randomly selecting a subset of the available distant blocks and creating a block pair for each of the selected distant blocks by adding the starting block to the selected distant blocks.
然后,图像800内的另一块可以用作起始块并且可以类似地执行上述步骤。这意味着使用新的起始块作为中心重新绘制第一圆和第二圆。从而,识别关于新的起始块的附近块和远处块。第一块集补充有基于新的起始块识别的附近块对,而第二块集补充有基于新的起始块识别的远处块对。Then, another block within image 800 can be used as a starting block, and the above steps can be performed similarly. This means redrawing the first and second circles using the new starting block as the center. Thus, nearby and distant blocks are identified with respect to the new starting block. The first block set is supplemented with pairs of nearby blocks identified based on the new starting block, while the second block set is supplemented with pairs of distant blocks identified based on the new starting block.
然后,可以选择图像800内的另一块作为起始块并且可以重复上述步骤,从而进一步用更多的块对补充第一块对集和第二块对集。可以执行新的起始块的选择,直到图像中的所有块都曾经选为起始块或者直到已经选择了预定数量的块作为起始块。Then, another block within image 800 can be selected as the starting block, and the above steps can be repeated to further supplement the first and second block sets with more block pairs. New starting block selections can be performed until all blocks in the image have been selected as starting blocks or until a predetermined number of blocks have been selected as starting blocks.
针对第一集中的块对中的每一个,例如对812,分配“相似”标签。针对第二集中的块对中的每一个,例如对814,分配“相异”标签。For each block pair in the first set, such as 812, assign the label "similar". For each block pair in the second set, such as 814, assign the label "dissimilar".
图8B示出了由数字组织样品图像800的x轴和y轴以及对应于彼此对齐的图像800、832、834的堆叠高度的z轴定义的3D坐标系中的块的空间距离根据由训练图像800、832、834分别描绘的组织块的组织切片的相对位置。训练图像分别描绘从特定患者的单个组织块导出的组织样品。所描绘的组织样品属于一堆多个相邻的组织切片。例如,该堆组织切片可以从FFPET组织块离体制备。将组织块切片,并将切片放置在显微镜载玻片上。然后,对切片进行染色,如参考图8A针对图像800所述。Figure 8B illustrates the spatial distances of blocks in a 3D coordinate system defined by the x and y axes of digital tissue sample image 800 and the z-axis corresponding to the stacking height of images 800, 832, and 834 aligned with each other, according to the relative positions of tissue slices from tissue blocks depicted by training images 800, 832, and 834, respectively. The training images depict tissue samples derived from individual tissue blocks from a specific patient. The depicted tissue samples belong to a stack of multiple adjacent tissue slices. For example, this stack of tissue slices can be prepared ex vivo from an FFPET tissue block. The tissue block is sliced, and the slices are placed on a microscope slide. The slices are then stained, as described with reference to Figure 8A for image 800.
由于该堆内的组织样品从单个组织块导出,因此可以在公共3D坐标系内对齐数字图像800、832、834,由此z轴与组织切片正交。z轴是与组织切片正交的轴。图像在z方向上的距离对应于所述图像所描绘的组织切片的距离。如果一对的两个块从同一图像导出,则在2D空间内计算块对的块距离。此外,可以创建块对,该块对的块从在公共3D坐标系中彼此对齐的不同图像导出。在这种情况下,一对中两个块的距离是使用3D坐标系计算的。Since the tissue samples within this stack are derived from a single tissue block, digital images 800, 832, and 834 can be aligned in a common 3D coordinate system, whereby the z-axis is orthogonal to the tissue slices. The z-axis is an axis orthogonal to the tissue slices. The distance in the z-direction of the images corresponds to the distance of the tissue slices depicted by the images. If two blocks in a pair are derived from the same image, the block distance of the pair is calculated in 2D space. Furthermore, block pairs can be created where the blocks are derived from different images aligned with each other in a common 3D coordinate system. In this case, the distance between the two blocks in the pair is calculated using the 3D coordinate system.
将对齐的数字图像中的每一个拆分成多个块。出于说明目的,图8B中的块尺寸大于通常的块尺寸。Each of the aligned digital images is divided into multiple blocks. For illustrative purposes, the block size in Figure 8B is larger than the usual block size.
可以通过以下方法自动用标签标记训练数据集:首先,选择起始块802。然后,如下所述识别的和用标签标记的包括起始块和附近块的块对以及包括起始块和远处块的块对。The training dataset can be automatically labeled using the following method: First, select the starting block 802. Then, identify and label the block pairs including the starting block and nearby blocks, as well as the block pairs including the starting block and distant blocks, as described below.
确定围绕该起始块的第一3D球体。出于说明目的,仅显示了第一球体的横截面。第一球体的半径也称为第一空间接近度阈值836。第一球体内的所有块,例如图像800中的块806以及图像834中的块840被认为是起始块802的“附近”块。此外,还确定了围绕该起始块的第二球体。第二球体的半径也称为第二空间接近度阈值838。第二球体之外的所有块,例如图像800的块804以及图像834的块842是相对于开始块802的“远处”块。A first 3D sphere is determined surrounding the starting block. For illustrative purposes, only a cross-section of the first sphere is shown. The radius of the first sphere is also referred to as a first spatial proximity threshold 836. All blocks within the first sphere, such as block 806 in image 800 and block 840 in image 834, are considered “nearby” blocks of the starting block 802. Furthermore, a second sphere is determined surrounding the starting block. The radius of the second sphere is also referred to as a second spatial proximity threshold 838. All blocks outside the second sphere, such as block 804 in image 800 and block 842 in image 834, are “far” blocks relative to the starting block 802.
创建第一块对集,其中第一集的每个块对包括起始块和起始块的“附近”块。例如,该步骤可以包括创建与第一球体中包含的附近块一样多的块对。替代性地,该步骤可包括随机选择可用附近块的子集并通过将起始块添加到所选附近块来为所选附近块中的每一个创建块对。Create a first set of block pairs, where each block pair in the first set includes a starting block and a "nearby" block of the starting block. For example, this step may include creating as many block pairs as the number of nearby blocks contained in the first sphere. Alternatively, this step may include randomly selecting a subset of available nearby blocks and creating a block pair for each of the selected nearby blocks by adding the starting block to the selected nearby blocks.
创建第二块对集。第二集的每个块对包括起始块和相对于起始块的“远处”块。例如,该步骤可以包括创建与在第二球体外的图像800、832、834中包含的远处块一样多的块对。替代性地,该步骤可包括随机选择可用的远处块的子集并通过将起始块添加到所选远处块来为所选远处块中的每一个创建块对。Create a second set of block pairs. Each block pair in the second set includes a starting block and a “distant” block relative to the starting block. For example, this step may include creating as many block pairs as the distant blocks contained in images 800, 832, 834 outside the second sphere. Alternatively, this step may include randomly selecting a subset of available distant blocks and creating a block pair for each of the selected distant blocks by adding the starting block to the selected distant blocks.
然后,图像800内或图像832、834内的另一块可以用作起始块并且可以类似地执行上述步骤。这意味着第一球体和第二球体使用新的起始块作为中心重新绘制。从而,识别关于新的起始块的附近块和远处块。第一块集补充有基于新的起始块识别的附近块对,而第二块集补充有基于新的起始块识别的远处块对。Then, another block within image 800 or images 832, 834 can be used as a starting block and the above steps can be performed similarly. This means that the first and second spheres are redrawn using the new starting block as the center. Thus, nearby and distant blocks with respect to the new starting block are identified. The first block set is supplemented with pairs of nearby blocks identified based on the new starting block, while the second block set is supplemented with pairs of distant blocks identified based on the new starting block.
可以重复上述步骤,直到接收到的图像800、832、834中的每一个的每个块都选为起始块(或直到满足另一终止标准),从而用进一步的块对进一步补充第一块对集和第二块对集。The above steps can be repeated until each block of each of the received images 800, 832, 834 is selected as the starting block (or until another termination criterion is met), thereby further supplementing the first block pair set and the second block pair set with further block pairs.
针对第一集中的块对中的每一个,例如对812和813,分配“相似”标签。针对第二集中的块对中的每一个,例如对814和815,分配“相异”标签。For each block pair in the first set, such as 812 and 813, assign the label "similar". For each block pair in the second set, such as 814 and 815, assign the label "dissimilar".
图8A和8B中所示的基于圆和球的距离计算只是用于计算基于距离的相似性标签的示例,在这种情况下,二进制标签应该是“相似”或“相异”。可能会使用其他方法,例如计算2D或3D坐标系中两个块之间的欧几里得距离,并计算与两个块的欧几里德距离呈负相关的数值相似度性。The circle- and sphere-based distance calculations shown in Figures 8A and 8B are merely examples for calculating distance-based similarity labels, in which case the binary labels should be "similar" or "dissimilar". Other methods could be used, such as calculating the Euclidean distance between two blocks in a 2D or 3D coordinate system and calculating the numerical similarity that is negatively correlated with the Euclidean distance between the two blocks.
由于一毫米组织对应的像素数量取决于各种因素,例如图像捕获设备的放大倍数和数字图像的分辨率,本文将针对所描绘的真实物理对象指出所有距离阈值,即组织样品或组织样品覆盖的载玻片。Since the number of pixels corresponding to one millimeter of tissue depends on various factors, such as the magnification of the image capture device and the resolution of the digital image, this paper will specify all distance thresholds for the real physical object being depicted, namely the tissue sample or the slide covering the tissue sample.
图9描绘了根据本发明的实施例训练的孪生神经网络的架构,用于提供能够从适合执行基于特征向量的相似性搜索和/或从图像块中提取具有生物医学意义的特征向量的子网络基于特征向量的块聚集。孪生神经网络900基于自动用标签标记的训练数据集进行训练,该数据集包括具有基于接近度的相似性标签的块对,例如参照图8A和/或8B所描述的自动创建。Figure 9 illustrates the architecture of a Siamese neural network trained according to an embodiment of the invention, for providing feature vector-based block aggregation from a subnetwork suitable for performing feature vector-based similarity search and/or extracting biomedically significant feature vectors from image patches. The Siamese neural network 900 is trained on an automatically labeled training dataset comprising block pairs with proximity-based similarity labels, such as those automatically created with reference to Figures 8A and/or 8B.
孪生神经网络900由在其输出层924处连接的两个相同子网络902、903组成。每个网络包括输入层905、915,适于接收作为输入的单个数字图像(例如块)954、914。每个子网络包括多个隐藏层906、916、908、918。通过两个子网络中的相应一个从两个输入图像之一中提取一维特征向量910、920。因此,每个网络的最后隐藏层908、918适于计算特征向量并将特征向量提供给输出层924。输入图像的处理是严格分开的。这意味着,该子网络仅处理输入图像954,并且子网络仅处理输入图像914。当输出层比较两个向量以确定向量相似性时,两个输入图像中传达的信息唯一结合的点是在输出层中,并且从而确定两个输入图像中描绘的组织模式的相似性。The Siamese neural network 900 consists of two identical subnetworks 902 and 903 connected at their output layer 924. Each network includes input layers 905 and 915, adapted to receive a single digital image (e.g., a block) 954 and 914 as input. Each subnetwork includes multiple hidden layers 906, 916, 908, and 918. A one-dimensional feature vector 910 and 920 is extracted from one of the two input images through a corresponding subnetwork. Therefore, the final hidden layers 908 and 918 of each network are adapted to compute the feature vector and provide it to the output layer 924. The processing of the input images is strictly separate. This means that the subnetwork processes only the input image 954, and the subnetwork processes only the input image 914. When the output layer compares the two vectors to determine vector similarity, the only point where the information conveyed in the two input images is combined is in the output layer, thus determining the similarity of the organizational patterns depicted in the two input images.
根据实施例,每个子网络902、903基于修改的resnet-50架构(He等人,DeepResidual Learning for Image Recognition,2015,CVPR’15)。根据实施例,resnet-50预训练子网络902、903基于ImageNet预训练。最后一层(通常输出1,000个特征)用全连接层408、418替换,其尺寸具有特征向量的期望尺寸,例如尺寸128。例如,每个子网络的最后一层908、918可以被配置为从倒数第二层提取特征,由此倒数第二层可以提供比最后一层908、418多得多的特征数量(例如2048)。根据实施例,优化器,例如使用PyTorch中的默认参数(学习率为0.001,beta为0.9,0.999)的Adam优化器,并且在训练期间使用了256的批量尺寸。针对数据增强,随机水平和垂直翻转和/或高达20度的随机旋转,和/或亮度、对比度饱和度和/或色调值为0.075的颜色抖动增强可以应用于块以增加训练数据集。According to an embodiment, each subnetwork 902, 903 is based on a modified ResNet-50 architecture (He et al., Deep Residual Learning for Image Recognition, 2015, CVPR’15). According to an embodiment, the ResNet-50 pre-trained subnetworks 902, 903 are pre-trained on ImageNet. The last layer (typically outputting 1,000 features) is replaced with fully connected layers 408, 418, with dimensions equal to the desired size of the feature vectors, such as size 128. For example, the last layer 908, 918 of each subnetwork can be configured to extract features from the penultimate layer, whereby the penultimate layer can provide a significantly larger number of features than the last layer 908, 418 (e.g., 2048). According to an embodiment, the optimizer, for example, uses the Adam optimizer with default parameters in PyTorch (learning rate 0.001, beta 0.9, 0.999), and a batch size of 256 is used during training. For data augmentation, random horizontal and vertical flips and/or random rotations up to 20 degrees, and/or color jitter enhancements with brightness, contrast, saturation, and/or hue values of 0.075 can be applied to blocks to increase the training dataset.
当孪生神经网络基于自动用标签标记的图像对进行训练时,学习过程的目标是相似图像应该具有彼此相似的输出(特征向量),而相异的图像应该具有彼此相异的输出。这可以通过最小化损失函数来实现,例如衡量两个子网络提取的特征向量之间差异的函数。When a Siamese neural network is trained on automatically labeled pairs of images, the goal of the learning process is to ensure that similar images have similar outputs (feature vectors), while dissimilar images have dissimilar outputs. This can be achieved by minimizing a loss function, such as a function that measures the difference between the feature vectors extracted by the two sub-networks.
根据实施例,使用损失函数基于块对训练孪生神经神经元网络,使得由两个子网络针对该对的两个块提取的特征向量的相似性分别与该对的两个块中描绘的组织模式的相似性相关。According to an embodiment, a loss function is used to train a twin neural network based on block pairs, such that the similarity of feature vectors extracted by the two subnetworks for the two blocks of the pair is correlated with the similarity of the organizational patterns depicted in the two blocks of the pair.
例如,孪生神经网络可以是,如Bromley等人在“使用‘孪生神经’时间延迟神经网络的签名验证,1994,NIPS’1994”中所述的孪生神经网络。孪生神经网络的每个子网络适于从作为输入所提供的两个图像块中的相应一个提取多维特征向量。该网络基于多个已自动标注有基于邻近度的组织模式相似性标签的块对进行训练,目标是描绘相似组织模式的块对应具有彼此接近(相似)的输出(特征向量),以及描绘相异组织模式的块对应该具有彼此远离的输出。根据一个实施例,这是通过执行对比损失来实现的,例如Hadsell等人所描述的,通过学习不变映射进行降维,2006,CVPR`06。在训练期间,将对比损失最小化。对比损失CL可以计算,例如,根据For example, a Siamese neural network can be, as described by Bromley et al. in “Signature Verification Using a Time-Delayed Neural Network with ‘Siamese Neural’, 1994, NIPS’ 1994.” Each subnetwork of the Siamese neural network is adapted to extract a multidimensional feature vector from a corresponding one of two image patches provided as input. The network is trained on multiple patch pairs that have been automatically labeled with proximity-based organizational pattern similarity tags, with the goal that patches depicting similar organizational patterns should have outputs (feature vectors) that are close to each other, and patch pairs depicting dissimilar organizational patterns should have outputs that are far from each other. According to one embodiment, this is achieved by performing a contrastive loss, such as dimensionality reduction by learning an invariant mapping as described by Hadsell et al., 2006, CVPR’06. During training, the contrastive loss is minimized. The contrastive loss CL can be calculated, for example, according to…
CL=(1-y)2(f1-f2)+y*max(0,m-L2(f1-f2)),其中1,2是两个相同子网络的输出,并且y是块对的基本事实标签:如果它们用标签标记为“相似”(第一块对集),则为0,如果它们用标签标记为“相异”(第二块对集),则为1。CL = (1-y)²(f1-f2) + y*max(0, m-L²(f1-f2)), where 1 and 2 are the outputs of two identical subnetworks, and y is the basic fact label of the block pair: 0 if they are labeled "similar" (first block pair set), and 1 if they are labeled "dissimilar" (second block pair set).
孪生神经网络900的训练包括向网络900馈送多个自动用标签标记的相似812、813和相异814、815块对。每个输入训练数据记录928包括块对的两个块及其自动分配的、基于空间接近度的标签907。基于接近度的标签403作为“基本事实”提供。输出层924适合于计算针对两个输入图像904、914的预测相似性标签,作为两个相比较的特征向量908、918的相似性的函数。孪生神经网络的训练包括一个反向传播过程。预测标签926与输入标签907的任何偏差都被视为以损失函数的形式测量的“错误”或“损失”。孪生神经网络的训练包括通过迭代使用反向传播来最小化损失函数计算出的误差。例如,可以实现孪生神经网络900,如Bromley等人在“使用‘孪生神经’时间延迟神经网络的签名验证”,1994,NIPS’1994中所述。Training the Siamese neural network 900 involves feeding the network 900 multiple pairs of automatically labeled similar blocks 812, 813 and dissimilar blocks 814, 815. Each input training data record 928 includes the two blocks of the block pair and their automatically assigned spatial proximity-based labels 907. The proximity-based labels 403 are provided as “ground facts.” The output layer 924 is adapted to compute predicted similarity labels for two input images 904, 914 as a function of the similarity between two compared feature vectors 908, 918. Training the Siamese neural network includes a backpropagation process. Any deviation between the predicted label 926 and the input label 907 is considered an “error” or “loss” measured in the form of a loss function. Training the Siamese neural network involves iteratively minimizing the error computed by the loss function using backpropagation. For example, the Siamese neural network 900 can be implemented as described by Bromley et al. in “Signature Verification Using a Time-Delayed Neural Network with ‘Siamese Neural Networks’,” 1994, NIPS 1994.
图10描绘了例如参考图9所述的作为截短的孪生神经网络实现的特征提取MLL950。Figure 10 illustrates the feature extraction MLL950, implemented as a truncated Siamese neural network, for example, as described with reference to Figure 9.
特征提取MLL 950可以,例如,通过分别存储经训练的孪生神经网络900的子网络902、903之一来获得。与经训练的孪生神经网络相反,用作特征提取MLL的子网络90、903仅需要单个图像952作为输入,并且不输出相似性标签,而是输出特征向量910,该特征向量910选择性地包括限定特征集,该限定特征集在孪生神经网络900的训练期间被识别为针对特定组织模式具有特定特征,并且特别适合通过从两个图像中提取和比较该特定的特征集来确定两个图像中描绘的组织模式的相似性。The feature extraction MLL 950 can be obtained, for example, by storing one of the subnetworks 902 and 903 of a trained Siamese neural network 900, respectively. In contrast to the trained Siamese neural network, the subnetworks 90 and 903 used as the feature extraction MLL require only a single image 952 as input and do not output similarity labels. Instead, they output a feature vector 910 that selectively includes a defined set of features identified during the training of the Siamese neural network 900 as having specific characteristics for a particular organizational pattern, and is particularly suitable for determining the similarity of organizational patterns depicted in two images by extracting and comparing this specific set of features from the two images.
图11描绘了在图像数据库中使用基于特征向量的相似性搜索的计算机系统980。例如,相似性搜索可用于计算搜索块库,图4中描绘了一个示例。计算机系统980包括一个或多个处理器982和经训练的特征提取MLL 950,该MLL 950可以是经训练的孪生神经网络(“截短的孪生神经网络”)的子网络。系统980适用于使用特征提取MLL来执行图像相似性搜索,以分别从搜索图像和从搜索图像(块)中的每一个中提取特征向量。Figure 11 illustrates a computer system 980 using feature vector-based similarity search in an image database. For example, similarity search can be used to compute a search block library, an example of which is depicted in Figure 4. The computer system 980 includes one or more processors 982 and a trained feature extraction MLL 950, which may be a subnetwork of a trained Siamese neural network (“truncated Siamese neural network”). System 980 is adapted to perform image similarity search using the feature extraction MLL to extract feature vectors from the search image and from each of the search images (blocks), respectively.
例如,计算机系统可以是标准计算机系统或由数据库992组成或操作上与之耦合的服务器。例如,数据库可以是相关的BDSM,包括描绘多个患者的组织样品的成百上千的整个载玻片图像。优选地,针对数据库中的图像中的每一个,数据库包括已经由特征输出MLL950从数据库中的所述图像提取的相应特征向量。优选地,在接收任何此类请求之前,在单个预处理步骤中执行数据库中每个图像的特征向量的计算。然而,也可以响应于搜索请求动态地计算和提取数据库中图像的特征向量。搜索可以限于从特定数字图像导出的块,例如用于识别描绘与搜索图像986中描绘的组织模式相似的组织模式的单个整个载玻片图像内的块。搜索图像986可以是例如包含在由用户选择的报告块库中的块。For example, the computer system may be a standard computer system or a server comprised of or operationally coupled to the database 992. For example, the database may be a related BDSM, comprising hundreds or thousands of whole slide images depicting tissue samples from multiple patients. Preferably, for each image in the database, the database includes a corresponding feature vector that has been extracted from said image in the database by the feature output MLL 950. Preferably, the calculation of the feature vector for each image in the database is performed in a single preprocessing step before any such request is received. However, the feature vectors of images in the database may also be dynamically calculated and extracted in response to a search request. The search may be limited to blocks derived from a particular digital image, such as blocks within a single whole slide image used to identify tissue patterns similar to those depicted in the search image 986. The search image 986 may be, for example, a block included in a library of report blocks selected by the user.
计算机系统包括使用户984能够选择或提供用作搜索图像986的特定图像或图像块的用户界面。经训练的特征提取MLL 950适用于从输入图像中提取特征向量988(“搜索特征向量”)。搜索引擎990从特征输出MLL 950接收搜索特征向量988并在图像数据库中执行基于向量的相似性搜索。相似性搜索包括将搜索特征向量与数据库中图像的特征向量中的每一个进行比较,以计算作为两个相比较的特征向量的函数的相似性得分。相似性得分指示搜索特征向量与数据库中图像的特征向量的相似性程度,并且从而指示两个相比较图像中描绘的组织模式的相似性。搜索引擎990适用于向用户返回并输出搜索结果994。搜索结果可以是,例如,数据库的计算出最高相似性得分的一个或多个图像。The computer system includes a user interface that enables user 984 to select or provide a specific image or image patch for use as a search image 986. A trained feature extraction MLL 950 is adapted to extract feature vectors 988 (“search feature vectors”) from the input image. A search engine 990 receives the search feature vectors 988 from the feature output MLL 950 and performs a vector-based similarity search in an image database. The similarity search involves comparing the search feature vector with each of the feature vectors of the images in the database to compute a similarity score as a function of the two compared feature vectors. The similarity score indicates the degree of similarity between the search feature vector and the feature vectors of the images in the database, and thus indicates the similarity of the organizational patterns depicted in the two compared images. The search engine 990 is adapted to return and output search results 994 to the user. The search results may be, for example, one or more images in the database that compute the highest similarity score.
例如,如果搜索图像986是已知描绘乳腺癌组织的图像块,则系统980可用于识别描绘类似乳腺癌组织模式的多个其他块(或包括此类块的整个载玻片图像)。For example, if the search image 986 is a known image patch depicting breast cancer tissue, the system 980 can be used to identify multiple other patches (or an entire slide image including such patches) that depict similar breast cancer tissue patterns.
图12显示了两个块矩阵,每个矩阵由三列组成,每列包含六个块对。第一(上)矩阵显示了第一块对集(A),该第一块对集由彼此靠近的以及已自动分配了标签“相似”的块对的块组成。第二(下)矩阵显示了第二块对集(B),彼此相距很远,并且已自动分配了标签“相异”的块对。在某些情况下,用标签标记“相似”的块看起来相异,而用标签标记“相异”的块看起来相似。这种噪声的起因是:在两个不同组织模式相接的边界处,两个附近的块可能描绘不同的组织模式,并且甚至远处的组织区域也可能描绘相同的组织模式。这是数据集生成过程中预期的固有噪声。Figure 12 shows two block matrices, each consisting of three columns, with six block pairs per column. The first (top) matrix shows the first set of block pairs (A), which consists of blocks that are close to each other and have been automatically assigned the label "similar." The second (bottom) matrix shows the second set of block pairs (B), which consists of blocks that are far apart and have been automatically assigned the label "dissimilar." In some cases, blocks labeled "similar" appear dissimilar, while blocks labeled "dissimilar" appear similar. This noise arises because at the boundary where two different organizational patterns meet, two nearby blocks may depict different organizational patterns, and even distant organizational regions may depict the same organizational pattern. This is an inherent noise expected during the dataset generation process.
申请人已经观察到,尽管存在这种噪声,但基于自动用标签标记的数据集训练的特征提取MLL能够准确地识别和提取特征,从而明确区分相似和相异的块对。申请人假设所观察到的经训练的MLL对这种噪声的稳健性是基于这样一个事实:即区域边界的面积通常小于区域的非边界面积。The applicant has observed that, despite this noise, the feature extraction MLL trained on an automatically labeled dataset is able to accurately identify and extract features, thus clearly distinguishing between similar and dissimilar block pairs. The applicant hypothesizes that the observed robustness of the trained MLL to this noise is based on the fact that the area of a region boundary is typically smaller than the area of the region's non-boundary areas.
根据实施例,自动生成的训练数据集的质量是在第一步中使用先前经训练的相似性网络或ImageNet预训练网络来评估块对的相似性,然后在第二步中基于本文所述的针对本发明的实施例描述的块的空间接近度生成相似性标签,并且然后纠正成对标签,其中观察到一方面在第一步中确定的和另一方面在第二步中确定的两个块的相似性的强烈偏差。According to an embodiment, the quality of the automatically generated training dataset is determined by evaluating the similarity of block pairs in the first step using a previously trained similarity network or an ImageNet pre-trained network, then generating similarity labels based on the spatial proximity of the blocks described herein with respect to embodiments of the invention in the second step, and then correcting the pairwise labels, wherein a strong discrepancy is observed between the similarity of the two blocks determined in the first step and the similarity of the two blocks determined in the second step.
图13显示了基于相似性搜索结果的特征向量,该特征向量由基于接近度的相似性标签训练的特征提取MLL提取。这5个肿瘤查询块称为A、B、C、D和E。查询块用于图像检索任务,用于分别识别和检索除查询载玻片(A1-A5、B1-B5、C1-C5、D1-D5、E1-E5)以外的5个块,按从低到高的距离进行排序,使用特征提取MLL提取的特征向量,该特征提取MLL基于具有基于接近度的标签自动用标签标记的数据进行训练。目标类别(例如肿瘤)仅占搜索到的块的3%。即使某些检索到的块看起来与查询块(例如C3和C)非常不同,但除A4之外的所有检索到的块都已由专家病理学家验证包含肿瘤细胞(即正确的类别检索)。Figure 13 shows the feature vectors of the similarity-based search results, extracted by a feature extraction MLL trained with proximity-based similarity labels. The five tumor query blocks are designated A, B, C, D, and E. These query blocks are used for the image retrieval task to identify and retrieve five blocks excluding the query slides (A1-A5, B1-B5, C1-C5, D1-D5, E1-E5), sorted from lowest to highest distance, using feature vectors extracted using a feature extraction MLL trained on data automatically labeled with proximity-based tags. The target category (e.g., tumor) comprises only 3% of the searched blocks. Even though some retrieved blocks appear very different from the query blocks (e.g., C3 and C), all retrieved blocks except A4 have been verified by expert pathologists to contain tumor cells (i.e., the correct category retrieval).
参考编号列表Reference Number List
100 方法100 Methods
102-110 步骤Steps 102-110
200 图像分析系统200 Image Analysis System
202 处理器202 Processor
204 显示器204 Monitor
206 图像块库206 Image Tile Library
208 整个载玻片加热m达ap208 The entire glass slide is heated to map
210 存储介质210 Storage medium
212 数字图像212 Digital Images
214 拆分模块214 Module Splitting
216 用标签标记的块的包216 Packets of blocks marked with tags
218 特征提取模块218 Feature Extraction Module
220 特征向量220 Eigenvectors
222 注意力机器学习逻辑程序222 Attention Machine Learning Logic Program
224 特征向量权重224 Eigenvector Weights
226 多实例学习程序226 Multi-Instance Learning Program
228 块的数值相关度得分Numerical correlation score of 228 blocks
230 GUI生成模块230 GUI Generation Module
232 GUI232 GUI
300 包含报告块库的GUI300 A GUI containing the report block library
302 相似块的第一子集第一组织模式302 First Subset of Similar Blocks, First Organization Pattern
304 代表第二组织模式的相似块的第二子集304 represents the second subset of similar blocks representing the second organizational pattern.
306 代表第三组织模式的相似块的第三子集306 represents the third subset of similar blocks representing the third organizational pattern.
308 代表第四组织模式的相似块的第四子集308 represents the fourth subset of similar blocks in the fourth organizational pattern.
310 可选择GUI元素集310 Selectable GUI element set
312 整个载玻片图像312 Image of the entire slide
314 整个载玻片图像314 Image of the entire slide
316 整个载玻片图像316 Image of the entire slide
318 整个载玻片图像318 Image of the entire slide
322 相关度热图322 Relevance Heatmap
324 相关度热图324 Relevance Heatmap
326 相关度热图326 Relevance Heatmap
328 相关度热图328 Relevance Heatmap
400 包含相似性搜索块库的GUI400 A GUI containing a similarity search block library
402 相似块的第一子集第一组织模式402 First Subset of Similar Blocks, First Organization Pattern
404 代表第二组织模式的相似块的第二子集404 represents the second subset of similar blocks in the second organizational pattern.
406 代表第三组织模式的相似块的第三子集406 represents the third subset of similar blocks representing the third organizational pattern.
408 代表第四组织模式的相似块的第四子集408 represents the fourth subset of similar blocks representing the fourth organizational pattern.
410 可选择GUI元素集410 Selectable GUI element set
412 整个载玻片图像412 Image of the entire slide
414 整个载玻片图像414 Image of the entire slide
416 整个载玻片图像416 Image of the entire slide
418 整个载玻片图像418 Image of the entire slide
422 相似性热图422 Similarity Heatmap
424 相似性热图424 Similarity Heatmap
426 相似性热图426 Similarity Heatmap
428 相似性热图428 Similarity Heatmap
430 查询块430 Query block
950 特征提取MLL的网络架构950 Network Architecture for Feature Extraction MLL
602 用作输入的图像块602 Image block used as input
603 输入层603 Input Layer
604 多层604 Multi-story
606 瓶颈层606 Bottleneck Layer
800 数字组织图像切成多个块800 Digital tissue image cut into multiple blocks
802 块T1802 Block T1
804 块T2804 Block T2
806 块T3806 T3 blocks
808 第一空间接近度阈值(2D)808 First spatial proximity threshold (2D)
810 第二空间接近度阈值(2D)810 Second spatial proximity threshold (2D)
812 用标签标记“相似”的块对812 Label "similar" block pairs
813 用标签标记“相似”的块对813 Label "similar" block pairs
814 用标签标记“相异”的块对814 Label "dissimilar" block pairs
815 用标签标记“相异”的块对815 Label "dissimilar" block pairs
816 训练数据816 Training Data
832 与图像300对齐的数字组织图像832 Digital tissue image aligned with image 300
834 与图像332对齐的数字组织图像834 A digitally organized image aligned with image 332
836 第一空间接近度阈值(3D)836 First Spatial Proximity Threshold (3D)
838 第二空间接近度阈值(3D)838 Second Spatial Proximity Threshold (3D)
840 块T4840 T4 units
842 块T5842 T5 blocks
900 孪生神经网络900 Siamese Neural Network
902 子网络902 Subnetwork
903 子网络903 Subnetwork
904 第一输入块904 First Input Block
905 第一网络N1的输入层905 Input layer of the first network N1
906 隐藏层906 Hidden Layer
907 基于接近度(“测量”)的相似性标签907 Similarity tags based on proximity (“measurement”)
908 适用于针对第一输入块计算特征向量的隐藏层908 Suitable for hidden layers that compute feature vectors for the first input block
910 从第一输入块中904中提取的特征向量910 Feature vector extracted from 904 in the first input block
914 第二输入块914 Second Input Block
915 第二网络N2的输入层915 Input layer of the second network N2
916 隐藏层916 Hidden Layer
918 适用于针对第二输入块计算特征向量的隐藏层918 Suitable for hidden layers that compute feature vectors for the second input block
920 从第二输入块914中提取的特征向量920 Feature vector extracted from the second input block 914
922 输入块对922 Input block pairs
924 连接网络N1、N2的输出层924 Output layer connecting networks N1 and N2
926 预测相似度标签926 Predicted similarity tags
928 训练数据集的单独数据记录928 Individual data records of the training dataset
950 特征提取MLL950 Feature Extraction MLL
952 单独输入图像/块952 Individual input image/block
954 特征向量954 Eigenvectors
980 计算机系统980 Computer System
982 处理器982 Processor
984 用户984 users
986 单独输入图像/块986 Individual input image/block
988 搜索特征向量988 Search Feature Vector
990 基于特征向量的搜索引擎990 Feature Vector-Based Search Engine
992 包含多个图像或块的数据库992 A database containing multiple images or blocks.
994 返回的相似性搜索结果994 The returned similarity search results
Claims (25)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP19162244.8 | 2019-03-12 | ||
| EP19165967.1 | 2019-03-28 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| HK40051109A HK40051109A (en) | 2021-12-31 |
| HK40051109B true HK40051109B (en) | 2024-10-10 |
Family
ID=
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11901077B2 (en) | Multiple instance learner for prognostic tissue pattern identification | |
| CN114730463B (en) | Multiple instance learners for tissue image classification | |
| Lu et al. | Data-efficient and weakly supervised computational pathology on whole-slide images | |
| JP7504116B2 (en) | Machine learning using distance-based similarity labels | |
| US11756198B1 (en) | Systems and methods for training a statistical model to predict tissue characteristics for a pathology image | |
| KR102583103B1 (en) | Systems and methods for processing electronic images for computational detection methods | |
| US8831327B2 (en) | Systems and methods for tissue classification using attributes of a biomarker enhanced tissue network (BETN) | |
| KR102562708B1 (en) | Systems and methods for processing electronic images for generalized disease detection | |
| CN114600155A (en) | Weakly Supervised Multi-Task Learning for Cell Detection and Segmentation | |
| CN115715416A (en) | Medical data inspector based on machine learning | |
| Al-Thelaya et al. | HistoContours: a Framework for Visual Annotation of Histopathology Whole Slide Images. | |
| Foran et al. | Automated image interpretation and computer-assisted diagnostics | |
| EP4627523A1 (en) | Systems and methods for detecting tertiary lymphoid structures | |
| HK40051109B (en) | Multiple instance learner for prognostic tissue pattern identification | |
| Nanthini et al. | Cervical Cancer Cell Segmentation and Classification using ML Approach | |
| HK40051109A (en) | Multiple instance learner for prognostic tissue pattern identification | |
| HK40071857A (en) | Multiple instance learner for tissue image classification | |
| Ben Taieb | Analyzing cancers in digitized histopathology images | |
| Swain | Combining VGG16 with Random Forest and Capsule Network for Detecting Multiple Myeloma | |
| Koohbanani | Working with scarce annotations in computational pathology |