CN107203778A - PVR intensity grade detecting system and method - Google Patents
PVR intensity grade detecting system and method Download PDFInfo
- Publication number
- CN107203778A CN107203778A CN201710312327.6A CN201710312327A CN107203778A CN 107203778 A CN107203778 A CN 107203778A CN 201710312327 A CN201710312327 A CN 201710312327A CN 107203778 A CN107203778 A CN 107203778A
- Authority
- CN
- China
- Prior art keywords
- retinopathy
- mrow
- degree
- model
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种视网膜病变程度等级检测系统及方法,该方法包括:在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。本发明无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。
The invention discloses a system and method for detecting the degree of retinopathy. The method includes: after receiving the retinopathy picture to be identified, using a predetermined recognition model to identify the received retinopathy picture, and outputting the recognition result ; Wherein, the predetermined recognition model is a convolutional neural network model obtained by training a preset number of sample pictures marked with different grades of retinopathy in advance; according to the mapping between the predetermined recognition results and grades of retinopathy relationship to determine the degree of retinopathy corresponding to the output recognition result. The present invention does not need to perform complex feature extraction calculations on eye images, is simpler, and can determine the corresponding grades of different retinopathy degrees according to the recognition results, and can effectively perform fine-grained recognition of the degree of retinopathy of patients.
Description
技术领域technical field
本发明涉及计算机技术领域,尤其涉及一种视网膜病变程度等级检测系统及方法。The invention relates to the field of computer technology, in particular to a system and method for detecting the degree of retinal disease.
背景技术Background technique
根据对发达国家超过9300万劳动力人口的调查,糖尿病性视网膜病变是导致眼睛失明的一个首要因素,目前,针对糖尿病性视网膜病变的识别通常需要对眼部图像进行特征提取(例如,眼部血管结构,视神经盘,视网膜中心凹槽等特征的提取),特征提取的算法复杂运行性能差,同时,难以对患者的视网膜病变程度进行精细化识别,识别精度难以达到要求。According to a survey of more than 93 million labor force populations in developed countries, diabetic retinopathy is a leading cause of eye blindness. Currently, the identification of diabetic retinopathy usually requires feature extraction on eye images (for example, ocular vascular structure , optic disc, retinal central groove and other feature extraction), the algorithm of feature extraction is complex and has poor performance.
发明内容Contents of the invention
本发明的主要目的在于提供一种视网膜病变程度等级检测系统及方法,旨在简单有效地对患者的视网膜病变程度进行精细化识别。The main purpose of the present invention is to provide a system and method for detecting the degree of retinopathy, aiming at finely identifying the degree of retinopathy of a patient simply and effectively.
为实现上述目的,本发明提供的一种视网膜病变程度等级检测系统,所述视网膜病变程度等级检测系统包括:In order to achieve the above object, the present invention provides a grade detection system for the degree of retinopathy, which includes:
识别模块,用于在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;The recognition module is configured to, after receiving the retinopathy picture to be recognized, use a predetermined recognition model to recognize the received retinopathy picture, and output a recognition result; A convolutional neural network model obtained by training a preset number of sample pictures marked with different levels of retinopathy;
确定模块,用于根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。The determination module is configured to determine the degree of retinopathy corresponding to the output recognition result according to the predetermined mapping relationship between the recognition result and the degree of retinopathy.
优选地,所述预先确定的识别模型的训练过程如下:Preferably, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Set a corresponding preset number of sample pictures for each preset degree of retinopathy, and mark the corresponding degree of retinopathy for each sample picture;
B、将各个样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training image to be trained by the model;
C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C, divide all training pictures into the training set of the first proportion and the verification set of the second proportion;
D、利用所述训练集训练所述预先确定的识别模型;D. Using the training set to train the predetermined recognition model;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. Use the verification set to verify the accuracy of the trained recognition model. If the accuracy is greater than or equal to the preset accuracy, the training ends. Or, if the accuracy is less than the preset accuracy, then increase the degree of retinopathy corresponding to each grade. number of sample images and re-execute steps B, C, D, and E above.
优选地,所述步骤B包括:Preferably, said step B includes:
将各个样本图片的较短边长缩放到第一预设大小以获得对应的第一图片,在各个第一图片上随机裁剪出一个第二预设大小的第二图片;Scaling the shorter side length of each sample picture to a first preset size to obtain a corresponding first picture, and randomly cropping a second picture of a second preset size on each first picture;
根据各个预先确定的预设类型参数对应的标准参数值,将各个第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,以获得对应的第三图片;Adjusting each predetermined preset type parameter value of each second picture to a corresponding standard parameter value according to a standard parameter value corresponding to each predetermined preset type parameter, so as to obtain a corresponding third picture;
对各个第三图片进行预设方向的翻转操作,及按照预设的扭曲角度对各个第三图片进行扭曲操作,以获得各个第三图片对应的第四图片,将各个第四图片作为待模型训练的训练图片。Perform a flip operation in a preset direction on each third picture, and perform a twist operation on each third picture according to a preset twist angle to obtain a fourth picture corresponding to each third picture, and use each fourth picture as a model to be trained training pictures.
优选地,所述深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层,各个所述网络层对应的激活函数f(x)为:Preferably, the deep convolutional neural network model includes an input layer and a plurality of network layers, and the network layers include a convolution layer, a pooling layer, a fully connected layer and a classifier layer, and the corresponding activation functions of each of the network layers f(x) is:
f(x)=max(α*x,0)f(x)=max(α*x,0)
其中,α为预设的泄漏率,x表示所述深度卷积神经网络模型中神经元的一个数值输入。Wherein, α is a preset leakage rate, and x represents a numerical input of a neuron in the deep convolutional neural network model.
优选地,各个所述网络层对应的交叉熵H(P,Q)为:Preferably, the cross-entropy H(P, Q) corresponding to each of the network layers is:
H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-Σx∈XP(x)logP(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。Among them, P and Q are two probability distributions, H(P) is the expectation of probability distribution P, H(P)=-Σ x∈X P(x)logP(x), x is the sample space X of probability distribution P Any sample in , P(x) represents the probability that sample x is selected; the expression of D KL (P||Q) is x is any sample in the common sample space X of probability distribution P and Q, P(x) represents the probability that sample x is selected on probability distribution P, and Q(x) represents the probability that sample x is selected on probability distribution Q.
优选地,所述预先确定的识别模型的打分函数为:Preferably, the scoring function of the predetermined recognition model for:
其中,Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{0,1,2,3,4},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。in, O i, j represents the number of pictures that are actually predicted as i for the first time and j for the second time, O represents an N*N matrix, O i, j represents the matrix elements in matrix O, and N represents the number of pictures participating in the prediction Number of pictures, prediction result i, j ∈ {0, 1, 2, 3, 4}, E i, j represents the number of images that should appear for the first prediction as i and the second prediction as j, E is the expected prediction The N*N matrix of the result, E i,j represent the matrix elements in the matrix E.
优选地,各个所述网络层对应的交叉熵损失函数L(x,:W)为:Preferably, the cross-entropy loss function L(x, :W) for:
其中,x表示模型的输入,表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:Among them, x represents the input of the model, Represents the label corresponding to the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the output of the model after transforming the input x, ζ represents the normalization factor, ||W|| 2 means to sum the matrix elements:
Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,表示第i批输入对应的平均梯度。Among them, ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy item, β is the weight decay coefficient, γ is the learning rate of the model, and W i represents the state value of the weight matrix at time i , Di represents the i-th batch of input, Indicates the average gradient corresponding to the i-th batch of inputs.
优选地,所述预先确定的识别模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围进行随机采样确定,所述全连接层的连接权重被丢弃的概率设置为第一预设值,所述交叉熵损失函数中的权值衰减系数设置为第二预设值,所述交叉商损失函数中的势能项设置为第三预设值。Preferably, the predetermined identification model includes at least one fully connected layer, the initial value of each weight in the predetermined identification model is determined by random sampling from a preset weight range, and the connection weight of the fully connected layer The probability of being discarded is set to a first preset value, the weight decay coefficient in the cross-entropy loss function is set to a second preset value, and the potential energy item in the cross-quotient loss function is set to a third preset value.
此外,为实现上述目的,本发明还提供一种视网膜病变程度等级的检测方法,所述方法包括以下步骤:In addition, in order to achieve the above object, the present invention also provides a method for detecting the degree of retinopathy, the method comprising the following steps:
在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;After receiving the picture of retinopathy to be identified, use a predetermined recognition model to identify the received picture of retinopathy, and output the recognition result; The convolutional neural network model obtained by training the preset number of sample pictures of degree level;
根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。According to the predetermined mapping relationship between the recognition result and the degree of retinopathy, the degree of retinopathy corresponding to the output recognition result is determined.
优选地,所述预先确定的识别模型的训练过程如下:Preferably, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Set a corresponding preset number of sample pictures for each preset degree of retinopathy, and mark the corresponding degree of retinopathy for each sample picture;
B、将各个样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training image to be trained by the model;
C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C, divide all training pictures into the training set of the first proportion and the verification set of the second proportion;
D、利用所述训练集训练所述预先确定的识别模型;D. Using the training set to train the predetermined recognition model;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. Use the verification set to verify the accuracy of the trained recognition model. If the accuracy is greater than or equal to the preset accuracy, the training ends. Or, if the accuracy is less than the preset accuracy, then increase the degree of retinopathy corresponding to each grade. number of sample images and re-execute steps B, C, D, and E above.
本发明提出的视网膜病变程度等级检测系统及方法,通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。The retinopathy degree level detection system and method proposed by the present invention recognize the received retinopathy pictures through the deep convolutional neural network model obtained by training based on a preset number of sample pictures marked with different retinopathy degree levels, and According to the identification result, the corresponding degree of retinal disease is determined. Since it only needs to recognize the received retinal lesion pictures according to the pre-trained deep convolutional neural network model, it does not need to perform complex feature extraction operations on eye images, which is simpler and can determine the corresponding different retinal lesions according to the recognition results The level of degree can effectively identify the degree of retinopathy of the patient in a fine-grained manner.
附图说明Description of drawings
图1为本发明视网膜病变程度等级的检测方法一实施例的流程示意图;Fig. 1 is a schematic flow chart of an embodiment of the detection method of the degree of retinopathy of the present invention;
图2为本发明视网膜病变程度等级检测系统10较佳实施例的运行环境示意图;FIG. 2 is a schematic diagram of the operating environment of a preferred embodiment of the retinopathy degree grade detection system 10 of the present invention;
图3为本发明视网膜病变程度等级检测系统一实施例的功能模块示意图。FIG. 3 is a schematic diagram of functional modules of an embodiment of the system for detecting the degree of retinopathy according to the present invention.
本发明目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization of the purpose of the present invention, functional characteristics and advantages will be further described in conjunction with the embodiments and with reference to the accompanying drawings.
具体实施方式detailed description
为了使本发明所要解决的技术问题、技术方案及有益效果更加清楚、明白,以下结合附图和实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。In order to make the technical problems, technical solutions and beneficial effects to be solved by the present invention clearer and clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.
本发明提供一种视网膜病变程度等级的检测方法。The invention provides a detection method for grades of retinal lesions.
参照图1,图1为本发明视网膜病变程度等级的检测方法一实施例的流程示意图。Referring to FIG. 1 , FIG. 1 is a schematic flowchart of an embodiment of the method for detecting the degree of retinopathy in the present invention.
在一实施例中,该视网膜病变程度等级的检测方法包括:In one embodiment, the detection method of the degree of retinopathy includes:
步骤S10、在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型。Step S10, after receiving the retinopathy picture to be identified, use a predetermined recognition model to identify the received retinopathy picture, and output the recognition result; A deep convolutional neural network model obtained by training a preset number of sample pictures of different degrees of retinopathy.
本实施例中,视网膜病变程度等级检测系统接收用户发出的包含待识别的视网膜病变图片的视网膜病变程度等级检测请求,例如,接收用户通过手机、平板电脑、自助终端设备等终端发送的视网膜病变程度等级检测请求,如接收用户在手机、平板电脑、自助终端设备等终端中预先安装的客户端上发送来的视网膜病变程度等级检测请求,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的视网膜病变程度等级检测请求。In this embodiment, the retinopathy degree level detection system receives the retinopathy level detection request sent by the user including the retinopathy picture to be identified, for example, receives the retinopathy level level detection request sent by the user through a terminal such as a mobile phone, a tablet computer, and a self-service terminal device. Level detection request, such as receiving the level detection request of the degree of retinopathy sent by the user on the client pre-installed in the mobile phone, tablet computer, self-service terminal equipment, etc. The retinopathy degree grade detection request sent from the browser system.
视网膜病变程度等级检测系统在收到用户发出的视网膜病变程度等级检测请求后,利用预先训练好的识别模型对收到的待识别的视网膜病变图片进行识别,识别出收到的待识别的视网膜病变图片在识别模型中的识别结果。该识别模型可预先通过对大量标注有不同视网膜病变程度等级的预设数量样本图片进行识别来不断进行训练、学习、验证、优化等,以将其训练成能准确识别出不同视网膜病变程度等级对应的标注的模型。例如,该识别模型可采用深度卷积神经网络模型(Convolutional Neural Network,CNN)模型等。Retinopathy degree level detection system, after receiving the user's retinopathy degree level detection request, uses the pre-trained recognition model to identify the received retinopathy pictures to be identified, and recognizes the received retinopathy level to be identified The recognition result of the image in the recognition model. The identification model can be continuously trained, learned, verified, optimized, etc. by identifying a large number of preset sample pictures marked with different levels of retinal lesions in advance, so as to train it to accurately identify the corresponding levels of different levels of retinal lesions. Annotated model of . For example, the recognition model may adopt a deep convolutional neural network model (Convolutional Neural Network, CNN) model and the like.
步骤S20、根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。Step S20 , according to the predetermined mapping relationship between the recognition result and the degree of retinopathy, determine the degree of retinopathy corresponding to the output recognition result.
在利用预先训练好的深度卷积神经网络模型对收到的视网膜病变图片进行识别获取到识别结果后,可根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级,确定的视网膜病变程度等级即为收到的视网膜病变图片所对应的视网膜病变程度等级。例如,在一种实施方式中,所述识别结果包括第一识别结果(例如,标注为“0”)、第二识别结果(例如,标注为“1”)、第三识别结果(例如,标注为“2”)、第四识别结果(例如,标注为“3”)及第五识别结果(例如,标注为“4”),所述视网膜病变程度等级包括第一等级、第二等级、第三等级、第四等级及第五等级。可预先确定不同识别结果与视网膜病变程度等级的映射关系,如所述第一识别结果对应第一等级,第二识别结果对应第二等级,第三识别结果对应第三等级,第四识别结果对应第四等级,第五识别结果对应第五等级。例如,具体地,第一等级可以对应正常和轻度的非增殖性糖尿病视网膜病变,该第一等级对应的视网膜病变图片表现为仅有个别血管瘤,硬性渗出,视网膜出血等。第二等级可以对应无临床意义黄斑水肿的非增殖性糖尿病视网膜病变,该第二等级对应的视网膜病变图片表现有微血管瘤,硬性渗出,视网膜出血,袢状或串珠状静脉。第三等级可以对应有临床意义的黄斑水肿(C SME)的非增殖性糖尿病视网膜病,该第三等级对应的视网膜病变图片表现黄斑区及其附近有视网膜增厚,并有微血管瘤,软性渗出,视网膜出血。第四等级可以对应非高危险期的增生性视网膜病,该第四等级对应的视网膜病变图片表现为视乳头外区有新生血管形成,其他区域内视网膜微血管形成的增殖型改变。第五等级可以对应高危险期的增生性视网膜病,该第五等级对应的视网膜病变图片表现为视乳头区有新生血管形成,玻璃体或视网膜前出血。After the pre-trained deep convolutional neural network model is used to identify the received retinopathy pictures and obtain the recognition results, the corresponding output recognition results can be determined according to the predetermined mapping relationship between the recognition results and the degree of retinopathy. Retinopathy degree grade, the determined retinopathy degree grade is the retinopathy degree grade corresponding to the received retinopathy picture. For example, in one embodiment, the recognition results include a first recognition result (for example, marked as "0"), a second recognition result (for example, marked as "1"), a third recognition result (for example, marked as is "2"), the fourth recognition result (for example, marked as "3") and the fifth recognition result (for example, marked as "4"), the grades of the degree of retinopathy include the first grade, the second grade, the first grade Level 3, Level 4 and Level 5. The mapping relationship between different recognition results and grades of retinopathy can be determined in advance, for example, the first recognition result corresponds to the first grade, the second recognition result corresponds to the second grade, the third recognition result corresponds to the third grade, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level. For example, specifically, the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the picture of retinopathy corresponding to the first level shows only individual hemangiomas, hard exudates, retinal hemorrhages and the like. The second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema. This second grade corresponds to retinopathy pictures showing microangiomas, hard exudates, retinal hemorrhages, looped or beaded veins. The third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (C SME). Exudation, retinal hemorrhage. The fourth grade can correspond to proliferative retinopathy in the non-high-risk stage. The retinopathy picture corresponding to the fourth grade shows neovascularization in the outer area of the optic disc and proliferative changes in retinal microvascular formation in other areas. The fifth grade can correspond to the proliferative retinopathy in the high-risk stage, and the retinopathy picture corresponding to the fifth grade shows neovascularization in the optic disc region, vitreous body or preretinal hemorrhage.
这样,在利用预先训练好的识别模型对收到的视网膜病变图片进行识别获取到识别结果后,即可根据获取到的不同识别结果确定对应的不同视网膜病变程度等级,从而实现对多种细化的视网膜病变程度等级的准确识别。In this way, after using the pre-trained recognition model to recognize the received retinopathy pictures and obtain the recognition results, the corresponding different grades of retinopathy can be determined according to the obtained different recognition results, so as to achieve a variety of refined Accurate identification of grades of retinopathy.
本实施例通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。In this embodiment, a deep convolutional neural network model obtained by training based on a preset number of sample pictures marked with different grades of retinopathy is used to identify received retinopathy pictures, and determine the corresponding grade of retinopathy according to the recognition results . Since it only needs to recognize the received retinal lesion pictures according to the pre-trained deep convolutional neural network model, it does not need to perform complex feature extraction operations on eye images, which is simpler and can determine the corresponding different retinal lesions according to the recognition results The level of degree can effectively identify the degree of retinopathy of the patient in a fine-grained manner.
进一步地,在其他实施例中,所述预先确定的识别模型的训练过程如下:Further, in other embodiments, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级(如第一等级、第二等级、第三等级、第四等级及第五等级,或轻微、轻度、中度、重度等)准备对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Prepare corresponding presets for each preset grade of retinopathy (such as grade 1, grade 2, grade 3, grade 4 and grade 5, or mild, mild, moderate, severe, etc.) number of sample pictures, and mark the corresponding grade of retinal lesion for each sample picture;
B、将各个样本图片进行图片预处理以获得待模型训练的训练图片。通过对各个样本图片进行图片预处理如缩放、裁剪、翻转及/或扭曲等操作后才进行模型训练,以有效提高模型训练的真实性及准确率。例如在一种实施方式中,对各个样本图片进行图片预处理可以包括:B. Perform image preprocessing on each sample image to obtain a training image to be trained for the model. Model training is performed after image preprocessing such as scaling, cropping, flipping, and/or distorting operations are performed on each sample image to effectively improve the authenticity and accuracy of model training. For example, in one implementation manner, performing image preprocessing on each sample image may include:
将各个样本图片的较短边长缩放到第一预设大小(例如,640像素)以获得对应的第一图片,在各个第一图片上随机裁剪出一个第二预设大小(例如,256*256像素)的第二图片;Scale the length of the shorter side of each sample picture to a first preset size (for example, 640 pixels) to obtain a corresponding first picture, and randomly crop out a second preset size (for example, 256* 256 pixels) of the second image;
根据各个预先确定的预设类型参数(例如,颜色、亮度及/或对比度等)对应的标准参数值(例如,颜色对应的标准参数值为a1,亮度对应的标准参数值为a2,对比度对应的标准参数值为a3),将各个第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,获得对应的第三图片,以消除作为医学图片的样本图片在拍摄时外界条件导致的图片不清晰,提高模型训练的有效性;例如,将各个第二图片的亮度值调整为标准参数值a2,将各个第二图片的对比度值调整为标准参数值a3;According to the standard parameter values corresponding to each predetermined preset type parameter (for example, color, brightness and/or contrast, etc.) (for example, the standard parameter value corresponding to color is a1, the standard parameter value corresponding to brightness is a2, and the standard parameter value corresponding to contrast The standard parameter value is a3), each predetermined preset type parameter value of each second picture is adjusted to the corresponding standard parameter value, and the corresponding third picture is obtained, so as to eliminate the external conditions when taking the sample picture as a medical picture The resulting pictures are unclear, improving the effectiveness of model training; for example, adjusting the brightness value of each second picture to the standard parameter value a2, and adjusting the contrast value of each second picture to the standard parameter value a3;
对各个第三图片进行预设方向(例如,水平和垂直方向)的翻转,及按照预设的扭曲角度(例如,30度)对各个第三图片进行扭曲操作,获得各个第三图片对应的第四图片,各个第四图片即为对应的样本图片的训练图片。其中,翻转和扭曲操作的作用是模拟实际业务场景下各种形式的图片,通过这些翻转和扭曲操作可以增大数据集的规模,从而提高模型训练的真实性和实用性。Flip each third picture in a preset direction (for example, horizontal and vertical directions), and perform a twist operation on each third picture according to a preset twist angle (for example, 30 degrees), to obtain the third picture corresponding to each third picture Four pictures, each fourth picture is a training picture of the corresponding sample picture. Among them, the function of flipping and twisting operations is to simulate various forms of pictures in actual business scenarios. Through these flipping and twisting operations, the scale of the data set can be increased, thereby improving the authenticity and practicability of model training.
C、将所有训练图片分为第一比例(例如,50%)的训练集、第二比例(例如,25%)的验证集;C, divide all training pictures into the training set of the first proportion (for example, 50%), the verification set of the second proportion (for example, 25%);
D、利用所述训练集训练所述预先确定的识别模型;D. Using the training set to train the predetermined recognition model;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E,直至训练的识别模型的准确率大于或者等于预设准确率。E. Use the verification set to verify the accuracy of the trained recognition model. If the accuracy is greater than or equal to the preset accuracy, the training ends. Or, if the accuracy is less than the preset accuracy, then increase the degree of retinopathy corresponding to each grade. The number of sample pictures and re-execute the above steps B, C, D, E until the accuracy of the trained recognition model is greater than or equal to the preset accuracy.
进一步地,在其他实施例中,所述预先确定的识别模型即深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层,可选的,深度卷积神经网络模型还可以包括具有随机丢弃某些连接权重机制的网络层(即Dropout层),该网络层的作用是提升模型的识别精度。Further, in other embodiments, the predetermined recognition model, that is, the deep convolutional neural network model includes an input layer and a plurality of network layers, and the network layers include a convolutional layer, a pooling layer, a fully connected layer, and a classification Optionally, the deep convolutional neural network model can also include a network layer (i.e. Dropout layer) with a mechanism for randomly discarding certain connection weights. The function of this network layer is to improve the recognition accuracy of the model.
在一种具体的实施方式中,所述深度卷积神经网络模型由1个输入层,11个卷积层,5个池化层,1个具有随机丢弃某些连接权重机制的网络层(即Dropout层),1个全连接层,1个分类器层构成。该深度卷积神经网络模型的详细结构如下表1所示:In a specific embodiment, the deep convolutional neural network model consists of 1 input layer, 11 convolutional layers, 5 pooling layers, and 1 network layer with a mechanism of randomly discarding some connection weights (i.e. Dropout layer), a fully connected layer, and a classifier layer. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
表1Table 1
其中:Layer Name表示网络层的名称,Input表示网络的数据输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示第一个基于最大值池化层,Dropout表示具有随机丢弃某些连接权重机制的网络层,Avgpool5表示第5个池化层但采用取均值方式进行池化,Fc表示模型中的全连接层,Fc1表示第1个全连接层,Softmax表示Softmax分类器层;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度为3x 3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Output Size表示网络层输出特征映射的尺寸。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Among them: Layer Name represents the name of the network layer, Input represents the data input layer of the network, Conv represents the convolutional layer of the model, Conv1 represents the first convolutional layer of the model, MaxPool represents the maximum pooling layer of the model, and MaxPool1 represents the first A pooling layer based on the maximum value, Dropout means a network layer with a mechanism of randomly discarding some connection weights, Avgpool5 means the fifth pooling layer but uses the mean value for pooling, Fc means the fully connected layer in the model, Fc1 Indicates the first fully connected layer, Softmax indicates the Softmax classifier layer; Batch Size indicates the number of input images of the current layer; Kernel Size indicates the scale of the convolution kernel of the current layer (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel is 3x 3); Stride Size represents the moving step of the convolution kernel, that is, the distance to move to the next convolution position after one convolution; Output Size represents the size of the network layer output feature map. It should be noted that the pooling methods of the pooling layer in this embodiment include but are not limited to Mean pooling (average sampling), Max pooling (maximum sampling), Overlapping (overlapping sampling), L2pooling (mean square sampling), Local Contrast Normalization (normalized sampling), Stochasticpooling (random sampling), Def-pooling (deformation constraint sampling), etc.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的激活函数f(x)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolutional layer, a pooling layer, a network layer with a mechanism for randomly discarding certain connection weights, a fully connected layer, and a classifier) layer, etc.) corresponding to the activation function f(x) is:
f(x)=max(α*x,0)f(x)=max(α*x,0)
其中,α为泄漏率,x表示该深度卷积神经网络模型中神经元的一个数值输入。在本实施例的一个优选实施方式中,将α设定为0.5。经过相同测试数据集的对比测试,相较于其他现有的激活函数,通过本实施例的激活函数f(x),该深度卷积神经网络模型的识别准确率大约有3%的提升。Among them, α is the leakage rate, and x represents a numerical input of a neuron in the deep convolutional neural network model. In a preferred implementation of this embodiment, α is set to 0.5. After a comparative test of the same test data set, compared with other existing activation functions, the recognition accuracy of the deep convolutional neural network model is improved by about 3% through the activation function f(x) of this embodiment.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的交叉熵H(P,Q)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolutional layer, a pooling layer, a network layer with a mechanism for randomly discarding certain connection weights, a fully connected layer, and a classifier) The corresponding cross entropy H(P,Q) is:
H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)logP(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。Among them, P and Q are two probability distributions, H(P) is the expectation of probability distribution P, H(P)=-∑ x∈X P(x)logP(x), x is the sample space X of probability distribution P Any sample in , P(x) represents the probability that sample x is selected; the expression of D KL (P||Q) is x is any sample in the common sample space X of probability distribution P and Q, P(x) represents the probability that sample x is selected on probability distribution P, and Q(x) represents the probability that sample x is selected on probability distribution Q.
进一步地,为了保证模型训练的效率和准确性,各个所述网络层对应的交叉熵损失函数L(x,:W)为:Further, in order to ensure the efficiency and accuracy of model training, the cross-entropy loss function L(x, :W) for:
其中,x表示模型的输入,表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:Among them, x represents the input of the model, Represents the label corresponding to the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the output of the model after transforming the input x, ζ represents the normalization factor, ||W|| 2 means to sum the matrix elements:
Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,表示第i批输入对应的平均梯度。Among them, ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy item, β is the weight decay coefficient, γ is the learning rate of the model, and W i represents the state value of the weight matrix at time i , Di represents the i-th batch of input, Indicates the average gradient corresponding to the i-th batch of inputs.
本实施例中,交叉熵可在神经网络(机器学习)中作为损失函数,例如,P表示真实标记的分布,Q则为训练后的模型的预测标记分布,交叉熵损失函数可以衡量P与Q的相似性,以保证模型训练的准确性。而且,交叉熵作为损失函数在梯度下降时能避免均方误差损失函数学习速率降低的问题,因此,能保证模型训练的效率。In this embodiment, cross entropy can be used as a loss function in a neural network (machine learning). For example, P represents the distribution of real marks, and Q is the predicted mark distribution of the trained model. The cross entropy loss function can measure P and Q similarity to ensure the accuracy of model training. Moreover, cross entropy as a loss function can avoid the problem of reduced learning rate of the mean square error loss function during gradient descent, so it can ensure the efficiency of model training.
进一步地,在其他实施例中,所述深度卷积神经网络模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围(例如,(0,1)权重范围)进行随机采样确定,所述全连接层的连接权重被丢弃(Dropout)的概率设置为第一预设值(例如,0.5),所述交叉商损失函数中的权值衰减系数设置为第二预设值(例如,0.0005),所述交叉商损失函数中的势能项设置为第三预设值(例如,0.9)。Further, in other embodiments, the deep convolutional neural network model includes at least one fully connected layer, and the initial value of each weight in the predetermined recognition model ranges from a preset weight range (for example, (0, 1) weight range) is determined by random sampling, the probability of the connection weight of the fully connected layer being discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight decay coefficient in the cross quotient loss function It is set to a second preset value (for example, 0.0005), and the potential energy item in the cross quotient loss function is set to a third preset value (for example, 0.9).
进一步地,在其他实施例中,所述预先确定的识别模型的打分函数为:Further, in other embodiments, the scoring function of the predetermined recognition model for:
其中,Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{0,1,2,3,4},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。in, O i, j represents the number of pictures that are actually predicted as i for the first time and j for the second time, O represents an N*N matrix, O i, j represents the matrix elements in matrix O, and N represents the number of pictures participating in the prediction Number of pictures, prediction result i, j ∈ {0, 1, 2, 3, 4}, E i, j represents the number of images that should appear for the first prediction as i and the second prediction as j, E is the expected prediction The N*N matrix of the result, E i,j represent the matrix elements in the matrix E.
本实施例中通过打分函数来检测所述预先确定的识别模型的识别准确率,以保证训练出的所述预先确定的识别模型的识别准确率保持在较高水平,以保证对患者的视网膜病变程度进行准确地识别。In this embodiment, through the scoring function To detect the recognition accuracy of the predetermined recognition model, so as to ensure that the recognition accuracy of the trained predetermined recognition model is maintained at a high level, so as to ensure accurate recognition of the degree of retinopathy of the patient.
本发明进一步提供一种视网膜病变程度等级检测系统。请参阅图2,是本发明视网膜病变程度等级检测系统10较佳实施例的运行环境示意图。The present invention further provides a grade detection system for the degree of retinopathy. Please refer to FIG. 2 , which is a schematic view of the operating environment of a preferred embodiment of the system 10 for detecting the degree of retinopathy in the present invention.
在本实施例中,所述的视网膜病变程度等级检测系统10安装并运行于电子装置1中。该电子装置1可包括,但不仅限于,存储器11、处理器12及显示器13。图2仅示出了具有组件11-13的电子装置1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。In this embodiment, the retinopathy level detection system 10 is installed and operated in the electronic device 1 . The electronic device 1 may include, but not limited to, a memory 11 , a processor 12 and a display 13 . Figure 2 only shows the electronic device 1 with the components 11-13, but it is to be understood that implementation of all of the illustrated components is not required and that more or fewer components may instead be implemented.
所述存储器11在一些实施例中可以是所述电子装置1的内部存储单元,例如该电子装置1的硬盘或内存。所述存储器11在另一些实施例中也可以是所述电子装置1的外部存储设备,例如所述电子装置1上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器11还可以既包括所述电子装置1的内部存储单元也包括外部存储设备。所述存储器11用于存储安装于所述电子装置1的应用软件及各类数据,例如所述视网膜病变程度等级检测系统10的程序代码等。所述存储器11还可以用于暂时地存储已经输出或者将要输出的数据。The storage 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a hard disk or a memory of the electronic device 1 . The memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, flash memory card (Flash Card), etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 is used to store application software and various data installed in the electronic device 1 , such as program codes of the retinopathy level detection system 10 . The memory 11 can also be used to temporarily store data that has been output or will be output.
所述处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行所述存储器11中存储的程序代码或处理数据,例如执行所述视网膜病变程度等级检测系统10等。The processor 12 may be a central processing unit (Central Processing Unit, CPU) in some embodiments, a microprocessor or other data processing chips, for running the program code stored in the memory 11 or processing data, for example Execute the retinopathy degree grade detection system 10 and the like.
所述显示器13在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。所述显示器13用于显示在所述电子装置1中处理的信息以及用于显示可视化的用户界面,例如应用菜单界面、应用图标界面等。所述电子装置1的部件11-13通过系统总线相互通信。The display 13 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode, Organic Light-Emitting Diode) touch panel, etc. in some embodiments. The display 13 is used for displaying information processed in the electronic device 1 and for displaying a visualized user interface, such as an application menu interface, an application icon interface, and the like. The components 11-13 of the electronic device 1 communicate with each other via a system bus.
请参阅图3,是本发明视网膜病变程度等级检测系统10较佳实施例的功能模块图。在本实施例中,所述的视网膜病变程度等级检测系统10可以被分割成一个或多个模块,所述一个或者多个模块被存储于所述存储器11中,并由一个或多个处理器(本实施例为所述处理器12)所执行,以完成本发明。例如,在图3中,所述的视网膜病变程度等级检测系统10可以被分割成识别模块01、确定模块02。本发明所称的模块是指能够完成特定功能的一系列计算机程序指令段,比程序更适合于描述所述视网膜病变程度等级检测系统10在所述电子装置1中的执行过程。以下描述将具体介绍所述识别模块01、确定模块02的功能。Please refer to FIG. 3 , which is a functional block diagram of a preferred embodiment of the system 10 for detecting the degree of retinopathy in the present invention. In this embodiment, the retinopathy level detection system 10 can be divided into one or more modules, and the one or more modules are stored in the memory 11 and controlled by one or more processors (This embodiment is executed by the processor 12) to complete the present invention. For example, in FIG. 3 , the retinopathy level detection system 10 can be divided into an identification module 01 and a determination module 02 . The module referred to in the present invention refers to a series of computer program instruction segments capable of completing specific functions, which is more suitable than a program to describe the execution process of the retinopathy level detection system 10 in the electronic device 1 . The following description will specifically introduce the functions of the identification module 01 and the determination module 02.
参照图3,图3为本发明视网膜病变程度等级检测系统一实施例的功能模块示意图。Referring to FIG. 3 , FIG. 3 is a schematic diagram of functional modules of an embodiment of the system for detecting the degree of retinopathy in the present invention.
在一实施例中,该视网膜病变程度等级检测系统包括:In one embodiment, the retinopathy grade detection system includes:
识别模块01,用于在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型。The recognition module 01 is configured to, after receiving the retinopathy picture to be recognized, use a predetermined recognition model to recognize the received retinopathy picture, and output the recognition result; wherein, the predetermined recognition model is a pre-passed A convolutional neural network model obtained by training a preset number of sample pictures marked with different grades of retinopathy.
本实施例中,视网膜病变程度等级检测系统接收用户发出的包含待识别的视网膜病变图片的视网膜病变程度等级检测请求,例如,接收用户通过手机、平板电脑、自助终端设备等终端发送的视网膜病变程度等级检测请求,如接收用户在手机、平板电脑、自助终端设备等终端中预先安装的客户端上发送来的视网膜病变程度等级检测请求,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的视网膜病变程度等级检测请求。In this embodiment, the retinopathy degree level detection system receives the retinopathy level detection request sent by the user including the retinopathy picture to be identified, for example, receives the retinopathy level level detection request sent by the user through a terminal such as a mobile phone, a tablet computer, and a self-service terminal device. Level detection request, such as receiving the level detection request of the degree of retinopathy sent by the user on the client pre-installed in the mobile phone, tablet computer, self-service terminal equipment, etc. The retinopathy degree grade detection request sent from the browser system.
视网膜病变程度等级检测系统在收到用户发出的视网膜病变程度等级检测请求后,利用预先训练好的识别模型对收到的待识别的视网膜病变图片进行识别,识别出收到的待识别的视网膜病变图片在识别模型中的识别结果。该识别模型可预先通过对大量标注有不同视网膜病变程度等级的预设数量样本图片进行识别来不断进行训练、学习、验证、优化等,以将其训练成能准确识别出不同视网膜病变程度等级对应的标注的模型。例如,该识别模型可采用深度卷积神经网络模型(Convolutional Neural Network,CNN)模型等。Retinopathy degree level detection system, after receiving the user's retinopathy degree level detection request, uses the pre-trained recognition model to identify the received retinopathy pictures to be identified, and recognizes the received retinopathy level to be identified The recognition result of the image in the recognition model. The identification model can be continuously trained, learned, verified, optimized, etc. by identifying a large number of preset sample pictures marked with different levels of retinal lesions in advance, so as to train it to accurately identify the corresponding levels of different levels of retinal lesions. Annotated model of . For example, the recognition model may adopt a deep convolutional neural network model (Convolutional Neural Network, CNN) model and the like.
确定模块02,用于根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。The determination module 02 is configured to determine the degree of retinopathy corresponding to the output recognition result according to the predetermined mapping relationship between the recognition result and the degree of retinopathy.
在利用预先训练好的深度卷积神经网络模型对收到的视网膜病变图片进行识别获取到识别结果后,可根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级,确定的视网膜病变程度等级即为收到的视网膜病变图片所对应的视网膜病变程度等级。例如,在一种实施方式中,所述识别结果包括第一识别结果(例如,标注为“0”)、第二识别结果(例如,标注为“1”)、第三识别结果(例如,标注为“2”)、第四识别结果(例如,标注为“3”)、及第五识别结果(例如,标注为“4”),所述视网膜病变程度等级包括第一等级、第二等级、第三等级、第四等级及第五等级。可预先确定不同识别结果与视网膜病变程度等级的映射关系,如所述第一识别结果对应第一等级,第二识别结果对应第二等级,第三识别结果对应第三等级,第四识别结果对应第四等级,第五识别结果对应第五等级。例如,具体地,第一等级可以对应正常和轻度的非增殖性糖尿病视网膜病变,该第一等级对应的视网膜病变图片表现为仅有个别血管瘤,硬性渗出,视网膜出血等。第二等级可以对应无临床意义黄斑水肿的非增殖性糖尿病视网膜病变,该第二等级对应的视网膜病变图片表现有微血管瘤,硬性渗出,视网膜出血,袢状或串珠状静脉。第三等级可以对应有临床意义的黄斑水肿(C SME)的非增殖性糖尿病视网膜病,该第三等级对应的视网膜病变图片表现黄斑区及其附近有视网膜增厚,并有微血管瘤,软性渗出,视网膜出血。第四等级可以对应非高危险期的增生性视网膜病,该第四等级对应的视网膜病变图片表现为视乳头外区有新生血管形成,其他区域内视网膜微血管形成的增殖型改变。第五等级可以对应高危险期的增生性视网膜病,该第五等级对应的视网膜病变图片表现为视乳头区有新生血管形成,玻璃体或视网膜前出血。After the pre-trained deep convolutional neural network model is used to identify the received retinopathy pictures and obtain the recognition results, the corresponding output recognition results can be determined according to the predetermined mapping relationship between the recognition results and the degree of retinopathy. Retinopathy degree grade, the determined retinopathy degree grade is the retinopathy degree grade corresponding to the received retinopathy picture. For example, in one embodiment, the recognition results include a first recognition result (for example, marked as "0"), a second recognition result (for example, marked as "1"), a third recognition result (for example, marked as is "2"), the fourth recognition result (for example, marked as "3"), and the fifth recognition result (for example, marked as "4"), the grades of the degree of retinopathy include the first grade, the second grade, The third level, the fourth level and the fifth level. The mapping relationship between different recognition results and grades of retinopathy can be determined in advance, for example, the first recognition result corresponds to the first grade, the second recognition result corresponds to the second grade, the third recognition result corresponds to the third grade, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level. For example, specifically, the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the picture of retinopathy corresponding to the first level shows only individual hemangiomas, hard exudates, retinal hemorrhages and the like. The second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema. This second grade corresponds to retinopathy pictures showing microangiomas, hard exudates, retinal hemorrhages, looped or beaded veins. The third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (C SME). Exudation, retinal hemorrhage. The fourth grade can correspond to proliferative retinopathy in the non-high-risk stage. The retinopathy picture corresponding to the fourth grade shows neovascularization in the outer area of the optic disc and proliferative changes in retinal microvascular formation in other areas. The fifth grade can correspond to the proliferative retinopathy in the high-risk stage, and the retinopathy picture corresponding to the fifth grade shows neovascularization in the optic disc region, vitreous body or preretinal hemorrhage.
这样,在利用预先训练好的识别模型对收到的视网膜病变图片进行识别获取到识别结果后,即可根据获取到的不同识别结果确定对应的不同视网膜病变程度等级,从而实现对多种细化的视网膜病变程度等级的准确识别。In this way, after using the pre-trained recognition model to recognize the received retinopathy pictures and obtain the recognition results, the corresponding different grades of retinopathy can be determined according to the obtained different recognition results, so as to achieve a variety of refined Accurate identification of grades of retinopathy.
本实施例通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。In this embodiment, a deep convolutional neural network model obtained by training based on a preset number of sample pictures marked with different grades of retinopathy is used to identify received retinopathy pictures, and determine the corresponding grade of retinopathy according to the recognition results . Since it only needs to recognize the received retinal lesion pictures according to the pre-trained deep convolutional neural network model, it does not need to perform complex feature extraction operations on eye images, which is simpler and can determine the corresponding different retinal lesions according to the recognition results The level of degree can effectively identify the degree of retinopathy of the patient in a fine-grained manner.
进一步地,在其他实施例中,所述预先确定的识别模型的训练过程如下:Further, in other embodiments, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级(如第一等级、第二等级、第三等级、第四等级及第五等级,或轻微、轻度、中度、重度等)准备对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Prepare corresponding presets for each preset grade of retinopathy (such as grade 1, grade 2, grade 3, grade 4 and grade 5, or mild, mild, moderate, severe, etc.) number of sample pictures, and mark the corresponding grade of retinal lesion for each sample picture;
B、将各个样本图片进行图片预处理以获得待模型训练的训练图片。通过对各个样本图片进行图片预处理如缩放、裁剪、翻转及/或扭曲等操作后才进行模型训练,以有效提高模型训练的真实性及准确率。例如在一种实施方式中,对各个样本图片进行图片预处理可以包括:B. Perform image preprocessing on each sample image to obtain a training image to be trained for the model. Model training is performed after image preprocessing such as scaling, cropping, flipping, and/or distorting operations are performed on each sample image to effectively improve the authenticity and accuracy of model training. For example, in one implementation manner, performing image preprocessing on each sample image may include:
将各个样本图片的较短边长缩放到第一预设大小(例如,640像素)以获得对应的第一图片,在各个第一图片上随机裁剪出一个第二预设大小(例如,256*256像素)的第二图片;Scale the length of the shorter side of each sample picture to a first preset size (for example, 640 pixels) to obtain a corresponding first picture, and randomly crop out a second preset size (for example, 256* 256 pixels) of the second image;
根据各个预先确定的预设类型参数(例如,颜色、亮度及/或对比度等)对应的标准参数值(例如,颜色对应的标准参数值为a1,亮度对应的标准参数值为a2,对比度对应的标准参数值为a3),将各个第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,获得对应的第三图片,以消除作为医学图片的样本图片在拍摄时外界条件导致的图片不清晰,提高模型训练的有效性;例如,将各个第二图片的亮度值调整为标准参数值a2,将各个第二图片的对比度值调整为标准参数值a3;According to the standard parameter values corresponding to each predetermined preset type parameter (for example, color, brightness and/or contrast, etc.) (for example, the standard parameter value corresponding to color is a1, the standard parameter value corresponding to brightness is a2, and the standard parameter value corresponding to contrast The standard parameter value is a3), each predetermined preset type parameter value of each second picture is adjusted to the corresponding standard parameter value, and the corresponding third picture is obtained, so as to eliminate the external conditions when taking the sample picture as a medical picture The resulting pictures are unclear, improving the effectiveness of model training; for example, adjusting the brightness value of each second picture to the standard parameter value a2, and adjusting the contrast value of each second picture to the standard parameter value a3;
对各个第三图片进行预设方向(例如,水平和垂直方向)的翻转,及按照预设的扭曲角度(例如,30度)对各个第三图片进行扭曲操作,获得各个第三图片对应的第四图片,各个第四图片即为对应的样本图片的训练图片。其中,翻转和扭曲操作的作用是模拟实际业务场景下各种形式的图片,通过这些翻转和扭曲操作可以增大数据集的规模,从而提高模型训练的真实性和实用性。Flip each third picture in a preset direction (for example, horizontal and vertical directions), and perform a twist operation on each third picture according to a preset twist angle (for example, 30 degrees), to obtain the third picture corresponding to each third picture Four pictures, each fourth picture is a training picture of the corresponding sample picture. Among them, the function of flipping and twisting operations is to simulate various forms of pictures in actual business scenarios. Through these flipping and twisting operations, the scale of the data set can be increased, thereby improving the authenticity and practicability of model training.
C、将所有训练图片分为第一比例(例如,50%)的训练集、第二比例(例如,25%)的验证集;C, divide all training pictures into the training set of the first proportion (for example, 50%), the verification set of the second proportion (for example, 25%);
D、利用所述训练集训练所述预先确定的识别模型;D. Using the training set to train the predetermined recognition model;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E,直至训练的识别模型的准确率大于或者等于预设准确率。E. Use the verification set to verify the accuracy of the trained recognition model. If the accuracy is greater than or equal to the preset accuracy, the training ends. Or, if the accuracy is less than the preset accuracy, then increase the degree of retinopathy corresponding to each grade. The number of sample pictures and re-execute the above steps B, C, D, E until the accuracy of the trained recognition model is greater than or equal to the preset accuracy.
进一步地,在其他实施例中,所述预先确定的识别模型即深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层,可选的,深度卷积神经网络模型还可以包括具有随机丢弃某些连接权重机制的网络层(即Dropout层),该网络层的作用是提升模型的识别精度。Further, in other embodiments, the predetermined recognition model, that is, the deep convolutional neural network model includes an input layer and a plurality of network layers, and the network layers include a convolutional layer, a pooling layer, a fully connected layer, and a classification Optionally, the deep convolutional neural network model can also include a network layer (i.e. Dropout layer) with a mechanism for randomly discarding certain connection weights. The function of this network layer is to improve the recognition accuracy of the model.
在一种具体的实施方式中,所述深度卷积神经网络模型由1个输入层,11个卷积层,5个池化层,1个具有随机丢弃某些连接权重机制的网络层(即Dropout层),1个全连接层,1个分类器层构成。该深度卷积神经网络模型的详细结构如下表1所示:In a specific embodiment, the deep convolutional neural network model consists of 1 input layer, 11 convolutional layers, 5 pooling layers, and 1 network layer with a mechanism of randomly discarding some connection weights (i.e. Dropout layer), a fully connected layer, and a classifier layer. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
表1Table 1
其中:Layer Name表示网络层的名称,Input表示网络的数据输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示第一个基于最大值池化层,Dropout表示具有随机丢弃某些连接权重机制的网络层,Avgpool5表示第5个池化层但采用取均值方式进行池化,Fc表示模型中的全连接层,Fc1表示第1个全连接层,Softmax表示Softmax分类器层;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度为3x3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Output Size表示网络层输出特征映射的尺寸。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Among them: Layer Name represents the name of the network layer, Input represents the data input layer of the network, Conv represents the convolutional layer of the model, Conv1 represents the first convolutional layer of the model, MaxPool represents the maximum pooling layer of the model, and MaxPool1 represents the first A pooling layer based on the maximum value, Dropout means a network layer with a mechanism of randomly discarding some connection weights, Avgpool5 means the fifth pooling layer but uses the mean value for pooling, Fc means the fully connected layer in the model, Fc1 Indicates the first fully connected layer, Softmax indicates the Softmax classifier layer; Batch Size indicates the number of input images of the current layer; Kernel Size indicates the scale of the convolution kernel of the current layer (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel is 3x3); Stride Size represents the moving step of the convolution kernel, that is, the distance to move to the next convolution position after one convolution is completed; Output Size represents the size of the feature map output by the network layer. It should be noted that the pooling methods of the pooling layer in this embodiment include but are not limited to Mean pooling (average sampling), Max pooling (maximum sampling), Overlapping (overlapping sampling), L2pooling (mean square sampling), Local Contrast Normalization (normalized sampling), Stochasticpooling (random sampling), Def-pooling (deformation constraint sampling), etc.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的激活函数f(x)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolutional layer, a pooling layer, a network layer with a mechanism for randomly discarding certain connection weights, a fully connected layer, and a classifier) layer, etc.) corresponding to the activation function f(x) is:
f(x)=max(α*x,0)f(x)=max(α*x,0)
其中,α为泄漏率,x表示该深度卷积神经网络模型中神经元的一个数值输入。在本实施例的一个优选实施方式中,将α设定为0.5。经过相同测试数据集的对比测试,相较于其他现有的激活函数,通过本实施例的激活函数f(x),该深度卷积神经网络模型的识别准确率大约有3%的提升。Among them, α is the leakage rate, and x represents a numerical input of a neuron in the deep convolutional neural network model. In a preferred implementation of this embodiment, α is set to 0.5. After a comparative test of the same test data set, compared with other existing activation functions, the recognition accuracy of the deep convolutional neural network model is improved by about 3% through the activation function f(x) of this embodiment.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的交叉熵H(P,Q)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolutional layer, a pooling layer, a network layer with a mechanism for randomly discarding certain connection weights, a fully connected layer, and a classifier) The corresponding cross entropy H(P,Q) is:
H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)logP(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。Among them, P and Q are two probability distributions, H(P) is the expectation of probability distribution P, H(P)=-∑ x∈X P(x)logP(x), x is the sample space X of probability distribution P Any sample in , P(x) represents the probability that sample x is selected; the expression of D KL (P||Q) is x is any sample in the common sample space X of probability distribution P and Q, P(x) represents the probability that sample x is selected on probability distribution P, and Q(x) represents the probability that sample x is selected on probability distribution Q.
进一步地,为了保证模型训练的效率和准确性,各个所述网络层对应的交叉熵损失函数L(x,:W)为:Further, in order to ensure the efficiency and accuracy of model training, the cross-entropy loss function L(x, :W) for:
其中,x表示模型的输入,表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:Among them, x represents the input of the model, Represents the label corresponding to the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the output of the model after transforming the input x, ζ represents the normalization factor, ||W|| 2 means to sum the matrix elements:
Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,表示第i批输入对应的平均梯度。Among them, ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy item, β is the weight decay coefficient, γ is the learning rate of the model, and W i represents the state value of the weight matrix at time i , Di represents the i-th batch of input, Indicates the average gradient corresponding to the i-th batch of inputs.
本实施例中,交叉熵可在神经网络(机器学习)中作为损失函数,例如,P表示真实标记的分布,Q则为训练后的模型的预测标记分布,交叉熵损失函数可以衡量P与Q的相似性,以保证模型训练的准确性。而且,交叉熵作为损失函数在梯度下降时能避免均方误差损失函数学习速率降低的问题,因此,能保证模型训练的效率。In this embodiment, cross entropy can be used as a loss function in a neural network (machine learning). For example, P represents the distribution of real marks, and Q is the predicted mark distribution of the trained model. The cross entropy loss function can measure P and Q similarity to ensure the accuracy of model training. Moreover, cross entropy as a loss function can avoid the problem of reduced learning rate of the mean square error loss function during gradient descent, so it can ensure the efficiency of model training.
进一步地,在其他实施例中,所述深度卷积神经网络模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围(例如,(0,1)权重范围)进行随机采样确定,所述全连接层的连接权重被丢弃(Dropout)的概率设置为第一预设值(例如,0.5),所述交叉商损失函数中的权值衰减系数设置为第二预设值(例如,0.0005),所述交叉商损失函数中的势能项设置为第三预设值(例如,0.9)。Further, in other embodiments, the deep convolutional neural network model includes at least one fully connected layer, and the initial value of each weight in the predetermined recognition model ranges from a preset weight range (for example, (0, 1) weight range) is determined by random sampling, the probability of the connection weight of the fully connected layer being discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight decay coefficient in the cross quotient loss function It is set to a second preset value (for example, 0.0005), and the potential energy item in the cross quotient loss function is set to a third preset value (for example, 0.9).
进一步地,在其他实施例中,所述预先确定的识别模型的打分函数为:Further, in other embodiments, the scoring function of the predetermined recognition model for:
其中,Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{0,1,2,3,4},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。in, O i, j represents the number of pictures that are actually predicted as i for the first time and j for the second time, O represents an N*N matrix, O i, j represents the matrix elements in matrix O, and N represents the number of pictures participating in the prediction Number of pictures, prediction result i, j ∈ {0, 1, 2, 3, 4}, E i, j represents the number of images that should appear for the first prediction as i and the second prediction as j, E is the expected prediction The N*N matrix of the result, E i,j represent the matrix elements in the matrix E.
本实施例中通过打分函数来检测所述预先确定的识别模型的识别准确率,以保证训练出的所述预先确定的识别模型的识别准确率保持在较高水平,以保证对患者的视网膜病变程度进行准确地识别。In this embodiment, through the scoring function To detect the recognition accuracy of the predetermined recognition model, so as to ensure that the recognition accuracy of the trained predetermined recognition model is maintained at a high level, so as to ensure accurate recognition of the degree of retinopathy of the patient.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It should be noted that, in this document, the term "comprising", "comprising" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件来实现,但很多情况下前者是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本发明各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the methods in the above embodiments can be implemented by means of software plus a necessary general-purpose hardware platform, and of course can also be implemented by hardware, but in many cases the former is better implementation. Based on such an understanding, the essence of the technical solution of the present invention or the part that contributes to the prior art can be embodied in the form of software products, and the computer software products are stored in a storage medium (such as ROM/RAM, disk, CD) contains several instructions to make a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in various embodiments of the present invention.
以上参照附图说明了本发明的优选实施例,并非因此局限本发明的权利范围。上述本发明实施例序号仅仅为了描述,不代表实施例的优劣。另外,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤。The preferred embodiments of the present invention have been described above with reference to the accompanying drawings, and the scope of rights of the present invention is not limited thereto. The serial numbers of the above embodiments of the present invention are for description only, and do not represent the advantages and disadvantages of the embodiments. In addition, although a logical order is shown in the flowcharts, in some cases the steps shown or described may be performed in an order different from that shown or described herein.
本领域技术人员不脱离本发明的范围和实质,可以有多种变型方案实现本发明,比如作为一个实施例的特征可用于另一实施例而得到又一实施例。凡在运用本发明的技术构思之内所作的任何修改、等同替换和改进,均应在本发明的权利范围之内。Those skilled in the art can realize the present invention with many variants without departing from the scope and spirit of the present invention, for example, the features of one embodiment can be used in another embodiment to obtain another embodiment. All modifications, equivalent replacements and improvements made within the technical conception of the application of the present invention shall fall within the scope of rights of the present invention.
Claims (10)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201710312327.6A CN107203778A (en) | 2017-05-05 | 2017-05-05 | PVR intensity grade detecting system and method |
| PCT/CN2017/100044 WO2018201647A1 (en) | 2017-05-05 | 2017-08-31 | Method for detecting retinopathy degree level, device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201710312327.6A CN107203778A (en) | 2017-05-05 | 2017-05-05 | PVR intensity grade detecting system and method |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN107203778A true CN107203778A (en) | 2017-09-26 |
Family
ID=59905737
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201710312327.6A Pending CN107203778A (en) | 2017-05-05 | 2017-05-05 | PVR intensity grade detecting system and method |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN107203778A (en) |
| WO (1) | WO2018201647A1 (en) |
Cited By (28)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107680684A (en) * | 2017-10-12 | 2018-02-09 | 百度在线网络技术(北京)有限公司 | For obtaining the method and device of information |
| CN107729929A (en) * | 2017-09-30 | 2018-02-23 | 百度在线网络技术(北京)有限公司 | For obtaining the method and device of information |
| CN108230296A (en) * | 2017-11-30 | 2018-06-29 | 腾讯科技(深圳)有限公司 | Image feature recognition method and device, storage medium, electronic device |
| CN108392174A (en) * | 2018-04-19 | 2018-08-14 | 梁建宏 | A kind of automatic check method and system of retinopathy of prematurity |
| CN108564570A (en) * | 2018-03-29 | 2018-09-21 | 哈尔滨工业大学(威海) | A kind of method and apparatus of intelligentized pathological tissues positioning |
| CN108595583A (en) * | 2018-04-18 | 2018-09-28 | 平安科技(深圳)有限公司 | Dynamic chart class page data crawling method, device, terminal and storage medium |
| CN109034264A (en) * | 2018-08-15 | 2018-12-18 | 云南大学 | Traffic accident seriousness predicts CSP-CNN model and its modeling method |
| CN109427060A (en) * | 2018-10-30 | 2019-03-05 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus, terminal device and the medical system of image identification |
| CN109567797A (en) * | 2019-01-30 | 2019-04-05 | 浙江强脑科技有限公司 | Epilepsy method for early warning, device and computer readable storage medium |
| CN109726726A (en) * | 2017-10-27 | 2019-05-07 | 北京邮电大学 | Event detection method and device in video |
| WO2019136908A1 (en) * | 2018-01-12 | 2019-07-18 | 平安科技(深圳)有限公司 | Cancer identification method, device and storage medium |
| CN110097966A (en) * | 2018-01-30 | 2019-08-06 | 中国移动通信有限公司研究院 | A kind of information prompting method, device and terminal device |
| WO2019196268A1 (en) * | 2018-04-13 | 2019-10-17 | 博众精工科技股份有限公司 | Diabetic retina image classification method and system based on deep learning |
| CN110720888A (en) * | 2019-10-12 | 2020-01-24 | 杭州电子科技大学 | A method for predicting macular edema in fundus images based on deep learning |
| CN110739071A (en) * | 2019-10-10 | 2020-01-31 | 北京致远慧图科技有限公司 | Determination method and device for optic disc macula lutea combined positioning model and storage medium |
| WO2020061972A1 (en) * | 2018-09-27 | 2020-04-02 | 电子科技大学 | Convolutional neural network-based diabetic retinopathy diagnosing technique |
| CN110956628A (en) * | 2019-12-13 | 2020-04-03 | 广州达安临床检验中心有限公司 | Picture grade classification method and device, computer equipment and storage medium |
| CN110992364A (en) * | 2019-12-31 | 2020-04-10 | 重庆艾可立安医疗器械有限公司 | Retina image recognition method and device, computer equipment and storage medium |
| CN111212591A (en) * | 2017-10-17 | 2020-05-29 | 富士胶片株式会社 | Medical image processing apparatus and endoscope apparatus |
| CN111739616A (en) * | 2020-07-20 | 2020-10-02 | 平安国际智慧城市科技股份有限公司 | Eye image processing method, device, equipment and storage medium |
| WO2020200087A1 (en) * | 2019-03-29 | 2020-10-08 | Ai Technologies Inc. | Image-based detection of ophthalmic and systemic diseases |
| CN112365535A (en) * | 2019-07-26 | 2021-02-12 | 长佳智能股份有限公司 | Retinal lesion evaluation model establishing method and system |
| US10963757B2 (en) | 2018-12-14 | 2021-03-30 | Industrial Technology Research Institute | Neural network model fusion method and electronic device using the same |
| CN113569612A (en) * | 2021-02-09 | 2021-10-29 | 腾讯科技(深圳)有限公司 | Training method and device for image recognition neural network and image recognition method |
| CN114078112A (en) * | 2020-10-16 | 2022-02-22 | 武汉联影医疗科技有限公司 | Method and device for detecting liver degeneration degree, computer equipment and storage medium |
| US20220207732A1 (en) * | 2020-12-28 | 2022-06-30 | Seyed Ehsan Vaghefi Rezaei | Systems and methods for processing of fundus images |
| WO2022145541A1 (en) * | 2020-12-30 | 2022-07-07 | 주식회사 메디웨일 | Method and device for assisting renal disease diagnosis |
| CN117197053A (en) * | 2023-08-21 | 2023-12-08 | 浙江大学 | Automatic identification and treatment prompt method of phlebitis grading based on AI intelligent model |
Families Citing this family (21)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111612021B (en) * | 2019-02-22 | 2023-10-31 | 中国移动通信有限公司研究院 | An error sample identification method, device and terminal |
| CN110211109B (en) * | 2019-05-30 | 2022-12-06 | 西安电子科技大学 | Image change detection method based on deep neural network structure optimization |
| CN110517225B (en) * | 2019-07-19 | 2023-07-11 | 平安科技(深圳)有限公司 | AI image recognition method, apparatus, device and storage medium |
| CN110414607A (en) * | 2019-07-31 | 2019-11-05 | 中山大学 | Classification method, device, equipment and medium for capsule endoscopy images |
| CN110599451B (en) * | 2019-08-05 | 2023-01-20 | 平安科技(深圳)有限公司 | Medical image focus detection and positioning method, device, equipment and storage medium |
| CN110555478B (en) * | 2019-09-05 | 2023-02-03 | 东北大学 | Multi-fault diagnosis method for fan based on deep metric network based on hard sample mining |
| CN113191478B (en) * | 2020-01-14 | 2025-02-14 | 阿里巴巴集团控股有限公司 | Training method, device and system for neural network model |
| CN111242933B (en) * | 2020-01-15 | 2023-06-20 | 华南理工大学 | Device, equipment and storage medium for classifying retinal image arteries and veins |
| CN111275121B (en) * | 2020-01-23 | 2023-07-18 | 北京康夫子健康技术有限公司 | Medical image processing method and device and electronic equipment |
| CN111696101A (en) * | 2020-06-18 | 2020-09-22 | 中国农业大学 | Light-weight solanaceae disease identification method based on SE-Inception |
| CN111754486B (en) * | 2020-06-24 | 2023-08-15 | 北京百度网讯科技有限公司 | Image processing method, device, electronic equipment and storage medium |
| CN111914814A (en) * | 2020-09-01 | 2020-11-10 | 平安国际智慧城市科技股份有限公司 | Wheat rust detection method and device and computer equipment |
| CN112052904A (en) * | 2020-09-09 | 2020-12-08 | 陕西理工大学 | An identification method of pests and diseases based on transfer learning and convolutional neural network |
| CN112215239B (en) * | 2020-09-15 | 2024-11-12 | 浙江工业大学 | A method and device for fine-grained grading of retinal lesions |
| JP2023551899A (en) * | 2020-12-04 | 2023-12-13 | ジェネンテック, インコーポレイテッド | Automated screening of diabetic retinopathy severity using color fundus image data |
| CN112581448A (en) * | 2020-12-17 | 2021-03-30 | 中南大学 | Retina image hard exudate identification method and imaging method |
| CN112819797B (en) * | 2021-02-06 | 2023-09-19 | 国药集团基因科技有限公司 | Method, device, system and storage medium for analyzing diabetic retinopathy |
| TWI790572B (en) * | 2021-03-19 | 2023-01-21 | 宏碁智醫股份有限公司 | Detecting method and detecting apparatus related to image |
| CN113658178B (en) * | 2021-10-14 | 2022-01-25 | 北京字节跳动网络技术有限公司 | Tissue image identification method and device, readable medium and electronic equipment |
| CN115311496A (en) * | 2022-08-09 | 2022-11-08 | 长沙理工大学 | Method, system, device and storage medium for grading diabetic retinopathy |
| CN118345334B (en) * | 2024-06-17 | 2024-08-23 | 华兴源创(成都)科技有限公司 | Film thickness correction method and device and computer equipment |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106231990A (en) * | 2014-03-14 | 2016-12-14 | Lkc科技有限公司 | Retinopathy detecting system and method |
| CN106295547A (en) * | 2016-08-05 | 2017-01-04 | 深圳市商汤科技有限公司 | A kind of image comparison method and image comparison device |
| CN106570530A (en) * | 2016-11-10 | 2017-04-19 | 西南交通大学 | Extraction method for extracting hard exudates in ophthalmoscopic image |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10115194B2 (en) * | 2015-04-06 | 2018-10-30 | IDx, LLC | Systems and methods for feature detection in retinal images |
| US10405739B2 (en) * | 2015-10-23 | 2019-09-10 | International Business Machines Corporation | Automatically detecting eye type in retinal fundus images |
| CN106530295A (en) * | 2016-11-07 | 2017-03-22 | 首都医科大学 | Fundus image classification method and device of retinopathy |
| CN106934798B (en) * | 2017-02-20 | 2020-08-21 | 苏州体素信息科技有限公司 | Diabetic retinopathy classification and classification method based on deep learning |
-
2017
- 2017-05-05 CN CN201710312327.6A patent/CN107203778A/en active Pending
- 2017-08-31 WO PCT/CN2017/100044 patent/WO2018201647A1/en not_active Ceased
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106231990A (en) * | 2014-03-14 | 2016-12-14 | Lkc科技有限公司 | Retinopathy detecting system and method |
| CN106295547A (en) * | 2016-08-05 | 2017-01-04 | 深圳市商汤科技有限公司 | A kind of image comparison method and image comparison device |
| CN106570530A (en) * | 2016-11-10 | 2017-04-19 | 西南交通大学 | Extraction method for extracting hard exudates in ophthalmoscopic image |
Non-Patent Citations (2)
| Title |
|---|
| HARRY PRATT等: "《Convolutional Neural Networks for Diabetic Retinopathy》", 《PROCEDIA COMPUTER SCIENCE》 * |
| 冯裕忠: "《计算机导论》", 30 September 2016 * |
Cited By (46)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107729929A (en) * | 2017-09-30 | 2018-02-23 | 百度在线网络技术(北京)有限公司 | For obtaining the method and device of information |
| US10755410B2 (en) | 2017-09-30 | 2020-08-25 | Baidu Online Network Technology (Beijing) Co., Ltd. | Method and apparatus for acquiring information |
| CN107680684B (en) * | 2017-10-12 | 2021-05-07 | 百度在线网络技术(北京)有限公司 | Method and device for acquiring information |
| CN107680684A (en) * | 2017-10-12 | 2018-02-09 | 百度在线网络技术(北京)有限公司 | For obtaining the method and device of information |
| CN111212591B (en) * | 2017-10-17 | 2022-08-02 | 富士胶片株式会社 | Medical image processing apparatus and endoscope apparatus |
| US11574401B2 (en) | 2017-10-17 | 2023-02-07 | Fujifilm Corporation | Medical image processing apparatus and endoscope apparatus |
| CN111212591A (en) * | 2017-10-17 | 2020-05-29 | 富士胶片株式会社 | Medical image processing apparatus and endoscope apparatus |
| CN109726726A (en) * | 2017-10-27 | 2019-05-07 | 北京邮电大学 | Event detection method and device in video |
| CN109726726B (en) * | 2017-10-27 | 2023-06-20 | 北京邮电大学 | Event detection method and device in video |
| US11526983B2 (en) | 2017-11-30 | 2022-12-13 | Tencent Technology (Shenzhen) Company Limited | Image feature recognition method and apparatus, storage medium, and electronic apparatus |
| CN110349156B (en) * | 2017-11-30 | 2023-05-30 | 腾讯科技(深圳)有限公司 | Method, device, and storage medium for identifying lesion features in fundus images |
| CN108230296A (en) * | 2017-11-30 | 2018-06-29 | 腾讯科技(深圳)有限公司 | Image feature recognition method and device, storage medium, electronic device |
| CN108230296B (en) * | 2017-11-30 | 2023-04-07 | 腾讯科技(深圳)有限公司 | Image feature recognition method and device, storage medium and electronic device |
| CN110349156A (en) * | 2017-11-30 | 2019-10-18 | 腾讯科技(深圳)有限公司 | The recognition methods of lesion characteristics and device, storage medium in the picture of eyeground |
| WO2019105218A1 (en) * | 2017-11-30 | 2019-06-06 | 腾讯科技(深圳)有限公司 | Recognition method and device for image feature, storage medium and electronic device |
| WO2019136908A1 (en) * | 2018-01-12 | 2019-07-18 | 平安科技(深圳)有限公司 | Cancer identification method, device and storage medium |
| CN110097966B (en) * | 2018-01-30 | 2021-09-14 | 中国移动通信有限公司研究院 | Information reminding method and device and terminal equipment |
| CN110097966A (en) * | 2018-01-30 | 2019-08-06 | 中国移动通信有限公司研究院 | A kind of information prompting method, device and terminal device |
| CN108564570A (en) * | 2018-03-29 | 2018-09-21 | 哈尔滨工业大学(威海) | A kind of method and apparatus of intelligentized pathological tissues positioning |
| US11132799B2 (en) | 2018-04-13 | 2021-09-28 | Bozhon Precision Industry Technology Co., Ltd. | Method and system for classifying diabetic retina images based on deep learning |
| WO2019196268A1 (en) * | 2018-04-13 | 2019-10-17 | 博众精工科技股份有限公司 | Diabetic retina image classification method and system based on deep learning |
| CN108595583A (en) * | 2018-04-18 | 2018-09-28 | 平安科技(深圳)有限公司 | Dynamic chart class page data crawling method, device, terminal and storage medium |
| CN108392174A (en) * | 2018-04-19 | 2018-08-14 | 梁建宏 | A kind of automatic check method and system of retinopathy of prematurity |
| CN109034264B (en) * | 2018-08-15 | 2021-11-19 | 云南大学 | CSP-CNN model for predicting severity of traffic accident and modeling method thereof |
| CN109034264A (en) * | 2018-08-15 | 2018-12-18 | 云南大学 | Traffic accident seriousness predicts CSP-CNN model and its modeling method |
| WO2020061972A1 (en) * | 2018-09-27 | 2020-04-02 | 电子科技大学 | Convolutional neural network-based diabetic retinopathy diagnosing technique |
| US11410306B2 (en) | 2018-10-30 | 2022-08-09 | Tencent Technology (Shenzhen) Company Limited | Method, apparatus, system, and storage medium for recognizing medical image |
| CN109427060A (en) * | 2018-10-30 | 2019-03-05 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus, terminal device and the medical system of image identification |
| US11610310B2 (en) | 2018-10-30 | 2023-03-21 | Tencent Technology (Shenzhen) Company Limited | Method, apparatus, system, and storage medium for recognizing medical image |
| US10963757B2 (en) | 2018-12-14 | 2021-03-30 | Industrial Technology Research Institute | Neural network model fusion method and electronic device using the same |
| CN109567797A (en) * | 2019-01-30 | 2019-04-05 | 浙江强脑科技有限公司 | Epilepsy method for early warning, device and computer readable storage medium |
| WO2020200087A1 (en) * | 2019-03-29 | 2020-10-08 | Ai Technologies Inc. | Image-based detection of ophthalmic and systemic diseases |
| CN112365535A (en) * | 2019-07-26 | 2021-02-12 | 长佳智能股份有限公司 | Retinal lesion evaluation model establishing method and system |
| CN110739071B (en) * | 2019-10-10 | 2022-05-31 | 北京致远慧图科技有限公司 | Method and device for determining optic disc and yellow spot combined positioning model and storage medium |
| CN110739071A (en) * | 2019-10-10 | 2020-01-31 | 北京致远慧图科技有限公司 | Determination method and device for optic disc macula lutea combined positioning model and storage medium |
| CN110720888A (en) * | 2019-10-12 | 2020-01-24 | 杭州电子科技大学 | A method for predicting macular edema in fundus images based on deep learning |
| CN110956628A (en) * | 2019-12-13 | 2020-04-03 | 广州达安临床检验中心有限公司 | Picture grade classification method and device, computer equipment and storage medium |
| CN110992364A (en) * | 2019-12-31 | 2020-04-10 | 重庆艾可立安医疗器械有限公司 | Retina image recognition method and device, computer equipment and storage medium |
| CN110992364B (en) * | 2019-12-31 | 2023-11-28 | 重庆艾可立安医疗器械有限公司 | Retina image recognition method, retina image recognition device, computer equipment and storage medium |
| CN111739616B (en) * | 2020-07-20 | 2020-12-01 | 平安国际智慧城市科技股份有限公司 | Eye image processing method, device, equipment and storage medium |
| CN111739616A (en) * | 2020-07-20 | 2020-10-02 | 平安国际智慧城市科技股份有限公司 | Eye image processing method, device, equipment and storage medium |
| CN114078112A (en) * | 2020-10-16 | 2022-02-22 | 武汉联影医疗科技有限公司 | Method and device for detecting liver degeneration degree, computer equipment and storage medium |
| US20220207732A1 (en) * | 2020-12-28 | 2022-06-30 | Seyed Ehsan Vaghefi Rezaei | Systems and methods for processing of fundus images |
| WO2022145541A1 (en) * | 2020-12-30 | 2022-07-07 | 주식회사 메디웨일 | Method and device for assisting renal disease diagnosis |
| CN113569612A (en) * | 2021-02-09 | 2021-10-29 | 腾讯科技(深圳)有限公司 | Training method and device for image recognition neural network and image recognition method |
| CN117197053A (en) * | 2023-08-21 | 2023-12-08 | 浙江大学 | Automatic identification and treatment prompt method of phlebitis grading based on AI intelligent model |
Also Published As
| Publication number | Publication date |
|---|---|
| WO2018201647A1 (en) | 2018-11-08 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN107203778A (en) | PVR intensity grade detecting system and method | |
| CN113159147B (en) | Image recognition method and device based on neural network and electronic equipment | |
| CN107766809B (en) | Electronic device, ticket information identification method, and computer-readable storage medium | |
| US11259718B1 (en) | Systems and methods for automated body mass index calculation to determine value | |
| WO2019174130A1 (en) | Bill recognition method, server, and computer readable storage medium | |
| EP4024270A1 (en) | Gesture recognition method, electronic device, computer-readable storage medium, and chip | |
| WO2022105179A1 (en) | Biological feature image recognition method and apparatus, and electronic device and readable storage medium | |
| CN110276366A (en) | Carry out test object using Weakly supervised model | |
| WO2018205467A1 (en) | Automobile damage part recognition method, system and electronic device and storage medium | |
| WO2018166114A1 (en) | Picture identification method and system, electronic device, and medium | |
| US10748217B1 (en) | Systems and methods for automated body mass index calculation | |
| WO2019071660A1 (en) | Bill information identification method, electronic device, and readable storage medium | |
| CN108345641A (en) | A kind of method crawling website data, storage medium and server | |
| CN110033018A (en) | Graphic similarity judgment method, device and computer-readable storage medium | |
| EP4492331A1 (en) | Image processing method, neural network training method and related device | |
| CN113705460A (en) | Method, device and equipment for detecting opening and closing of eyes of human face in image and storage medium | |
| WO2021189856A1 (en) | Certificate check method and apparatus, and electronic device and medium | |
| CN114049568B (en) | Target object deformation detection method, device, equipment and medium based on image comparison | |
| WO2019056503A1 (en) | Store monitoring evaluation method, device and storage medium | |
| WO2020248848A1 (en) | Intelligent abnormal cell determination method and device, and computer readable storage medium | |
| CN113627394B (en) | Face extraction method, device, electronic equipment and readable storage medium | |
| WO2023123926A1 (en) | Artificial intelligence task processing method and apparatus, electronic device, and readable storage medium | |
| Tuyet et al. | A Deep Bottleneck U-Net Combined with Saliency Map for Classifying Diabetic Retinopathy in Fundus Images. | |
| CN114519401B (en) | Image classification method and device, electronic device, and storage medium | |
| US10896339B2 (en) | Detecting magnetic ink character recognition codes |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| REG | Reference to a national code |
Ref country code: HK Ref legal event code: DE Ref document number: 1244088 Country of ref document: HK |
|
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170926 |
|
| REG | Reference to a national code |
Ref country code: HK Ref legal event code: WD Ref document number: 1244088 Country of ref document: HK |