CN110458136B - Traffic sign identification method, device and equipment - Google Patents
Traffic sign identification method, device and equipment Download PDFInfo
- Publication number
- CN110458136B CN110458136B CN201910764738.8A CN201910764738A CN110458136B CN 110458136 B CN110458136 B CN 110458136B CN 201910764738 A CN201910764738 A CN 201910764738A CN 110458136 B CN110458136 B CN 110458136B
- Authority
- CN
- China
- Prior art keywords
- traffic sign
- trained
- image
- neural network
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
本申请公开了一种交通标志识别方法、装置和设备,其中方法包括:获取待识别交通标志图像;将待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待识别交通标志图像对应的第一特征向量;将第一特征向量转换成第一脉冲序列;将第一脉冲序列输入到训练好的脉冲神经网络中,获取训练好的脉冲神经网络输出的识别结果。本申请采用深度信念网络模型与脉冲神经网络相结合的方式进行交通标志识别,不需要进行人工特征提取,大大减少了人工干预,提高了识别速度,通过充分利用深度信念网络模型和脉冲神经网络的优点,提高了识别结果,解决了现有的交通标志识别准确度低、速度慢的技术问题。
The present application discloses a traffic sign recognition method, device and device, wherein the method includes: acquiring an image of a traffic sign to be recognized; inputting the image of the traffic sign to be recognized into a trained deep belief network model for feature extraction to obtain the traffic sign to be recognized Mark the first feature vector corresponding to the image; convert the first feature vector into a first pulse sequence; input the first pulse sequence into the trained spiking neural network, and obtain the recognition result output by the trained spiking neural network. This application uses the combination of the deep belief network model and the spiking neural network for traffic sign recognition, which does not require manual feature extraction, greatly reduces manual intervention, and improves the recognition speed. By making full use of the deep belief network model and the spiking neural network The invention has the advantages of improving the recognition result and solving the technical problems of low accuracy and slow speed of the existing traffic sign recognition.
Description
技术领域technical field
本申请涉及图像识别技术领域,尤其涉及一种交通标志识别方法、装置和设备。The present application relates to the technical field of image recognition, and in particular, to a traffic sign recognition method, device and device.
背景技术Background technique
交通标志是车辆行驶过程中获得信息的重要来源,对交通标志准确、快速识别对于保障交通安全、交通秩序,提高交通效率有着重要意义,也对目前正在兴起的无人驾驶研究有重要意义。Traffic signs are an important source of information obtained during vehicle driving. Accurate and rapid identification of traffic signs is of great significance for ensuring traffic safety, traffic order, and improving traffic efficiency.
传统的交通标志识别方法是利用图像匹配、特征提取与分类器结合等方法进行识别,人为干预较多,存在识别准确度低、速度慢的问题。The traditional traffic sign recognition method uses image matching, feature extraction and classifier combination for recognition, which requires more human intervention, and has the problems of low recognition accuracy and slow speed.
发明内容SUMMARY OF THE INVENTION
本申请提供了一种交通标志识别方法、装置和设备,用于解决现有的交通标志识别准确度低、速度慢的技术问题。The present application provides a traffic sign recognition method, device and device, which are used to solve the technical problems of low accuracy and slow speed of existing traffic sign recognition.
有鉴于此,本申请第一方面提供了一种交通标志识别方法,包括:In view of this, a first aspect of the present application provides a traffic sign recognition method, including:
获取待识别交通标志图像;Obtain the image of the traffic sign to be recognized;
将所述待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到所述待识别交通标志图像对应的第一特征向量;Inputting the traffic sign image to be recognized into the trained deep belief network model for feature extraction to obtain a first feature vector corresponding to the traffic sign image to be recognized;
将所述第一特征向量转换成第一脉冲序列;converting the first feature vector into a first pulse sequence;
将所述第一脉冲序列输入到训练好的脉冲神经网络中,获取所述训练好的脉冲神经网络输出的识别结果。The first pulse sequence is input into the trained spiking neural network, and the recognition result output by the trained spiking neural network is obtained.
优选地,还包括:Preferably, it also includes:
获取待训练交通标志图像集;Obtain the traffic sign image set to be trained;
将所述待训练交通标志图像集中的待训练交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到所述待训练交通标志图像对应的第二特征向量;Inputting the to-be-trained traffic-sign images in the to-be-trained traffic-sign image set into the trained deep belief network model for feature extraction to obtain a second feature vector corresponding to the to-be-trained traffic sign images;
基于时滞相位编码对所述第二特征向量进行转换,得到第二脉冲序列;Converting the second eigenvector based on time-delay phase coding to obtain a second pulse sequence;
将所述第二脉冲序列输入到脉冲神经网络中,对所述脉冲神经网络进行训练;inputting the second pulse sequence into a spiking neural network to train the spiking neural network;
计算所述脉冲神经网络对所述待训练交通标志图像的识别准确率,当所述识别准确率高于阈值时,训练完成,得到训练好的脉冲神经网络。Calculate the recognition accuracy rate of the spiking neural network for the traffic sign image to be trained, when the recognition accuracy rate is higher than a threshold, the training is completed, and a trained spiking neural network is obtained.
优选地,所述将所述待训练交通标志图像集中的待训练交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到所述待训练交通标志图像对应的第二特征向量,之前还包括:Preferably, the to-be-trained traffic sign images in the to-be-trained traffic sign image set are input into the trained deep belief network model to perform feature extraction, and the second feature vector corresponding to the to-be-trained traffic sign images is obtained. Also includes:
对所述待训练交通标志图像进行预处理。Perform preprocessing on the traffic sign image to be trained.
优选地,所述预处理包括:Preferably, the preprocessing includes:
基于双线性插值算法对所述待训练交通标志图像进行尺寸归一化处理,得到归一化处理后的待训练交通标志图像。The size of the traffic sign image to be trained is normalized based on the bilinear interpolation algorithm, and the normalized traffic sign image to be trained is obtained.
优选地,所述将所述第二脉冲序列输入到脉冲神经网络中,对所述脉冲神经网络进行训练,包括:Preferably, inputting the second pulse sequence into a spiking neural network, and training the spiking neural network, includes:
将所述第二脉冲序列输入到脉冲神经网络中,基于三脉冲STDP与阈值可塑性结合的学习方法对所述脉冲神经网络进行训练。The second pulse sequence is input into a spiking neural network, and the spiking neural network is trained based on a learning method combining three-pulse STDP and threshold plasticity.
本申请第二方面提供了一种交通标志识别装置,包括:A second aspect of the present application provides a traffic sign recognition device, comprising:
第一图像获取模块,用于获取待识别交通标志图像;a first image acquisition module, configured to acquire an image of a traffic sign to be recognized;
第一特征提取模块,用于将所述待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到所述待识别交通标志图像对应的第一特征向量;a first feature extraction module, configured to input the traffic sign image to be recognized into the trained deep belief network model for feature extraction, and obtain a first feature vector corresponding to the traffic sign image to be recognized;
第一转换模块,用于将所述第一特征向量转换成第一脉冲序列;a first conversion module for converting the first feature vector into a first pulse sequence;
识别模块,用于将所述第一脉冲序列输入到训练好的脉冲神经网络中,获取所述训练好的脉冲神经网络输出的识别结果。The identification module is configured to input the first pulse sequence into the trained spiking neural network, and obtain the recognition result output by the trained spiking neural network.
优选地,还包括:Preferably, it also includes:
第二图像获取模块,用于获取待训练交通标志图像集;The second image acquisition module is used to acquire the traffic sign image set to be trained;
第二特征提取模块,用于将所述待训练交通标志图像集中的待训练交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到所述待训练交通标志图像对应的第二特征向量;The second feature extraction module is used to input the to-be-trained traffic sign images in the to-be-trained traffic sign image set into the trained deep belief network model for feature extraction to obtain the second feature corresponding to the to-be-trained traffic sign images vector;
第二转换模块,用于基于时滞相位编码对所述第二特征向量进行转换,得到第二脉冲序列;a second conversion module, configured to convert the second eigenvector based on time-delay phase coding to obtain a second pulse sequence;
训练模块,用于将所述第二脉冲序列输入到脉冲神经网络中,对所述脉冲神经网络进行训练;a training module, configured to input the second pulse sequence into the spiking neural network, and train the spiking neural network;
计算模块,用于计算所述脉冲神经网络对所述待训练交通标志图像的识别准确率,当所述识别准确率高于阈值时,训练完成,得到训练好的脉冲神经网络。The calculation module is used for calculating the recognition accuracy rate of the spiking neural network for the traffic sign image to be trained. When the recognition accuracy rate is higher than a threshold, the training is completed and a trained spiking neural network is obtained.
优选地,还包括:Preferably, it also includes:
预处理模块,用于对所述待训练交通标志图像进行预处理。The preprocessing module is used for preprocessing the traffic sign image to be trained.
本申请第三方面提供了一种交通标志识别设备,其特征在于,包括:所述设备包括处理器以及存储器;A third aspect of the present application provides a traffic sign recognition device, characterized in that: the device includes a processor and a memory;
所述存储器用于存储程序代码,并将所述程序代码传输给所述处理器;the memory is used to store program code and transmit the program code to the processor;
所述处理器用于根据所述程序代码中的指令执行第一方面任一项所述的交通标志识别方法。The processor is configured to execute the traffic sign recognition method according to any one of the first aspects according to the instructions in the program code.
本申请第四方面提供了一种计算机可读存储介质,其特征在于,所述计算机可读存储介质用于存储程序代码,所述程序代码用于执行第一方面任一项所述的交通标志识别方法。A fourth aspect of the present application provides a computer-readable storage medium, wherein the computer-readable storage medium is used to store program codes, and the program codes are used to execute the traffic sign according to any one of the first aspect recognition methods.
从以上技术方案可以看出,本申请实施例具有以下优点:As can be seen from the above technical solutions, the embodiments of the present application have the following advantages:
本申请中,提供了一种交通标志识别方法,包括:获取待识别交通标志图像;将待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待识别交通标志图像对应的第一特征向量;将第一特征向量转换成第一脉冲序列;将第一脉冲序列输入到训练好的脉冲神经网络中,获取训练好的脉冲神经网络输出的识别结果。本申请利用训练好的深度信念网络模型对交通标志图像进行特征提取,不需要进行人工特征提取,通过深度信念网络模型对输入的交通标志图像进行特征降维和特征选择,大大减少了人工干预,从原始的交通标志图像中提取到高层次的特征,并筛选掉了冗余的噪声信息,有助于提高后续脉冲神经网络的识别结果,采用深度信念网络模型与脉冲神经网络相结合的方式进行交通标志识别,提高了识别速度,通过充分利用深度信念网络模型和脉冲神经网络的优点,提高了识别结果,解决了现有的交通标志识别准确度低、速度慢的技术问题。In this application, a traffic sign recognition method is provided, which includes: acquiring an image of a traffic sign to be recognized; inputting the image of the traffic sign to be recognized into a trained deep belief network model for feature extraction, and obtaining a corresponding image of the traffic sign to be recognized. The first feature vector; converting the first feature vector into a first pulse sequence; inputting the first pulse sequence into the trained spiking neural network to obtain the recognition result output by the trained spiking neural network. This application uses the trained deep belief network model to perform feature extraction on traffic sign images, and does not require manual feature extraction. The deep belief network model is used to perform feature dimension reduction and feature selection on the input traffic sign images, which greatly reduces manual intervention. High-level features are extracted from the original traffic sign image, and redundant noise information is filtered out, which helps to improve the recognition results of the subsequent spiking neural network. The combination of the deep belief network model and the spiking neural network is used to conduct traffic. Sign recognition improves the recognition speed. By making full use of the advantages of the deep belief network model and the spiking neural network, the recognition results are improved, and the existing technical problems of low accuracy and slow speed of traffic sign recognition are solved.
附图说明Description of drawings
图1为本申请提供的一种交通标志识别方法的一个实施例的流程示意图;1 is a schematic flowchart of an embodiment of a traffic sign recognition method provided by the application;
图2为本申请提供的一种交通标志识别方法的另一个实施例的流程示意图;2 is a schematic flowchart of another embodiment of a traffic sign recognition method provided by the present application;
图3为本申请提供的一种交通标志识别装置的一个实施例的结构示意图。FIG. 3 is a schematic structural diagram of an embodiment of a traffic sign recognition device provided by the present application.
具体实施方式Detailed ways
为了使本技术领域的人员更好地理解本申请方案,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make those skilled in the art better understand the solutions of the present application, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are only It is a part of the embodiments of the present application, but not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.
为了便于理解,请参阅图1,本申请提供的一种交通标志识别方法的一个实施例,包括:For ease of understanding, please refer to FIG. 1, an embodiment of a traffic sign recognition method provided by the present application, including:
步骤101,获取待识别交通标志图像。
需要说明的是,由于可能会存在获取到的图像中有不符合要求的图像,即存在不含有交通标志的图像,为了不对识别的结果产生影响,可以对获取的交通标志图像进行筛选,筛选掉一些不包含交通标志的图像以及一些模糊不清楚的交通标志图像,将符合要求的交通标志图像作为最终获取到的待识别交通标志图像。It should be noted that, since there may be images that do not meet the requirements in the acquired images, that is, there are images that do not contain traffic signs, in order not to affect the recognition results, the acquired images of traffic signs can be screened and filtered out. For some images that do not contain traffic signs and some blurred and unclear traffic sign images, the traffic sign images that meet the requirements are used as the final acquired traffic sign images to be recognized.
步骤102,将待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待识别交通标志图像对应的第一特征向量。Step 102: Input the traffic sign image to be recognized into the trained deep belief network model for feature extraction, and obtain a first feature vector corresponding to the traffic sign image to be recognized.
需要说明的是,本实施例中的深度信念网络模型(deep belief network,DBN)由两层深度玻尔兹曼机模型(deep Boltzmann machine,DBM)和两层DBN模型融合构成,其中,DBM模型和DBN模型均以受限玻尔兹曼机(Restricted Boltzmann Machine,RBM)为基本构成单元,不同之处在于,DBM层间是无向连接,而DBN层间是有向连接。It should be noted that the deep belief network (DBN) in this embodiment is composed of a two-layer deep Boltzmann machine (DBM) model and a two-layer DBN model, wherein the DBM model Both the Restricted Boltzmann Machine (RBM) model and the DBN model are based on the basic unit, the difference is that the DBM layers are undirected connections, while the DBN layers are directed connections.
深度信念网络模型中的低层采用信息还原度强的两层DBM对待识别交通标志图像进行初步的降维,得到去噪声后的和完整度较高的特征,将得到的特征作为两层DBN的输入,然后通过两层DBN来提取更高层特征。通过对DBM和DBN分别进行无监督训练和有监督微调,最终得到训练好的深度信念网络模型,通过训练好的深度信念网络模型,提取待识别交通标志图像的高层特征,有助于提高后续脉冲神经网络的识别结果。The lower layer in the deep belief network model adopts the two-layer DBM with strong information reduction degree to perform preliminary dimensionality reduction on the traffic sign image to be recognized, and obtains the features after denoising and high integrity, and uses the obtained features as the input of the two-layer DBN. , and then extract higher-level features through a two-layer DBN. Through unsupervised training and supervised fine-tuning of DBM and DBN, respectively, a trained deep belief network model is finally obtained. Through the trained deep belief network model, the high-level features of the traffic sign image to be recognized are extracted, which helps to improve the subsequent pulse rate. The recognition result of the neural network.
步骤103,将第一特征向量转换成第一脉冲序列。Step 103: Convert the first feature vector into a first pulse sequence.
需要说明的是,脉冲神经网络的输入用脉冲序列表示,所以需要将提取的第一特征向量转换成第一脉冲序列,从而能够与后续采用脉冲神经网络进行交通标志识别相适应,使得脉冲神经网络能够更好地进行识别。It should be noted that the input of the spiking neural network is represented by a pulse sequence, so it is necessary to convert the extracted first feature vector into a first pulse sequence, so as to be compatible with the subsequent use of the spiking neural network for traffic sign recognition, so that the spiking neural network can be used for traffic sign recognition. better identification.
步骤104,将第一脉冲序列输入到训练好的脉冲神经网络中,获取训练好的脉冲神经网络输出的识别结果。Step 104: Input the first pulse sequence into the trained spiking neural network, and obtain the recognition result output by the trained spiking neural network.
需要说明的是,传统的人工神经网络是对生物神经元的脉冲发放频率进行编码,神经元的输出一般为给定区间的模拟,其运算能力和生物真实性弱于脉冲神经网络,而采用训练好的脉冲神经网络进行交通标志识别有利于提高识别结果,因此,本实施例采用训练好的脉冲神经网络进行交通标志识别。It should be noted that the traditional artificial neural network encodes the pulse firing frequency of biological neurons. The output of the neuron is generally a simulation of a given interval, and its computing power and biological authenticity are weaker than those of the pulse neural network. A good spiking neural network for traffic sign recognition is beneficial to improve the recognition result. Therefore, in this embodiment, a trained spiking neural network is used for traffic sign recognition.
申请人发现,现有技术中采用图像匹配、特征提取与分类器结合的方法,人工干预较多,存在识别准确度低、速度慢的问题。因此,为解决现有技术中存在的这些问题,申请人提出本申请实施例中提供的交通标志识别方法,该方法达到了以下技术效果:The applicant found that the method of combining image matching, feature extraction and classifier in the prior art requires a lot of manual intervention, and has the problems of low recognition accuracy and slow speed. Therefore, in order to solve these problems existing in the prior art, the applicant proposes the traffic sign recognition method provided in the embodiment of the present application, and the method achieves the following technical effects:
本申请实施例提供的交通标志识别方法,通过利用训练好的深度信念网络模型对交通标志图像进行特征提取,不需要进行人工特征提取,通过深度信念网络模型对输入的交通标志图像进行特征降维和特征选择,大大减少了人工干预,从原始的交通标志图像中提取到高层次的特征,并筛选掉了冗余的噪声信息,有助于提高后续脉冲神经网络的识别结果,采用深度信念网络模型与脉冲神经网络相结合的方式进行交通标志识别,提高了识别速度,通过充分利用深度信念网络模型和脉冲神经网络的优点,提高了识别结果,解决了现有的交通标志识别准确度低、速度慢的技术问题。The traffic sign recognition method provided by the embodiment of the present application uses the trained deep belief network model to perform feature extraction on the traffic sign image without manual feature extraction, and performs feature dimension reduction and summation on the input traffic sign image through the deep belief network model. Feature selection, which greatly reduces manual intervention, extracts high-level features from the original traffic sign image, and filters out redundant noise information, which helps to improve the recognition results of subsequent spiking neural networks, using a deep belief network model The method of combining with the spiking neural network for traffic sign recognition improves the recognition speed. By making full use of the advantages of the deep belief network model and the spiking neural network, the recognition results are improved, and the existing traffic sign recognition accuracy is low and the speed is improved. Slow technical issues.
为了便于理解,请参阅图2,本申请提供的一种交通标志识别方法的另一个实施例,包括:For ease of understanding, please refer to FIG. 2, another embodiment of a traffic sign recognition method provided by the present application, including:
步骤201,获取待训练交通标志图像集。Step 201: Obtain a traffic sign image set to be trained.
需要说明的是,本实施例中的待训练交通标志图像集的待训练交通标志图像来自于德国交通标志识别数据库(German traffic sign recognition benchmark,GTSRB)。It should be noted that the to-be-trained traffic sign images of the to-be-trained traffic sign image set in this embodiment come from a German traffic sign recognition benchmark (GTSRB).
步骤202,对待训练交通标志图像集中的待训练交通标志图像进行预处理。
需要说明的是,为了方便深度信念网络进行特征提取,可以对待训练交通标志图像集中的待训练交通标志图像进行尺寸归一化处理,可以采用双线性插值算法对待训练交通标志图像进行尺寸归一化,归一化后的待训练交通标志图像为同样大小,例如大小为48×48或28×28像素。It should be noted that, in order to facilitate the feature extraction of the deep belief network, the size of the to-be-trained traffic sign images in the training traffic sign image set can be normalized, and the bilinear interpolation algorithm can be used to normalize the size of the to-be-trained traffic sign images. The normalized traffic sign images to be trained are of the same size, for example, the size is 48×48 or 28×28 pixels.
GTSRB数据库中有部分模糊、大面积水印的交通标志图像,可以对其进行筛选,筛选掉质量低的交通标志图像,留下高质量的交通标志图像,有利于深度信念网络提取到有益的特征信息,从而提高脉冲神经网络的识别结果。There are some blurred and large-area watermarked traffic sign images in the GTSRB database, which can be screened to filter out low-quality traffic sign images, leaving high-quality traffic sign images, which is conducive to the deep belief network to extract useful feature information , so as to improve the recognition results of the spiking neural network.
步骤203,将预处理后的待训练交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待训练交通标志图像对应的第二特征向量。Step 203: Input the preprocessed traffic sign image to be trained into the trained deep belief network model for feature extraction to obtain a second feature vector corresponding to the traffic sign image to be trained.
步骤204,基于时滞相位编码对第二特征向量进行转换,得到第二脉冲序列。Step 204: Convert the second eigenvector based on the time-delay phase encoding to obtain a second pulse sequence.
需要说明的是,常用的编码方式有时滞编码、相位编码,本实施例中采用时滞编码和相位编码相结合的方式对提取的第二特征向量进行编码,即时滞相位编码,采用时滞相位编码可以生成更好的脉冲序列,有助于提高脉冲神经网络的交通标志识别结果。It should be noted that the commonly used encoding methods are time-delay encoding and phase encoding. In this embodiment, a combination of time-delay encoding and phase encoding is used to encode the extracted second eigenvector, namely, time-delay phase encoding, using time-delay phase encoding. Encoding can generate better spike trains, helping to improve the traffic sign recognition results of spiking neural networks.
步骤205,将第二脉冲序列输入到脉冲神经网络中,对脉冲神经网络进行训练。
需要说明的是,本实施例中的脉冲神经网络采用LIF(Leaky Integrate-and-Fire)神经元,其中,第一层为竞争层,由多个LIF神经元构成,竞争层中神经元之间采用侧抑制的方式达到竞争性学习的目的,其侧抑制过程具体为:每当一个神经元迸发一个动作电位,该神经元会立即复位到初始状态,并进入不应期,而其他所有神经元复位到静息膜电位,进入抑制期。每当有神经元发放脉冲,所有神经元都会复位,然后重新开始竞争,通过竞争来学习,从而对脉冲神经网络进行训练;脉冲神经网络的第二层为输出层,输出层输出每个类别的相似度值,相似度值越小,说明相似度越高,相似度值最小的类的标签为最终的识别结果,相似度值的计算的具体步骤为:It should be noted that the spiking neural network in this embodiment uses LIF (Leaky Integrate-and-Fire) neurons, wherein the first layer is a competition layer, which is composed of multiple LIF neurons, and the neurons in the competition layer are The purpose of competitive learning is achieved by lateral inhibition. The lateral inhibition process is as follows: whenever a neuron bursts an action potential, the neuron will immediately reset to the initial state and enter the refractory period, while all other neurons Reset to the resting membrane potential and enter the inhibitory phase. Whenever a neuron emits a pulse, all neurons will reset, and then start to compete again, and learn through competition, so as to train the spiking neural network; the second layer of the spiking neural network is the output layer, and the output layer outputs the output of each category. Similarity value, the smaller the similarity value, the higher the similarity. The label of the class with the smallest similarity value is the final recognition result. The specific steps for calculating the similarity value are:
假设输入图像矩阵为I=xij∈Rn×n,对输入图像矩阵进行标准化处理,即:Assuming that the input image matrix is I=x ij ∈R n×n , normalize the input image matrix, namely:
x′ij=(xij-xmin)/(xmax-xmin)x′ ij =(x ij -x min )/(x max -x min )
其中,xmax、xmin分别为I中最大、最小像素值,x′ij为标准化处理后的输入图像。Among them, x max and x min are the maximum and minimum pixel values in I respectively, and x′ ij is the input image after normalization.
假设标签为L的神经元分别记为其中,为标签为L的第m个神经元,ML为神经元的个数,L=0,1,…,9;标签为L的第m个神经元对输入图像对应的脉冲序列发放的脉冲个数为标签为L的第m个神经元对应的感受野权值矩阵为将标签为L的ML个神经元的脉冲个数与感受野权值进行乘累加,得到标签为L的神经元对输入图像的重构图像,即:Suppose that the neurons labeled L are denoted as in, is the mth neuron labelled L, ML is the number of neurons, L=0,1,...,9; the number of pulses emitted by the mth neuron labelled L to the pulse sequence corresponding to the input image number is The receptive field weight matrix corresponding to the mth neuron labeled L is: Multiply and accumulate the number of pulses of the M L neurons labeled L and the weight of the receptive field to obtain the reconstructed image of the input image by the neuron labeled L, namely:
假设RL=rij∈Rn×n,对其同样进行标准化处理,得到标准化处理后的重构图像r′ij,计算标准化处理后的输入图像与标准化处理后的重构图像的相似度值SL,具体相似度值计算公式如下所示:Suppose R L =r ij ∈R n×n , perform normalization processing on it as well to obtain the reconstructed image r′ ij after normalization processing, and calculate the similarity value between the input image after normalization processing and the reconstructed image after normalization processing S L , the specific similarity value calculation formula is as follows:
得到10个相似度数值S0,S1,…,S9,比较S0,S1,…,S9的大小,相似度值最小的类的标签为最终的识别结果,假设S6最小,则标签6所代表的类别为最终识别结果。Obtain 10 similarity values S 0 , S 1 ,…, S 9 , compare the sizes of S 0 , S 1 ,…, S 9 , the label of the class with the smallest similarity value is the final recognition result, assuming S 6 is the smallest, Then the category represented by label 6 is the final recognition result.
脉冲神经网络采用三脉冲STDP与阈值可塑性结合的学习方法进行训练,三脉冲STDP学习方法是针对突触的,而阈值可塑性学习方法是针对神经元的阈值电位,采用三脉冲STDP与阈值可塑性结合的学习方法,使得神经元的发放频率得到了限制。The spiking neural network is trained by a learning method combining three-pulse STDP and threshold plasticity. The three-pulse STDP learning method is for synapses, while the threshold plasticity learning method is based on the threshold potential of neurons. Three-pulse STDP combined with threshold plasticity is used for training. The learning method limits the firing frequency of neurons.
步骤206,计算脉冲神经网络对待训练交通标志图像的识别准确率,当识别准确率高于阈值时,训练完成,得到训练好的脉冲神经网络。Step 206: Calculate the recognition accuracy rate of the traffic sign image to be trained by the spiking neural network. When the recognition accuracy rate is higher than the threshold, the training is completed, and the trained spiking neural network is obtained.
需要说明的是,识别准确率通过正确识别的待训练交通标志图像数目与所有待训练图像数目的比值计算得到,当识别准确率高于预先设定的阈值时,则认为训练完成,停止训练,得到训练好的脉冲神经网络。It should be noted that the recognition accuracy is calculated by the ratio of the number of correctly identified traffic sign images to be trained to the number of all images to be trained. When the recognition accuracy is higher than the preset threshold, the training is considered complete and the training is stopped. Get the trained spiking neural network.
步骤207,获取待识别交通标志图像。
步骤208,将待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待识别交通标志图像对应的第一特征向量。Step 208: Input the traffic sign image to be recognized into the trained deep belief network model for feature extraction to obtain a first feature vector corresponding to the traffic sign image to be recognized.
步骤209,将第一特征向量转换成第一脉冲序列。Step 209: Convert the first feature vector into a first pulse sequence.
步骤210,将第一脉冲序列输入到训练好的脉冲神经网络中,获取训练好的脉冲神经网络输出的识别结果。Step 210: Input the first pulse sequence into the trained spiking neural network, and obtain the recognition result output by the trained spiking neural network.
需要说明的是,本申请实施例中的步骤207至步骤210与上一实施例中的步骤101至步骤104一致,在此不再进行赘述。It should be noted that, steps 207 to 210 in this embodiment of the present application are consistent with
为了便于理解,请参阅图3,本发明提供的一种交通标志识别装置的一个实施例,包括:For ease of understanding, please refer to FIG. 3 , an embodiment of a traffic sign recognition device provided by the present invention includes:
第一图像获取模块301,用于获取待识别交通标志图像;a first image acquisition module 301, configured to acquire an image of a traffic sign to be recognized;
第一特征提取模块302,用于将待识别交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待识别交通标志图像对应的第一特征向量;The first feature extraction module 302 is configured to input the traffic sign image to be recognized into the trained deep belief network model for feature extraction to obtain the first feature vector corresponding to the traffic sign image to be recognized;
第一转换模块303,用于将第一特征向量转换成第一脉冲序列;a first conversion module 303, configured to convert the first feature vector into a first pulse sequence;
识别模块304,用于将第一脉冲序列输入到训练好的脉冲神经网络中,获取训练好的脉冲神经网络输出的识别结果。The identification module 304 is configured to input the first pulse sequence into the trained spiking neural network, and obtain the recognition result output by the trained spiking neural network.
进一步,还包括:Further, it also includes:
第二图像获取模块305,用于获取待训练交通标志图像集;The second image acquisition module 305 is configured to acquire the traffic sign image set to be trained;
第二特征提取模块306,用于将待训练交通标志图像集中的待训练交通标志图像输入到训练好的深度信念网络模型中进行特征提取,得到待训练交通标志图像对应的第二特征向量;The second feature extraction module 306 is configured to input the to-be-trained traffic sign images in the to-be-trained traffic sign image set into the trained deep belief network model for feature extraction, and obtain a second feature vector corresponding to the to-be-trained traffic sign images;
第二转换模块307,用于采用时滞相位编码对第二特征向量进行转换,得到第二脉冲序列;The second conversion module 307 is configured to convert the second eigenvector using time-delay phase coding to obtain a second pulse sequence;
训练模块308,用于将第二脉冲序列输入到脉冲神经网络中,对脉冲神经网络进行训练;A training module 308, configured to input the second pulse sequence into the spiking neural network to train the spiking neural network;
计算模块309,用于计算脉冲神经网络对待训练交通标志图像的识别准确率,当识别准确率高于阈值时,训练完成,得到训练好的脉冲神经网络。The calculation module 309 is used to calculate the recognition accuracy rate of the traffic sign image to be trained by the spiking neural network. When the recognition accuracy rate is higher than the threshold, the training is completed and the trained spiking neural network is obtained.
进一步,还包括:Further, it also includes:
预处理模块310,用于对待训练交通标志图像进行预处理。The preprocessing module 310 is used for preprocessing the traffic sign images to be trained.
在本申请所提供的几个实施例中,应该理解到,所揭露的装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed apparatus and method may be implemented in other manners. For example, the apparatus embodiments described above are only illustrative. For example, the division of the units is only a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit. The above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以通过一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(英文全称:Read-OnlyMemory,英文缩写:ROM)、随机存取存储器(英文全称:Random Access Memory,英文缩写:RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The integrated unit, if implemented in the form of a software functional unit and sold or used as an independent product, may be stored in a computer-readable storage medium. Based on this understanding, the technical solutions of the present application can be embodied in the form of software products in essence, or the parts that contribute to the prior art, or all or part of the technical solutions, and the computer software products are stored in a storage medium , including several instructions for executing all or part of the steps of the methods described in the various embodiments of the present application through a computer device (which may be a personal computer, a server, or a network device, etc.). The aforementioned storage media include: U disk, mobile hard disk, read-only memory (full English name: Read-Only Memory, English abbreviation: ROM), random access memory (English full name: Random Access Memory, English abbreviation: RAM), magnetic disks Or various media such as optical discs that can store program codes.
以上所述,以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围。As mentioned above, the above embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand: The technical solutions described in the embodiments are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions in the embodiments of the present application.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910764738.8A CN110458136B (en) | 2019-08-19 | 2019-08-19 | Traffic sign identification method, device and equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910764738.8A CN110458136B (en) | 2019-08-19 | 2019-08-19 | Traffic sign identification method, device and equipment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110458136A CN110458136A (en) | 2019-11-15 |
CN110458136B true CN110458136B (en) | 2022-07-12 |
Family
ID=68487609
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910764738.8A Active CN110458136B (en) | 2019-08-19 | 2019-08-19 | Traffic sign identification method, device and equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110458136B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113395331B (en) * | 2021-05-25 | 2022-03-15 | 郑州信大捷安信息技术股份有限公司 | Safety traffic sign error surveying method and system based on Internet of vehicles |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103235937A (en) * | 2013-04-27 | 2013-08-07 | 武汉大学 | A Traffic Sign Recognition Method Based on Pulse-Coupled Neural Network |
CN109522448A (en) * | 2018-10-18 | 2019-03-26 | 天津大学 | A method of robustness speech Gender Classification is carried out based on CRBM and SNN |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8990132B2 (en) * | 2010-01-19 | 2015-03-24 | James Ting-Ho Lo | Artificial neural networks based on a low-order model of biological neural networks |
CN102262728B (en) * | 2011-07-28 | 2012-12-19 | 电子科技大学 | Road traffic sign identification method |
US9098811B2 (en) * | 2012-06-04 | 2015-08-04 | Brain Corporation | Spiking neuron network apparatus and methods |
CN109816026B (en) * | 2019-01-29 | 2021-09-10 | 清华大学 | Fusion device and method of convolutional neural network and impulse neural network |
-
2019
- 2019-08-19 CN CN201910764738.8A patent/CN110458136B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103235937A (en) * | 2013-04-27 | 2013-08-07 | 武汉大学 | A Traffic Sign Recognition Method Based on Pulse-Coupled Neural Network |
CN109522448A (en) * | 2018-10-18 | 2019-03-26 | 天津大学 | A method of robustness speech Gender Classification is carried out based on CRBM and SNN |
Non-Patent Citations (2)
Title |
---|
A handwritten numeral recognition method based on STDP based with unsupervised learning;Yonghong Xie,et.al;《2017 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET)》;20170324;全文 * |
类脑机的思想与体系结构综述;黄铁军等;《计算机研究与发展》;20190615(第006期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110458136A (en) | 2019-11-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109615582B (en) | A Face Image Super-resolution Reconstruction Method Based on Attribute Description Generative Adversarial Network | |
US20230059877A1 (en) | Deep learning-based splice site classification | |
CN110555368B (en) | Fall behavior recognition method based on three-dimensional convolutional neural network | |
CN108985252B (en) | Improved spiking deep neural network for image classification | |
Xu et al. | Denoising convolutional neural network | |
CN113344933B (en) | Glandular cell segmentation method based on multi-level feature fusion network | |
CN110458903B (en) | Image processing method of coding pulse sequence | |
CN104268568A (en) | Behavior recognition method based on intelligent sub-space networks | |
CN109783682A (en) | A Deep Unrelaxed Hash Image Retrieval Method Based on Point Pair Similarity | |
CN111681188B (en) | Image Deblurring Method Based on Combining Image Pixel Prior and Image Gradient Prior | |
CN104866855A (en) | Image feature extraction method and apparatus | |
CN112818764A (en) | Low-resolution image facial expression recognition method based on feature reconstruction model | |
CN107301396A (en) | Video fraud convolutional neural networks training method and human face in-vivo detection method | |
CN110674774A (en) | An improved deep learning facial expression recognition method and system | |
CN115100107B (en) | Method and system for dividing skin mirror image | |
CN108182475A (en) | It is a kind of based on automatic coding machine-the multi-dimensional data characteristic recognition method of the learning machine that transfinites | |
CN109255381A (en) | A kind of image classification method based on the sparse adaptive depth network of second order VLAD | |
CN111951220A (en) | An Unsupervised Intracerebral Hemorrhage Segmentation Method Based on Multi-level Domain Adaptive Technology | |
Aouayeb et al. | Micro-expression recognition from local facial regions | |
CN110458136B (en) | Traffic sign identification method, device and equipment | |
CN107038419A (en) | A kind of personage's behavior method for recognizing semantics based on video sequence deep learning | |
CN117455774A (en) | Image reconstruction method and system based on differential output | |
CN107239827B (en) | A spatial information learning method based on artificial neural network | |
CN116189272A (en) | Facial expression recognition method and system based on feature fusion and attention mechanism | |
Huang et al. | Human emotion recognition based on face and facial expression detection using deep belief network under complicated backgrounds |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |