[go: up one dir, main page]

CN113469277A - Image recognition method and device - Google Patents

Image recognition method and device Download PDF

Info

Publication number
CN113469277A
CN113469277A CN202110827593.9A CN202110827593A CN113469277A CN 113469277 A CN113469277 A CN 113469277A CN 202110827593 A CN202110827593 A CN 202110827593A CN 113469277 A CN113469277 A CN 113469277A
Authority
CN
China
Prior art keywords
convolutional neural
neural network
scaling factor
factor sets
sets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110827593.9A
Other languages
Chinese (zh)
Inventor
郑哲
王盟
陈斐洋
吴立
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Dahua Technology Co Ltd
Original Assignee
Zhejiang Dahua Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Dahua Technology Co Ltd filed Critical Zhejiang Dahua Technology Co Ltd
Priority to CN202110827593.9A priority Critical patent/CN113469277A/en
Publication of CN113469277A publication Critical patent/CN113469277A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

本发明提供了一种图像识别方法及装置,包括:获取原始卷积神经网络的N组初始缩放因子集合,其中,每组初始缩放因子集合中包括M个缩放因子,每个初始缩放因子对应于原始卷积神经网络的至少一层卷积层,N和M大于或等于1;使用进化算法对N组初始缩放因子集合进行优化,得到一组优化缩放因子集合,其中,一组优化缩放因子集合中的每个优化缩放因子用于表示对应卷积层的卷积核剪枝比例;按照一组优化缩放因子集合对原始卷积神经网络的卷积核进行剪枝,得到轻量卷积神经网络。通过本发明,解决了对于资源受限的终端,图像识别准确率较低的问题,进而达到了提高卷积神经网络的剪枝效率的效果。

Figure 202110827593

The present invention provides an image recognition method and device, comprising: acquiring N groups of initial scaling factor sets of an original convolutional neural network, wherein each group of initial scaling factor sets includes M scaling factors, and each initial scaling factor corresponds to At least one convolution layer of the original convolutional neural network, N and M are greater than or equal to 1; use an evolutionary algorithm to optimize N groups of initial scaling factor sets to obtain a set of optimized scaling factor sets, wherein, a set of optimized scaling factor sets Each optimized scaling factor in is used to represent the convolution kernel pruning ratio of the corresponding convolutional layer; prune the convolution kernel of the original convolutional neural network according to a set of optimized scaling factors to obtain a lightweight convolutional neural network . The invention solves the problem of low image recognition accuracy for terminals with limited resources, thereby achieving the effect of improving the pruning efficiency of the convolutional neural network.

Figure 202110827593

Description

Image recognition method and device
Technical Field
The invention relates to the field of image recognition, in particular to an image recognition method and device.
Background
In the related technology, the convolutional neural network has high accuracy in image recognition. However, most convolutional neural networks have the characteristics of large calculation amount and many model parameters, and at present, images are usually identified on a large computer through the convolutional neural networks, but for resource-limited mobile terminals (such as mobile phones, wearable devices and other devices), it is difficult to directly deploy the convolutional neural networks to the resource-limited mobile terminals, and the mobile terminals cannot identify the images by using the convolutional neural networks.
Aiming at the problem that the accuracy rate of image recognition is low for a resource-limited terminal in the related art, an effective solution does not exist at present.
Disclosure of Invention
The embodiment of the invention provides an image identification method and device, which at least solve the problem of low image identification accuracy rate of a resource-limited terminal in the related art.
According to an embodiment of the present invention, there is provided an image recognition method including: acquiring N groups of initial scaling factor sets of an original convolutional neural network, wherein each group of initial scaling factor sets comprises M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1; optimizing the N groups of initial scaling factor sets by using an evolutionary algorithm to obtain a group of optimized scaling factor sets, wherein each optimized scaling factor in the group of optimized scaling factor sets is used for representing the convolution kernel pruning proportion of a corresponding convolution layer; pruning the convolution kernels of the original convolution neural network according to the set of optimized scaling factor sets to obtain the lightweight convolution neural network.
Optionally, optimizing the N sets of initial scaling factor sets by using an evolutionary algorithm to obtain a set of optimized scaling factor sets, including: pruning the original convolutional neural network by using each group of initial scaling factor sets in the N groups of initial scaling factor sets to obtain N first convolutional neural networks; and under the condition that a convolutional neural network meeting a preset convergence condition exists in the N first convolutional neural networks, determining the convolutional neural network meeting the convergence condition as the light-weight convolutional neural network, wherein the preset convergence condition is used for indicating that the output of the fitness function is within a threshold range.
Optionally, the method further comprises: under the condition that no convolutional neural network meeting a preset convergence condition exists in the N first convolutional neural networks, crossing scaling factors at preset cross points in any two groups of initial scaling factor sets, and/or carrying out variation on one or more initial scaling factors at preset variation points in the scaling factor sets to obtain N groups of first scaling factor sets; pruning the original convolutional neural network by using each group of first scaling factor sets in the N groups of first scaling factor sets respectively to obtain N second convolutional neural networks; determining a convolutional neural network of the N second convolutional neural networks that satisfies the convergence condition as the lightweight convolutional neural network.
Optionally, pruning the convolution kernel of the original convolutional neural network according to the set of optimized scaling factors, including: determining a T value according to an optimized scaling factor corresponding to each convolutional layer in the original convolutional neural network and the number of convolutional kernels in each convolutional layer, wherein the T value is the number of convolutional kernels needing pruning, and T is an integer; sorting the convolution kernels in each convolution layer according to the norm of the convolution kernels; and pruning the convolution kernels ranked as the first T in each convolution layer.
Optionally, obtaining N sets of initial scaling factor sets of the original convolutional neural network, including: and randomly generating the N groups of initial scaling factor sets according to the number of convolution layers of the original convolutional neural network, wherein each initial scaling factor is a binary code, and each binary code corresponds to at least one convolution layer of the original convolutional neural network.
Optionally, the method further comprises: generating image feature data of a specified category using a conditional generation countermeasure network; training the lightweight convolutional neural network using the specified class of image feature data.
According to another embodiment of the present invention, there is provided an image recognition apparatus including: an obtaining module, configured to obtain N sets of initial scaling factors of an original convolutional neural network, where each set of initial scaling factors includes M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1; the optimization module is used for optimizing the N groups of initial scaling factor sets by using an evolutionary algorithm to obtain a group of optimized scaling factor sets, wherein each optimized scaling factor in the group of optimized scaling factor sets is used for representing the convolution kernel pruning proportion of a corresponding convolution layer; the processing module is used for pruning the convolution kernels of the original convolution neural network according to the set of optimized scaling factor sets to obtain a light-weight convolution neural network; and the identification module is used for identifying the target image by using the light-weight convolutional neural network to obtain the identification result of the target image.
Optionally, the optimization module comprises: a first processing unit, configured to prune the original convolutional neural network using each initial scaling factor set of the N initial scaling factor sets to obtain N first convolutional neural networks; and a determining unit, configured to determine, when a convolutional neural network that satisfies a preset convergence condition exists in the N first convolutional neural networks, the convolutional neural network that satisfies the convergence condition as the lightweight convolutional neural network, where the preset convergence condition is used to indicate that an output of the fitness function is within a threshold range.
According to a further embodiment of the present invention, there is also provided a storage medium having a computer program stored therein, wherein the computer program is arranged to perform the steps of any of the above method embodiments when executed.
According to yet another embodiment of the present invention, there is also provided an electronic device, including a memory in which a computer program is stored and a processor configured to execute the computer program to perform the steps in any of the above method embodiments.
According to the invention, N groups of initial scaling factor sets of the original convolutional neural network are obtained; optimizing the N groups of initial zoom factor sets by using an evolutionary algorithm to obtain a group of optimized zoom factor sets; pruning the convolution kernel of the original convolution neural network according to the set of optimized scaling factor sets to obtain a lightweight convolution neural network, deploying the lightweight convolution neural network at a resource-limited mobile terminal, and identifying the target image by using the lightweight convolution neural network at the mobile terminal. The problem that the image recognition accuracy rate is low for a terminal with limited resources can be solved, and the effect of improving the image recognition accuracy rate is achieved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the invention without limiting the invention. In the drawings:
fig. 1 is a block diagram of a hardware configuration of a mobile terminal of an image recognition method according to an embodiment of the present invention;
FIG. 2 is a flow chart of an image recognition method according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a model structure of a non-block type deep convolutional neural network according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a model structure of a block-type deep convolutional neural network according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a convolutional neural network model structure in accordance with an alternative embodiment of the present invention;
fig. 6 is a block diagram of the structure of an image recognition apparatus according to an embodiment of the present invention.
Detailed Description
The invention will be described in detail hereinafter with reference to the accompanying drawings in conjunction with embodiments. It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order.
The method provided by the first embodiment of the present application may be executed in a mobile terminal, a computer terminal, or a similar computing device. Taking an example of the method performed in a mobile terminal, fig. 1 is a block diagram of a hardware structure of the mobile terminal according to an embodiment of the present invention. As shown in fig. 1, the mobile terminal 10 may include one or more (only one shown in fig. 1) processors 102 (the processor 102 may include, but is not limited to, a processing device such as a microprocessor MCU or a programmable logic device FPGA) and a memory 104 for storing data, and optionally may also include a transmission device 106 for communication functions and an input-output device 108. It will be understood by those skilled in the art that the structure shown in fig. 1 is only an illustration, and does not limit the structure of the mobile terminal. For example, the mobile terminal 10 may also include more or fewer components than shown in FIG. 1, or have a different configuration than shown in FIG. 1.
The memory 104 may be used to store a computer program, for example, a software program and a module of application software, such as a computer program corresponding to the image recognition method in the embodiment of the present invention, and the processor 102 executes various functional applications and data processing by running the computer program stored in the memory 104, so as to implement the method described above. The memory 104 may include high speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some instances, the memory 104 may further include memory located remotely from the processor 102, which may be connected to the mobile terminal 10 via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The transmission device 106 is used for receiving or transmitting data via a network. Specific examples of the network described above may include a wireless network provided by a communication provider of the mobile terminal 10. In one example, the transmission device 106 includes a Network adapter (NIC), which can be connected to other Network devices through a base station so as to communicate with the internet. In one example, the transmission device 106 may be a Radio Frequency (RF) module, which is used for communicating with the internet in a wireless manner.
In the present embodiment, an image recognition method operating in the mobile terminal is provided, and fig. 2 is a flowchart of the image recognition method according to the embodiment of the present invention, as shown in fig. 2, the flowchart includes the following steps:
step S202, obtaining N groups of initial scaling factor sets of an original convolutional neural network, wherein each group of initial scaling factor sets comprises M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1;
step S204, optimizing the N groups of initial scaling factor sets by using an evolutionary algorithm to obtain a group of optimized scaling factor sets, wherein each optimized scaling factor in the group of optimized scaling factor sets is used for representing the convolution kernel pruning proportion of the corresponding convolution layer;
step S206, pruning the convolution kernel of the original convolution neural network according to the set of optimized scaling factors to obtain a light-weight convolution neural network;
and S208, identifying the target image by using the lightweight convolutional neural network to obtain an identification result of the target image.
Through the steps, N groups of initial scaling factor sets of the original convolutional neural network are obtained; optimizing the N groups of initial zoom factor sets by using an evolutionary algorithm to obtain a group of optimized zoom factor sets; pruning the convolution kernel of the original convolution neural network according to the set of optimized scaling factor sets to obtain a lightweight convolution neural network, deploying the lightweight convolution neural network at a resource-limited mobile terminal, and identifying the target image by using the lightweight convolution neural network at the mobile terminal. The problem that the image recognition accuracy rate is low for a terminal with limited resources can be solved, and the effect of improving the image recognition accuracy rate is achieved.
Alternatively, the execution subject of the above steps may be a terminal or the like, but is not limited thereto.
As an optional implementation mode, the deep convolutional network is divided according to the composition characteristics of the deep convolutional network and can be divided into a non-block type and a block type. The non-block type network, such as early AlexNet, VGGNet, etc., is a schematic model structure diagram of a non-block type deep convolutional neural network according to an optional embodiment of the present invention, as shown in fig. 3, and the non-block type network is formed by stacking and connecting single-layer convolutional networks in series. The block types such as ResNet, MobileNet, ShuffleNet and the like are shown in fig. 4, which is a schematic diagram of a model structure of a block type deep convolutional neural network according to an optional embodiment of the present invention, the block type deep neural network is formed by stacking a plurality of stages containing blocks, each block is composed of several convolutional layers, and the blocks are connected by residual blocks.
As an optional implementation, when the convolutional neural network CNN is of a non-block type, a scaling factor α e [0, 1] is introduced into each convolutional layer of the convolutional neural network to form a scaling factor set, where each scaling factor in the scaling factor set corresponds to one convolutional layer. For example, a convolutional neural network comprises M layers, then M scaling factors constitute a set of scaling factors. When the convolutional neural network CNN is a block type, a scaling factor is introduced to each block of the convolutional neural network, where each scaling factor corresponds to one block, and each block includes a plurality of convolutional layers. For example, the convolutional neural network includes M blocks, and then scaling factors corresponding to the M blocks constitute a scaling factor set.
As an optional implementation, taking a non-block type convolutional neural network as an example, assume that the original convolutional neural network model CNNorigThe number of convolution layers of (1) is M. The number of convolution kernels of each convolution layer is NlThe number of channels of the convolution kernel is ClHeight of two-dimensional convolution is HlWidth of the two-dimensional convolution is Wl. Each layer of the convolution kernel may be parameterized by
Figure BDA0003174237140000071
And (4) showing. An evolutionary algorithm may be employed to search for an optimal scaling factor, which searches forThe rope process comprises the following steps:
in step S21, performing population initialization, which may be preset to set the number N of individuals in the population, where each individual is used to represent a set of scaling factor sets, and a set of scaling factor sets is performed to initialize a scaling factor for each convolutional layer in the convolutional neural network. And initializing scaling factors for N times on each convolution layer in the convolutional neural network to obtain N scaling factor sets, wherein the number of individuals in the population is N. Since the convolutional neural network model has M convolutional layers, M scaling factors are included in each scaling factor set.
Step S22, performing cross variation on the scaling factors by using an evolutionary algorithm to obtain a set of optimized scaling factor sets, where each set of optimized scaling factor sets includes M scaling factors, and each scaling factor is used to represent a scaling ratio of a convolution kernel in each convolution layer of the convolutional neural network. For example, a1 in a set of scaling factors is 0.3, and there are 10 convolution sums in the convolution layer corresponding to the scaling factor a1, then the convolution layer needs to prune 3 convolution kernels.
And step S23, according to each scaling factor in the set of optimized scaling factors, pruning the convolution sum in the convolution layer corresponding to each scaling factor. Assume that the convolutional neural network includes two convolutional layers, a first convolutional layer with 10 convolutional sums and a second convolutional layer with 20 convolutional kernels. One set of optimized scaling factors includes a 1-0.3 and a 2-0.5, where a1 corresponds to the first convolutional layer and a2 corresponds to the second convolutional layer, and the first convolutional layer needs to prune 3 convolutional kernels and the second convolutional layer needs to prune 10 convolutional kernels. And pruning each convolutional layer to obtain the light convolutional neural network.
Optionally, optimizing the N sets of initial scaling factor sets by using an evolutionary algorithm to obtain a set of optimized scaling factor sets, including: pruning the original convolutional neural network by using each group of initial scaling factor sets in the N groups of initial scaling factor sets to obtain N first convolutional neural networks; and under the condition that a convolutional neural network meeting a preset convergence condition exists in the N first convolutional neural networks, determining the convolutional neural network meeting the convergence condition as the light-weight convolutional neural network, wherein the preset convergence condition is used for indicating that the output of the fitness function is within a threshold range.
As an optional embodiment, optimizing the N sets of initial scaling factor sets using an evolutionary algorithm may include the steps of:
step S31, for each individual, i.e. each set of initial scaling factor sets, for the original convolutional neural network CNNorigPruning is carried out, and a lightweight model CNN is extracted from the original neural networklight_preExtracting the N times of lightweight models by N groups of initial scaling factor sets to obtain N lightweight models CNNlight_preI.e. N first convolutional neural networks. For the selection of the convolution kernels needing pruning in each convolution layer, the norm of each convolution kernel in each convolution layer can be obtained, the convolution kernels of each convolution layer are sorted according to the magnitude of the norm of the convolution kernels, and the convolution kernels can be sorted in the order of the norm from small to large. Pruning the convolution kernel of the convolution layer according to the corresponding scaling factor of the convolution layer. For example, the convolutional layer includes 30 convolutional kernels, the 30 convolutional kernels are sorted from small to large according to the norm of the 30 convolutional kernels, if the scaling factor corresponding to the convolutional layer is 0.1, 3 convolutional kernels need to be pruned, the convolutional kernels in the first three of the sorting are removed, and the remaining convolutional kernels are retained.
Step S32, determining whether the N first convolution neural networks satisfy a preset convergence condition, where the preset convergence condition may be convergence of an output value of a fitness function f ═ γ accumy + (1- γ) cost, where accumy is an accuracy rate of output of the first convolution neural network, and specifically, inputting the verification data set into the first convolution neural network, and comparing the estimated classification result output by the first convolution neural network with a known classification result of the verification data set, so as to obtain the accuracy rate accumy. cost is the ratio of the total parameters of the pruning to the total parameters of the model, and can also be the ratio of the number of the convolution kernels of the pruning to the total convolution kernels of the original convolution neural network model. Gamma is a hyper-parameter used for adjusting the proportion of the accuracy rate in evaluating the individual. The larger the fitness function is, the better the individual is. And if the first convolution nerves meeting the preset convergence condition exist in the N first convolution neural networks, determining the first convolution neural network meeting the convergence condition as a light-weight convolution neural network model. If there is no first convolutional neural network satisfying the convergence condition among the N first convolutional neural networks, step 32 is performed.
Step S33, when the preset crossover prob is satisfiedcrossWhen in probability, carrying out multipoint matching and crossing on the N groups of initial scaling factor sets, and presetting the crossing probability probcrossCan be a numerical value between 0 and 1, and the specific size can be determined according to actual conditions; when the preset cross probability prob is satisfiedmutateThen, a single-point variation is performed on one or more initial scaling factor sets in the N initial scaling factor sets to preset a cross probability probmutateMay be a value between 0 and 1. Pairing original convolutional neural network CNN with N sets of scale factors after interleaving and/or mutationorigAnd performing pruning, and repeatedly executing the steps S31 and S32 until the fitness function of the convolutional neural network model obtained by pruning converges.
Optionally, the method further comprises: under the condition that no convolutional neural network meeting a preset convergence condition exists in the N first convolutional neural networks, crossing scaling factors at preset cross points in any two groups of initial scaling factor sets, and/or carrying out variation on one or more initial scaling factors at preset variation points in the scaling factor sets to obtain N groups of first scaling factor sets; pruning the original convolutional neural network by using each group of first scaling factor sets in the N groups of first scaling factor sets respectively to obtain N second convolutional neural networks; determining a convolutional neural network of the N second convolutional neural networks that satisfies the convergence condition as the lightweight convolutional neural network.
As an optional implementation manner, in the case that there is no convolutional neural network that satisfies the preset convergence condition, the following steps may be iteratively performed until the convolutional neural network satisfies the preset convergence condition:
in step S41, a preset intersection and a variation point are set. The scaling factor may be a binary number of eight bits. Specifically, all the scaling factors α may be binary-coded, assuming that the number of coded bits is b-8, the coding space is code ∈ [00000000, …,11111111], when the scaling factor is 0 in decimal, the binary is 00000000, and when the scaling factor is 1 in decimal, the binary is 11111111, and the decimal a may be represented by an 8-bit binary code. The preset intersection and the variation point may be determined according to actual situations, for example, the first bit of the 8-bit binary code may be selected as the intersection, the second bit may be selected as the variation point, or the first bit and the second bit may be selected as the intersection, and the third bit and the fourth bit may be selected as the variation point. The number of the cross points and the variation points can be determined according to actual conditions, and the positions can also be determined according to actual conditions.
Step S42, performing intersection and/or mutation processing on the N sets of initial scaling factor sets according to preset intersection points and mutation points to obtain N sets of first scaling factor sets. For example, assume that the first scaling factor in the first set of scaling factors is 11100000 and the first scaling factor in the second set of scaling factors is 00000111. In the case of the first bit and the second bit at the intersection, after the intersection is performed, the first scaling factor in the first set of scaling factors is 00100000, and the first scaling factor in the second set of scaling factors is 11000111. Specifically, the crossed scaling factor sets may be determined according to actual situations, and may be obtained by crossing corresponding scaling factors in any two sets of scaling factor sets in the N sets of scaling factor sets. Taking the variation fork as an example, assuming that one of the scaling factors in a set of scaling factors is 11100000, and the predetermined variation point is the third bit of the first bit, the variation of 11100000 is 01000000.
And step S43, pruning the original convolutional neural network by using the scaling factor set obtained after the cross mutation processing. And judging whether the pruned convolutional neural network meets a preset convergence condition, if so, the pruned convolutional neural network is a light convolutional neural network. If not, the above steps S41 and S42 are repeatedly executed until satisfied.
Optionally, pruning the convolution kernel of the original convolutional neural network according to the set of optimized scaling factors, including: determining a T value according to an optimized scaling factor corresponding to each convolutional layer in the original convolutional neural network and the number of convolutional kernels in each convolutional layer, wherein the T value is the number of convolutional kernels needing pruning, and T is an integer; sorting the convolution kernels in each convolution layer according to the norm of the convolution kernels; and pruning the convolution kernels ranked as the first T in each convolution layer.
As an optional embodiment, for the selection of the convolution kernels that need pruning in each convolution layer, the norm of each convolution kernel in each convolution layer may be obtained, the convolution kernels of each convolution layer may be sorted according to the magnitude of the norm of the convolution kernel, and the sorting may be performed in the order of the norm from small to large. Pruning the convolution kernel of the convolution layer according to the corresponding scaling factor of the convolution layer. For example, the convolutional layer includes 30 convolutional kernels, the 30 convolutional kernels are sorted from small to large according to the norm of the 30 convolutional kernels, if the scaling factor corresponding to the convolutional layer is 0.1, 3 convolutional kernels need to be pruned, the convolutional kernels in the first three of the sorting are removed, and the remaining convolutional kernels are retained.
Optionally, obtaining N sets of initial scaling factor sets of the original convolutional neural network, including: and randomly generating the N groups of initial scaling factor sets according to the number of convolution layers of the original convolutional neural network, wherein each initial scaling factor is a binary code, and each binary code corresponds to at least one convolution layer of the original convolutional neural network.
As an alternative embodiment, the initial scaling factors in the N sets of initial scaling factors may be randomly generated, and the scaling factors may be binary numbers of eight bits. Specifically, all the scaling factors α may be binary-coded, assuming that the number of coded bits is b-8, the coding space is code ∈ [00000000, …,11111111], when the scaling factor is 0 in decimal, the binary is 00000000, and when the scaling factor is 1 in decimal, the binary is 11111111, and the decimal a may be represented by an 8-bit binary code.
Optionally, the method further comprises: generating image feature data of a specified category using a conditional generation countermeasure network; training the lightweight convolutional neural network using the specified class of image feature data.
As an alternative implementation, fig. 5 is a schematic diagram illustrating a convolutional neural network model structure according to an alternative embodiment of the present invention, and the depth model is composed of a feature extractor and a classifier. Aiming at the classifier, the false features of the class can be generated by the conditional generation countermeasure network according to the specified class, the class image feature data is expanded, and the classifier is further trained, optimized and improved in performance. The "data" enhancement can be done from a hierarchy of features. A Conditional Generation Adaptive Network (CGAN) may be used to input class information so that the generating network generates pseudo features from the noise that may be used to train the classifier. Suppose the ith class of classification is CiFirst, a countermeasure network is generated using conditions such that it is in accordance with a given class CiGenerating image feature data F having such non-discriminativityfake(i)Combining the true features Freal(i)And constructing image characteristic data of the category, and finally finely adjusting the classifier by using the SGD.
As an optional implementation manner, for a block-type convolutional neural network model, the same processing manner as for a non-block-type convolutional neural network model is used, except that a scaling factor is introduced into each block, the other processes are the same, after the scaling factor of each block is solved, pruning is performed on all convolutional layers under the block in the same proportion, and the size of the scaling factor is the size of the scaling factor.
The following description of the present application is made by specific examples, taking non-block types as examples:
setting original convolution neural network model CNNorigThe number of convolution layers of (a) is L. Introducing a scaling factor a to each convolutional layer in the networkl∈[0,1]Let each layer of convolution kernel be composed of parameters
Figure BDA0003174237140000121
Is shown byIn, NlRepresenting the number of convolution kernels, ClNumber of channels, H, representing convolution kernellHigh, W representing a two-dimensional convolutionlRepresenting the width of the two-dimensional convolution. An evolutionary algorithm is adopted to search for the optimal scaling factor, and the searching process is as follows:
a. population initialization: all the scaling factors α are binary coded, and assuming that the number of coded bits is b equal to 8, the coding space is code e [00000000, 111111111, when α is 0, the coding space is 00000000, and when α is 1, the coding space is 11111111, that is, α is represented by 8 binary codes. The transformation relationship is as follows: α is decimal (code) × 1/255. Then the population matrix p is e {0, 1}P×(b×L)Wherein P is the number of population individuals, and the coding length of each individual is b multiplied by L.
b. Constructing a fitness function: for each individual, from CNNorigMiddle-extracted lightweight model CNNlight_pre. For example, let i-th layer of convolution layer ai=0.72,ai-1=0.51,Ni=Ni-1=32,Hi=WlIf 3, the parameter matrix is
Figure BDA0003174237140000122
It should be noted that, here, the convolution kernel selection is not random selection, but the convolution kernel is firstly selected according to norm l1norm sorts, picks top int (N)iai) Small convolution kernel composition CNNlight_pre
For the obtained CNNligh_tprAnd respectively obtaining the accuracy accuracuacy of Cifar10 on the proxy data set, and the calculated quantity flops and the parameter quantity params of Cifar10, and constructing a fitness function f ═ gamma accuracyaccy + (1-gamma) cost. The cost function is a flops parameter or params parameter calculated by searching out the model, and here refers to a parameter removal ratio of the model, i.e. a ratio of a total parameter of pruning to a total parameter of the model. Gamma is a hyper-parameter used for adjusting the proportion of the accuracy rate in evaluating the individual. The larger the fitness function is, the better the individual is.
c. And (3) a crossover operator: setting the crossover probability probcross∈[0,1]Performing multipoint matching intersection when the intersection probability is met;
d. mutation operator: setting the probability of variation probmutate∈[0,1]Performing single-point mutation when the mutation probability is met;
e. selecting: selecting individuals meeting the calculated amount and the parameter number at the same time according to the performance parameters of the mobile terminal equipment to be deployed for next iteration;
f. and repeating the steps b to e until a termination condition (iteration times or fitness function convergence) is reached.
After the optimal search factor is solved on the proxy data set, pruning operation is executed, and finally fine adjustment is carried out on the target data set to obtain CNNlight_pre
According to the method, the resource parameters of the hardware are fused into the automatic searching process, and under the limiting condition, the lightweight model meeting the equipment resource requirement can be directly obtained by adopting the automatic searching method, so that the complicated and tedious workload of algorithm personnel is avoided. A novel data enhancement method is provided, image characteristic data related to the generation category of the countermeasure network is directly generated by using conditions, data expansion is carried out at a characteristic level, and performance gain is obtained by finely adjusting a classifier.
Through the above description of the embodiments, those skilled in the art can clearly understand that the method according to the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but the former is a better implementation mode in many cases. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which is stored in a storage medium (e.g., ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, or a network device) to execute the method according to the embodiments of the present invention.
In this embodiment, an image recognition apparatus is further provided, and the apparatus is used to implement the foregoing embodiments and preferred embodiments, and the description of the apparatus is omitted for brevity. As used below, the term "module" may be a combination of software and/or hardware that implements a predetermined function. Although the means described in the embodiments below are preferably implemented in software, an implementation in hardware, or a combination of software and hardware is also possible and contemplated.
Fig. 6 is a block diagram of the structure of an image recognition apparatus according to an embodiment of the present invention, as shown in fig. 6, the apparatus including: an obtaining module 62, configured to obtain N sets of initial scaling factors of an original convolutional neural network, where each set of initial scaling factors includes M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1; an optimizing module 64, configured to optimize the N sets of initial scaling factor sets by using an evolutionary algorithm to obtain a set of optimized scaling factor sets, where each optimized scaling factor in the set of optimized scaling factor sets is used to represent a convolution kernel pruning proportion of a corresponding convolution layer; the processing module 66 is configured to prune the convolution kernel of the original convolution neural network according to the set of optimized scaling factor sets to obtain a lightweight convolution neural network; and the identification module 68 is configured to identify the target image by using the lightweight convolutional neural network, so as to obtain an identification result of the target image.
Optionally, the optimization module comprises: a first processing unit, configured to prune the original convolutional neural network using each initial scaling factor set of the N initial scaling factor sets to obtain N first convolutional neural networks; and a determining unit, configured to determine, when a convolutional neural network that satisfies a preset convergence condition exists in the N first convolutional neural networks, the convolutional neural network that satisfies the convergence condition as the lightweight convolutional neural network, where the preset convergence condition is used to indicate that an output of the fitness function is within a threshold range.
Optionally, the apparatus is further configured to, in a case that there is no convolutional neural network that satisfies a preset convergence condition in the N first convolutional neural networks, intersect scaling factors at a preset intersection in any two sets of the initial scaling factor sets, and/or vary one or more initial scaling factors at a preset variation point in the scaling factor sets, to obtain N sets of first scaling factor sets; pruning the original convolutional neural network by using each group of first scaling factor sets in the N groups of first scaling factor sets respectively to obtain N second convolutional neural networks; determining a convolutional neural network of the N second convolutional neural networks that satisfies the convergence condition as the lightweight convolutional neural network.
Optionally, the apparatus is further configured to determine a T value according to an optimized scaling factor corresponding to each convolutional layer in the original convolutional neural network and the number of convolutional kernels in each convolutional layer, where the T value is the number of convolutional kernels that need pruning, and T is an integer; sorting the convolution kernels in each convolution layer according to the norm of the convolution kernels; and pruning the convolution kernels ranked as the first T in each convolution layer.
Optionally, the apparatus is further configured to randomly generate the N sets of initial scaling factors according to the number of convolutional layers of the original convolutional neural network, where each initial scaling factor is a binary code, and each binary code corresponds to at least one convolutional layer of the original convolutional neural network.
Optionally, the apparatus is further configured to generate image feature data of a specified category using the conditional generation countermeasure network; training the lightweight convolutional neural network using the specified class of image feature data.
It should be noted that, the above modules may be implemented by software or hardware, and for the latter, the following may be implemented, but not limited to: the modules are all positioned in the same processor; alternatively, the modules are respectively located in different processors in any combination.
Embodiments of the present invention also provide a storage medium having a computer program stored therein, wherein the computer program is arranged to perform the steps of any of the above method embodiments when executed.
Alternatively, in the present embodiment, the storage medium may be configured to store a computer program for executing the steps of:
s1, obtaining N groups of initial scaling factor sets of the original convolutional neural network, wherein each group of initial scaling factor sets comprises M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1;
s2, optimizing the N groups of initial scaling factor sets by using an evolutionary algorithm to obtain a group of optimized scaling factor sets, wherein each optimized scaling factor in the group of optimized scaling factor sets is used for representing the convolution kernel pruning proportion of the corresponding convolution layer;
s3, pruning the convolution kernels of the original convolution neural network according to the set of optimized scaling factors to obtain a light-weight convolution neural network;
and S4, identifying the target image by using the lightweight convolutional neural network to obtain the identification result of the target image.
Optionally, in this embodiment, the storage medium may include, but is not limited to: various media capable of storing computer programs, such as a usb disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic disk, or an optical disk.
Embodiments of the present invention also provide an electronic device comprising a memory having a computer program stored therein and a processor arranged to run the computer program to perform the steps of any of the above method embodiments.
Optionally, the electronic apparatus may further include a transmission device and an input/output device, wherein the transmission device is connected to the processor, and the input/output device is connected to the processor.
Optionally, in this embodiment, the processor may be configured to execute the following steps by a computer program:
s1, obtaining N groups of initial scaling factor sets of the original convolutional neural network, wherein each group of initial scaling factor sets comprises M scaling factors, each initial scaling factor corresponds to at least one convolutional layer of the original convolutional neural network, and N and M are greater than or equal to 1;
s2, optimizing the N groups of initial scaling factor sets by using an evolutionary algorithm to obtain a group of optimized scaling factor sets, wherein each optimized scaling factor in the group of optimized scaling factor sets is used for representing the convolution kernel pruning proportion of the corresponding convolution layer;
s3, pruning the convolution kernels of the original convolution neural network according to the set of optimized scaling factors to obtain a light-weight convolution neural network;
and S4, identifying the target image by using the lightweight convolutional neural network to obtain the identification result of the target image.
Optionally, the specific examples in this embodiment may refer to the examples described in the above embodiments and optional implementation manners, and this embodiment is not described herein again.
It will be apparent to those skilled in the art that the modules or steps of the present invention described above may be implemented by a general purpose computing device, they may be centralized on a single computing device or distributed across a network of multiple computing devices, and alternatively, they may be implemented by program code executable by a computing device, such that they may be stored in a storage device and executed by a computing device, and in some cases, the steps shown or described may be performed in an order different than that described herein, or they may be separately fabricated into individual integrated circuit modules, or multiple ones of them may be fabricated into a single integrated circuit module. Thus, the present invention is not limited to any specific combination of hardware and software.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1.一种图像识别方法,其特征在于,包括:1. an image recognition method, is characterized in that, comprises: 获取原始卷积神经网络的N组初始缩放因子集合,其中,每组初始缩放因子集合中包括M个缩放因子,每个所述初始缩放因子对应于所述原始卷积神经网络的至少一层卷积层,N和M大于或等于1;Obtaining N groups of initial scaling factor sets of the original convolutional neural network, wherein each group of initial scaling factor sets includes M scaling factors, and each initial scaling factor corresponds to at least one layer of the original convolutional neural network. Layers, N and M are greater than or equal to 1; 使用进化算法对所述N组初始缩放因子集合进行优化,得到一组优化缩放因子集合,其中,所述一组优化缩放因子集合中的每个优化缩放因子用于表示对应卷积层的卷积核剪枝比例;Using an evolutionary algorithm to optimize the N groups of initial scaling factor sets to obtain a set of optimized scaling factor sets, wherein each optimized scaling factor in the set of optimized scaling factor sets is used to represent the convolution of the corresponding convolutional layer nuclear pruning ratio; 按照所述一组优化缩放因子集合对所述原始卷积神经网络的卷积核进行剪枝,得到轻量卷积神经网络;Prune the convolution kernel of the original convolutional neural network according to the set of optimized scaling factors to obtain a lightweight convolutional neural network; 使用所述轻量卷积神经网络对目标图像进行识别,得到所述目标图像的识别结果。Recognizing the target image by using the lightweight convolutional neural network to obtain the recognition result of the target image. 2.根据权利要求1所述的方法,其特征在于,使用进化算法对所述N组初始缩放因子集合进行优化,得到一组优化缩放因子集合,包括:2. The method according to claim 1, wherein an evolutionary algorithm is used to optimize the N groups of initial scaling factor sets to obtain a set of optimized scaling factor sets, comprising: 使用所述N组初始缩放因子集合中的每组初始缩放因子集合对所述原始卷积神经网络进行剪枝,得到N个第一卷积神经网络;Using each group of initial scaling factor sets in the N groups of initial scaling factor sets to prune the original convolutional neural network to obtain N first convolutional neural networks; 在所述N个第一卷积神经网络中存在满足预设收敛条件的卷积神经网络的情况下,确定满足所述预设收敛条件的卷积神经网络为所述轻量卷积神经网络,其中,所述预设收敛条件用于表示适应度函数的输出在阈值范围内。In the case where a convolutional neural network that satisfies a preset convergence condition exists in the N first convolutional neural networks, it is determined that the convolutional neural network that satisfies the preset convergence condition is the lightweight convolutional neural network, Wherein, the preset convergence condition is used to indicate that the output of the fitness function is within the threshold range. 3.根据权利要求2所述的方法,其特征在于,所述方法还包括:3. The method according to claim 2, wherein the method further comprises: 在所述N个第一卷积神经网络中不存在满足预设收敛条件的卷积神经网络的情况下,将任意两组所述初始缩放因子集合中在预设交叉点的缩放因子进行交叉,和/或,将所述缩放因子集合中在预设变异点的一个或多个初始缩放因子进行变异,得到N组第一缩放因子集合;In the case where there is no convolutional neural network that satisfies the preset convergence condition in the N first convolutional neural networks, the scaling factors at the preset intersection points in any two sets of the initial scaling factor sets are crossed, And/or, mutating one or more initial scaling factors at preset variation points in the scaling factor set to obtain N groups of first scaling factor sets; 分别使用所述N组第一缩放因子集合中的每组第一缩放因子集合对原始卷积神经网络进行剪枝,得到N个第二卷积神经网络;The original convolutional neural network is pruned by using each group of first scaling factor sets in the N groups of first scaling factor sets, respectively, to obtain N second convolutional neural networks; 确定所述N个第二卷积神经网络中满足所述收敛条件的卷积神经网络为所述轻量卷积神经网络。It is determined that the convolutional neural network satisfying the convergence condition among the N second convolutional neural networks is the light-weight convolutional neural network. 4.根据权利要求1至3中任一项所述的方法,其特征在于,按照所述一组优化缩放因子集合对所述原始卷积神经网络的卷积核进行剪枝,包括:4. The method according to any one of claims 1 to 3, wherein pruning the convolution kernel of the original convolutional neural network according to the set of optimized scaling factors, comprising: 根据所述原始卷积神经网络中每层卷积层对应的优化缩放因子,以及每层卷积层中的卷积核数量,确定T值,其中,所述T值是需要剪枝的卷积核数量,T为整数;The T value is determined according to the optimized scaling factor corresponding to each convolutional layer in the original convolutional neural network and the number of convolution kernels in each convolutional layer, where the T value is the convolution that needs to be pruned The number of cores, T is an integer; 对所述每层卷积层中的卷积核按照卷积核范数进行排序;Sorting the convolution kernels in the convolutional layers of each layer according to the convolution kernel norm; 对所述每层卷积层中排序为前T个的卷积核进行剪枝。Prune the top T convolution kernels in each convolutional layer. 5.根据权利要求1至3中任一项所述的方法,其特征在于,获取原始卷积神经网络的N组初始缩放因子集合,包括:5. The method according to any one of claims 1 to 3, wherein obtaining N groups of initial scaling factor sets of the original convolutional neural network, comprising: 根据所述原始卷积神经网络的卷积层数随机生成所述N组初始缩放因子集合,其中,每个所述初始缩放因子为二进制编码,每个所述二进制编码对应于所述原始卷积神经网络的至少一层卷积层。The N groups of initial scaling factor sets are randomly generated according to the number of convolutional layers of the original convolutional neural network, wherein each initial scaling factor is a binary code, and each binary code corresponds to the original convolutional At least one convolutional layer of the neural network. 6.根据权利要求1所述的方法,其特征在于,所述方法还包括:6. The method of claim 1, wherein the method further comprises: 使用条件生成对抗网络生成指定类别的图像特征数据;Use conditional generative adversarial network to generate image feature data of specified categories; 使用所述指定类别的图像特征数据对所述轻量卷积神经网络进行训练。The lightweight convolutional neural network is trained using the specified class of image feature data. 7.一种图像识别装置,其特征在于,包括:7. An image recognition device, characterized in that, comprising: 获取模块,用于获取原始卷积神经网络的N组初始缩放因子集合,其中,每组初始缩放因子集合中包括M个缩放因子,每个所述初始缩放因子对应于所述原始卷积神经网络的至少一层卷积层,N和M大于或等于1;an acquisition module, configured to acquire N groups of initial scaling factor sets of the original convolutional neural network, wherein each group of initial scaling factor sets includes M scaling factors, and each initial scaling factor corresponds to the original convolutional neural network At least one convolutional layer of , N and M are greater than or equal to 1; 优化模块,用于使用进化算法对所述N组初始缩放因子集合进行优化,得到一组优化缩放因子集合,其中,所述一组优化缩放因子集合中的每个优化缩放因子用于表示对应卷积层的卷积核剪枝比例;An optimization module, configured to use an evolutionary algorithm to optimize the N groups of initial scaling factor sets to obtain a set of optimized scaling factor sets, wherein each optimized scaling factor in the set of optimized scaling factor sets is used to represent a corresponding volume The convolution kernel pruning ratio of the multi-layer; 处理模块,用于按照所述一组优化缩放因子集合对所述原始卷积神经网络的卷积核进行剪枝,得到轻量卷积神经网络;a processing module, configured to prune the convolution kernel of the original convolutional neural network according to the set of optimized scaling factors to obtain a lightweight convolutional neural network; 识别模块,用于使用所述轻量卷积神经网络对目标图像进行识别,得到所述目标图像的识别结果。The recognition module is used to recognize the target image by using the light-weight convolutional neural network to obtain the recognition result of the target image. 8.根据权利要求7所述的装置,其特征在于,所述优化模块包括:8. The device according to claim 7, wherein the optimization module comprises: 第一处理单元,用于使用所述N组初始缩放因子集合中的每组初始缩放因子集合对所述原始卷积神经网络进行剪枝,得到N个第一卷积神经网络;a first processing unit, configured to prune the original convolutional neural network using each group of initial scaling factor sets in the N groups of initial scaling factor sets to obtain N first convolutional neural networks; 确定单元,用于在所述N个第一卷积神经网络中存在满足预设收敛条件的卷积神经网络的情况下,确定满足所述收敛条件的卷积神经网络为所述轻量卷积神经网络,其中,所述预设收敛条件用于表示适应度函数的输出在阈值范围内。A determination unit, configured to determine that the convolutional neural network that satisfies the convergence condition is the lightweight convolutional neural network when there is a convolutional neural network that satisfies a preset convergence condition in the N first convolutional neural networks A neural network, wherein the preset convergence condition is used to indicate that the output of the fitness function is within a threshold range. 9.一种存储介质,其特征在于,所述存储介质中存储有计算机程序,其中,所述程序可被终端设备或计算机运行时执行所述权利要求1至6任一项中所述的方法。9. A storage medium, wherein a computer program is stored in the storage medium, wherein the program can be executed by a terminal device or a computer to execute the method according to any one of claims 1 to 6 . 10.一种电子装置,包括存储器和处理器,其特征在于,所述存储器中存储有计算机程序,所述处理器被设置为运行所述计算机程序以执行所述权利要求1至6任一项中所述的方法。10. An electronic device comprising a memory and a processor, wherein a computer program is stored in the memory, and the processor is configured to run the computer program to execute any one of claims 1 to 6 method described in.
CN202110827593.9A 2021-07-21 2021-07-21 Image recognition method and device Pending CN113469277A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110827593.9A CN113469277A (en) 2021-07-21 2021-07-21 Image recognition method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110827593.9A CN113469277A (en) 2021-07-21 2021-07-21 Image recognition method and device

Publications (1)

Publication Number Publication Date
CN113469277A true CN113469277A (en) 2021-10-01

Family

ID=77881747

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110827593.9A Pending CN113469277A (en) 2021-07-21 2021-07-21 Image recognition method and device

Country Status (1)

Country Link
CN (1) CN113469277A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113850365A (en) * 2021-07-28 2021-12-28 浙江大华技术股份有限公司 Method, device, equipment and storage medium for compressing and transplanting convolutional neural network
CN114065920A (en) * 2021-11-18 2022-02-18 山东海量信息技术研究院 A method and system for image recognition based on channel-level pruning neural network
CN117313799A (en) * 2023-08-31 2023-12-29 佛山职业技术学院 Deep learning model compression method and device based on LSMOF

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108830292A (en) * 2018-05-08 2018-11-16 西北大学 Data classification model optimization method and classification method
CN110033083A (en) * 2019-03-29 2019-07-19 腾讯科技(深圳)有限公司 Convolutional neural networks model compression method and apparatus, storage medium and electronic device
CN111105035A (en) * 2019-12-24 2020-05-05 西安电子科技大学 Neural network pruning method based on combination of sparse learning and genetic algorithm
CN111507383A (en) * 2020-04-03 2020-08-07 厦门大学 A neural network automatic pruning method based on evolutionary algorithm
CN111667054A (en) * 2020-06-05 2020-09-15 北京百度网讯科技有限公司 Method and device for generating neural network model, electronic equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108830292A (en) * 2018-05-08 2018-11-16 西北大学 Data classification model optimization method and classification method
CN110033083A (en) * 2019-03-29 2019-07-19 腾讯科技(深圳)有限公司 Convolutional neural networks model compression method and apparatus, storage medium and electronic device
CN111105035A (en) * 2019-12-24 2020-05-05 西安电子科技大学 Neural network pruning method based on combination of sparse learning and genetic algorithm
CN111507383A (en) * 2020-04-03 2020-08-07 厦门大学 A neural network automatic pruning method based on evolutionary algorithm
CN111667054A (en) * 2020-06-05 2020-09-15 北京百度网讯科技有限公司 Method and device for generating neural network model, electronic equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
尚荣华等: "计算智能导论", 30 September 2019, 西安:西安电子科技大学出版社, pages: 303 - 308 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113850365A (en) * 2021-07-28 2021-12-28 浙江大华技术股份有限公司 Method, device, equipment and storage medium for compressing and transplanting convolutional neural network
CN113850365B (en) * 2021-07-28 2025-10-28 浙江大华技术股份有限公司 Convolutional neural network compression and transplantation method, device, equipment and storage medium
CN114065920A (en) * 2021-11-18 2022-02-18 山东海量信息技术研究院 A method and system for image recognition based on channel-level pruning neural network
CN117313799A (en) * 2023-08-31 2023-12-29 佛山职业技术学院 Deep learning model compression method and device based on LSMOF

Similar Documents

Publication Publication Date Title
CN111667022B (en) User data processing method, device, computer equipment and storage medium
CN109783582B (en) Knowledge base alignment method, device, computer equipment and storage medium
Cao et al. Automatic selection of t-SNE perplexity
CN112818690B (en) Semantic recognition method and device combined with knowledge graph entity information and related equipment
CN110442702B (en) Searching method and device, readable storage medium and electronic equipment
CN113657421B (en) Convolutional neural network compression method and device, image classification method and device
CN113469277A (en) Image recognition method and device
EP0672995A2 (en) Construction classification weights matrices for pattern recognition systems using reduced element feature subsets
JP2013206187A (en) Information conversion device, information search device, information conversion method, information search method, information conversion program and information search program
KR101183391B1 (en) Image comparison by metric embeddings
CN112766288A (en) Image processing model construction method and device, electronic equipment and readable storage medium
CN111126626A (en) Training method, device, server, platform and storage medium
CN117786705B (en) A sentence-level vulnerability detection method and system based on heterogeneous graph transformation network
Chattopadhyay et al. Feature selection using differential evolution with binary mutation scheme
CN112598062A (en) Image identification method and device
CN111126501A (en) Image identification method, terminal equipment and storage medium
CN112200275B (en) Artificial neural network quantification method and device
CN113963201A (en) Bone action recognition method and device, electronic equipment and storage medium
CN117033546B (en) Similar code searching method and system
WO2019220608A1 (en) Information processing device, information processing method, and information processing program
CN115759192A (en) A neural network acceleration method, device, equipment, chip and storage medium
Rahimpour et al. Class-discriminative feature embedding for meta-learning based few-shot classification
Kim et al. Automatic rank selection for high-speed convolutional neural network
CN111797973A (en) Method, device and electronic system for determining model structure
CN111191065A (en) Homologous image determining method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20211001

RJ01 Rejection of invention patent application after publication