[go: up one dir, main page]

CN112801897A - Image denoising method based on wide convolution neural network - Google Patents

Image denoising method based on wide convolution neural network Download PDF

Info

Publication number
CN112801897A
CN112801897A CN202110071024.6A CN202110071024A CN112801897A CN 112801897 A CN112801897 A CN 112801897A CN 202110071024 A CN202110071024 A CN 202110071024A CN 112801897 A CN112801897 A CN 112801897A
Authority
CN
China
Prior art keywords
sub
image
noise
network
wcnn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110071024.6A
Other languages
Chinese (zh)
Other versions
CN112801897B (en
Inventor
刘晶
刘润川
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Technology
Original Assignee
Xian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Technology filed Critical Xian University of Technology
Priority to CN202110071024.6A priority Critical patent/CN112801897B/en
Publication of CN112801897A publication Critical patent/CN112801897A/en
Application granted granted Critical
Publication of CN112801897B publication Critical patent/CN112801897B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20064Wavelet transform [DWT]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)

Abstract

本发明公开了一种基于宽卷积神经网络的图像去噪方法,通过将子网部署在每个小波子带上,负责捕捉和学习特定方向和特定尺度的图像特征,每个子网络不仅有自己独立的卷积神经网络结构,包含较少的卷积层,还具有自己的损失函数,这些损失函数被用来监督每个子网络的学习过程,使得WCNN能够使用一组学习参数来处理一定范围的噪声,当所有的损失函数都达到最优值时,利用小波逆变换从粗到细的方式获取干净的图像,提高卷积神经网络的去噪性能,同时减少训练时间。

Figure 202110071024

The invention discloses an image denoising method based on a wide convolutional neural network. By deploying a sub-network on each wavelet sub-band, it is responsible for capturing and learning image features of a specific direction and a specific scale. Each sub-network not only has its own A stand-alone convolutional neural network structure, containing fewer convolutional layers, also has its own loss function, which is used to supervise the learning process of each sub-network, enabling WCNN to use a set of learning parameters to handle a range of Noise, when all loss functions reach the optimal value, use the wavelet inverse transform to obtain clean images in a coarse-to-fine manner, improve the denoising performance of convolutional neural networks, and reduce training time.

Figure 202110071024

Description

Image denoising method based on wide convolution neural network
Technical Field
The invention belongs to the technical field of neural network image denoising, and relates to an image denoising method based on a wide convolution neural network.
Background
At present, image denoising methods based on a Convolutional Neural Network (CNN) are diversified, and the existing CNN structures for image denoising capture image features and improve denoising performance by deepening the CNN network structures. However, for the CNN model adopting a single stream structure, it is difficult to obtain image detail features from multiple directions, and if the network depth is increased blindly, the information flow of the image features will be weakened, making network training very difficult. Because the CNN of image denoising requires that the generated feature map in the network has the same size as the input image and the output image, the operation of the network of large-size feature mapping inevitably consumes a large amount of memory and time of a GPU even if the number of convolution layers is small, and the depth of the CNN network of image denoising is limited, namely, the optimal balance between the image denoising performance and the network training time cannot be obtained by increasing the number of convolution layers on a single chain.
Disclosure of Invention
The invention aims to provide an image denoising method based on a wide convolution neural network, and solves the problems of low denoising performance and long training time of the conventional neural network image denoising method.
The technical scheme adopted by the invention is that the image denoising method based on the wide convolution neural network is characterized by comprising the following steps of;
step 1, constructing a network WCNN;
step 2, training a network WCNN;
step 2.1, setting a data set comprising a training set, a verification set and a test set;
step 2.2, setting parameters for training the WCNN;
and 2.3, setting a training platform of the network WCNN.
The present invention is also characterized in that,
the network WCNN constructed in step 1 includes 10 subnets, which are ResNet1, ResNet2, ResNet3, ResNet4, ResNet5, ResNet6, UNet1, UNet2, UNet3, and densneen 1, respectively; ten wavelet sub-bands are obtained by wavelet tri-layer decomposition of the image, the ten wavelet sub-bands are HH1, LH1, HL1, HH2, LH2, HL2, HH3, LH3, HL3 and LL3 respectively, and 10 sub-nets correspond to feature maps of the ten wavelet sub-bands respectively responsible for learning the image.
The specific steps of each subnet in the network WCNN constructed in step 1 are as follows:
step 1.1, firstly, six subnets of ResNet1, ResNet2, ResNet3, ResNet4, ResNet5 and ResNet6 are designed; the six sub-networks are correspondingly responsible for training HH1, LH1, HL1, HH2, LH2 and HL2 fine sub-bands obtained by decomposing the first layer and the second layer of the wavelet; adopting a ResNet structure, utilizing residual learning to directly estimate noise, and estimating a denoised wavelet sub-band through jumping; wherein three subnets, ResNet1, ResNet2, and ResNet3, are comprised of 6 standard convolutional layers, and ResNet4, ResNet5, and ResNet6 are comprised of 8 standard convolutional layers;
step 1.2, then designing three subnets of UNet1, UNet2 and UNet 3; the three subnets are responsible for training HH3, LH3 and HL3 fine subbands obtained by wavelet third-layer decomposition, and adopt a UNet structure, and have 6 convolutional layers in total, wherein 4 convolutional layers are formed by convolution obtained by expansion convolutional and standard convolution operation;
step 1.3, designing a DenseNet subnet; it is responsible for training LL3 rough sub-band obtained by wavelet third-layer decomposition, adopts DenseNet structure, and is composed of 4 dense blocks containing 3-layer convolution;
step 1.4, designing a loss function of each subnet;
and step 1.5, performing inverse wavelet transform on the ten wavelet sub-bands processed by each sub-network when the loss function of each sub-band reaches an optimal value, and obtaining an image with clear details and cleanness.
The step 1.4 is specifically as follows:
step 1.4.1, loss function of wavelet transform coarse sub-band adopts mean square error measurement MSEl
Figure BDA0002905781310000031
Wherein x (i, j) and y (i, j) represent the estimated image and corresponding, respectively, wavelet coefficient values for the net image, and c, w, and h represent the channel, width, and height, respectively, of the input subband pair;
step 1.4.2, calculating a loss function of a wavelet transform fine sub-band, introducing a weight factor delta and an adjustment factor beta into a mean square error measurement index (1) formula, and calculating a loss function MSE of the fine sub-bandhThe following were used:
Figure BDA0002905781310000032
wherein the weighting factor δ is calculated by:
Figure BDA0002905781310000033
here, the ave table is an average value of wavelet coefficients of each fine subband, the average value ave of each subband coefficient is calculated by equation (4), and the adjustment factor β is calculated by equation (5):
Figure BDA0002905781310000034
Figure BDA0002905781310000035
where σ represents the noise intensity.
As the noise level increases in step 1.4, the amplitude of the noise in the subband increases and may be greater than the average of the subband coefficients; to prevent these noise coefficients, which are larger than the average of the subband coefficients, from being enhanced, an adjustment factor β is used to intervene; if the variance σ of the noise level is above 45, then the subband coefficient value is not less than 1.2 times the average value and is considered as the image detail coefficient, giving a weight of δ 1.1; thereby suppressing coefficients less than 1.2 times the average, which are considered to represent noise information; the ave of each fine subband is different and is closely related to the noise coefficient and the feature coefficient of each subband.
The training set in step 2.1 consists of 800 images of the data set DIV2K, 200 images of the data set BSD, and 4744 images of the data set WED.
The validation set in step 2.1 consists of the images in dataset RNI5 and 300 images of dataset DIV 2K.
The test Set in step 2.1 consists of images in the data Set CSet8 and images in Set 12.
The size of the images in the training set in step 2.2 is set to 256 × 256, gaussian noise with a specific noise level, i.e., σ ═ 5, 15, 25, 35, 45, 55, 65, and 75, is added to the clean images, generating 256 × 8000 image pairs, with the network model labeled WCNN1 for the former and WCNN2 for the latter, respectively, trained using noise images with low noise intensity superimposed, i.e., σ ≦ 45, and noise images with high noise level superimposed, i.e., 45< σ ≦ 75; during testing, when the variance of the noise intensity of the tested noise image is not more than 45, denoising by using a network WCNN 1; if the variance of the noise intensity of the test noise image is greater than 45, the noise is removed by the WCNN2 network.
Step 2.3, building a WCNN network on a TensorFlow framework, updating the WCNN network by using an Adam optimizer, wherein an activation function is a ReLU, and the learning rate of all subnets is initially set to be 9 multiplied by 10-4And after every 16 cycles, the learning rate is reduced by one third, and the NVIDIA RTX 2080Ti is used for training the WCNN network.
The invention has the beneficial effects that:
1. the invention places a CNN on the sub-bands of different scales and different directions of the image, which is beneficial to fully learning the image characteristics and details of each scale and each direction, thereby inhibiting speckle noise and simultaneously maintaining high resolution of the image.
2. Each sub-network constituting the network WCNN has its own structure and loss function, ensuring that each wavelet sub-band of the noise image is most similar to the corresponding sub-band of the clean image after network training.
3. Each wavelet sub-band concentrates image features of a specific dimension in a specific direction, and therefore, having a small number of convolution layers or simple sub-nets is sufficient to capture and learn the image features contained in each wavelet sub-band, and can effectively suppress noise in each sub-band.
4. The sub-networks used for training the wavelet sub-bands are independent from each other, and the sub-networks can run on a single computer or a plurality of computers in parallel, so that the network training time is shortened.
Drawings
FIG. 1 is a diagram of the WCNN network framework of the present invention;
FIG. 2 is a network architecture diagram of the first and second layers of wavelet fine sub-bands of the present invention;
FIG. 3 is a network structure diagram of a third layer wavelet fine sub-band of the present invention;
FIG. 4 is a network structure diagram of a third layer wavelet coarse sub-band according to the present invention;
FIG. 5(a) is a graph comparing the impact of different subnet structures on PSNR network performance;
fig. 5(b) is a graph comparing the impact of different subnet structures on SSIM network performance;
FIG. 5(c) is a graph comparing the impact of different subnet structures on IFC network performance;
FIG. 6(a) is a graph comparing the impact of different loss functions on PSNR network performance;
FIG. 6(b) is a graph comparing the impact of different loss functions on SSIM network performance;
FIG. 6(c) is a graph comparing the impact of different loss functions on IFC network performance;
FIG. 7(a) is a visual effect diagram obtained by denoising a gray image;
FIG. 7(b) is a visual effect diagram obtained by denoising a gray scale image, BM 3D;
FIG. 7(c) is a visual effect diagram obtained by DnCNN with de-noising of gray level images;
FIG. 7(d) is a graph of visual effect obtained by denoising a gray image and MWCNN;
FIG. 7(e) is a diagram of visual effects obtained by UDNet in denoising gray images;
FIG. 7(f) is a diagram of the visual effect obtained by denoising and FFDNet of a gray image;
FIG. 7(g) is a visual effect diagram obtained by the WCNN after the gray image is denoised;
FIG. 8(a) is a visual effect diagram obtained by denoising a color image;
FIG. 8(b) is a diagram of the visual effect obtained by the color image denoising, BM 3D;
FIG. 8(c) is a diagram of the visual effect obtained by DnCNN for color image denoising;
FIG. 8(d) is a diagram of the visual effect obtained by MWCNN for denoising color images;
FIG. 8(e) is a diagram of the visual effect obtained by UDNet in denoising color images;
FIG. 8(f) is a diagram of the visual effect obtained by color image denoising and FFDNet;
fig. 8(g) is a visual effect diagram obtained by WCNN in denoising color images.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.
The invention relates to an image denoising method based on a wide convolution neural network, which is characterized in that when an image is denoised, the image is converted into a plurality of wavelet sub-bands for denoising through wavelet decomposition, each wavelet sub-band is denoised by using a CNN to learn the characteristic mapping of the wavelet sub-band with a specific direction, a specific scale and a small size and inhibiting the noise coefficient contained in the wavelet sub-band, so that a plurality of CNNs with different independent structures are arranged on each wavelet sub-band of the image, the image detail characteristic with a certain direction and a certain scale in each sub-band is captured, and the noise in a certain intensity range is removed by using a group of learning parameters, so that the best balance is obtained between the image denoising performance and the network training time.
The invention relates to an image denoising method based on a wide convolution neural network, which is implemented by the following steps:
step 1, constructing a network WCNN;
the network WCNN constructed in step 1 includes 10 subnets, as shown in fig. 1, namely ResNet1, ResNet2, ResNet3, ResNet4, ResNet5, ResNet6, UNet1, UNet2, UNet3 and densneet 1; ten wavelet sub-bands are obtained through wavelet three-layer decomposition of the image, the ten wavelet sub-bands are HH1, LH1, HL1, HH2, LH2, HL2, HH3, LH3, HL3 and LL3 respectively, and 10 sub-nets correspond to feature mapping of the ten wavelet sub-bands respectively responsible for learning the image;
the specific steps of each subnet in the network WCNN constructed in step 1 are as follows:
step 1.1, firstly, six subnets of ResNet1, ResNet2, ResNet3, ResNet4, ResNet5 and ResNet6 are designed; the six sub-networks are correspondingly responsible for training HH1, LH1, HL1, HH2, LH2 and HL2 fine sub-bands obtained by decomposing the first layer and the second layer of the wavelet; adopting a ResNet structure, utilizing residual learning to directly estimate noise, and estimating a denoised wavelet sub-band through jumping; wherein the three subnets ResNet1, ResNet2, and ResNet3 are comprised of 6 standard convolutional layers, and ResNet4, ResNet5, and ResNet6 are comprised of 8 standard convolutional layers, as shown in FIG. 2;
step 1.2, then designing three subnets of UNet1, UNet2 and UNet 3; the three sub-networks are responsible for training HH3, LH3 and HL3 fine sub-bands obtained by wavelet third-layer decomposition, and 6 convolutional layers are formed in total by adopting a UNet structure, wherein 4 convolutional layers are formed by convolution obtained by extended convolutional and standard convolution operation, and are shown in FIG. 3;
step 1.3, designing a DenseNet subnet; it is responsible for training the coarse LL3 subband obtained by the third-layer decomposition of the wavelet, and is composed of 4 dense blocks containing 3-layer convolution by adopting a DenseNet structure, as shown in fig. 4;
step 1.4, a loss function of each subnet is designed.
Step 1.4.1, loss function of wavelet transform coarse sub-band adopts mean square error measurement MSEl
Figure BDA0002905781310000081
Wherein x (i, j) and y (i, j) represent the estimated image and corresponding, respectively, wavelet coefficient values for the net image, and c, w, and h represent the channel, width, and height, respectively, of the input subband pair;
step 1.4.2, calculating a loss function of a wavelet transform fine sub-band, introducing a weight factor delta and an adjustment factor beta into a mean square error measurement index (1) formula, and calculating a loss function MSE of the fine sub-bandhThe following were used:
Figure BDA0002905781310000082
wherein the weighting factor δ is calculated by:
Figure BDA0002905781310000083
here, the ave table is an average value of wavelet coefficients of each fine subband, the average value ave of each subband coefficient is calculated by equation (4), and the adjustment factor β is calculated by equation (5):
Figure BDA0002905781310000084
Figure BDA0002905781310000085
where σ represents the noise intensity;
as the noise level increases, the amplitude of the noise in the subband increases and may be greater than the average of the subband coefficients; to prevent these noise coefficients, which are larger than the average of the subband coefficients, from being enhanced, an adjustment factor β is used to intervene; if the variance σ of the noise level is above 45, then the subband coefficient value is not less than 1.2 times the average value and is considered as the image detail coefficient, giving a weight of δ 1.1; thereby suppressing coefficients less than 1.2 times the average, which are considered to represent noise information; it is noted that the ave of each fine subband is different and is closely related to the noise coefficient and the feature coefficient of each subband.
And step 1.5, performing inverse wavelet transform on the ten wavelet sub-bands processed by each sub-network when the loss function of each sub-band reaches an optimal value, and obtaining an image with clear details and cleanness.
Each subnet of the WCNN network in step 1 is a wavelet of the respective independent training image, and the number of convolution layers of each subnet is small, so that the WCNN network has 10 independent feature extraction and learning channels. Thus, the performance of the WCNN network in suppressing noise and capturing image features is improved by extending the network width rather than deepening the network depth; because each subnet of the WCNN is independent, the subnets can be trained on a plurality of computers in parallel, so that the training time of the WCNN can be obviously shortened under the condition of not influencing the network performance, the loss function of each subnet is related to the characteristic of the wavelet subband coefficient trained by the subnet, and the contradiction between image detail preservation and noise removal is favorably coordinated.
Step 2, training a network WCNN;
and 2.1, setting a data set comprising a training set, a verification set and a test set.
The training set consists of 800 images of the data set DIV2K, 200 images of the data set BSD, and 4744 images of the data set WED;
the validation set consists of the images in dataset RNI5 and 300 images of dataset DIV 2K;
and (3) test set: consisting of images in the dataset CSet8 and images in Set 12;
step 2.2, setting parameters for training the WCNN;
the size of the images in the training set is set to 256 × 256, gaussian noise with a certain noise level, i.e., σ ═ 5, 15, 25, 35, 45, 55, 65, and 75, is added to the clean images, generating 256 × 8000 image pairs, with the network model labeled WCNN1 and WCNN2, respectively, trained using noise images with low noise intensity, i.e., σ ≦ 45, and using noise images with high noise level, i.e., 45< σ ≦ 75, respectively;
during testing, when the variance of the noise intensity of the tested noise image is not more than 45, denoising by using a network WCNN 1; if the variance of the noise intensity of the test noise image is larger than 45, denoising by using a WCNN2 network;
step 2.3, setting a training platform of the network WCNN;
building a WCNN network on a TensorFlow framework, updating by using an Adam optimizer, setting an activation function to be ReLU, and initially setting the learning rate of all subnets to be 9 multiplied by 10-4After every 16 cycles, the learning rate is reduced by one third, and the NVIDIA RTX 2080Ti is used for training the WCNN network, which takes about 7 hours.
Examples
The invention aims to provide an image denoising method based on a wide convolutional neural network, which aims to improve the denoising performance of the convolutional neural network and reduce the training time, so that the performance of the WCNN is tested and verified through experiments. Firstly, the advantages brought by independent training of each subnet are inspected; secondly, researching the influence of different sub-networks forming the WCNN on the image denoising quality; third, the impact of the loss function on the WCNN performance was investigated.
The effect of the basic components on the WCNN performance was demonstrated by ablation studies of three experiments. And finally, selecting five representative denoising methods as comparison baselines, and comparing and comprehensively analyzing the denoising effect of the WCNN method: one is a wavelet-based CNN denoising method (MWCNN 1), three CNN-based methods (DnCNN 2, UDNet 3 and FFDNet 4), and a representative conventional method (BM3D 5).
Ablation experiment
1. Independent training subnet study
In the part, it is verified that training each subnet independently can not only shorten training time, but also ensure denoising quality. One advantage of the WCNN is that it can be divided into several sub-networks, which learn the feature mapping of the sub-bands in parallel on different computers, respectively, and the trained sub-bands are integrated by Haar inverse wavelet transform to obtain a clear and clean image, where the WCNN trained in this way is denoted as WCNN-1; the WCNN is a model for learning all wavelet subband feature mappings on a single computer, and here, the WCNN is compared with BM3D, DnCNN, MWCNN, FFDNet and UDNet denoising benchmark methods, and table 1 shows watermark transparency performance index results of these methods.
TABLE 1 comparison of different method run times and PSNR (dB)/SSIM/IFC index
Figure BDA0002905781310000111
Table 1 shows the GPU run time and PSNRs/SSIMs/IFCs when WCNN and the comparison method are applied to 200 grayscale images from the DIV2K dataset, adding σ 25 noise, and both WCNN-1 and WCNN achieve the best performance at relatively low execution time compared to the most advanced denoising method, from which it can be seen that the training time and execution time of WCNN are slightly more than WCNN-1, because each subnet of WCNN-1 can learn their feature maps in parallel on multiple computers, while each subnet of WCNN can only run on one computer; due to the multi-scale multi-directional decomposition of the wavelet, the sub-bands of the wavelet not only have common directional characteristics, such as larger coefficients of horizontal directional edges in the LH sub-band, but also have smaller sizes, and each sub-band does not lose any detail characteristics due to clipping. Thus, there is no need for a deep CNN with multiple convolutional layers to capture the characteristics of these subbands, and each subnet has its own loss function, and the parameters of the subnets can be controlled and adjusted to ensure that each estimated subband is very similar to the subband of a clean image, which ensures that the WCNN can obtain higher PSNRs/SSIM at a relatively fast speed than other comparative methods, even if the WCNN is running on one computer.
2. Subnet structure study
As shown in fig. 5(a) -5(c), the positive impact of the subnet structure on the WCNN performance is shown; WCNN-2 represents a variation of WCNN in which all subnets are designed with the structure of ResNet, i.e., the structure shown in fig. 2, with ten convolutional layers per subnet. The performance of MWCNN is used as a base line, the image is subjected to multi-layer wavelet decomposition, and all obtained wavelet sub-bands are input into one CNN for feature learning and training; FIGS. 5(a) -5(c) show a comparison of proposed WCNN method, WCNN-2 method, and MWCNN method in terms of PSNRs/SSIMs/IFCs; these data are from the average of 200 denoised images; by comparing fig. 5(a) -5(c), it is shown that the performance of WCNN is significantly better than that of MWCNN, and the performance of WCNN-2 is slightly higher than that of MWCNN method; this shows that the strategy of training different sub-bands by adopting different CNN structures can significantly improve the denoising performance of the WCNN.
3. Study of loss function
The following experiments were performed: each subnet uses the same loss function, namely equation (1) learns the feature mapping of the subnet, and the WCNN is expressed as WCNN-3; UDNet is considered herein as a comparison baseline because it is able to process images with a range of noise levels using a single network; three networks WCNN-3, WCNN and UDNet were trained using 1000 images from BSD and WED datasets with an image plus noise variance σ of 45; the 200 images were then tested using these three networks, with a noise variance σ of 5 for 40 images, a noise variance σ of 15 for 40 images, a noise variance σ of 25 for 40 images, a noise variance σ of 35 for 40 images, and a noise variance σ of 45 for 40 images.
As shown in fig. 6(a) -6(c), the probability distributions of PSNR, SSIM and IFC gains for these 200 images are shown; the white histogram represents the distribution of index values obtained by the method WCNN-3 compared with the standard method UDNet, and one part of the index values obtained from 200 images is lower than the value obtained by the standard method UDNet and is distributed on the left half part of an abscissa 0; the black bar graph represents the distribution of the index values obtained by the WCNN method compared to the UDNet method, the index values obtained from 200 images are substantially higher than those obtained by UDNet, and almost all the values are distributed in the right half of the abscissa 0; these gain values are obtained from the baseline of WCNN and WCNN-3 relative to UDNet, and the PSNR/SSIM/IFC gains in FIGS. 6(a) -6(c) illustrate that the performance of WCNN far exceeds that of UDNet, and that WCNN-3 performs slightly less than UDNet, indicating that different loss functions can significantly improve the performance of WCNN when WCNN uses a trained set of parameters to handle a range of noise.
Second, Performance comparison with comparison methods
To fully verify the performance of WCNN, the quality of WCNN and other methods when processing images with σ ═ 5, 15, 25, 35, 45, 55, 65, or 75 noise were investigated. Furthermore, the performance of the WCNN +, which is another variation of the WCNN, in which the number of convolutional layers in each subnet is increased to 15, three evaluation values of PSNR, SSIM and IFC obtained by these methods, i.e., an average value of 20 gray images and an average value of 20 color images, as shown in table 2, visual qualities of the denoised images, the gray maps Parrot and the color map Comic, respectively, as shown in fig. 7(a) -7(g) and fig. 8(a) -8(g), are compared for effects, and a target region of interest (ROI) enlarged by bicubic interpolation (x 2) is displayed at a corner for comparing detailed features of the denoised images.
TABLE 2 PSNR (dB)/SSIM/IFC index obtained by different methods
Figure BDA0002905781310000141
The WCNN in Table 2 gives the best numerical results, and the method not only has the highest average PSNRs, but also has relatively high SSIM and IFCs; high PSNRs indicate that the denoised image is closest to the original clean image, and high SSIM and IFC values indicate that: these methods can recover the edge and texture details, as shown in fig. 7(a) -7(g) and fig. 8(a) -8(g), the visual quality obtained from the WCNN is quite excellent, some minor artifacts appear only at certain edges, and furthermore, the evaluation results based on PSNRs/SSIMs/IFCs are better when more convolutional layers are used per subnet of the WCNN, and the method of the present invention is superior to the current state-of-the-art denoising method.
The invention relates to an image denoising method based on a wide convolution neural network, which improves the image denoising performance by training and learning wavelet sub-bands for each sub-network in parallel so that the WCNN network expands the width of the network instead of the depth; each subnet can run on different computers, thereby shortening the network training time; each subnet captures image characteristics and noise with a specific scale and a specific direction, so that each subnet has a simple structure and requires fewer convolution layers; each subnet has a loss function suitable for itself, so that each trained noise subband and a clean image subband can be ensured to be most similar; and calculating a loss function of the fine sub-band, and enhancing the influence of the image characteristic coefficient, so that the image characteristic details of the de-noised image are more reserved.

Claims (10)

1.一种基于宽卷积神经网络的图像去噪方法,其特征在于,具体包括以下步骤实施;1. an image denoising method based on wide convolutional neural network, is characterized in that, specifically comprises the following steps to implement; 步骤1,构建网络WCNN;Step 1, construct the network WCNN; 步骤2,训练网络WCNN;Step 2, train the network WCNN; 步骤2.1,设置包含训练集,验证集和测试集的数据集;Step 2.1, set the dataset including training set, validation set and test set; 步骤2.2,设置训练WCNN网络的参数;Step 2.2, set the parameters for training the WCNN network; 步骤2.3,设置网络WCNN的训练平台。Step 2.3, set the training platform of the network WCNN. 2.根据权利要求1所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤1中构建的网络WCNN包含10个子网,分别为ResNet1、ResNet2、ResNet3、ResNet4、ResNet5、ResNet6、UNet1、UNet2、UNet3和DenseNet1;通过图像经小波三层分解获得十个小波子带,十个小波子带分别为HH1、LH1、HL1、HH2、LH2、HL2、HH3、LH3、HL3和LL3,10个子网对应分别负责学习图像的十个小波子带的特征映射。2. a kind of image denoising method based on wide convolutional neural network according to claim 1, is characterized in that, the network WCNN constructed in described step 1 comprises 10 sub-networks, respectively ResNet1, ResNet2, ResNet3, ResNet4 , ResNet5, ResNet6, UNet1, UNet2, UNet3 and DenseNet1; ten wavelet subbands are obtained by decomposing the image through three layers of wavelet, and the ten wavelet subbands are HH1, LH1, HL1, HH2, LH2, HL2, HH3, LH3, HL3 and LL3, 10 sub-networks are responsible for learning the feature maps of the ten wavelet sub-bands of the image respectively. 3.根据权利要求2所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤1中构建的网络WCNN中每个子网的具体步骤为:3. a kind of image denoising method based on wide convolutional neural network according to claim 2, is characterized in that, the concrete steps of each subnet in the network WCNN constructed in described step 1 are: 步骤1.1,首先设计ResNet1、ResNet2、ResNet3、ResNet4、ResNet5、ResNet6六个子网;这六个子网对应负责训练小波第一层和第二层分解得到的HH1,LH1,HL1,HH2,LH2,HL2精细子带;采用ResNet结构,利用残差学习直接估计噪声,通过跳接估计去噪后小波子带;其中ResNet1、ResNet2、和ResNet3三个子网由6个标准卷积层组成,ResNet4、ResNet5和ResNet6由8个标准卷积层组成;Step 1.1, first design ResNet1, ResNet2, ResNet3, ResNet4, ResNet5, ResNet6 six subnets; these six subnets correspond to the HH1, LH1, HL1, HH2, LH2, HL2 fine-tuned for the first and second layers of the wavelet decomposition Sub-band; adopt ResNet structure, use residual learning to directly estimate noise, and estimate denoised wavelet sub-band by jumping; among them, ResNet1, ResNet2, and ResNet3 three sub-networks are composed of 6 standard convolutional layers, ResNet4, ResNet5 and ResNet6 It consists of 8 standard convolutional layers; 步骤1.2,然后设计UNet1、UNet2和UNet3三个子网;这三个子网负责训练小波第三层分解得到的HH3,LH3,HL3精细子带,采用UNet结构,共有6个卷积层,其中4个是由扩展卷积和标准卷积操作得到的卷积构成混合卷积层;Step 1.2, then design three sub-networks UNet1, UNet2 and UNet3; these three sub-networks are responsible for training the HH3, LH3, HL3 fine sub-bands obtained by the third layer of wavelet decomposition, using the UNet structure, there are 6 convolutional layers, of which 4 are It is a hybrid convolution layer composed of convolution obtained by extended convolution and standard convolution operations; 步骤1.3,再设计DenseNet子网;它负责训练小波第三层分解得到的LL3粗糙子带,采用DenseNet结构,由4个包含3层卷积的稠密块组成;Step 1.3, then design the DenseNet sub-network; it is responsible for training the LL3 rough sub-band obtained by the third layer of wavelet decomposition, adopts the DenseNet structure, and consists of 4 dense blocks containing 3 layers of convolution; 步骤1.4,设计每个子网的损失函数;Step 1.4, design the loss function of each subnet; 步骤1.5,经过每个子网处理过的十个小波子带,在每个子带的损失函数都达到最优值时,进行小波逆变换,获得细节清晰且干净的图像。Step 1.5: After ten wavelet sub-bands processed by each sub-network, when the loss function of each sub-band reaches the optimal value, perform wavelet inverse transformation to obtain clear and clean images with clear details. 4.根据权利要求3所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤1.4具体为:4. a kind of image denoising method based on wide convolutional neural network according to claim 3, is characterized in that, described step 1.4 is specifically: 步骤1.4.1,小波变换粗子带的损耗函数采用均方误差度量MSElStep 1.4.1, the loss function of the coarse subband of the wavelet transform adopts the mean square error metric MSE l ;
Figure FDA0002905781300000021
Figure FDA0002905781300000021
其中x(i,j)和y(i,j)分别表示估计的图像和相应的,净图像的小波系数值,c、w和h分别表示输入子带对的信道、宽度和高度;where x(i,j) and y(i,j) denote the estimated image and the corresponding, net image wavelet coefficient values, respectively, and c, w, and h denote the channel, width, and height of the input subband pair, respectively; 步骤1.4.2,计算小波变换精细子带的损耗函数,在均方误差度量指标(1)式中引入权重因子δ和调整因子β,计算精细子带的损耗函数MSEh如下:Step 1.4.2, calculate the loss function of the fine subband of the wavelet transform, introduce the weight factor δ and the adjustment factor β into the mean square error metric index (1), and calculate the loss function MSE h of the fine subband as follows:
Figure FDA0002905781300000022
Figure FDA0002905781300000022
其中,权重因子δ由下式计算:Among them, the weight factor δ is calculated by the following formula:
Figure FDA0002905781300000031
Figure FDA0002905781300000031
这里的ave表,每个精细子带的小波系数的平均值,每个子带系数平均值ave由式(4)计算获得,调整因子β由式(5)计算得到:The ave table here, the average value of the wavelet coefficients of each fine subband, the average value ave of each subband coefficient is calculated by the formula (4), and the adjustment factor β is calculated by the formula (5):
Figure FDA0002905781300000032
Figure FDA0002905781300000032
Figure FDA0002905781300000033
Figure FDA0002905781300000033
其中,σ表示噪声强度。where σ represents the noise intensity.
5.根据权利要求4所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,在所述步骤1.4中当噪声级增加时,子带中的噪声的振幅增大并且可能大于子带系数的平均值;为了防止这些大于子带系数平均值的噪声系数被增强,使用调整因子β来干预;如果噪声级的方差σ在45以上,那么子带系数值不小于平均值的1.2倍,才被认为是图像细节系数,赋予一个δ=1.1的权重;从而抑制了小于平均值1.2倍的系数,而这些系数被认为表示的是噪声信息;每个精细子带的ave是不同的,它与每个子带的噪声系数和特征系数密切相关。5. An image denoising method based on a wide convolutional neural network according to claim 4, wherein in the step 1.4, when the noise level increases, the amplitude of the noise in the sub-band increases and may is greater than the average value of the subband coefficients; in order to prevent these noise figures larger than the average value of the subband coefficients from being enhanced, an adjustment factor β is used to intervene; if the variance σ of the noise level is above 45, then the subband coefficient value is not less than the average value of 1.2 times, it is considered as the image detail coefficient, giving a weight of δ=1.1; thus suppressing coefficients less than 1.2 times the average value, and these coefficients are considered to represent noise information; the ave of each fine subband is different , which is closely related to the noise figure and eigencoefficient of each subband. 6.根据权利要求1所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤2.1中训练集由数据集DIV2K的800幅图像,数据集BSD的200幅图像,以及数据集WED中的4744幅图像构成。6. A kind of image denoising method based on wide convolutional neural network according to claim 1, it is characterized in that, in described step 2.1, the training set consists of 800 images of data set DIV2K, 200 images of data set BSD , and the 4744 images in the dataset WED. 7.根据权利要求6所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤2.1中验证集由数据集RNI5中的图像和数据集DIV2K的300张图像构成。7. A kind of image denoising method based on wide convolutional neural network according to claim 6, is characterized in that, in described step 2.1, the verification set is made up of the image in data set RNI5 and 300 images of data set DIV2K . 8.根据权利要求7所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤2.1中测试集由数据集CSet8中的图像和Set12中的图像构成。8 . The image denoising method based on a wide convolutional neural network according to claim 7 , wherein the test set in the step 2.1 is composed of the images in the dataset CSet8 and the images in the Set12. 9 . 9.根据权利要求8所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤2.2中训练集中的图像的大小设置为256×256,将具有特定噪声级的高斯噪声,即σ=5、15、25、35、45、55、65和75添加到干净的图像中,生成256×8000个图像对,分别使用叠加低噪声强度,即σ≤45的噪声图像和使用叠加高噪声水平,即45<σ≤75的噪声图像训练网络WCNN,前者得到的网络模型标记为WCNN1,后者标记为WCNN2;测试时,当测试噪声图像的噪声强度方差不大于45,用网络WCNN1进行除噪;如果测试噪声图像的噪声强度方差大于45,则用WCNN2网络除噪。9. An image denoising method based on a wide convolutional neural network according to claim 8, wherein the size of the images in the training set in the step 2.2 is set to 256×256, and the Gaussian noise, i.e. σ = 5, 15, 25, 35, 45, 55, 65, and 75 is added to the clean image, resulting in 256 × 8000 image pairs, each using superimposed noise images with low noise intensity, i.e. σ ≤ 45 and using the superimposed high noise level, that is, 45<σ≤75 noise image training network WCNN, the former obtained network model is marked as WCNN1, the latter is marked as WCNN2; during testing, when the noise intensity variance of the test noise image is not greater than 45, The network WCNN1 is used for denoising; if the noise intensity variance of the test noise image is greater than 45, the WCNN2 network is used for denoising. 10.根据权利要求9所述的一种基于宽卷积神经网络的图像去噪方法,其特征在于,所述步骤2.3在TensorFlow框架搭建WCNN网络,并用Adam优化器进行更新,激活函数是ReLU,所有子网的学习速率初始设置为9×10-4,每16个周期后学习速率降低三分之一,用NVIDIARTX 2080Ti训练WCNN网络。10. a kind of image denoising method based on wide convolutional neural network according to claim 9, is characterized in that, described step 2.3 builds WCNN network in TensorFlow framework, and uses Adam optimizer to update, and activation function is ReLU, The learning rate of all subnets is initially set to 9 × 10-4 , and the learning rate is reduced by one third after every 16 epochs, and the WCNN network is trained with NVIDIA RTX 2080Ti.
CN202110071024.6A 2021-01-19 2021-01-19 Image denoising method based on wide convolution neural network Active CN112801897B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110071024.6A CN112801897B (en) 2021-01-19 2021-01-19 Image denoising method based on wide convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110071024.6A CN112801897B (en) 2021-01-19 2021-01-19 Image denoising method based on wide convolution neural network

Publications (2)

Publication Number Publication Date
CN112801897A true CN112801897A (en) 2021-05-14
CN112801897B CN112801897B (en) 2024-02-09

Family

ID=75810485

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110071024.6A Active CN112801897B (en) 2021-01-19 2021-01-19 Image denoising method based on wide convolution neural network

Country Status (1)

Country Link
CN (1) CN112801897B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114219820A (en) * 2021-12-08 2022-03-22 苏州工业园区智在天下科技有限公司 Neural network generation method, denoising method and device
CN116385280A (en) * 2023-01-09 2023-07-04 爱芯元智半导体(上海)有限公司 Image noise reduction system and method and noise reduction neural network training method
CN116703772A (en) * 2023-06-15 2023-09-05 山东财经大学 Image denoising method, system and terminal based on adaptive interpolation algorithm

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090324117A1 (en) * 2008-06-27 2009-12-31 Microsoft Corporation Image denoising techniques
US10032256B1 (en) * 2016-11-18 2018-07-24 The Florida State University Research Foundation, Inc. System and method for image processing using automatically estimated tuning parameters
CN110276726A (en) * 2019-05-13 2019-09-24 南昌大学 An image deblurring method guided by multi-channel network prior information
CN110599409A (en) * 2019-08-01 2019-12-20 西安理工大学 Convolutional neural network image denoising method based on multi-scale convolutional groups and parallel

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090324117A1 (en) * 2008-06-27 2009-12-31 Microsoft Corporation Image denoising techniques
US10032256B1 (en) * 2016-11-18 2018-07-24 The Florida State University Research Foundation, Inc. System and method for image processing using automatically estimated tuning parameters
CN110276726A (en) * 2019-05-13 2019-09-24 南昌大学 An image deblurring method guided by multi-channel network prior information
CN110599409A (en) * 2019-08-01 2019-12-20 西安理工大学 Convolutional neural network image denoising method based on multi-scale convolutional groups and parallel

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
段立娟;武春丽;恩擎;乔元华;张韵东;陈军成;: "基于小波域的深度残差网络图像超分辨率算法", 软件学报, no. 04 *
陈清江;石小涵;柴昱洲;: "基于小波变换与卷积神经网络的图像去噪算法", 应用光学, no. 02 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114219820A (en) * 2021-12-08 2022-03-22 苏州工业园区智在天下科技有限公司 Neural network generation method, denoising method and device
CN116385280A (en) * 2023-01-09 2023-07-04 爱芯元智半导体(上海)有限公司 Image noise reduction system and method and noise reduction neural network training method
CN116385280B (en) * 2023-01-09 2024-01-23 爱芯元智半导体(上海)有限公司 Image noise reduction system and method and noise reduction neural network training method
CN116703772A (en) * 2023-06-15 2023-09-05 山东财经大学 Image denoising method, system and terminal based on adaptive interpolation algorithm
CN116703772B (en) * 2023-06-15 2024-03-15 山东财经大学 An image denoising method, system and terminal based on adaptive interpolation algorithm

Also Published As

Publication number Publication date
CN112801897B (en) 2024-02-09

Similar Documents

Publication Publication Date Title
CN102663702B (en) Natural image denoising method based on regional division
CN113808042B (en) SAR image denoising method based on wavelet transformation and generation countermeasure network
CN101950414B (en) Non-local mean de-noising method for natural image
CN112801897B (en) Image denoising method based on wide convolution neural network
CN101477681B (en) Wavelet image denoising process based on adaptive sliding window adjacent region threshold
CN111612695B (en) Super-resolution reconstruction method for low-resolution face image
CN111311508B (en) Noise reduction method for pavement crack image with noise
CN101847257A (en) Image denoising method based on non-local means and multi-level directional images
CN101477680A (en) Wavelet image denoising process based on sliding window adjacent region data selection
CN112288652B (en) A guided filter-Retinex low-light image enhancement method based on PSO optimization
CN111815550B (en) A method of infrared and visible light image fusion based on gray level co-occurrence matrix
CN113222853B (en) Progressive infrared image noise reduction method based on noise estimation
CN104182939B (en) Medical image detail enhancement method
CN103093428A (en) Space-time united image sequence multi-scale geometric transformation denoising method
CN114881081A (en) Interference phase optimization method based on adaptive space-time filtering fusion
CN104881847A (en) Match video image enhancement method based on wavelet analysis and pseudo-color processing
CN102222327A (en) Image denoising method based on Treelet transformation and minimum mean-square error estimation
DE102021103189A1 (en) Systems and methods for training networks for real super resolution with unknown degradations
CN104616259B (en) A kind of adaptive non-local mean image de-noising method of noise intensity
CN101957984B (en) Image Denoising Method Based on Parameter Estimation of Nonlocal Atrophy Factor
CN113793280A (en) A Real Image Noise Reduction Method Combining Local Noise Variance Estimation and BM3D Block Matching
CN104754183B (en) A kind of real-time monitor video adaptive filter method and its system
CN110706186B (en) Underwater information anti-interference method based on compressed sensing and Renyi entropy
CN116612032A (en) Sonar image denoising method and device based on self-adaptive wiener filtering and 2D-VMD
CN100417191C (en) A Noise Reduction Method for Image Combination

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant