Research on Image Super-Resolution Reconstruction Technology Based on Unsupervised Learning

A ﬀ ected by the movement of drones, missiles, and other aircraft platforms and the limitation of the accuracy of image sensors, the obtained images have low-resolution and serious loss of image details. Aiming at these problems, this paper studies the image super-resolution reconstruction technology. Firstly, a natural image degradation model based on a generative adversarial network is designed to learn the degradation relationship between image blocks within the image; then, an unsupervised learning residual network is designed based on the idea of image self-similarity to complete image super-resolution reconstruction. The experimental results show that the unsupervised super-resolution reconstruction algorithm is equivalent to the mainstream supervised learning algorithm under ideal conditions. Compared to mainstream algorithms, this algorithm has signi ﬁ cantly improved its various indicators in real-world environments under nonideal conditions.


Introduction
Deploying deep learning-based target detection algorithms on infrared-guided weapon platforms can effectively improve guidance accuracy.However, deep learning-based object detection algorithms require high-quality dataset to train the network model.Usually affected by the movement of the aircraft platform and the accuracy limit of the infrared image sensor, the quality of the infrared image obtained is poor, which is manifested with low resolution, lots of noise, and artifacts.Based on this, this paper studies the superresolution reconstruction technology of infrared images and visible light images.
Image super-resolution reconstruction technology is a technology that migrates images from low-resolution domains to high-resolution domains.It enhances image details and restores part of image information while improving image resolution.Traditional image superresolution algorithms need to pair high-resolution (HR) images and low-resolution (LR) images and realize the conversion from LR images to HR images by learning the mapping relationship between them.Since it is difficult to collect real HR and LR image pairs, the existing HR-LR image dataset construction method is mainly to establish a degradation model, and the corresponding LR image is obtained from the HR image through the degradation model.Currently, the performance of super-resolution algorithms in nonideal dataset largely depends on the accuracy of the degradation model.
In order to solve the above problems, this paper designs an image super-resolution algorithm based on unsupervised learning.Compared with the existing image superresolution algorithm, this algorithm has the following advantages: Firstly, an image degradation model based on the generation countermeasure network is constructed.The network establishes a degradation model by learning the degradation relationship between image blocks and truly simulates the process of image degradation.Secondly, based on the idea of image self-similarity, a super division model of unsupervised learning is designed, which can realize the super-resolution reconstruction of input images without matching HR and LR datasets.

Literature Review
Image super-resolution reconstruction technology has important research significance and application value in both civilian and military fields [1].Traditional methods are interpolation-based super-resolution algorithms, such as nearest neighbor interpolation and bicubic interpolation [2], which have simple principle and fast operation speed, but the image quality after super-resolution is poor, there are serious fuzzy phenomena, and the indicators after subjective and objective evaluations are relatively general.Later, modeling-based super-resolution algorithms were developed, such as the maximum a posteriori estimation method [3] and the iterative back-projection algorithm [4], which have better reconstruction effects than interpolation, but require a large amount of prior knowledge, high cost, and low operational efficiency.
In recent years, there has been significant progress in the application of deep learning in the field of super-resolution reconstruction, solving many problems in traditional algorithms.In 2014, the SRCNN algorithm proposed by Dong et al. applied the convolutional neural network structure to the super-resolution reconstruction task, and the obtained image quality was greatly improved [5].With the superposition of the network depth, the parameters increase a lot, which may lead to the disappearance of the gradient, and the network is difficult to converge.Kim et al. introduced the residual network into the super-resolution algorithm, which alleviated the problems of gradient disappearance and gradient explosion while increasing the network depth [6].He et al. introduced a dense model on the basis of the residual network and connected each convolutional layer in the network to the subsequent convolutional layer, so that each layer can learn the features of the previous layer, effectively improving the quality of the reconstructed image [7].In 2018, Zhang et al. proposed the RCAN algorithm, introduced the attention mechanism, and enhanced the ability of the network to learn features by learning the importance of different channels in the network and assigning corresponding weights to them [8].
The proposal of adversarial networks has pushed the research on image super-resolution reconstruction to a new level, making the high-frequency details of reconstructed images more accurate, edge blur control better, and texture details more refined.Early researchers focused their research on reducing the minimum mean square error of the reconstructed image, resulting in excellent objective indicators of the reconstructed image, but the image is too smooth and lacks high-frequency details.In 2017, Ledig et al. first applied adversarial networks to image superresolution reconstruction projects and proposed the SRGAN algorithm [9], which added L 1 loss to the basis of adversarial loss to make the reconstructed image visually more natural.In 2018, Wang et al. proposed the ESRGAN algorithm, which incorporated dense block connections into deep residual networks, used relative discriminators, and removed the batch normalization layer from SRGAN, and used the deep network to enhance the quality of reconstructed images, effectively improving the artifact problem [10].
The current image super-resolution reconstruction algorithm is usually based on supervised learning, which requires a training set composed of high-resolution (HR) and low-resolution (LR) images during training.However, it is difficult to collect images of different resolutions in the same scene, so usually performing predefined degradation on high-resolution images to generate low-resolution images leads to poor generalization of the trained network.In 2018, Shocher et al. proposed the ZSSR algorithm, which only relies on the internal information of a single image to complete network training and has achieved remarkable results in processing image super-resolution reconstruction under nonideal conditions [11].In 2022, Karwowska and Wierzbicki conducted an extensive review of current deep learning-based image super-resolution methods, especially pointing out that generative adversarial networks show great potential in super-resolution satellite image reconstruction [12].

Construction of Degradation Model Based on Generative
Adversarial Network.Most super-resolution algorithms usually assume that the low-resolution image is obtained by downsampling the high-resolution image in a fixed way, such as bicubic downsampling.However, the degradation process of low-resolution images under real conditions is different from others.When the set downscaling kernel deviates from the real downscaling kernel, the performance of the super-resolution reconstruction model will drop significantly, resulting in poor algorithm performance under nonideal conditions.The degradation model of real low-resolution images will be affected by sensor optics and jitter, and the use of incorrect degradation models has the greatest impact on algorithm performance [13,14].Therefore, establishing an accurate degradation model is of great significance for improving the performance of superresolution algorithms.
3.1.1.Overall Network Design.The natural degradation model we designed is based on the generative adversarial network and uses the cross scale cooccurrence property [15] to realize the training of the degradation kernel of a single image and restore the degradation process of the image to the greatest extent.The cross scale cooccurrence property can be understood as the consistency of overall pixel distribution between individual downsampled images of different sizes in natural scenes, although the sizes of image blocks (patches) and downsampled image blocks may be different.At the same time, since the information entropy inside a single image is the lowest, that is, it contains the most information, the patch set inside a single image is sufficient for network training.The overall architecture of the network is shown in Figure 1.The image Real to be reconstructed is input, the degraded output image Fake is generated by the generator, and the patches of the same size in Real and Fake are paired as real samples and fake samples.Finally, the two crops are sent to the discriminator for discrimination.
The degradation network we designed can establish corresponding degradation models for each image by 2 International Journal of Aerospace Engineering learning the degradation methods of different image blocks within a single image.The goal of this network is to generate a downscaled image such that the pixels in each image patch match the real LR image as much as possible.First, the input image is downsampled to generate a downscaled image, and then the downscaled image and the input image are cropped to obtain real image blocks and samples for generating downscaled image blocks, and then, both are sent to the discriminator Discriminate to optimize the generator.Finally, the downscaled image generated by the generator is consistent with the input image in distribution.At this time, the confrontation between the discriminator and the generator also reaches a balance, and the generator G can be regarded as the degradation model of the input image.
3.1.2.Generative Network Design.The basic assumption of the super-resolution model is that the low-resolution input image is the result of n-fold downscaling of the HR image.
Here, we use generative adversarial networks to learn the degradation relationship of images to establish a more accurate image degradation model.Some influencing factors such as Gaussian noise and Poisson's noise can be ignored and do not affect the overall effect of the experiment.In most cases, the conversion process from HR images to LR images is described as the convolution of HR images and degradation kernels, as shown in In the equation, I LR is a low-resolution image, I HR is a high-resolution image, k n is a degenerated kernel, and n is a downscaling multiple.It can be approximated that lowresolution images are mainly generated by linear transformation of high-resolution images, so the generation network is designed as a deep and completely linear network, which does not contain batch normalization layers and nonlinear activation functions.Afterwards, the relative loss function and regularization constraint are added to the overall network to avoid the tendency of degenerate kernel towards error and improve the anti-interference performance of the designed degenerate model.In addition, if the generator has no linear display constraints, it may cause unconstrained changes in the overall pixel distribution of the image block, making its pixel distribution deviate from the highresolution image [16], so it is necessary to use a linear network as the generator.The generative network designed in this paper is shown in Figure 2, which is a completely linear structure and adopts a deep network structure to enhance the network's feature extraction ability and optimal solution solving ability.
The input to the generated network is a single-channel image.When the input is infrared image, it enters the generation network directly as a single-channel image; when the input is visible light image, it cannot be directly input into the generation network as a three-channel image, so it needs to be merged into a single-channel image by using the merge function.Before entering the network, the image is randomly cropped, and the 2n × 2n image blocks are cut out and sent to the network, and the output of the network is n × n downscaled image blocks.When n = 32, the parameters of each layer of the network are shown in Table 1.
In Table 1, c, h, and w represent the number of channels, height, and width of the input and output features, respectively, while k, n, s, and p are the convolution kernel size, number of convolution kernels, step, and padding, respectively.It can be seen from the table that the first convolutional layer is a downsampling layer, and the feature size is reduced from 64 × 64 to 32 × 32; the second to fifth convolutional layers are equal-scale convolution, and the feature size remains unchanged; the last layer of the network reduces the number of channels to 1 and outputs 32 × 32 image blocks.
3.1.3.Discriminative Network Design.The goal of the discriminative network is to learn the distribution of individual patches of the input image and to discriminate between real patches belonging to this distribution and fake patches generated by the generative network.The real samples input by the discriminative network are cropped from real LR images, and the fake samples come from low-resolution images generated by the generator.The discriminator designed in this paper is a fully convolutional network, as shown in Figure 3.
Five convolutional layers are used in the discriminant network.The parameters of each convolutional layer are shown in Table 2, which does not contain any pooling layer.After each convolutional layer, the IN layer and the ReLU layer are connected, and spectral normalization is performed.The activation function in the last layer is replaced by the sigmoid function.The convolution kernel size of the first layer of convolution is 7 × 7, and the convolution kernel sizes of the second to sixth layers are all 1 × 1.The input of the discriminative network is 32 × 32 image blocks.Through the calculation and processing of the network, the final 3 International Journal of Aerospace Engineering output is a discriminative matrix of size 26 × 26.After multiple trainings, we calculated that the best discrimination effect is achieved in this size.
In Table 2, c, h, and w represent the number of channels, height, and width of the input and output features, respectively, while k, n, s, and p are the convolution kernel size, number of convolution kernels, step, and padding, respectively.From the above table analysis, the discriminant network only reduces the feature size in the first layer, while expanding the number of channels to 64; use 1 for the second to fifth floors × 1 convolutional kernel to extract features; the number of contraction channels in the last layer is 1, and the output is limited between 0 and 1 using the sigmoid function to output the final discrimination matrix.The final output required by the network is a degraded kernel that can be directly applied to the image rather than a downscaling network, because the degraded kernel can be directly applied to any super-resolution algorithm, and a priori operation can be performed on the degraded kernel such as artificially constrained degraded kernel, making it more in line with the natural degradation process of the image.The degenerate kernel is extracted from the trained generator, and its specific operation is to input the matrix whose element values are all 1 into the generating network for calculation.
In addition to the usual adversarial loss, the loss function of the generator adds regularization restrictions to limit the possible error tendency of the degenerate kernel, as shown in the following equations: where k i,j is the value of each point in the degenerate kernel, m is a constant weight matrix whose weight increases exponentially with the distance from the center of k, x 0 , y 0 represents the center index, l sum to 1 ensures that the overall sum of the degenerate kernel is as close to 1 as possible, l boundaries punishes nonzero values close to the boundary to ensure that the value of the degenerate kernel boundary is small, l sparse guarantees the sparsity of the coefficients to prevent excessive smoothing of the degenerate kernel causing the image generated by degeneration to be too blurry, and l center ensures that the pixel value in the center of the degeneration kernel is as large as possible.In the first round of training, the output of the generator network will be trained with the bicubic downsampled image as a constraint at the same time, as the initialization of subsequent training.The total network loss parameter is shown in In the equation, R = αl sum to 1 + βl boundaries + γl sparse + δl center , 7 where α, β, γ, and δ are the weight parameters, which are calculated through training.Here, we take the values 0.5, 0.5, 5, and 1, respectively, after training the network.

Image Super-Resolution Algorithm Based on
Unsupervised Learning.In recent years, the development of deep learning has greatly improved the performance of super-resolution reconstruction (SR) algorithms.However, due to the limitation of supervised learning, these SR algorithms are restricted to be used in a specific training set, which requires the existence of both high-resolution images and matching low-resolution images in the training set.It is very difficult to collect paired high-and low-resolution images in reality, and it is more difficult to collect infrared dataset.Therefore, the effect of super-resolution reconstruction of infrared images using traditional super-resolution algorithms with supervised learning is not ideal.Therefore, we propose an unsupervised super-resolution algorithm based on the theory of image self-similarity, which realizes the training of the super-resolution network using the training set of a single image.
3.2.1.Algorithm Overall Design.Shocher et al. [11] argue that supervised training overly focuses on the differences between large amounts of data, making it impossible to recover this image-specific information.Compared with the external entropy of the regional blocks in the general natural image collection, the information entropy of the internal region blocks of a single image is much smaller.According to information theory, parts with low entropy are richer in information [17], so internal image statistics tend to have stronger predictive power than external information obtained from general image collections.Based on this idea, this paper designs and builds a single-image super-resolution model, using the internal recursion of a single-image information, while training and testing a residual network, whose training set is extracted from the image to be super-resolution itself.Among them, the principle structure of supervised learning and unsupervised learning super-resolution algorithms is shown in Figure 4.The super-resolution reconstruction algorithm based on supervised learning requires a training set composed of a large number of paired HR and LR images.The LR images are input into the network, and the training network outputs the corresponding HR images.The image super-resolution algorithm we designed based on unsupervised learning requires only one image for the training set.Using the similarity of the image itself, the dataset is constructed from the image itself, and then, the self-supervised network is trained, and finally, the converged network is applied to the original lowresolution image to complete the super-resolution reconstruction.The overall process of the algorithm is as follows: Step 1. Perform data expansion and image enhancement on the input single image.Data expansion is mainly because the image obtained by the aircraft platform in a nonstationary state may be an incomplete image, so it needs to be filled into a complete image.Image enhancement mainly includes contrast adjustment, probability flipping, and other preprocessing.Then, the image is divided into multiple images with smaller sizes as HR samples and fed into the degradation model built in the previous section to obtain LR samples, thereby matching specific HR-LR image pairs.The network can be trained on these image pairs randomly.

International Journal of Aerospace Engineering
Step 2. Send the LR image to the network for super-resolution, generate an HR image, and compare it with the HR image in the dataset to calculate the loss.
Step 3. Backpropagation, the loss value obtained in step 2, is forwarded to update the parameters of the network.
Step 4. Repeat steps 2 and 3 until the loss approaches the lowest value and converges.At this time, the image generated by the network is closest to the HR image.
Step 5. Send the original image to the network to obtain a high-resolution image whose resolution is enlarged by n times.

Network Structure Design.
The supervised convolutional network (CNN) trained on a large external set of diverse HR-LR images [18] must capture all possible diversity in the weights of its network, learning as much as possible from all HR-LR images in the training set.The mapping relationship of LR image pairs, therefore, the number of layers of the supervised learning network, is more, and the structure is more complex.In the unsupervised learning designed in this paper, the diversity of the HR-LR relationship of a single image is much smaller, so a smaller and simpler dedicated network can be used for encoding.The designed network structure is shown in Figure 5.
The network module includes an upsampling layer and four residual blocks.The first layer is an upsampling layer, which upsamples the image by n times according to the scaling factor n and restores it to a high-resolution image.The second to fifth layers are the residual connection layers, in which the design structure of the residual block is shown in the Bottleneck structure at the bottom of Figure 5, in which the 1 × 1 convolution kernel of the first layer performs downsampling operation on the input features, reducing the number of channels of the matrix is a quarter, the second layer performs feature learning, and the 1 × 1 convolution of the third layer performs a dimension-boosting operation on the feature matrix, restoring the depth of the feature matrix to the number of channels of the input residual block.Through the design of the Bottleneck residual block, the   In this paper, the FLIR thermal infrared imaging dataset and the DIV2K high-definition visible light dataset are selected as the experimental dataset for super-resolution reconstruction and test the effect of the algorithm.Among them, the FLIR thermal infrared imaging dataset has obvious characteristics and rich detailed information, so it is suitable for use as a dataset for unsupervised super-resolution algorithms; DIV2K is classic super-resolution algorithm training set, which includes 1000 high-definition images with 2 K resolution and high clarity, making it suitable for comparative experimental verification of the algorithm.An example of FLIR and DIV2K datasets is shown in Figure 6.

Verification and Analysis
The algorithm in this paper is based on the network design of unsupervised learning, and the establishment of the degradation model and the super-resolution image reconstruction can be completed through a single image.Among them, the establishment of the training dataset is mainly for the comparison experiment with the supervised image super-resolution algorithm.We firstly perform 4 times bicubic downsampling on the FLIR dataset to obtain the infrared LR image under ideal conditions and form an HR-LR image pair with the original image; then, process the

Degradation Model Comparison Experiment Analysis.
In this section, the degradation model is firstly analyzed by comparative experiments and tested on the infrared dataset FLIR and the visible light dataset DIV2K, respectively.Figure 7 shows the results of testing on infrared images, and Figure 7(a) shows the infrared images from the original FLIR dataset.Figure 7(b) shows the degradation results of the degradation model on the original image in this paper.The left side shows the degradation kernel calculated by the algorithm in this paper.The degradation kernel is used as a filter to process the original image and obtain the degradation image on the right side. Figure 7(c) shows the degraded effect of bicubic and Gaussian blurring on the original image.It can be seen from the figure that the algorithm in this paper can learn the degradation mode of the image and construct different degradation kernels according to different images.Bicubic downsampling only scales the image on the pixel, without adding blur, resulting in the loss of blur information.Gaussian blur applies the same blur kernel to all images, which easily leads to the problem of excessively high generated blur in Figure 7(c).
Next, experiments were conducted on the visible light dataset DIV2K, which provided both 2 times and 4 times downsampling degraded image dataset under the composite degradation model.In this section, the  It can be seen from Figure 8 that the degraded image in the DIV2K dataset is very similar to the image processed by the Gaussian blur kernel and the algorithm blur kernel proposed in this paper in terms of blur.It can also be seen from the degraded core extracted by the degradation network that the distribution of degraded pixels is very similar to the traditional isotropic Gaussian blur kernel, as shown in Figure 9.
Then, it is objectively evaluated and analyzed for PSNR and SSIM indicators.The experimental results are shown in Table 3.
It can be seen from the table that the image after Gaussian blur kernel processing has the highest score, because the LR dataset of DIV2K was produced using Gaussian blur like check images for degradation.The score of the algorithm in this paper is slightly lower than that of the Gaussian blur kernel, but compared with the bicubic method, the PSNR index is improved by 14.1%, and the SSIM index is improved     10.Among them, the first line in the picture (a) and picture (b) is the full-scale image after the infrared image super-resolution enhancement, and the image in the second line is the reconstruction detail display of the red frame part.It can be seen from Figure 10(a) that the algorithm in this paper can better restore the contour information of the vehicle and the details of the pedestrian on the right, and the overall sensory performance is slightly inferior to the ESRGAN [10] algorithm based on the generative confrontation network.Compared to the traditional bicubic method and the deep network structured EDSR [20] algorithm, the restoration degree is higher, and the suppression of blur and noise is better, resulting in clearer generated images.From Figure 10(b), it can be seen that the algorithm in this paper effectively restores the detailed information of the vehicle, such as tires, rearview mirrors, and doors, while effectively suppressing the fuzzy effects present in the bicubic method; due to the rich detailed information of the vehicle in this test image, the algorithm in this paper is very close to the supervised super-resolution algorithms ESRGAN and EDSR in terms of restored visual clarity, but there are some jagged effects on the edges of the object.Under ideal conditions, the algorithm in this paper achieves similar results to EDSR and ESRGAN algorithms in terms of visual effects.Then, the above-mentioned algorithms are analyzed objectively, and PSNR and SSIM are used as specific indicators to evaluate the reconstruction effect.The experimental results are shown in Table 4.
Combined with the experimental data in Figure 10 and the above table, it can be concluded that (1) the bicubic interpolation algorithm performs well in the objective indicators of the dataset, even surpassing the EDSR, because the HR-LR image pair of the dataset itself is obtained by bicubic downsampling, and then, the bicubic interpolation is performed to achieve super-resolution reconstruction; that is, the reconstruction model coincides with the regression   At the same time, due to the characteristics of unsupervised learning, the network only needs to be trained on a small sample on a single image, which makes the network more generalizable and robust Then, test on the nonideal DIV2K dataset, and use the traditional bicubic method downsampling LR dataset for supervised network training.In order to test the network reconstruction effect under real conditions, the test set uses the LR dataset processed by the composite degradation model provided by DIV2K.The test results are shown in Figure 11.
Figure 10 shows the enhancement effect of different algorithms on the DIV2K dataset under nonideal conditions, with scaling factors of 4. From Figure 11(a), it can be seen that when the degradation model is unknown, the perfor-mance of the bicubic algorithm decreases sharply, becoming the most blurry among all results.At the same time, the performance of the EDSR and ESRGAN trained in the ideal training set also decreases significantly, as it can recover some details of the window but cannot remove blurring and artifacts; thanks to the ability to construct a degradation network for a single image, the algorithm in this paper can learn the degradation model of a single image, perform super-resolution reconstruction for its specific degradation mode, and generate images that effectively reconstruct details such as windows, roofs, and tower cranes.From Figure 11(b), it can be seen that the details of characters, horses, and other details generated by the algorithm in this paper are significantly better than EDSR and ESRGAN, far exceeding bicubic, and the generated images almost completely filter out blurring effects.The experimental results show that the super-resolution image generated by our algorithm can reconstruct and restore the image defects such as blur, offset, and artifact and perform well in the nonideal dataset.The super-resolution reconstruction algorithms EDSR and ESRGAN based on supervised learning show excellent results in the ideal dataset, but the results are general when processing real low-resolution images.
The performance of the above algorithms is analyzed by objective indicators, and PSNR and SSIM are used for evaluation.The experimental results are shown in Table 5.
After analysis, it can be concluded that

Conclusions
In this paper, image super-resolution enhancement technology based on unsupervised learning is studied.The construction of traditional image super-resolution reconstruction algorithm dataset needs to degrade high-resolution images to obtain low-resolution images, and the degradation model greatly affects the performance of the algorithm.Existing super-resolution reconstruction algorithms usually use bicubic downsampling to construct datasets, resulting in a sharp decline in performance of some algorithms that perform well during training when applied to nonideal datasets.
In response to the above issues, this paper first constructs a degradation model based on generative confrontation network and establishes a related degradation model by learning the degradation relationship between image blocks in the image, so as to simulate the real degradation process of the image to the greatest extent.Then, based on the principle of image self-similarity, an unsupervised learning residual network is designed, which realizes the super-resolution reconstruction of small sample training of a single image, which can be applied to any image.
The experimental results show that the performance of the proposed algorithm on the ideal test set is basically the same as that of the excellent mainstream super-resolution reconstruction algorithms such as EDSR and SRGAN, and the performance on the nonideal test set greatly surpasses these algorithms.

Figure 1 :
Figure 1: Overall structure of degenerate networks.
DIV2K dataset, in order to verify the effect of the proposed algorithm under nonideal conditions, the training set of visible light is set to the image pair composed of the original image and its 4 times bicubic downsampled image, and the test set is set to the image pair composed of the original image and the image after its composite degradation model.The algorithm model training in this paper uses an RTX3060Ti GPU, CUDA version 11.4, and the program is written based on the PyTorch deep learning framework.Use the Adam optimizer to optimize the training, and the parameters are set to β 1 = 0 5, β 2 = 0 99, and ε = 0 5.The initial learning rate is set to 0.0002, which is reduced to 0.0001 after training for 150 epochs, and a total of 300 epochs are trained.When training on the FLIR dataset, the batch size is set to 2. When training on the DIV2K training set, the batch size is set to 1, because of its large image resolution and large memory usage.

Figure 5 :
Figure 5: Schematic diagram of unsupervised learning network structure.
The blur kernel generated by the algorithm in this paper and the degraded image generated by it Bicubic Gaussian Bicubic Gaussian Bicubic Gaussian (c) Degraded image generated by bicubic downsampling and Gaussian blur
Small target super-resolution comparison experiment results REAL OURS ESRGAN EDSR Bicubic (b) Normal target super-resolution comparison test results
Restoration and reconstruction of low-light environment dataset REAL OURS ESRGAN EDSR Bicubic (b) Restoration and reconstruction of well-lit environment dataset
3.1.4.Degenerate Kernel Constraint and Loss Function Design.At any iteration during network training, the generator G performs an estimated downscaling operation on a

Table 3 :
Comparison of objective evaluation indexes of different degradation models.The degradation network designed in this paper can effectively learn the degradation relationship of the image itself, thereby simulating the degradation process of the image, and has good performance on both infrared and visible light image datasets.Compared with the traditional bicubic algorithm, it can more realistically simulate the image degradation process

Table 4 :
Comparison of objective evaluation indexes of each algorithm in FLIR dataset.International Journal of Aerospace Engineering model, resulting in the high objective indicators of the bicubic method in the dataset.From 10, it can also be seen that the image reconstructed by the bicubic method has the worst visual clarity among the above algorithms, indicating that objective indicators cannot completely replace subjective visual perception in visual tasks

Table 5 :
(2)parison of objective evaluation indexes of each algorithm in DIV2K dataset. ) the performance of EDSR and ESRGAN models trained with ideal dataset is not ideal on nonideal dataset, because the models obtained network training do not learn the parameter settings under nonideal conditions, resulting in a sharp decline in performance when applied to nonideal test set(2)due to its unsupervised nature, the algorithm proposed in this paper can theoretically adapt to images with any degradation mode in reality and reconstruct them.Among them, the performance under ideal conditions is basically consistent with the algorithm based on supervised learning; under nonideal conditions, the PSNR value increased by 19.4% and 18.6% compared to mainstream EDSR and ESRGAN algorithms, respectively, while the SSIM value increased by 38.07%and 25% compared to mainstream EDSR and ESRGAN algorithms, respectively