Hazardous Gas Emission Monitoring Based on High-Resolution Images

Air pollution presents unprecedentedly severe challenges to humans today. Various measures have been taken to monitor pollution from gas emissions and the changing atmosphere, of which imaging is of crucial importance. By images of target scenes, intuitional judgments and in-depth data are achievable. However, due to the limitations of imaging devices, effective and efficient monitoring work is often hindered by low-resolution target images. To deal with this problem, a superresolution reconstruction method was proposed in this study for high-resolution monitoring images. It was based on the idea of sparse representation. Particularly, multiple dictionary pairs were trained according to the gradient features of samples, and one optimal pair of dictionaries was chosen to reconstruct by judging the weighting of the information in different directions. Furthermore, the K-means singular value decomposition algorithm was used to train the dictionaries and the orthogonal matching pursuit algorithm was employed to calculate the sparse coding coefficients. Finally, the experiment’s results demonstrated its advantages in both visual fidelity and numerical measures.


Introduction
Today, humans are facing the most severe situation of air pollution due to industrial emission, fuel combustion, motor vehicle exhaust emission, and so on [1][2][3][4].It adversely affects human health and daily life.Various monitoring measures have been taken to provide real-time information of the environment [5].For example, the monitoring process focuses on the number of chimneys that are working and the size of the emissions.As an important method, imaging technology has made the monitoring task easy and efficient, although it often suffers from low-resolution problems originating from the imaging devices.Details are not available in the captured images, which further limits the judgment of the conditions [6][7][8].
To improve the resolution of monitoring images, chargecoupled devices (CCDs) with high resolution are exploited, which will certainly lead to a higher cost.To improve the effect of monitoring images without replacing the original detection equipment, people gradually turn to imageprocessing methods, of which the superresolution reconstruction method is an effective way to improve the quality of images including hyperspectral, visible light, and infrared images.Many significant superresolution (SR) methods have been proposed to deal with SR reconstruction problems in past decades [9][10][11][12].Among them, the conventional interpolationrestoration approach is computational.One improved SR algorithm is the generalized nonlocal means algorithm, which generalized the successful video denoising nonlocal means (NLM) technique [13,14].
In addition, in the study of improving image resolution, single-frame image SR reconstruction algorithms based on dictionary learning show their unique advantage in combining the prior information of images in this research area [15,16], so these reconstruct images with higher resolution than other methods.Inspired by compressed sensing (CS) theory, Yang et al. firstly proposed the SR reconstruction method based on sparse coding [17,18], named sparse coding-based superresolution (SCSR), which could reconstruct a HR image by global dictionary learning.Zeyde et al. made an improvement to Yang et al.'s method and proposed the single image scale-up method using sparse representation (SISR) [19].The SCSR algorithm proposed by Yang et al. and the SISR method proposed by Zeyde et al. are both effective learning-based SR reconstruction methods, which try to obtain a HR image by training a high-resolution dictionary and a corresponding low-resolution (LR) dictionary.
When reconstructing the LR images of different kinds, the mentioned methods choose the only couple of the dictionary.But the information in different gas observation images or different parts in one image may vary a lot [20].If we separate a whole LR image into many patches and then choose the dictionary couple with the highest similarity according to the features in the LR patches in the process of reconstructing HR patches, the resolution of the reconstructed monitoring image could be improved by a large margin.Hence, a novel SR reconstruction based on adaptive dictionary learning of gradient features is proposed in this paper according to the analyses above.In particular, the features of training samples are clustered into several groups by the K-means method, and multidictionary pairs are trained offline by the K-means singular value decomposition (K-SVD) algorithm [21].And then, the dictionary pair which has the highest similarity with the LR image is selected in the process of online image reconstruction.Finally, several groups of experiments are completed and the results show that the images reconstructed by the proposed method are excellent on both subjective vision perception and objective evaluation value.

Principle of SR Image Reconstruction
Dictionary learning is particularly important in the process of image SR reconstruction.By means of sparse representation, more information in the dictionaries can be contained by fewer atoms, which leads to the higher quality of reconstructed images and less time cost.In this section, the calculations of image sparse representation and dictionary pair learning are described in detail.
In the process of imaging, if the original ideal HR image is x, which is affected by the blurring of the optical imaging systems and the downsampling of the CCD when displayed on the device, the traditional imaging process can be modeled as [22] x = SHx + v, where x is the observed image, S represents the downsampling operator, H is a blurring filter, and v is the additive noise caused by the system.Based on the theory of CS, to obtain an ideal HR image using a fast SR reconstruction algorithm, both the methods of SISR and SCSR require that the ideal HR image x be sparsely represented by the HR dictionary D h , which can be written as [23] x = D h α 2 The observed LR image can be expressed by As ( 2) and (3) show, α ∈ R is a sparse matrix with n elements.The sparse representation of the HR ideal image is shown in Figure 1.But this equation is not satisfied by many images.The natural image statistics prove that usually the areas of image primitives have very low internal dimensions and thus they can be represented by a small amount of training samples [24].The primitives mentioned here refer to the high-frequency feature areas such as the edge and inflection points in images.
In this study, an assumption was made that the image primitives can be represented in the sparse form by the dictionary which is trained with a large number of primitive patches.A HR image can be obtained by the following steps, and Figure 2 shows the process of SR reconstruction.
(i) The primitive areas of training samples are extracted and divided into four subsets according to the direction of their gradient features.The four directions are 0 °, 45 °, 90 °, and 135 °, which indicate the angles with respect to the vertical direction.
(ii) Divide the primitive areas into four subsets according to the gradient features, which are used as the data source to train four subdictionary sets, each including a HR dictionary and a corresponding LR dictionary by the K-SVD algorithm.
(iii) According to the weightings of the gradient feature information in each direction within a LR patch, the certain HR subdictionary and the corresponding LR subdictionary are chosen to reconstruct a HR subpatch.
(iv) The HR subpatches reconstructed by different dictionaries are combined into the final HR image according to the weight of the features in four directions.
Typically, the primitive areas in the images can be obtained by (linear) feature extraction operators.For the training sample sets x, the HR primitive areas in one direction x k h and the corresponding LR primitive areas x k l can be obtained by where k = 1, 2, 3, 4 and S is the downsampling operator.The features in the horizontal gradient (0 °) can be obtained after being filtered by the operator F 1 The features in the vertical gradient (90 °) can be obtained after being filtered by F 2 The 2 Journal of Spectroscopy features in the 45 °and 135 °directions can be obtained by F 3 and F 4 , respectively.The four filters are described in The HR and LR primitive areas obtained in (3) and ( 4) can be sparely represented by the corresponding HR and LR dictionaries, respectively.The process of obtaining the dictionaries can be mathematically expressed by Here, D k h refers to the kth subdictionary of HR and D k l refers to the corresponding LR dictionary.Because the HR patches and the LR patches have the same sparse representation coefficients, ( 6) is expressed as , s t ∀i, α i 0 < T 0 , 7 where The N and M in the equation above mean the size of the input image.Given an initial dictionary D 0 , the dictionary pair D k h and D k l can be solved by the K-SVD algorithm [21].Figure 3 shows some parts of the four HR feature patches with the same training samples.
For a LR image y to be reconstructed, at first, a simple segmentation method was employed to divide it into p patches y i ∈ R n , n = 1, 2, … , p For a certain LR patch y i , its corresponding HR patch is ŷi According to (1), the LR patches can be written as y i = S Hŷ i + v, where v refers to the noise.Our purpose is to get ŷk i by y i and a certain couple of subdictionary D k l and D k h ; the process can be expressed as The symbol β k i in the above equations means the sparse representation coefficient of y i under a certain subdictionary D k l Equation ( 9)can be solved by the orthogonal matching pursuit (OMP) method [25] and by calculating β k i .In the process of SR reconstructing the HR patch ŷi , we need to combine the weight of the information in four different directions.The HR patch ŷi can be written as where q j represents the weighting coefficient of the information in four different directions.To guarantee the compatibility between the LR patches and the chosen dictionary, the weighting coefficient is employed which can be expressed by the amount of the information in different directions, which is Derived from ( 9), the final HR image ŷi is The symbols M and N in (15) represent the width and height of the image, respectively.x i and x i mean the pixel values corresponding to two images.The result of x i − x i is close to zero when these two images are similar.In other words, a small MSE value indicates two analogous images.In addition, the parameter MAX I in ( 16) means the value of the signal peak.For example, the value equals 255, when we use an 8-bit sample image.Furthermore, MSE is an important parameter for PSNR.A better reconstruction image always has a higher value of PSNR.
In our experiments, the growth of the value of PSNR of the reconstructed HR image is close to zero when the sample number increases to a degree.In order to guarantee the comprehensiveness of the dictionary information and the efficiency of the algorithm, 100 pictures of different kinds were chosen as the training samples.For the LR color images, we transform the RGB images to a YCbCr space.Since the human visual system is more sensitive to the luminance channel, we use the superresolution reconstruction method based on dictionary learning proposed in this study to reconstruct the images in the luminance channel, and the images in the Cb and Cr chrominance channels are simply magnified by the bicubic interpolation method.
The original LR images in these experiments are all downsampled from HR images by a factor of 1/3 in this experiment.The number of atoms in each subdictionary is 1000, and the iteration number of the dictionary training is set to 40.For a better comparison, the LR images are also reconstructed by the bicubic interpolation method and Zeyde et al.'s method, respectively.Figures 4 and 5 show the results of reconstructed gas monitoring images.In addition, Figure 5 takes an image of a scene full of smog as a sample.With the reconstructed HR images, it is more practical to determine the exact pollution sources and their locations.As for the detailed information in the images, all of the three reconstructed HR images show a better human visual effect than the LR image, which shows obvious mosaic blocks.Furthermore, the HR image reconstructed by In general, because four different dictionaries contain different features of the training samples and show different similarities with the LR images, the reconstructed images by the proposed SR algorithm of adaptive dictionary learning all present a better image quality than raw LR images do.While the HR reconstruction image will be obtained from the LR image, the HR patches are utilized based on the weight of the features in different directions, which consider the differences in various LR patches and guarantee the best matching results with the dictionaries in the process of SR reconstruction.
To evaluate the reconstruction results, the objective evaluation parameters PSNR and MSE are analyzed and calculated for reconstructed images in both Figures 4 and 5.The results are shown in Table 1.Indeed, PSNR and MSE are two basic evaluation standards.Higher PSNR and lower MSE mean better image quality.As Table 1 shows, the PSNR of two groups of images reconstructed by the proposed method are both higher than the PSNR of images reconstructed by the other two methods while the value of MSE are significantly reduced.Compared with the images reconstructed by Zeyde et al.'s method, the PSNR of the reconstructed chimney image by the proposed method increases by 0.8602 dB and the PSNR of the reconstructed smog image by the proposed method increases by 0.3369 dB.Different LR images will end with different reconstruction results due to the association with the image content.The images with more obvious gradient changes are more possible to be better reconstructed, benefitting from the higher matching accuracy with the various gradient dictionaries.

Conclusion
In conclusion, the proposed superresolution reconstruction method based on adaptive dictionary learning of gradient features can be used to obtain HR monitoring images of gas emission with more detailed information.It starts by training four couples of different dictionaries according to the directions of gradient information and reconstructing a HR gas monitoring image by combining the HR patches according to the weight of the information in different directions.Finally, actual images of pollution sources are tested by the proposed method.Experimental results show the effectiveness of the proposed method in reconstructing HR images.Furthermore, our further study will focus on increasing the efficiency of the whole SR reconstruction procedure and improving the dictionaries for some particular images with pertinence to make the monitoring task more effective and efficient.

Figure 4 (
a) shows the original LR chimney image, and Figures 4(b)-4(d) are the reconstructed images based on the bicubic method, Zeyde et al.'s method, and the proposed method, respectively.For the enlarged parts of these images, the construction of the chimney in Figures 4(b)-4(d) is more clear and smooth than that in Figure 4(a).The image in Figure 4(d) is smoother and shows better quality including better contrast, improved resolution, clearer texture, and detailed information.

Figure 4 :
Figure 4: LR chimney image and reconstructed HR image by different methods: (a) LR image, (b) reconstructed image by the bicubic method, (c) reconstructed image by Zeyde et al.'s method, and (d) reconstructed image by the proposed method.

Figure 5 :
Figure 5: LR smog image and reconstructed HR image by different methods: (a) LR image, (b) reconstructed image by the bicubic method, (c) reconstructed image by Zeyde et al.'s method, and (d) reconstructed image by the proposed method.

Table 1 :
Objective evaluation of the reconstructed images in Figures4 and 5.