An Artistic Image Fusion Method with Improved Cartoon-Texture Decomposition

When the art images are restored by the virtual restoration method, there are problems such as insufficient clarity and more noise in the reference image. An improved cartoon-texture decomposition method for art image fusion is proposed. The nonlinear local total variation component is used as the indicator function of image decomposition to obtain the image cartoon structure component and texture oscillation component. According to the oscillation component’s strong repetitiveness and structural directionality, the image texture part is filtered by combining the improved directional diffusion algorithm. Using the sparse coefficients of the fused cartoon component and the sparse coefficients of the texture component, the cartoon and texture of the image is inverse transformed and weighted and summed to obtain the recovered image after fusion. The experimental results show that this paper has a good effect after image fusion, and the recovered clarity is higher, which can better express the basic information of the source image; compared with several decomposition fusion methods commonly used at present, this paper has better recovery performance and detail processing ability and preserves the edge information of essential details in the image while filtering and denoising and is more excellent in objective performance evaluation indexes such as PSNR and SSIM. It can be used as a reference basis in the restoration process of art images.


Introduction
ere is a potential correspondence between the restored artwork and the viewer's appreciation. e performance before and after restoration can create a feeling that there is a huge di erence between the viewers. erefore, artworks need to be photographed and archived before restoration to witness future comparison [1,2]. In addition, the process of restoring artworks often requires the original image of the work as an auxiliary basis to grasp the color, style, composition, and other important information of the work, which not only avoids the risk of "secondary damage" that may result from direct restoration but also provides the necessary data resources for digital display and dissemination [3]. However, in the process of art image formation, dissemination, and storage, due to the in uence of imaging equipment or shooting environment and other factors, a certain degree of distortion will inevitably occur, resulting in blurred or severely noisy art images, where clarity cannot be guaranteed. Detailed information cannot be well presented, which brings greater trouble to the restoration of artworks and seriously a ects the subsequent restoration work [4]. erefore, it is of good research value and practical signi cance to restore the degraded art images to obtain clear and complete images.
Image restoration technology processes the image by the computer. It reconstructs the image without improving the accuracy of imaging equipment to obtain the important information lost in the image, improve the image quality, and has low application cost and good practicability [5]. e early research results of image restoration technology mainly introduced some technologies and methods used in digital signal processing, such as inverse ltering technology. In addition, it is found that the decline of image quality can be solved by the state-space model, autoregressive moving average model, nonlinear parameter identi cation, complex adaptive theory, and other methods. ese methods use modern control technology and greatly promote the development of image restoration technology [6][7][8].
erefore, more and more scholars are committed to finding methods with strong decomposition ability, fast operation speed, and conducive to the fusion of results. Afonso et al. [9] proposed an improved variable splitting method based on the augmented Lagrangian method (ALM) algorithm and studied the image restoration problems based on framework and constraints.
e image restoration quality and processing speed are improved, and the restoration effect is relatively good. Li et al. [10] proposed a multicomponent decomposition method of learning images based on a discriminant dictionary.
is method uses weighted kernel norm regularization and sparse constraints to characterize the coarse structure and fine components. It uses weighted Schatten sparse kernel norm regularization to extract rough structure from the separation model. e fusion result of this method can retain more detailed information and perform well in the objective evaluation index. Zhou et al. [11] proposed a learning method for medical image fusion dictionary, which uses multilayer details of the image to enhance weak information, introduces a multiscale sampling method to realize multiscale representation of patches, completes image patch clustering through neighbourhood energy index and multiscale spatial frequency index, and uses K-SVD combined with subdictionary to construct fusion dictionary. e calculation efficiency is high, and the objective evaluation standard is excellent. Niknejad and Figueiredo [12] used the ALM algorithm to restore the image polluted by Poisson noise and achieve a good restoration effect. Li et al. [13] proposed an image fusion method based on three-layer decomposition and sparse representation, which decomposes the source image into high-frequency and low-frequency components by using the high-pass characteristics of noise and realizes the fusion and denoising of high-frequency components by reconstructing error parameters. A structured texture decomposition model is designed according to the details and energy of low-frequency components. e corresponding fusion rules are used to complete the fusion of the two. Experimental results show that this method can effectively suppress the noise problem of images and has better performance in subjective vision and quantitative evaluation. At the same time, Li et al. [14] introduced the double-layer decomposition image fusion method through the joint bilateral filter, designed a local gradient energy operator based on structural tensor and neighbourhood energy to fuse the energy layer, and verified the reliability of the fusion method through multiple groups of images. e results show that this method has the excellent visual quality and quantitative evaluation performance, good real-time performance, and high fusion efficiency. Ono et al. [15] proposed a new matrix block kernel norm BNN to describe the prior information of globally dissimilar textures but highly repeated local texture patterns.
e TV + BNN model uses the image cartoon total variational regularization method. After rotation and overlap, the image texture components have a low rank and can better decompose the image textures in different directions.
Most decomposition methods have more or less false boundaries in the smooth region, loss of details in the texture part, and ladder effect in the cartoon part after image processing, resulting in poor image definition after fusion. erefore, this paper proposes an art image fusion method based on improved cartoon texture decomposition. e nonlinear local total variation is used as the indicator function of image decomposition to obtain the cartoon structure component and texture oscillation component. e improved directional diffusion algorithm filters the texture part of the image. Based on maintaining the structural integrity of the cartoon part and the fine structure of the texture part, the influence of noise on the fusion result is reduced. Finally, the cartoon component sparse coefficient and texture component sparse coefficient are used to transform the cartoon and texture of the image, and the clear art image is obtained after weighted fusion. is method is suitable for complex art images and can effectively solve many practical problems such as image denoising and restoration, edge detection and texture recognition. It is superior to the traditional fusion method in signal-to-noise ratio and recognition rate. It can be used as a reference in art image restoration.

Cartoon-Texture Decomposition.
e art image after cartoon texture decomposition gives a texture part containing the source image oscillations and noise and a cartoon part containing the source image structure and geometry [16]. e texture part can represent the texture of the clear part of the source image containing the noise and oscillation in the image, and the cartoon part is the difference between the source image and the texture part, so the clear part of the cartoon part becomes blurred after the image cartoon texture decomposition [16,17]. Compared with image decomposition methods such as wavelet decomposition and curvilinear transform, cartoon texture decomposition can preserve the information in the source image without producing a lack of description of the features in the image, and it has better results in image denoising, restoration, distortion processing, and detail reorganization. Figure 1 shows the source image and the decomposed cartoon and texture images.
By observing Figure 1, it can be seen that although the decomposed cartoon image and the textured image describe the same scene, the degree of blurring and the texture features differ greatly, and thus the clear part of the two images need to be fused to come to enhance the quality of the image. ere are more methods of image fusion, such as, fusion algorithms based on sparse representation and nonlocal regularized image fusion based on variational framework. is type of algorithm generally treats the image as a single signal [18]. However, under realistic conditions, a single artistic image often consists of several different components, such as the cartoon component and the texture component, which contain information content with different characteristics from each other. If the image is decomposed into different parts and then different fusion rules are applied according to the characteristics of each part, the quality of the fused image will be further improved.

Local Full Variational Cartoon-Texture Decomposition.
According to the theory of variation and scale space, the cartoon part of the image has a small total variation (TV), and the texture part has a small parametric number. e higher the frequency of the texture component oscillation, the smaller the magnitude. e essence of the total local variation (LTV) cartoon-texture decomposition is to use nonlinear low-pass and high-pass filters to solve the approximate solution of the original variational problem. By calculating the LTV value around each pixel in the graph and comparing it with the LTV after the low-pass filter, the LTV is used as a local indicator function of the pixel to determine the type and attribution of the pixel [19]. After low-pass filter processing, if the LTV of the pixel point changes weakly, it is a cartoon point type classified as a cartoon component; if the LTV of the pixel point decays rapidly after convolution operation, it is a texture point type. And then, keep the cartoon component U unchanged and classify the value of texture pixel point after low-pass filter as cartoon component U. In contrast, the texture component V takes the difference between cartoon component and source image H. e LTV expression for any pixel p in the source image H is where J σ denotes the Gaussian kernel with standard deviation σ and * denotes the convolution operation. ∇ is the gradient operator. Mapping p to η σ (p), the relative rate of change of LTV of pixel p is where L σ is the low-pass filter and determines the decomposition scale, if the LTV changes under the low-pass filter condition, its local oscillation can be derived from η σ (p).
When η σ tends to 0, it indicates that the LTV of the pixel point does not change much, and the low-pass filter has little effect on the LTV of the point, which belongs to the cartoon component U; conversely, when η σ tends to 1, it indicates that the LTV of the pixel point has a large relative rate of change and decays fast, which belongs to the texture component V. A set of nonlinear high-pass and low-pass filters can be obtained by solving the weighted average of L σ * H and H according to the relative rate of change of LTV of the pixel point.
where w(p) denotes a segmented, nondecreasing soft threshold function taking values in the range [0, 1]. To facilitate later calculations, the soft threshold function is set to where the parameters n 1 and n 2 take 0.3 and 0.6, respectively. When η σ (p) is small, the oscillation of function H(p) near pixel p is small, indicating that the point belongs to the cartoon part, that is, U(p) � H(p). Conversely, when η σ (p) is large, indicating that the function H(p) has a large oscillation near pixel p, then U(p) is replaced by With this image decomposition method, it is impossible to obtain a cartoon texture separation that can simultaneously take into account the extraction of texture components and keep the edges of the structured part unblurred, which is suitable for contaminated noise-laden images. However, the information contained in the cartoon component U cannot meet the actual needs because some key details are still retained in the texture part. e decomposed texture must be filtered and then weighted and synthesized with the decomposed cartoon component U to obtain the recovered image H.

Improved Texture Filtering by Directional Diffusion
Algorithm. For the filtering correction of the texture part after the decomposition of the art image, the traditional interpolation and amplification are used first. en, the image is filtered and smoothed by the improved directionbased diffusion algorithm, and the edge information is Advances in Multimedia retained. Finally, the low-frequency information of the acquired image is used as the basis for image fusion. e directional diffusion model belongs to the partial differential equation of nonlinear diffusion along the tangential direction parallel to the edges, i.e., perpendicular to the direction of the gradient vector [20]. e diffusivity of an image is the rate of change of the image gradient, and the expression is where t is the time and ζ denotes the unit vector perpendicular to the image gradient ∇H.
Since the art image is a two-dimensional plane, the image gradient component calculation can be expressed as where z|∇H|/zκ is the κ component of gradient ∇(|∇H|) and x and y represent low resolution image and source image, respectively. According to the p-m diffusion equation [21], the selection of edge function significantly impacts the diffusion behavior of texture. When the value of the edge function is unreasonable, the image edge gradient becomes more extensive, resulting in the occurrence of reverse diffusion. erefore, this paper solves this problem by introducing the local coordinate system (κ, ζ). e unit vector κ is parallel to the gradient direction in the coordinate system, and ζ is perpendicular to the gradient direction. Due to zH/zζ � 0, there are By substituting and deriving, we can get After solving the texture part of the image in different coordinate systems, the final expression of directional diffusion can be obtained: is method is used to filter the image texture. After several iterations, a smooth image can be obtained, ensuring that the texture part's critical information is preserved. e image remains sensitive to the edges of the image while filtering.

Cartoon Component Sparsity Factor Fusion.
e cartoon component of an image mainly reflects some structure and color information, which is sensitive to the human eye. At the same time, the cartoon component of natural images often reflects a large amount of primary information in the original image, so the purpose of fusing cartoon components is to make such information more prominent [22]. Some scholars use the absolute maximum fusion rule to fuse the cartoon component sparse coefficient of the decomposed image. However, this method does not reflect the proportion of the original information in the image and will make the resulting image too sharp in the information representation. In Figure 2, the cartoon part of the water bucket is not ideal. e absolute proportion of the fused data cannot be reflected by simply using the "relative-absolute maximum" fusion rule. erefore, this paper uses the "absolute maximum" and "relative-absolute maximum" fusion rules to fuse the sparse coefficients of the cartoon components of the fused image.
M and N is the row and column values of the sparse coefficient matrix, respectively. en, according to the "relative-absolute maximum" fusion rule, the sparse coefficients of the combined cartoon components are e "absolute maximum" fusion rule is used to obtain Also, since the cartoon component of the image partly reflects the primary energy information of the original image, the information entropy is used to determine the weights of ϑ 1 (h,c) and ϑ 2 (h,c) . Finally, the sparsity coefficient of the cartoon component after fusion can be obtained as where IE is the information entropy of the cartoon component of the image to be fused.

Texture Component Sparse Coefficient Fusion.
Since the texture component reflects the details such as edge information within the original image, in general, the weight of edge information in the final fused image is different between different fused images. So, it is crucial to calculate the weight of the texture component in the final fused image for each image to be fused. e same sparse coefficient fusion is used for texture component fusion. However, since the texture component of the image partially reflects the edge information of the original image, the edge intensity is used as the basis for weight judgment [23]. en, the sparse coefficients of the fused texture components are where IE is the information entropy of the cartoon component of the image to be fused.

Decomposition and Fusion
Process. e art image decomposition and fusion process is shown in Figure 3. e process of art image fusion with improved cartoon texture decomposition is briefly described. First, the input image H is decomposed using the local full-variance cartoon texture algorithm to obtain the cartoon image H c and the texture image H t . e texture image is filtered using the improved directional diffusion algorithm to obtain the optimized image texture image H t . en, the sparse coefficients of the fused cartoon component and the sparse coefficients of the texture component are inverted. Finally, the results of the inverted conversion are weighted and summed, and the fused image F is output.

Experimental Preparation.
e experimental environment of this paper is Intel Core i7 3.6 g CPU, 16 g SDRAM, and MATLAB 2019. In order to verify the feasibility and effectiveness of the proposed method, two standard art images (Vang. BMP, bri. BMP) are selected for experimental simulation. e texture of image Vang is apparently, while image bri is relatively fuzzy. Use a size of 7 × 7. e Gaussian check source image with the mean value of 0 and standard deviation of 0.6 is blurred, and the white noise with a standard deviation of 10 −3 normal distribution is added to obtain the blurred high noise image. e source image and blurred image are shown in Figure 4.
is paper adopts the general image evaluation standard's peak signal to noise ratio (PSNR). e structural similarity index measurement (SSIM) index is used to evaluate the main features of the image, such as comprehensive brightness, contrast, and structure. e closer the SSIM value is to 1, the closer the fused image is to the actual image, and the better the algorithm performance [24,25]. Objective evaluation indicators are defined as follows: In the formula, MSE is the mean square error between source image H and fused image F. μ H and μ F are the average values of image H and F, and μ 2 H and μ 2 F are the variances of image H and F.

Experimental Results and Analysis.
e blurred and noisy images (Figures 4(b) and 4(d)) are decomposed into cartoon textures. e value of the decomposition scale parameter is 5, and the value of the soft threshold function is 0.3 and 0.6, respectively. e fused cartoon component sparse coefficient and texture component sparse coefficient are used for inverse transformation, and the weighted sum is used to output the result: Among them, α and β are the weights of cartoon and texture parts, respectively. Considering the nature of the selected image, the image Vang BMP weight α � 0.7, β � 0.15; image bri weight of BMP α � 0.6, β � 0.25. e image cartoon texture decomposition and fusion results are shown in Figures 5 and 6. Among them, (a) and (b) are the decomposed texture part and texture filtered image, respectively, (c) is the decomposed cartoon part, and (d) is the image fusion result.
From Figures 5 and 6, we can see that, after the cartoon texture decomposition of the blurred image by the local total variation method and the texture filtering by the improved directional diffusion algorithm, the effect of image fusion is relatively good, which can better represent the basic information of the source image, and the image performs well in detail and brightness, with high definition.

Performance Comparison of Different Algorithms.
In order to further verify the performance of the method in this paper, the traditional cartoon-texture decomposition method, the nonlocal regularization iterative method, the global TV iterative method, and the method in this paper As shown in Figure 7, the fused images from the traditional cartoon-texture decomposition method and the global TV iteration method contain more noise. ere will   Figure 3: Art image decomposition fusion process. 6 Advances in Multimedia be different degrees of fuzziness and poor definition in some parts. e nonlocal regularization iterative method has dramatically improved the brightness, noise, and clarity, but there is a problem of local block effect. Because the cartoon texture decomposition method is improved and the directional diffusion algorithm is used to filter the image texture, the fusion result has a high definition, less noise, richer texture, no local block effect, and relatively excellent performance. Table 1 shows the comparison results of performance indexes of different methods.
Comparing the related indexes in Table 1, we can see that this method is superior to the other three methods in PSNR and SSIM performance indicators, has better recovery performance and detail processing ability, and retains important edge details in the image while filtering. Compared with the current image fusion methods, this method is shorter, better in real-time, more efficient in calculation, and more suitable for complex artistic images. It can be used as a reference in the process of art image repair.

Conclusion
According to the different sensitivity of cartoon components and oscillation components to Gaussian noise, this paper proposes an improved cartoon-texture decomposition method for artistic image fusion. By analyzing the geometric structure characteristics of the image, the variational problem of the source image is quickly approached by using the nonlinear low-pass filter to avoid the generation of false boundary and step effect and ensure that the decomposed cartoon section maintains a high level of detailed information. e texture part is also filtered using an improved directional diffusion algorithm to avoid the destruction of the image structure by medium and large-scale noise and to preserve the segmental smoothness and small-scale detail information of the image structure. e cartoon and texture components are inverted, respectively, using the cartoon component sparsity coefficients and texture component sparsity coefficients. Finally, a clear recovered image is obtained after weighted fusion. e experimental results show that the image fusion performance of the proposed method is relatively reasonable and can express the basic information of the source image. e image performs well in detail and luminance, with high definition. Compared with the traditional cartoon-texture decomposition, nonlocal regularization iteration, and global TV iteration, the PSNR and SSIM indexes of the algorithm of this paper have been greatly improved. erefore, the method in this paper has good adaptability and robustness and can be used as a reference basis and method for art image-assisted restoration.

Data Availability
e data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e author declares no conflicts of interest. decomposition using blockwise low-rank texture