Animation Design Based on 3D Visual Communication Technology

. The depth synthesis of image texture is neglected in the current image visual communication technology, which leads to the poor visual eﬀect. Therefore, the design method of ﬁlm and TV animation based on 3D visual communication technology is proposed. Collect ﬁlm and television animation videos through 3D visual communication content production, server processing, and client processing. Through stitching, projection mapping, and animation video image frame texture synthesis, 3D vision conveys animation video image projection. In order to ensure the continuous variation of scaling factors between adjacent triangles of animation and video images, the scaling factor ﬁeld is constructed. Deep learning is used to extract the deep features and to reconstruct the multiframe animated and animated video images based on visual communication. Based on this, the frame feature of video image under gray projection is identiﬁed and extracted, and the animation design based on 3D visual communication technology is completed. Experimental results show that the proposed method can enhance the visual transmission of animation video images signiﬁcantly and can achieve high-precision reconstruction of video images in a short time.


Introduction
3D visual communication technology is immersive, interactive, and conceivable for users to create high-quality visual experience.Video based on 3D visual communication technology needs to go through different processes such as content making, coding compression, network transmission, and terminal display.In recent years, visual communication technology has been widely used in various fields because of its function of fast and real-time measurement of information [1].With the development of multimedia technology, visual communication technology is widely used in the field of video and animation image management, such as network media operation, animation release, etc. Film and television animation and video images are developing constantly in the field of film and television animation and video images.e core of film and television animation and video is visual communication, and it is difficult to obtain target film and television animation and video images in complex background in real time and accurately.It is a common problem in multiframe film and television animation and video visual communication [2].
Reference [3] proposes an innovative design of visual communication system for animated figures and images in virtual reality environment.In the hardware design, the adaptation parameters of the main board of the renderer are designed to optimize the experience of visual communication.In the software design, by introducing Sobel edge operator, the gray function is established to solve the gradient amplitude, and the threshold is selected to compare it to complete the recognition and thinning of the edge data of animated character graphics and images.Design the motion capture module, establish the behavior control model, generate and manage the motion capture files, and complete the overall design of the system.Reference [4] proposes a new optimization method-plane visual communication effect optimization method based on wavelet change.e image is decomposed by wavelet, and the wavelet is reconstructed.In the process of reconstruction, the modulus diagram and phase angle diagram are calculated, and the edge images of each scale are extracted.
e corresponding edge points of the semireconstructed image are enhanced through the edge image.On the above basis, the graphic beautification vector of slip model is used for operation.In order to simplify the operation, the above operation method is transformed into simple mathematical operation, and the visual communication effect is optimized through the reflected light graphic mode.Reference [5] proposes a new automatic graphic language arrangement algorithm in visual communication design.e display size of buffer image in visual communication design is calculated by fixed value method, and the minimum number of layout and the maximum surface utilization are taken as the layout objective function.Ant colony algorithm is used to solve it, get the optimal solution, and obtain the best display position of graphics in visual communication design.e description of graphic language is realized by rule-based syntax description and ASM (abstract state machine) semantic description.e parallel process and selection process of graphic language in visual communication design are described by using parallel and selection marks, and the parallel process and selection process are used to realize the automatic arrangement of graphic language.e proposed algorithm is programmed and run in C language to achieve the purpose of automatic arrangement of graphic language.
However, when the above image visual communication technologies ignore the depth synthesis of image texture, the image visual effect is not ideal.erefore, a film and television animation design method based on 3D visual communication technology is proposed.Build the video processing process under 3D visual communication technology, and mainly process film and television animated video through 3D video content production, and server and client links; based on the visual distance between 2D image and scene, calculate the zoom factor of 3dimensional scene transformation; build a deep learningbased deep feature extraction model PCANet, filter images and extract features of film and television animation video images, use PCAN deep network to create a low-resolution feature dictionary, use high-resolution video image reconstruction method, realize LR animated video images and HR animation video images, and introduce nonlocal similarity empirical constraints to optimize HR images; analyze gray and detail features based on wavelet transform, extract the video image features, separate the prospect and background of animated video images, transform K-L (Karhunen-Loeve) as the core, dynamically identify the similarity between evaluation features, and complete the animation design based on three-dimensional visual communication technology.
e experimental results show that the proposed method can significantly improve the visual transmission of animated video images and improve the speed of high-precision video image reconstruction.

Film and Television Animation Design Based on 3D Visual Communication Technology
3D visual communication technology mainly uses the computer to simulate the environment, to make people feel immersive.ree-dimensional visual communication technology, on the basis of 2D images, reads the main information, compresses and decomposes, and transforms it into three-dimensional images, increasing the measurement and display function of arbitrary sections.In the humancomputer interaction module, we can zoom in the threedimensional image to facilitate users to browse and change the image.

Video Processing Process under 3D Visual Communication
Technology. Figure 1 shows the flowchart of video processing under 3D visual communication technology.
e 3D visual communication technology can be divided into three parts: 3D visual communication content production, server and client, and 3D visual communication content production which is the core of the whole video processing process [6].In the production of 3D visual communication content, a set of cameras or a set of camera devices and audio sensors containing multiple cameras and sensors are used to capture the sound and visual scene in the real physical world.Cameras and sensors, for example, produce a set of digital video and audio signals.In general, the camera captures content within 360 °of the periphery of the unit [7,8].
Aiming at the video processing under the 3D visual communication technology, the 3D visual communication video image at the same time is combined with mosaic, projection mapping, and frame texture to generate a package, as shown in Figure 2.

2.2.
e Scaling Factor of the 3D Scene Transformation.After determining the triangular meshes on a two-dimensional plane through screen projection, a continuous scaling field shall be assigned to each triangular facet; that is, a scaling factor shall be assigned to each triangular facet, and the scaling factor between adjacent triangular facets shall be continuously variable.e specific process is as follows [9,10]: S i is used to represent the distance from the triangle to the viewpoint in 3D scene.S i is defined as visual distance, and all S i can be integrated into a threshold of discrete function S i .
Determine the apparent distance S i and S i+1 between any adjacent triangles i and i + 1, and the difference z i between the two can be expressed as follows: . (1) Obtain the upper limit value max z of z i through formula (1).
When the value of min Scale is 0.4, the sample size changes gradually from 0.4-1 to 0.4.Based on the value of min Scale, pyramid shrinkage is applied to a given texture sample drawing to obtain a multilevel progressive shrinkage sample drawing, which is used as a scaling factor in the process of texture synthesis to obtain the basis for the corresponding sample drawing [11,12].
In order to make the neighboring triangle mesh correspond to the neighboring layer, based on max z, set up: the value of S function can only be different one layer within the sample layer if the difference is max z that is, according to max z to divide the value range of S function, the value of the function is divided into different layers according to max z, and the layers corresponding to the neighboring triangle mesh are also adjacent, thus ensuring the continuity of texture in the process of composition.
e value of S i for each triangle can obtain a value of Scale (min Scale ≤ Scale ≤ 1) according to the above setting, which is the value of Scale corresponding to the scaling of the metatexture sample image [13].Based on this, each Scalevalue corresponds to a Scale-fold shrunk layer, and a Scale-fold shrunk image can be mapped to a Scale-fold shrunk film and TV animation image, where each triangle has a scaling factor.

Deep Feature Extraction Model Based on Deep Learning.
PCANet is a relatively simple method of deep learning based on the theory of deep learning and convolutional neural networks [14,15].e PCANet structure is composed of two PCA (principal component analysis) filter layers, a hash layer and a histogram calculation layer, which can extract the deep-seated features of high-resolution animation and video images.But unlike the usual networks, PCANet's filters are more computationally efficient.PCANet is not obtained through training, but through obtaining the mapping results of parts of high-resolution film and television animation video images, and then adopts the PCA (principal component analysis) method to extract the principal components of the high-resolution film and television animation video images, each of which is an independent filter [16].When the neural network of the filter is used to extract the features, the optimal weights can be obtained without iterative operation; thus, the calculation time is reduced [17].In order to further discuss the advantages of PCANet depth network in extracting the features of high-resolution animated cartoon video images, a mapping eigenvalue of N, 8 × 8 and 1 moving distance is set up for the input of animated video images [18,19].e main parts of the mapping features L 1 were computed using the PCANet algorithm, and a two-layer filter filters L 1 .e steps are as follows: In formula, each column of the matrix A i represents a block of image, and each column contains k1 × k2 elements.
ere are m1 × m2 columns in total.After performing the above operations separately for all N training images, a new data matrix A was obtained: , and A has N × m1 × m2 columns.en, the A matrix was PCA and the ahead L i eigenvectors of the eigenvector matrix W were taken as the convolutional kernel of the filter group.
In formula, q l (AA T ) represents the l-th eigenvector of the matrix AA T .ma k1k2 is a function, representing rearranging each of the L i eigenvectors into the matrix.Each eigenvector contains k1 × k2 elements, amount to get L i convolutional nuclei of length k1 and width k2.Finally, for each input image L, the L i convolutional nuclei were filtered separately.us, the first layer of the PCA filter group is completed.

Second Layer: PCA Filter Layer 2.
e second layer of the PCA filter group is the same principle as the first layer.
is yields the convolutional kernel of the second layer: For each input image L 2 i in the second layer, the PCA filter group is all complete through the filter W 2 L i output features.
Finally, the L i and L 2 i feature mappings are obtained in the same way as CNN.PCANet is characterized by these L 1 mapping features, so it can provide a reliable data for the film and television animation video image processing [20].PCANet uses learning to get multilayer network filters.After the input video image is filtered through two layers of filters, it can have very high-dimensional data output and can be regarded as the input of video image characteristics of the film and television animation and be used for the reconstruction of high-resolution video image of the film and television animation [21].
erefore, when PCANet extracts the features of film and television animation video images, it is actually a direct processing of the pixels of the film and television animation video images, the operation stage has joined the process of blocking, and the number of film and television animation video image block output from the deep network has increased [22].PCANet uses the deep network learning process to extract the features of the animation video image, and the video image features extracted by PCANet and the features obtained by artificial rules are more abundant in detailed information and more prominent in texture structure, which provides rich prior knowledge for subsequent reconstruction and fills in the details of the animation video image of low-resolution film and television so as to facilitate the reconstruction of the animation video image of super-resolution film and television.

Image Reconstruction Method of Multiframe Animation
Video Based on Visual Communication.Combining the advantages of PCANet depth network and sparse representationbased reconstruction method, this paper proposes a visual communication-based multiframe animation video image reconstruction method [23,24].For the multiframe animation video image reconstruction method, it is assumed that the high-resolution image and the low-resolution image are sparse expressions for their respective dictionaries, and then, the image sample features are obtained by PCANet depth network.D h represents the dictionary feature of a high-resolution image, and D 1 represents the dictionary feature of a low-resolution image.In the stage of high-resolution image reconstruction, the low-resolution image is processed in the same way, the deepseated feature is extracted by PCANet, the sparse representation of D 1 above the low-resolution dictionary feature of the high-resolution image is obtained, the coefficients expressed by the sparse D 1 feature of the low-resolution image are directly applied on D h , and the corresponding high-resolution feature image is obtained [25].Low-resolution video animation images are used to achieve high-resolution reconstruction.rough the PCANet depth network, we can get better film and TV animation video image features than the nondepth network.
e deep feature dictionary can also improve the description ability of the feature dictionary and improve the quality of film and TV animation video image reconstruction significantly [26].
K blurred film and TV animation video image need to be sampled in the training data set, which is adjusted to the same size to get the corresponding low-resolution film and TV animation video image, which is combined into a model sample pair: represents low-resolution features.Compute the matrix into blocks for all samples in the data set, and select a sliding window of size k 1 × k 2 , (normally, the pixel square window of the film and television animation video image used is 3, 5, or 7).After feature extraction for all the film and television animation video images is carried out through the aforesaid sliding window, the new data matrix X of column N × m × n can be obtained, and each column of the matrix represents a film and television animation video image block with a total of k 1 × k 2 elements.
To obtain the training sample of i high-resolution animation video image, the formula is Features were extracted with the above data matrix X, and the extracted features were regarded as feature samples in SCSR model and substituted into dictionary features of PCANet.
Suppose that T L hi represents the results of the high-resolution animated video image after sparse coding, the H(•) function is regarded as a unit step function, the results of sparse coding are quantized, the histogram coding is completed, and the deep detail features of the high-resolution animated video image are extracted, that is, [27,28] Similarly, it is inferred that through the same processing process as the high-resolution film and television animation video image, the results of extracting the deep-seated detail features of the low-resolution film and television animation video image are given, namely, Here, F hi and F li on behalf of the film and television animation video image feature extraction results; Bhist on behalf of the histogram coding; and B on behalf of the film and television animation video image segmentation block number of samples.
In this paper, the dictionary is trained in the SCSR framework by combining the sparse coding method.e goal is to obtain a set of dictionary pairs D h and D 1 that can represent complex feature samples.Make K have the same sparse representation on D h and D 1 for the deep-level detail features F hi and F li generated by film, television, animation, and video images, and F hi and F li have the same description coefficient, that is, In the above formula, a � a i   k i�1 represents sparse matrix and λ i represents equilibrium coefficient [29].In order to make the high-resolution film and television animation video image features and low-resolution film and television animation video image features have the same sparse description method as their respective dictionaries, the combined training method is adopted through formula ( 5) and (6), that is, In the above formula, N and M, respectively, represent the dimension rearrangement of the column vectors by the element values of high-and low-resolution feature blocks of film, television, and animation video images, and 1/N and 1/M are the cost between D h and D 1 of balanced formula (9).In order to facilitate subsequent calculation, formula (10) is reconstituted: Equation ( 11) is solved by iteration.Given dictionary D C , calculate the sparse representation coefficient a i of each data sample F C on D C , obtain matrix a � a i   k i�1 , and finally update dictionary D C through a.
After obtaining the dictionary pair D C , the reconstruction of LR film and television animation video image Y and HR film and television animation video image X can be obtained by using the high-resolution film and television animation video image reconstruction method based on sparse regular model.
In the stage of reconstructing the video image of film and television animation, some noise exists in the video image of film and television animation due to the interference of external environment noise, and block effect and blurred artifact will appear in the reconstructed video image of film and television animation [30].Considering that the conventional back-projection model cannot guarantee the quality of the reconstructed image, the details of the image can be better preserved by matching the similar blocks between the image and the image without the prior constraints of partial similarity.Relative to the image block effect and blurred artifacts after the reconstruction, the nonlocal similarity empirical constraint is introduced to optimize the HR image based on the global optimization of back-projection.e global and nonglobal constraint models in this paper are as follows [31,32]: In the above formula, S represents sampling operation, H represents fuzzy filtering, ‖X − X 0 ‖ 2 2 represents global constraint term, ‖(I − W)X‖ 2  2 represents nonlocal self-similarity constraint term, I represents identity matrix, c and c represent normalization parameters, W represents nonlocal weight matrix, and the element formula in W is as follows: In the above formula, N(x i ) represents the i-th film and television animation video image block in the film and television animation video image, x j represents the searched film and television animation video image block similar to N(x i ), h represents the attenuation factor, and z(i) represents the normalized value.

Frame Feature Extraction of Video Image under Gray
Projection. e first step of frame feature extraction is to analyze the features of gray level and detail in the video image based on wavelet transform and then to extract the feature of the video image by gray projection [33].Finally, the problem of frame feature extraction of film and television animation video image evolves into the problem of foreground and background classification of film and television animation video image, and the separation coefficient is determined by the ratio function of the variance of feature distribution of film and television animation video image in the foreground and background region.
e process of extracting frame feature of film and television animation video image is as follows: e input of preprocessed grayscale animation video images shall be represented by f k (x, y), the grayscale image shall be decomposed by Mallat algorithm with wavelet, and a grayscale pyramid L k (s) shall be established.According to the high-frequency component, the decomposed grayscale image shall be built into detail pyramid O k (s, d), the size shall be represented by s ∈ [0, N], and the three detail directions of the video image shall be represented by d ∈ [HL, LH, HH]. e establishment of the original video grayscale image shall be completed by the pyramid L k (s) and detail pyramid O k (s, d) calculated in the above process, and the multiple features of two detail features can be obtained simultaneously [34].
In a video image conveyed visually, the local contrast of a certain image at different positions within a certain range is determined by the difference between scales, so as to improve the contrast effect of the animation and video images.
e formula is as follows: where r represents the feature contrast measure, σ c represents the range of feature area, and σ s represents the range of film, television, and animation video image suppression area.F(c) represents the coarse-scale film and television animation video image, which can suppress the features other than the target area [35].F(s) represents the fine-scale film and television animation image, which can describe the detail features of the target area, and then the scale difference between F(c) and F(s) is shown in the following equation: F(s) adjusts the difference value to the film and television animation video image area with the same scale as F(c), which is represented by L k (c) and L k (s), and subtracts each point to obtain the corresponding absolute values O k (c, d) and O k (s, d), so as to obtain the gray-level features and detail features of the film and television animation video image.e formula is as follows: G k (i, j) represents the gray value of the (i, j)-th pixel in the film and television animation video image,  dx and  dy represent the evaluation estimator between the film and television animation video images, and the image feature display diagram is calculated according to the translation estimator, as shown in the following equation: In the formula, G k−m (•) represents the translation difference, the normalized function (No) is used to unify the scale of the image, and then a group of weighting factors α, β, and c are determined to complete the weighted fusion of the video image through the above method.
e formula is shown in the following equation: where N(  L k ) represents grayscale, N(  O k ) represents detail, and N(  M k ) represents the normalization function of motion feature image and represents the feature value of film and television animation video image frame extracted at last.

Dynamic Identification. Dynamic recognition methods mainly include global and local recognition.
Because the dimension of video image is high, it is difficult to identify the whole video image by global method, and there are noise interference and information occlusion in the image after dynamic recognition.erefore, in order to improve the efficiency of dynamic recognition, local recognition method is adopted [36].
is method takes K-L (Karhunen-Loeve) transformation as the main core and uses the similarity between evaluation features to realize dynamic recognition.
K-L transform is based on the statistical characteristics of the transformation, through the incoming vector orthogonal transform and the outgoing vector to decouple the relevance of the data, which is the best way in the sense of variance.
Suppose, f 1 represents the N × M-dimensional matrix, Γ represents the sample statistic Γ � Γ 1 , Γ 2 , Γ 3 . . .Γ L  , NM × 1 represents the dimension vector, and the average value μ of Γ is calculated and obtained from the set of L sample values, as shown in the following equation: Its variance matrix C(Γ) is shown in the following equation: λ i represents the eigenvalue of the variance matrix, and Φ i represents the eigenvector of the variance matrix.L − 1 eigenvectors and their corresponding eigenvalues are obtained, as shown in (19): e order of feature vector Φ i is based on the order of the size of feature values.e first m eigenvectors are determined and then normalized to form a change matrix R.
en, the transformation matrix is used to transform vector Γ i into vector I in row m through (20): us, achieve the effect of dimensionality reduction, and effectively complete the film and television animation video image enhancement design.

Experimental Results and Analysis
e simulation environment is Pentium m1.60 Ghz CPU and 760 m RAM. e dual-core Intel second-generation core processor OptiPlex 3010 is used as the main frequency, the system is win10 flagship 32 bits, and the simulation software is MATLAB 2020 B. In order to prove the effect of the method in this paper, it is compared with current advanced visual communication methods of animation images (the plane visual communication design method based on graphic beautification technology and the visual communication design method of animated character image in virtual reality environment).

Comparative Analysis of Visual Communication Image
Reconstruction Effects under Different Methods.To make the experimental results more intuitive, two animated video images of 512×512 pixels shall be selected as reference animated video images, as shown in Figures 3 and 4(a), and the simulation experiment shall be carried out with the relevant simulation experimental tools.If we use Gaussian fuzzy model, set the Gaussian filter of 3×3 region invariant, set its sampling factor to 4, add Gaussian noise to all low-resolution images, and satisfy the signal-to-noise ratio of 30 dB.Figures 3(a According to Figures 3 and 4, it can be seen that the visual communication design method of VR environment animated character image cannot avoid blurring edges and the reconstruction effect is not good.e method of plane visual communication design based on graphic beautification technology has a large error when matching some pixels, which leads to the lack of rich and clear details of the final reconstructed animation video image, and the reconstructed effect is inferior to the visual communication design method of virtual reality animation image.Compared with the other two methods, the proposed method has obvious advantages in the reconstruction of high-resolution animation video images.

Comparative Analysis of Recognition Rate.
Compared with the visual communication design method based on VR environment animation and graphic landscaping technology, the effectiveness of visual communication video image dynamic recognition is tested.First, according to the actual situation, 50 kinds of license plate Chinese characters are segmented into two kinds of film and TV animation video images, which are divided into 10 sets of fuzzy similar Chinese characters according to the special structure features of Chinese characters, and then recognized.ere are 270 fuzzy film and TV animation video images, and the recognition rates of the two methods are shown in Table 1.
As can be seen from Table 1, the dynamic recognition rate of the proposed method is 91%, which is higher than other methods.e dynamic recognition rate of the graphic beautification technology-based graphic visual communication design method is 62.2%, and the dynamic recognition rate of the virtual reality environment animation figure visual communication design method is 67.5% for the Chinese character images with the same resolution and blurring degree of film and television animation video images.is is mainly due to the fact that the graphic visual communication design method based on graphic beautification technology and the virtual reality environment animation figure visual communication design method do not take into account the influence of illumination color on film and television animation video images, so that the feature dimension is reduced.At the same time, in order to prevent overlapping areas, 2 × 1 block is used as a block at the time of establishing a block, which also reduces the dimension, but when constructing a histogram, the gradient of each cell is divided into 9 parts.According to the experimental analysis of Dalal, if the area is reduced at this time, the feature dimension can be reduced, but the description ability is also reduced, and the computational complexity is increased.e advantage of the proposed method in recognition rate is that it arranges the eigenvalues in an orderly way and determines the eigenvectors of the eigenvalues as Chinese characters.
is method not only reduces the feature dimension, but also retains the key information of the animation video image.In the process of feature extraction, it highlights the main details of the animation video image, deepens the description of the feature, and effectively improves the recognition rate.

Space-Time Performance Test.
Taking the visual communication design method of animated character images in virtual reality environment proposed in reference [3] and the plane visual communication design method based on graphic beautification technology proposed in reference [4] as the comparison method, and taking the grassland texture sample synthesis as an example, the temporal and spatial performance of this method and the comparison method are compared.e results are shown in Tables 2 and 3.
Analysis of Table 2 shows that the length of time of the three methods used in 3D visual communication is different.
e length of time of the suggested method is below 3.8 s, the length of time of the visual communication design method of animated character image in virtual reality environment is up to 14.93 s, and the length of time of the graphic visual communication design method based on graphic beautification technology is up to 20.06 s. e length of time of the suggested method is smaller than that in VR environment.Compared with the two comparison methods, this method has obvious advantages in time. is is because this 3D visual communication animation video images through stitching, projection mapping, and animation video image frame texture synthesis projection, resulting in packaging.ere is no need to preprocess the texture sample to improve the efficiency of texture synthesis.
Analysis of Table 3 shows that the size of the animation video image after the three methods is quite different.e size of the synthesized image that is determined by this method is below 3.8 MB, the image size under visual communication design method of animated character image in virtual reality environment is up to 14.93 MB, and the image size under image synthesis based on graphic beautification technology is up to 20.06 MB.
e size of the 8 Scientific Programming synthesized image that is determined by the proposed method is smaller than other methods.is is because the present method constructs the scale factor for 3-dimensional Based on the identified and extracted video image feature scale is smaller and occupies less space.

Real Example Experiment of Video Image Reconstruction.
In order to reflect the performance of this method in practice, the video image reconstruction performance is tested.e 500 images from the Lion King animated video were selected as subjects and iterated 200 times to test the peak signal-to-noise ratio of the HD reconstructed image, and the results are shown in Figure 5: As shown in Figure 5, the image peak signal-to-noise ratio of the present method is 31.2dB-40.9dB, which is higher than the original image, proving that this method meets the practical needs of high-resolution image reconstruction.

Conclusion
A new animation design method based on 3D visual communication technology is designed to solve the problem of poor visual effect in existing image visual nication technology.Build a video processing process under visual communication technology; based on the visual distance a 2D images to a 3D scene, compute the scaling factor for the 3D scene transformation; building a deep learning-based PCANet model, images were used to filter for processing by using the model and to extract the characteristics of the film and television animation video images; deep features of the image samples were acquired using the PCAN deep network; create a low-resolution feature dictionary for high-resolution images; using a highresolution video image reconstruction method based on a sparse rule model, realize the reconstruction of animated video images; nonlocal similarity empirical constraints are also introduced to optimize the HR images; analyze grayscale and detail features based on wavelet transform; extract the features of the video images; with the K-L (Karhunen-Loeve) transformation as the core, the similarity between the picture features is used to dynamically identify the film and television animation video images, and to complete the high-precision film and television animation design.e experimental results show that the proposed method can improve the visual communication effect of animation and video images, with small time cost, small space capacity, and good application performance.

2. 3 . 1 .
First Layer: PCA Filter Layer 1.For the N training picture L N i with size of m1m2, PCANet selects a k1 × k2 window, and each training image is compared by row and column to obtain the local features of the image.Get m1 × m2 picture blocks of k1 × k2 size.Dislodge the mean value of the picture block, the output behind dislodging the mean value of L i :
) and 4(a) are the reference film and television animation video images, Figures3(b) and 4(b) are the reconstructions of film and television animation video images by using the visual communication design method of virtual reality environment animation character images, Figures3(c) and 4(c) are the reconstructions of film and television animation video images by using the graphic beautification technology-based graphic visual communication design method, and Figures3(d) and 4(d) are the reconstructions of this method.

Figure 3 :Figure 4 :
Figure 3: Simulation results.(a) Video image sample of film and television animation.(b) Visual communication design method of animated character image in virtual reality environment.(c) Graphic visual communication design method based on graphic beautification technology.(d) Paper method.

Figure 5 :
Figure 5: Peak SNR of the HD reconstructed images.

Table 1 :
Comparison and analysis of recognition rate.

Table 2 :
Time performance simulation results.

Table 3 :
Space performance simulation results.