Classification of Artistic Styles of Chinese Art Paintings Based on the CNN Model

People's appreciation needs of Chinese paintings have gradually increased. The research on automatic classification and recognition of Chinese painting artistic style and its authors have great practical value. This study presents a Chinese painting classification algorithm with higher classification accuracy and better robustness. Using a convolutional neural network (CNN) to extract the features of Chinese painting, the image features of Chinese painting are extracted by fine-tuning the pretrained VGG-F model. The mutual information theory is introduced into embedded machine learning, so that the embedded principle is affected by feature selection and feature importance. An embedded classification algorithm based on mutual information is proposed, and Chinese painting is classified.


Introduction
In the field of Chinese painting feature extraction and classification, there have been a lot of related research studies [1][2][3][4][5][6][7]. With the improvement of living standards, people's pursuit of spiritual life is getting higher and higher, and the appreciation demand for Chinese painting is also gradually increasing. A large number of digital Chinese paintings have appeared on the Internet or in digital museums, but how to effectively use and manage these paintings has become an urgent problem to be solved [5][6][7][8]. erefore, the research on automatic classification and recognition of Chinese painting artistic style and its authors has great practical value. e existing literature adopts the traditional feature extraction method to extract the features of Chinese painting, and the obtained features are relatively single and one-sided. In addition, the existing literature does not deal with the extracted features and directly uses the features for classification, without considering the relationship between features and categories and the influence of feature importance on the classification results. e DEFC (data embedding framework for classification) [9][10][11][12][13][14][15] takes the relevant features obtained through the selective conversion of data samples as the input of the algorithm. Although it is much improved than the existing algorithm, it does not consider the influence of feature importance on the calculation of similar features between objects in the process of calculating similar features between objects, resulting in the restriction of classification results.
In view of this, this paper uses a convolutional neural network (CNN) to extract the features of Chinese painting, and the obtained features are more comprehensive and detailed. In order to solve the problems mentioned above, this study extracts the image features of Chinese paintings by fine-tuning the pretrained VGG-F model, introduces the mutual information theory into embedded machine learning, makes the embedded principle affected by feature selection and feature importance, proposes an embedded classification algorithm based on mutual information, and classifies Chinese paintings.

Fundamental eory.
e DEFC embedded mapping framework [9][10][11][12][13][14][15] follows the principle of "friendship is close, hostility is far," and it introduces two parameters, C F and E D , to control the embedding process; then, where X i and X j are the ith and jth objects in the dataset X; C F and D E represent the intraclass object distance and the interclass object distance in the dataset X under the influence of weight parameters, respectively; d F is the function of calculating the distance between objects of the same type; d E is the function of calculating the distance between objects of different types; W ij is the weight parameter obtained by calculating the similarity between objects X i and X j ; n is the total number of objects in the dataset. e DEFC algorithm mainly includes 3 modules. e first module is a data preprocessing module, including 5 optional processes, namely, scaling, centering, normalization, principal component analysis, and whitening. e second module is the relevant feature calculation module; it calculates the correlation features between objects through the original dataset or the preprocessed dataset and performs the above data preprocessing on the calculated correlation features again to obtain the correlation feature matrix RX. e correlation feature captures the mutual influence between samples, which is beneficial to further discovering the nonlinear structure between the data. Since the dimension of the correlation feature matrix is related to the number of samples and has nothing to do with the feature dimension, dimensionality reduction is achieved when the number of samples is much smaller than the feature dimension. e third module is the embedded computing module, which solves C F is the smallest or D E is the largest to get the mapping matrix P, and then, X final � X R P is calculated by the matrix transformation X final , that is, the new dataset after the mapping transformation of the original dataset.
Another fundamental theory in this study is mutual information theory (MIT) [16][17][18][19][20][21][22][23][24][25][26]. In 1948, Shannon proposed that "information is something used to remove random uncertainty." Information entropy refers to the probability of a certain information appearing, and the greater the uncertainty of the variable, the greater the information entropy. Information can be measured by information entropy. Among them, the self-information and its mean can measure the information and the amount of information contained in the information set itself. e mutual information and its mean can measure the information and the amount of the information supplied by each other between sets of information.
Mutual information describes the correlation between two variables and can be used to measure the correlation between features and classification results. To calculate the mutual information of variables X and Y, first, obtain their respective probability distributions p(x), p(y) and the joint probability distribution p(x, y). According to H � − m i�1 P i lbP i , the information entropy H(X) and H(Y) of variables X and Y and the joint information entropy H(X, Y) of variables X and Y can be calculated, where H represents the information entropy of the feature, P i is the probability when the feature takes the ith eigenvalue, and m is the number of all different eigenvalues. e mutual information between X and Y is 2.2. Feature Extraction. In recent years, deep learning imitates the mechanism of the human brain, combines lowlevel features, obtains high-level features that represent the distributed features of data, and is widely used in object recognition, tracking, and other aspects. e traditional CNN is generally used directly for image classification. is study proposes to take advantage of the enormous advantages of the CNN in feature extraction. Compared with explicit feature extraction, more detailed feature data can be obtained for Chinese paintings in extracting digital features.
Based on the CNN theoretical framework, this study fine-tunes the VGG-F model pretrained on the ImageNet dataset to extract the image features of Chinese paintings. e model is a feed-forward neural network with 5 convolutional layers and 3 fully connected layers. e model adjusts the input image to 224 * 224 pixels and obtains a 4096-dimensional feature vector.
Using the VGG-F model pretrained on natural images is feasible to extract features from Chinese paintings. First, art comes from life, and the painter completes the work through the analysis and understanding of all things in nature. Chinese painting is an artistic reprocessing of natural scenes and another embodiment of a natural image. Second, the deep structure of the VGG-F model can extract complex structures from rich perceptual information and build intrinsic representations in the data. More than 10 million natural images are participating in training; the extracted features will directly or indirectly contain feature information similar to Chinese paintings; this way, helpful information for feature extraction of Chinese paintings can be learned. Finally, although, the number of Chinese paintings in the training dataset in this study is not enough to complete the training of the VGG-F model. But this study adds it as a primer to the VGG-F model; the model is finetuned so that the features extracted by the model can better express the artistic style of Chinese painting.

Mutual Information-Based Data Embedded Classification
Algorithm. Solve the problem that the DEFC algorithm cannot reflect the influence of feature importance on the relationship between computational objects, resulting in restricted classification results. is study introduces mutual information theory into embedded machine learning. A mutual information-based data embedded classification algorithm is proposed. is algorithm calculates the mutual information between each feature vector and the classification category vector. Determine the feature importance and select features with mutual information value greater than zero to form a new dataset to participate in the classification calculation. erefore, the classification of Chinese paintings is affected by feature selection and feature importance under the embedded principle of "friendship is close, hostility is far." erefore, formulas (1) and (2) become where C F and D E represent the intraclass object distance and the interclass object distance of dataset X under the influence of weight and feature importance, respectively; d F is the function of calculating the distance between objects of the same type; d E is the function of calculating the distance between objects of different types; W ij is the weight parameter obtained by calculating the similarity between objects X i and X j ; T feature represents the feature and its feature importance. e image features of paintings are extracted by finetuning the pretrained VGG-F model, and similar feature sets between paintings are embedded under the influence of mutual information.
e details of the proposed mutual information-based data embedded classification algorithm are as follows.
Input: training dataset G and test dataset G′. Output: the new training dataset Z after the embedded mapping is computed and the test dataset Z′.
Its specific process is as follows.
Step 1. Calculate the mutual information value set I between each feature vector in the dataset G and the classification classes vector Y. I i represents the mutual information value between the feature vector G i and the classification classes vector Y, and its calculation formula is Step 2. Normalize the mutual information value set I to obtain the feature weight set Step 3.
e calculation formulas of the training dataset X, the test dataset X9, and the feature weight set θ after feature ′ are the eigenvalue vectors of the ith feature of the training dataset G and the test dataset G9, respectively.
Step 4. Define the category indicator matrix L. en, 1, X i , X j , the same objects, −1, X i , X j , no the same objects, where X i and X j are the two objects of the training dataset X calculated in Step 3, when the objects X i and X j are the same objects, L ij � 1; otherwise, L ij � −1.
Step 5. Calculate the similarity matrix W between the objects in the set X. Common methods for calculating distance is the dot product method, polynomial kernel function method, Cosine similarity method, Euclidean distance, and Gaussian kernel function. In this study, various similarity algorithms are used to conduct a large number of experimental calculations. rough experiments, it is found that the classification effect is the best when the cosine similarity method is used. In this study, the cosine similarity method is used to calculate the similarity between objects, and the calculation formula of the similarity matrix W is where W ij represents the similarity between the two objects X i and X j of the training dataset X calculated in Step 3; X i and X j are the modulus of the eigenvectors of objects ‖X i ‖ and ‖X j ‖, respectively.
Step 6. Calculate the similar feature set S between the objects in the set X, the similar feature set S9 between the objects in the set X9 and the objects in the set X.
where X i and X j are the ith and jth objects of the training dataset X, φ(X i , X j ) is the similar eigenvalue calculation function of the objects X i and X j , the matrix S is an n × n dimensional matrix, and n is the set X number of objects.
Let S ij ′ � φ(X i ′ , X j ), X i ′ is the ith object in the test dataset X9, X j is the jth object in the training dataset X, φ(X i ′ , X j ) is the calculation function of similar eigenvalues of objects X9 and X j , the matrix S9 is an n ′ × n dimensional matrix, n and n ′ are the number of objects in the sets X and X9, respectively.
is study uses the distance between two objects as their similar feature. When formulas (1) and (2) calculate the distance, the weight of each feature is the same. However, for Chinese painting, each feature expresses different styles differently. erefore, this study uses the feature weight set θ calculated in Step 3 to calculate the weighted feature distance and use it as the similar feature value between each object.
is study calculates the weighted feature distance based on the weighted norm theory.
To sum up, the similar eigenvalues of objects X i and X j in the training dataset X calculated in Step 3 is where S ij is the similar eigenvalues between objects X i and X j ; m is the feature Computational Intelligence and Neuroscience dimension; θ k is the feature weight of the kth feature; X ik and X jk are the eigenvalues of the kth feature of the objects X i and X j , respectively.
Step 7. Calculate the optimal projection matrix P.
Take the similar feature set S calculated in Step 6 as the training sample; the set Z represents the new dataset generated by the mapping projection of S; let Z � SP represents the mapping transformation from the set S to Z, where P is the projection matrix, PP T � I, and I is the unity matrix.
Using the category indicator matrix L and similarity matrix W calculated in Steps 4 and 5, set the intraclass object distance C F and interclass object distance D E of set Z as where m is the feature dimension of the set Z. Set up: Reconstructing formulas (11) and (12) in matrix form, then A ′ � L · L + I m×m · W, (12) where I m×m is an all-one matrix with m rows and m columns. Substitute formula (6) into formula (3). Minimizing C F , then Substitute formula (7) into formula (4); maximizing D E , then where diag(B ′ ) represents the diagonal matrix of matrix B ′ , and diag(A ′ ) represents the diagonal matrix of matrix A ′ ; then, the Laplacian matrices of matrices A ′ and B ′ are, respectively, as Because the matrix L and the W are the symmetric matrices, the matrices A, B, ZAZ T , and ZBZ T are also symmetric matrices. According to the definition of symmetric matrices, it can be seen that established. Substitute Z = SP and formulas (15) and (16) into formula (13); then min Substitute Z � SP and formulas (16) and (17) into formula (14); then max e optimal mapping matrix P is obtained by solving the formula (11) or (12).
e experimental results show that although the mapping matrices obtained by solving formula (11) or (12) are different, the classification results of the datasets obtained after the mapping calculation are the same.
Step 8. Calculate the new training dataset Z and test dataset Z′ generated by the embedded mapping.
A new training dataset Z can be obtained by mapping the set S calculated in Step 6, and a new test dataset Z′ can be obtained by mapping the set S′, that is, Z � SP, Z ′ � S ′ P.

Experiments
e number of works created by Chinese painting writers in their lifetime is relatively small; therefore, compared with the classification of ordinary images, the amount of data on Chinese painting works is relatively small. At present, there is no standard dataset of Chinese paintings. Based on the above situation, the experimental dataset in this study contains 100 paintings of 10 Chinese painters from ancient times to the present, and 70% of the paintings are used as the training sample set. e remaining part is used as the test sample set, and the data in the training and test sample sets do not overlap. e ten painters are CDY, CSF, FZ, HYY, HZ, LDZ, LKL, TY, XBH, and BDSR. In this study, SVM [27][28][29][30] is used for classification experiments.
To test the algorithm in this study, the experiments will be designed from the following four aspects: to verify the influence of the change in the number of painters on the classification accuracy, the classification experiment was conducted to compare the works of 5 painters and 10 painters, respectively; to verify the effectiveness of the algorithm, it is compared with the existing literature such as fusion, MHMM, and other algorithms; to verify the effectiveness of the feature extraction of the VGG-F model, the algorithm in this study is compared with the traditional feature extraction algorithms based on color (HSV), texture (Gabor), and shape (HOG); and to verify the impact of the embedded learning introducing mutual information on the classification, the algorithm in this study is compared with the DEFC algorithm and the pure SVM algorithm.
is study uses the precision rate P (precision) and the recall rate R (recall) to evaluate the execution results of the algorithm, which are defined as P � a/a + b and R � a/N, respectively. Among them, a is the number of samples that are correctly classified into this class, b is the number of samples that are wrongly classified into this class, and N is the number of samples of this class in the test dataset. Considering the first aspect, this study selects the works of 5 painters and 10 painters for experimental comparison, and the comparison results are shown in Tables 1 and 2. It can be seen that the increase in the number of painters has a certain impact on the classification results. e precision and recall rates of the MIDEC algorithm in this study are 94.88% and 87.49% in the classification of the two cases. Although the accuracy rate has decreased, the magnitude is not very large.
e experimental results show that the algorithm in this study has certain robustness. e algorithm in this study is compared with the fusion and MHMM algorithms experimentally. It can be seen from Tables 1 and 2 that the MIDEC algorithm in this study has the highest classification accuracy on the two sets of works; among them, the precision and recall rates of LKL and XBH's paintings are the highest because the two painters have unique painting styles. XBH's paintings are rough, vigorous, and majestic; they pay attention to the anatomical structure of the object, the accurate grasp of the bones, and pay great attention to the modeling. It has a swaying style that is unrestrained without being arrogant and subtle without being trivial; LKL perfectly reproduces the realistic spirit of Western painting with the tools and materials of Chinese  painting. e works are slender and realistic, with a natural look. e art styles of the 2 painters differed the most from the other painters, resulting in the highest accuracy. It can be seen from Table 2 that the recall rate of HZ in this algorithm is 32.99%, which is the lowest recall rate among the 10 painters. is is because HZ copied paintings in the Forbidden City in Beijing for a long time, drawing on the strengths of many families, making his style unpredictable, resulting in a low recognition rate of paintings. As shown in Table 1, compared with the HSV + Gabor + HOG + MIDEC algorithm, the precision and recall of the algorithm in this study are increased by 17% and 18%, respectively. It is proved that the features extracted by the VGG-F model can more effectively reflect the different artistic styles of painters.
As shown in Table 1, the precision and recall of the algorithm in this study are higher than those of the pure SVM and DEFC algorithms, which prove that the introduction of mutual information and embedded can further improve the classification accuracy.
It is known from experiments that for Chinese paintings with different themes and styles that are similar in style and with the same theme and different styles, the algorithm in this study has good robustness in classification.

Conclusions
Most of the existing classifications of Chinese paintings only consider the correlation between paintings and class labels, ignoring the correlation between paintings and the influence of feature importance. In this study, the image features of Chinese painting are extracted by fine-tuning the pretrained VGG-F model. Based on the DEFC algorithm, the mutual information theory is introduced, and an embedded classification algorithm based on mutual information is proposed, which makes the embedded principle affected by feature selection and feature importance. Experiments show that this algorithm has strong robustness.
In addition to analyzing the style characteristics of the painting itself, the identification of Chinese painting painters should also include the analysis of paper color, ink color, color components, style identification, and seals Adding other factor analysis can help overcome the decline of recall and precision, which will be part of the future work. In addition, the existing image emotion research mainly focuses on natural images and face images, lacking the digital description system of Chinese painting emotion, so the next work will also be carried out on the emotional classification of Chinese painting.

Data Availability
e dataset used to support this study can be accessed upon request.

Conflicts of Interest
e authors declare that they have no conflicts of interest.