Deep Learning Model of Image Classification Using Machine Learning

the


Introduction
Not only is using manual method to select image features time-consuming and laborious, but also the feature quality depends on professional knowledge and practical experience to a certain extent, which limits the application of manual feature selection method [1]. With the advent of the era of big data and arti cial intelligence, the traditional arti cial way to obtain image information has been unable to meet the needs of applications in related elds. In recent years, the method of obtaining images by computer has made some progress, but, for the massive image information processing, the traditional computer vision is still ine cient. For example, traditional machine vision methods have low accuracy in detecting low-level features and high-level features of images [2]. Classi cation and processing of massive image information is the prerequisite for in-depth mining of image information. It is also of great signi cance for the expansion of computer vision in the eld of image processing application. At present, the research of image classi cation by computer has been gradually applied to agriculture, industry, aerospace, and other elds. e research of image classi cation mainly includes e cient and accurate classication of large-scale image information and accurate classi cation of image semantic similar feature information.
In order to recognize and classify the image, we need to use the image acquisition tool to obtain the original image information, then clean the image data and extract and lter the features, and nally use the machine learning algorithm to realize the image classi cation. With the deepening of the application of machine vision in many elds, people can use computer vision technology to quickly realize the processing operations of image information, such as noise reduction and feature extraction [3]. e research of image classification based on machine learning has made some progress. From the adoption of artificial recognition to the continuous application of computer vision technology in image classification, scholars have made some achievements in the research of image classification and recognition. Machine learning is an important branch of artificial intelligence. Although machine learning has experienced half a century of development, there are still some unsolved problems, for example, complex image understanding and recognition, natural language translation, and recommendation system [4]. Deep learning is an important branch developed on the basis of machine learning. It makes full use of the hierarchical characteristics of artificial neural network and biological neural system to process information and obtains high-level features by learning low-level features and adopting feature combination method, so as to realize image classification or regression. Different from traditional machine learning, deep learning uses multilayer neural network to automatically learn the image and extract the deep-seated features of the image. Different depth learning models can be formed according to different feature learning and its combination. However, the accuracy of image classification is not high and the operation efficiency of the existing deep learning model is low. erefore, based on the existing basic theory of convolution neural network, this paper establishes a deep learning model of image classification by improving the structure of convolution neural network, which provides a basis for image classification under complex environmental conditions.

Related Works
Image classification provides an important basis for image depth processing and the application of computer vision technology in related fields. Traditional image classification mainly goes through different stages, such as image proprocessing, feature extraction, classifier construction, and learning training [5]. Traditional image classification methods mainly use the extracted basic image features to realize image classification, which can provide a basis for further obtaining the semantic information of images by computer. Traditional image classification generally uses image color, texture, and other information to calculate image features and uses support vector machine and logistic regression to realize image classification [6]. e results of image classification not only depend on the extracted features to a great extent but also are affected by the knowledge and experience of relevant fields.
Not only are the manually acquired features difficult to apply to image classification but also a lot of time is spent in analyzing feature data. At the same time, the traditional machine learning cannot be applied to the processing of large datasets, and it is difficult to realize the optimization of feature design, feature selection, and model training, which makes the classification effect of the model poor. erefore, image classification methods using traditional machine learning are affected in many application fields [7]. Research shows that because texture, shape, and color features can be used for image classification and recognition, low-level basic features can be used as the basis of image classification. Traditional image classification methods generally use single feature extraction or feature combination and take the extracted features as the input value of support vector machine. In recent years, some progress has been made in image classification using artificial neural network classifier. In order to improve the accuracy of image classification, we can focus on the standardized design of low-level features such as texture, shape, and color.
Deep learning realizes the training of large-scale datasets through multilevel network model and adopts the method of layer-by-layer feature extraction to obtain the high-level features of the image. Not only is the deep learning network model used to extract the basic features of the image but also it can obtain the deep features of the image through multiple hidden layers. Compared with traditional machine learning methods, the features obtained by deep learning method are not only accurate but also conducive to image classification. In the process of image recognition and classification, the way of feature learning and combination is mainly determined by the deep learning model [8]. At present, the commonly used deep learning models are sparse model, restricted Boltzmann machine model, and convolution neural network model. Although these models have some differences in feature extraction, they have similarities in image classification and recognition. ey all go through the steps of image information input, data proprocessing, feature extraction, model training, and classification output.
In image classification, some scholars have carried out a lot of research work on image feature representation and classifier selection. For example, the deep learning model based on feature representation can be effectively applied to the recognition and classification of various images. Some scholars use deep convolution neural networks (DCN) to deeply extract image features and apply them to large-scale dataset ImageNet [9]. Experiments show that the model can effectively classify large data image sets. In addition, the deep learning model can effectively learn and describe image features. For example, the deep learning model can better describe the hierarchical features through unsupervised learning, and the features extracted by the model not only have strong expression ability but also improve the efficiency of image classification. A large number of research results show that, with the deepening of the research of image classification methods in related fields, deep learning model has gradually replaced the traditional artificial feature extraction and machine learning methods and will be widely concerned by scholars in the field of image recognition and classification [10].

Basic eory of Neural
Network. e traditional neural network, referred to as artificial neural network (ANN), is a hot spot in the field of early artificial intelligence [5]. Artificial neural network mainly uses the neurons of network model to abstract the characteristics of external things, so as to be used by computer to complete information processing.
Artificial neural network generally establishes the corresponding network structure according to the different construction methods of neurons. Neural network is an operation model composed of several different nodes or neurons connected with each other. Each node in the model is a processing function, and the connection between different nodes uses weight to represent the memory ability of artificial neural network. e output of neural network depends on the connection form, weight value, and excitation function of different nodes. At the same time, the neural network model is mainly constructed according to some algorithm or function to express some specific logical operation.
A basic neural network model usually includes information input layer, hidden layer, and calculation result output layer. Different layers can contain several neurons [11]. Neurons represent a transformation or operation, which is completed by the activation function of neurons. Two adjacent layers of neurons are connected to each other, as shown in Figure 1.
As can be seen from Figure 1, the neural network model includes 11 neurons: 3 input layers, 5 hidden layers, and 3 output layers. e structure belongs to a two-layer neural network model, where W 1 and W 2 are the weight matrices of the hidden layer and the output layer, respectively.
Deep learning method is a part of machine learning. It is widely used in natural language recognition and image detection and classification. Moreover, deep learning comes from the theory of artificial neural network. By referring to the human brain for hierarchical processing of information, different levels of neural networks are established. Deep learning effectively extracts multilevel feature information by simulating human brain, so as to obtain the key feature information of image, text, and other data. Deep learning mainly describes the specific object characteristics through hierarchical processing according to a large amount of edge feature information. It is a process from low-level feature extraction to high-level feature combination. As an important method of machine learning, deep learning is an effective method to process big data and obtain abstract features by using neural network model.
Deep learning is a multilayer deep neural network model. According to the connection law of human brain neurons, the sample features are processed in different layers of the model, and the deep features of the sample data are obtained in turn. Similar to deep neural network, artificial neural network belongs to hierarchical structure. e model is composed of multilayer perceptron, including input layer, hidden layer, and output layer. Different from the deep neural network, the artificial neural network model only contains two to three layers of forward neural network, and the number of neurons in each layer is small, so the processing ability of large datasets is limited. Because the deep neural network model contains many layers and each layer contains a large number of neurons, the neural network model not only can realize the abstract expression of data but also has strong learning function. Compared with the traditional machine learning methods, the deep learning model does not need to rely on manual design and feature extraction. It not only solves the shortage of manual or knowledge but also avoids the preference problem in feature extraction. At the same time, the deep learning model can obtain representative high-level features through the organic combination of low-level features [12]. As shown in Figure 2, the working processes of deep learning and traditional machine learning are compared.
In the neural network model, the activation function is used to perform nonlinear operation on the input data of neurons in order to extract effective feature information from the original input data. Activation function is a nonlinear function. With the increase of the number of layers of neural network model, the most effective feature information can be obtained after many iterations and data training.
In order to realize the classification of features, Softmax function is often used as the activation function in the neural network model and used in the output layer of the model [13]. Softmax function is generally used as a classifier. e calculation formula is as follows: (1) In the above formula, n is the number of neurons in the current layer and R k is the nonlinear transformation value of the k-th neuron in this layer.
Softmax is a classifier that can output different feature categories. e value of each neuron contained in Softmax can be considered as the probability of the corresponding category.
e calculation process of Softmax function is shown in Figure 3.   [14]. Different from traditional machine learning, convolution neural network can be better used for image and time series data processing, especially for image classification and language recognition. e basic structure of convolution neural network is shown in Figure 4.
Convolution neural network generally includes three different types of processing layers: convolution layer, pooling layer, and full connection layer. Among them, the feature extraction task is completed by the convolution layer, and the pooling layer is used for feature mapping. e full connection layer is similar to the general neural network structure. All nodes in this layer are not connected to each other but completely connected to the nodes of the previous layer. In addition, like other neural networks, convolution neural networks also have data input layer and result output layer. e calculation task of convolution neural network is mainly completed through the convolution layer, and the convolution kernel in the convolution layer is the core of the convolution neural network model. e convolution layer uses convolution check to convolute the input image and extract the characteristic information of the image. e images processed by convolution operation will gradually become smaller, and the pixels at the edge of the image have little effect on the output results.
As shown in Figure 5, assuming that the original input image is a 3 × 3 matrix, the original image is convoluted through a convolution check with a size of 2 × 2, and the corresponding feature map is output.
Generally, there is a strong correlation between adjacent pixels in the image. Convolution kernel mainly extracts   features from the local area of the image and sends the extracted local features to the high level for integration processing. Because the bottom feature of the image is independent of its position, it can not only use the same convolution check to extract the relevant features but also reduce the number of parameters of the neural network through the shared weight characteristic of the convolution kernel, so as to improve the training efficiency of the network model. For complex images, in order to reduce the amount of parameter training of the model, the pooling layer in convolution neural network can be used to reduce the size of feature map. During pooling, the depth and size of the image can remain unchanged. e operation of pooling layer generally includes max pooling and average pooling [15], as shown in Figure 6.

Convolution Neural Network Model.
In the existing image detection and recognition models, such as ResNet, Mask-RCNN, and Faster R-CNN models, they are usually based on common network models [16]. LeNet is the most basic convolution neural network model [9]. After transforming the LeNet model, the LeNet-5 model is established to classify ordinary images. Because the LeNet-5 model is not deep enough, it cannot extract enough image features during model training. erefore, it cannot be applied to the classification of complex images. erefore, some people put forward the AlexNet model based on the LeNet structure, applied the convolution neural network to the processing of complex images, and provided a theoretical basis for the application of deep learning model in the field of computer vision [17]. e network structure of AlexNet is shown in Figure 7.
AlexNet is a network structure with 8 layers. e model includes 5 convolution layers and 3 full connection layers. e model uses the ReLU function as the activation function to avoid the gradient dispersion phenomenon caused by the large number of layers of the network model. In order to reduce network training time, AlexNet uses multiple GPUs for training. In order to suppress neurons with small response ability, AlexNet uses LRN (local response normalization) processing layer to establish a competition mechanism for neurons, so as to make the value with large response ability increase continuously, so as to increase the generalization ability of the model. In addition, in order to prevent neurons from forward propagation and back

Advances in Multimedia
propagation, the dropout mechanism is adopted in the full connection layer of AlexNet model, so that the output results of all hidden layer neurons are 0, so as to avoid the complex interaction between neurons. At present, VggNet is a widely used deep convolution neural network model. Compared with other models, VggNet not only has better generalization ability but also can be effectively used for the recognition of different types of images [11]. For example, convolution neural networks such as FCN, UNet, and SegNet are based on VggNet model. In recent years, Vgg-16 and Vgg-19 networks have been commonly used for VggNet models [18]. e network structure of Vgg-16 is shown in Table 1.
e Vgg-16 network structure has 16 layers in total, excluding the pooling layer and Softmax layer. e convolution core size is 3 × 3, the pooling layer size is 2 × 2, and the pooling layer adopts the maximum pooling operation with step size of 2.    Vgg-16 network uses convolution blocks instead of convolution layers, in which each convolution block contains 2∼3 convolution layers, which is conducive to reducing the network model parameters. At the same time, Vgg-16 network adopts ReLU activation function to enhance the training ability of the model. Although Vgg model has more layers than AlexNet model, the convolution kernel of Vgg model is smaller than that of AlexNet model. erefore, the number of training iterations of Vgg model is less than that of AlexNet model.

Improved Image Classification Model Framework.
Because image classification algorithms are usually used in systems with high real-time requirements, image classification algorithms need to consider real-time performance. For complex neural network models, image classification needs to consume a lot of time. erefore, this paper simplifies VggNet model and takes it as the model basis of image classification.
Considering the distribution characteristics of datasets used for model training, a typical dataset can be selected as the weight of the model to initialize the training dataset. When the model is pretrained and reaches a certain accuracy, the number of nodes in Softmax layer is reduced by ten times, and then the dataset is used for weight training. Considering that the data processed by the model may be affected by various noises, a noise reduction automatic encoder is added to the model to eliminate the noise interference, and the existing dataset is extended through the data enhancement method to enhance the generalization ability of the model.
Considering that the image classification algorithm needs to meet certain real-time performance, the corresponding image classification model is established and optimized based on VggNet model. Among them, the algorithm combining convolution neural network and noise reduction automatic encoder can be used. Because there may be overfitting problem in image classification, it can be optimized by data enhancement. Compared with other algorithms, this classification algorithm has certain generalization performance in the case of small amount of data. In addition, the algorithm also adds a noise reduction automatic encoder, which can effectively reduce the impact of data noise on the performance of the model, so as to ensure that the model has good generalization ability. Since the improved algorithm is based on VggNet network, the training time of the model may increase [19]. e improved image classification algorithm is shown in Figure 8.
In order to solve the problem of automatic noise reduction of complex image structure, the normalized network of encoder is used for classification in this paper. Based on the existing convolution neural network model, the noise reduction automatic encoder and sparse automatic encoder are organically combined, and the input original image information is normalized on the sparse automatic encoder. en, the improved convolution neural network model is used to extract the image feature information, and the Softmax classifier is used to classify the features [14]. When the improved convolution neural network model is used to classify images, it is very necessary to proprocess the image such as noise reduction and grayscale, select a certain number of training sets and test sets from the dataset, and then take the training set as the input object of the model after unsupervised learning processing. Secondly, the hidden layer of the noise reduction automatic encoder is used to encode and decode the input object, and the processing results are output to the sparse automatic encoder of the next layer for normalization. e data is trained layer by layer through the hidden layer of sparse automatic encoder, and finally the training results of sparse automatic encoder are output to Softmax classifier. In order to improve the classification accuracy, gradient descent method can be used to strengthen the training of classifier model parameters in order to improve the performance of image classification depth learning model. Finally, the network model is verified by using the image test set, and the effectiveness of the image classification method is tested according to the classification results output by the model. e improved convolution neural network model can overcome the problem that the traditional neural network is only limited to some features in image classification.
rough the normalization of sparse automatic encoder, the overfitting phenomenon of model in data processing can be avoided, and more abstract representative features can be obtained by using the hidden layer of sparse automatic encoder to train the data layer by layer. e improved model adopts Softmax classifier, which can make the classification result closer to the real value. e improved deep learning network model is mainly divided into two stages: training and testing. e training stage is mainly used to build an effective image classification model, and the testing stage is mainly to evaluate and analyze the model according to the experimental classification results. Figure 9 shows the workflow of the improved deep learning network model.

Image Classification Model Optimization.
It is known from the existing research that the convolution neural network model can be optimized from the aspects of data enhancement and adjusting training methods, and the optimization of convolution neural network model is related to the type of model used. For example, for the deep learning model with more layers, the training parameters can be optimized.
According to the structure of convolution neural network, convolution layer is used to extract features, and the size of convolution kernel determines the extraction quality of image features. From the perspective of image composition, adjacent pixels can generally form the edge lines of the image. Several edge lines form the image texture, and the image texture is combined into several local patterns. e local pattern is the basic element of the image. rough the convolution layer of the network model, different types of features can be extracted and the local pattern of the image can be formed. When the convolution kernel is smaller, Advances in Multimedia although the convolution layer extracts more features, there may be overfitting problems in the model. On the contrary, the larger the convolution kernel is, the fewer features extracted by the convolution layer are, and the worse the image classification effect may be. erefore, the reasonable optimization of the size of convolution kernel can improve the accuracy of image classification.
Because the convolution neural network model mainly extracts the image features layer by layer through different convolution layers, the number of convolution layers will affect the feature extraction quality of the model to a certain extent. Similar to the number of convolution kernels contained in the convolution layer, the more convolution layers, the finer the features obtained by the model classifier, which may lead to overfitting phenomenon, while the less convolution layers, the coarser the features obtained by the model classifier, which may lead to the decline of image classification accuracy. erefore, the optimization of convolution layers can improve the classification accuracy of the model. In order to improve the accuracy of image classification and recognition, the depth learning model proposed in this paper needs to be optimized. Firstly, a smaller convolution kernel is selected in the first convolution layer in order to extract more image feature details. Secondly, the maximum pool sampling operation is adopted in the model to avoid the overfitting problem. As shown in Figure 10, the optimized image classification model consists of three convolution layers, in which the convolution kernel of each convolution layer decreases in turn. At the same time, after each convolution layer, the features are processed by ReLU activation function, and the generated features are used as the input of the maximum pooling layer. e model adopts three full connection layers, takes the processing result of the last full connection layer as the input of Softmax classifier, and then generates the classification result of the image. e first convolution layer is close to the image input layer and is mainly used to extract the basic features of the image, so the parameters of the first convolution layer have a great influence on the features. In order to facilitate the further processing of features in the subsequent convolution layer, a smaller convolution kernel needs to be used to extract the attribute information such as shadow, boundary, and light of the image. e convolution layer maps the obtained features through the activation function. erefore, the optimized convolution neural network model adopts ReLU activation function [20], which can be expressed by mathematical function as follows: When the traditional convolution neural network model uses ReLU activation function to train features, it may lose useful feature information in the process of image classification. In order to prevent the loss of useful features during image classification, it can be improved on the basis of the existing ReLU activation function [20]. e optimized activation function can be expressed as From the improved calculation formula of activation function, when the input feature is less than zero, it can not only retain the negative value information in the feature map but also increase the reinforcement learning of effective features.
e optimized convolution neural model uses Softmax function to classify the images, and Softmax function uses supervised learning algorithm to regress the features [21]. In the classification process, category y of the image target can have M different values. If the image training set is (x 1 , y 1 ), · · · , (x i , y i ) , where x i represents the image training sample, y i is the image classification category, and y i ∈ 1, 2, · · · , M { }, the cost function of Softmax regression algorithm can be expressed as Assuming that M markers are accumulated in the cost function, the probability calculation of training sample x as category k can be obtained, which is expressed as

Selection of Datasets and Experimental Methods.
In order to verify the effectiveness of the image classification depth learning model proposed in this paper, the Flower dataset provided by Oxford University was used in the experiment [22]. ese images describe the proportion, shape, and light changes of different types of flowers, and the images of some flower categories change greatly. e dataset contains 17 categories of flower datasets; each category contains 80 pictures, for a total of 1360 pictures. In the experiment, the dataset is randomly divided into three parts as the training set, verification set, and test set of the model. A partial image of the Flower dataset is shown in Figure 11. e experiment is based on Matlab, Python, and deep learning framework.
rough the proprocessing, feature extraction, training, and classification of Flower dataset, the effectiveness and accuracy of the model in image classification are verified.

Advances in Multimedia
In the experiment, the classification accuracy is used to evaluate the robustness of the deep learning network model. According to different evaluation angles, classification accuracy generally includes overall accuracy and category classification accuracy [23]. e overall accuracy AC is expressed by the ratio of the number of correctly classified samples to the total number of samples, while the classification accuracy CCA i is expressed by the ratio of the number of correctly classified samples to the number of such tests. e calculation formula is as follows: In the above formula, t ni represents the number of correctly classified samples, t n denotes the number of test samples, t ri indicates the number of correctly classified samples of type i, and t ni expresses the number of test samples of type i.

Results and Analysis.
In order to facilitate comparative analysis, three common network models, LeNet, AlexNet, and VggNet, are selected for comparison in the experiment. According to the experimental comparison results, the relationship between the classification accuracy and the number of iterations of the model when using different network models to classify flower images is shown in Figure 12. When iterating 50 times, the accuracy of LeNet model is 28%, that of AlexNet model is 17%, and that of VggNet model is 48%. When the number of iterations is 100, the accuracy of LeNet model is 45%, that of AlexNet model is 39%, and that of VggNet model is 72%. In addition, when the VggNet model converges, the highest accuracy rate is as high as 75%, the highest accuracy rate of LeNet model is 58%, and the highest accuracy rate of AlexNet model is 41%.
In order to test the effect of image classification after the model is optimized to a certain extent, the accuracy of flower image classification before and after the model optimization is compared in the experiment, as shown in Figure 13. e comparison results show that, for the training dataset, the optimized model converges faster in the early stage of training and slower in the middle stage of training, while the two models in the later stage of training are basically the same. For the test dataset, the optimized model is higher than the nonoptimized model in terms of convergence speed and image classification accuracy. erefore, the model optimization method proposed in this paper can effectively improve the accuracy of image classification.
In addition, in order to test the relationship between the loss value function of the optimized model and the number of iterations, the training set and test set are used to compare the models before and after optimization, as shown in Figure 14.
e loss value function of the nonoptimized model shows an upward trend with the increase of the number of iterations, indicating that the nonoptimized model has the phenomenon of overfitting, while the loss value function of the optimized model shows a downward trend with the increase of the number of iterations. It can be seen that the cost of parameter training can be reduced through model optimization. From the above experimental comparison results of the relationship between the accuracy of common network models in image classification and the number of iterations, it is known that the model proposed in this paper is superior to other models in classification accuracy. By comparing the classification accuracy of the deep learning model on the training set and the test set before and after optimization, it is known that the accuracy of image classification can be significantly improved after a certain degree of optimization.

Conclusion
Aiming at the problems of large time overhead and low classification accuracy in traditional image classification methods, a deep learning model of image classification based on machine learning was proposed in this paper. By analyzing the basic theory of neural network, this paper expounded the types of convolution neural network and its application in image classification. Using the existing convolution neural network model, through noise reduction and parameter adjustment in the feature extraction process, an image classification depth learning model based on improved convolution neural network was proposed. In order to improve the classification efficiency and accuracy of the model, this paper optimized the proposed deep learning model. Finally, the accuracy of several common network models in image classification was compared through experiments. e results showed that the proposed model was better than other models in classification accuracy. At the same time, the classification accuracy before and after the optimization of the deep learning model was compared and analyzed. e results showed that the optimized model had a great improvement in the accuracy of image classification. How to classify dynamic targets in complex environment is a problem that needs further research in the future.

Data Availability
e labeled dataset used to support the findings of this study is available from the corresponding author upon request.