Deep Learning Hash for Wireless Multimedia Image Content Security

With the explosive growth of the wireless multimedia data on the wireless Internet, a large number of illegal images have been widely disseminated in wireless networks, which seriously endangers the content security of wireless networks. However, how to identify and classify illegal images quickly, accurately, and in real time is a key challenge for wireless multimedia networks. To avoid illegal images circulating on the Internet, each image needs to be detected, extracted features, and compared with the image in the feature library to verify the legitimacy of the image. An improved image deep learning hash (IDLH)method to learn compact binary codes for image search is proposed in this paper. Specifically, there are three major processes of IDLH: the feature extraction, deep secondary search, and image classification. IDLH performs image retrieval by the deep neural networks (DNN) as well as image classification with the binary hash codes. Different from other deep learning-hash methods that often entail heavy computations by using a conventional classifier, exemplified by K nearest neighbor (K-NN) and support vector machines (SVM), our method learns classifiers using binary hash codes, which can be learned synchronously in training. Finally, comprehensive experiments are conducted to evaluate IDLH method by using CIFAR-10 and Caltech 256 image library datasets, and the results show that the retrieval performance of IDLH method can effectively identify illegal images.


Introduction
With the rapid development of multimedia technology, the retrieval of wireless multimedia data has become very convenient and easy.The security of multimedia data has attracted more and more attention from researchers.Multimedia data content security belongs to the branch of information security and requires direct understanding and analysis of the information content transmitted in the network.Judging rapidly from the massive information, filtering, and monitoring the abnormal information in the network are the key to ensure the security of the wireless network content.Meanwhile, a large number of illegal images are disseminated in the network, which seriously endangers the security of network content.So, it is very important practical significance to research the content security-oriented image recognition technology and identify and supervise illegal image information in the network.Although there are a lot of image retrieval methods currently, due to the various disadvantages, such as the low expression ability of image feature, high dimension of feature, and low precision of image retrieval, the retrieval results are not always effective.
How to retrieve the large-scale image resources quickly and effectively meet the needs is urgently to solve.Since the hash-based learning algorithm can effectively preserve the similarity between the original feature spaces and the hash code spaces, more and more scholars have drawn attention to it.Particularly, learning deep hash algorithm has greatly improved the retrieval performance.
Besides the widely used text-based search methods, content-based image retrieval (CBIR) has attracted widespread attention of more and more scholars in the past decade [1].As we all know, the nearest neighbor search algorithm is an effective method for searching for similar data samples [2].We should not only consider the scalability issue, but also consider most practical large-scale applications which are affected by dimensional disasters [3].The high-level encoding complexity prevents widespread adoption in real-time multimedia systems [4].One of the many practical applications, the approximate nearest neighbors (ANN), is very efficient.However, the method requires huge storage costs and hard to handling high-dimensional data.Hence, quantization techniques have been proposed to encode high-dimensional data vectors recently.Due to the fact that hashing-based ANN search techniques can reduce the storage via storing the compact binary codes, hash technology is widely used in computer vision, machine learning, information retrieval, and other related fields, in view of the retrieval of large-scale data.The goal of hash is to turn high-dimensional data into the low-dimensional compact binary codes [5].For example,  dimension data is turned into  dimension (d>>r), and generally  dimension data is between dozens of bits and hundreds of bits.After turning high-dimensional data into hash code, we calculate the distance or similarity between the data rapidly.Because of the high efficiency of binary hash code in Hamming distance calculation and the advantages of storage space, hash code is very efficient in large-scale image retrieval.
One of the most famous jobs in hash and the most applied job in the industry was locality sensitive hashing (LSH) [3], which was put forward by Gionis, Indyk, and Motwani in 1999 [1].LSH is one of the most popular data independent methods, which generates hash functions by random projections [6].In addition to traditional Euclidean distance LSH has been generalized to accommodate other distance and similarity measures such as p-norm distance [3], Mahalanobis metric [7], and kernel similarity [8].And L. Qi, X. Zhang, W. Dou, and Q. Ni put forward another application direction of LSH in 2017 [9].They proposed a privacypreserving and scalable service recommendation approach based on distributed LSH (SerRecdistri-LSH).The purpose of this application is to use SerRecdistri-LSH method to handle service recommendation in distributed cloud environment.
A disadvantage of the LSH family is that LSH usually needs long bit length (≥ 1000) to achieve both high precision and recall.This may leads to a huge storage overhead and thus limits the sale at which an LSH algorithm may be applied.In order to achieve the desired search accuracy, LSH often requires the long hash codes, thereby reducing the recall rate.This problem can be alleviated by using multiple hash tables, but it greatly increases storage costs and query time.[10].
So, learning-based data-dependent hashing methods have become increasingly popular because of the benefit that learned compact binary codes can effectively and highly efficiently index and organize massive data [3].The goal of the data-dependent hashing methods is to generate short hash codes (typically ≤ 200) using available training data.Various hashing algorithms have been proposed in the literature, of which a large category focuses on linear hashing algorithms which learn a set of hyperplanes as linear hash functions.The representative algorithms in this category include unsupervised PCA Hashing [11], Iterative Quantization (ITQ) [12], and Isotropic Hashing [13] and supervised Minimal Loss Hashing (MLH) [14], Semisupervised Hashing (SSH) [11], Supervised Discrete Hashing (SDH) [6], Ranking-Based Supervised Hashing [15], FastHash [16], etc.A bilinear form of hash functions was introduced by [17,18].
As an extension of linear hash functions, a variety of algorithms have been proposed to generate nonlinear hash functions in a kernel space, including Binary Reconstructive Embedding (BRE) [10], Random Maximum Margin Hashing (RMMH) [14], Kernel-Based Supervised Hashing (KSH) [12], and the kernel variant of ITQ [13].In parallel, harnessing nonlinear manifold structures has been shown to be effective in producing compact neighborhood-preserving hash codes.The early algorithm in this fashion is Spectral Hashing (SH) [19], which produces hash codes through solving a continuously relaxed mathematical program similar to Laplacian Eigenmaps.More recently, Anchor Graph Hashing (AGH) [20] leveraged anchor graphs for solving the eigenfunctions of the resulting graph Laplacians, making hash code training and out-of-sample extension to novel data both tractable and efficient for large-scale datasets.Shen et al. [6,21,22] proposed a general induction.
The rest of this paper is organized as follows.Section 2 briefly reviews the related works about deep-learninghashing.And we present framework and steps of the improved deep learning-based hash, IDLH, algorithm in detail in Section 3. Section 4 evaluates the effectiveness of the IDLH through a series of contrastive experiments and carefully analyzes the experimental results, followed by the conclusion in Section 5.

Related Work
Recently, deep-learning-hashing, as a popular research topic, has drawn increasing attention and research efforts in information retrieval, computer vision, and machine learning.Semantic hashing is recognized as the earliest starting deep learning hash [23].This method establishes a deep generative model to discover hidden binary features.Such a deep model is made as a stack of restricted Boltzmann machines (RMBs) [24].After learning a multilayer RBM by pretraining and fine-tuning the document collection, the hash code of any document is obtained by simply thresholding the deepest output.Such a hash code provided by deep RBM is shown to maintain a semantically similar relationship of input documents into code space, where each hash code (or hash key) is used as a memory address to locate the corresponding document.In this way, semantically similar documents are mapped to adjacent memory addresses, enabling efficient searching through hash table lookups.In order to improve the performance of deep RBMS, a supervised version was proposed in [25] and the idea of nonlinear neighborhood component analysis (NCA) embedding in [26] was adopted.The supervised information is derived from training a given neighbor/nonneighbor relationship between samples.Then, based on the depth RBM, the objective function of the NCA is optimized so that the depth RBM yields a hash code.Note that supervised deep RBMS can be applied to wide data fields other than text data.In [25], the depth RBMs are supervised using a Gaussian distribution, and the models of the visible units in the first layer are successfully applied to the processing of massive image data.
In [27], a deep neural network was developed to learn multilevel nonlinear transformations, mapping the original image to a compact binary hash code to support large-scale image retrieval for learning binary image representations.A deep hashing model is established under the three-layer constraint on the deep network: (1) the reconstruction error between the original real-valued image feature vector and the resulting binary code is minimized; (2) each bit of the binary code has a balance; (3) all bits are independent of each other.Similar constraints were used in previous unsupervised hash or binary coding methods, such as iterative quantization (ITQ) [28].In [27], a supervised version is called supervised depth hash 3, in which a discriminative item containing two pairs of supervised information is added to the objective function of the deep hash model.
It is worth noting that, in the training of sparse neural networks, in addition to the sparse similarity maintaining hash, depth hash, and supervised depth hash, the pretraining phase is not included.Instead, the hash codes are learned from scratch using a set of training data.However, no pretraining can make the general hash code less efficient.In particular, the sparse similarity keep hash method is found to be inferior to the existing supervised hashing method, i.e., kernel-based supervised hashing (KSH) [29], in terms of search accuracy on some image datasets [30]; the deep hash method and its supervised version are slightly better than ITQ and its supervised version CCA+ ITQ, respectively [31].Note that KSH, ITQ, and CCA+ITQ develop a shallow learning framework.
One of the main purposes of deep learning is to learn the robust and powerful representation of complex data.It is very natural to use deep learning to explore compact hash codes, which can be thought of as binary representation of the data.The deployed CNN consists of three convolution pools, including rectified linear activation, maximum pool merging, and local contrast normalization, a standard fully connected layer, and an output layer with softmax activation functions.In [32], a new method called deep semantic sorting hashing is proposed to learn hash values, thereby maintaining multilevel semantic similarity between multilabel images.This method is combined with convolutional neural network hashing method, taking image pixels as input, training depth CNN, and jointly learning image feature representation and hash value by this method.The deployed CNN consists of five convolution-pooling layers, two fully connected layers, and a hash layer (i.e., output layer).
Indexing massive amounts of multimedia data, such as images and videos, is a natural application based on learning hashing.In particular, due to the well-known semantic divide, hashing methods have been extensively studied for image search and retrieval, as well as mobile product search [33,34].Although hashing techniques have been applied to the active learning framework to cope with big data applications, these hash algorithms are rarely applied to image classification.For the image recognition problem with many categories, the computational and memory overhead primarily stems from the large number of classifiers to be learned.Table 1 show that the complexities can be high at the stages of both training and deploying these classifiers.
Inspired by [35,36], we proposed a combination classifiers with DNN and binarizing classifiers to solve some classic security calculation problems [37][38][39].Different from other deep learning-hash methods that often entails heavy computations by using a conventional classifier, exemplified by K-NN and SVM, our method learns classifiers using binary hash codes, which are simultaneously learned from the training data.The combination classifiers can provide both image features and accelerate image classification, and thus it make the large-scale image recognition faster.The advantages of the extending hashing techniques from fast image search to image classification inspire us to apply them to deep learning hash framework.
With the development of Internet technology, the spreading form of illegal information on the Internet is changing gradually.Traditionally, the form of communication based on word description has been transformed into a diversified form of communication based on video and image.Therefore, the original keyword blocking, web content grading, blacklist restriction, and other filtering methods have not been able to block illegal information dissemination.At present, network illegal image recognition is mainly divided into the following categories.
(1) Erotic Image Recognition.The identification of online pornographic images mainly includes methods based on skin color feature extraction detection and judgment and methods based on limb judgment.The limb state is determined by extracting the divided skin color information and the connection feature of the human body posture and further determining whether the image transmitted on the network contains pornographic content.
(2) Images Involving State Secrets or Military Secrets.The recognition of secret-related images mainly involves steganography, which hides secret information in image, video, and other carriers for secret transmission.At present, the high-order statistical features based on modeling the complex correlation of image neighborhood become the mainstream features in the field of steganalysis.SRM (Spatial Rich Model), PSRM (Projection Speciation Rich Model), and other models are based on such high-order, high-dimensional features and have achieved good detection results.Steganalysis based on depth learning is a hotspot in the field of information hiding in order to identify the illegal secret-related images accurately and quickly.
(3) Images Containing Antihuman Content such as Terrorist Violence.The identification of such images is mainly based on image contrast techniques.Image comparison includes techniques such as image feature extraction, high-dimensional spatial feature index establishment, and similarity measure.It is a very worthwhile to study how to quickly compare the massive network images to the illegal target images, so that the recall rate and the precision rate can be taken into account.

The Proposed Method
In this section, we will present the notations, as summarized in Table 2 firstly.The concept of deep learning stems from the field of artificial neural networks.Deep learning is deep neural network learning and is a learning structure with multiple hidden layers.In the process of deep learning, the network is trained layer by layer.Each layer of the learning network extracts certain features and information and takes the training result as a deeper input.Finally, the entire network is fine-tuned with a top-down algorithm.
Through deep learning, complex function expressions can be learned, thereby completing the concept of high-level abstraction from the underlying information.It has been widely used in language understanding, target recognition, and speech perception.
Figure 1 shows that the proposed framework IDLH includes three components.The first component is preprocessing layer on the image dataset.The second component is the training self-coding network with a layer-by-layer greedy learning algorithm to obtain the feature expression function of the image.The third is hash layer which retrieves images similar to the query image with compact binary codes and categorizes the query one by the majority semantic label within the hashing bucket.

Preprocessing.
Since the depth learning algorithm used in this paper is an unsupervised learning algorithm, it can automatically learn the deep features of the image from the original pixel information of the image.Therefore, the original pixel value of the whole image can be directly used as input data for the deep learning model.In order to facilitate the training of the network, it is necessary to preprocess the image.Through preprocessing, the image is simply scaled, sample-by-sample mean-value reduction, and whitening is processed to reduce the redundant information in the image and facilitate the deep learning network for training and calculation.Preprocessing can be further grouped into three suboperations.
(1) Normalization.Normalization can prevent neuron output saturation caused by excessive net input absolute value.We use the sigmoid function to do normalization, as shown as follows: is the gray value of the image pixels.
(2) Pointwise Mean Reduction.This process mainly is to get rid of the redundant information of the image, the mean value is eliminated for each point of the image, the average brightness of the image is removed, and the DC component of the data is eliminated.Assuming that  () ∈   is the gray value of each pixel of image I, we use formulae ( 2) and ( 3) to zero-mean image.
(3) Whitening.Whitening is an important pretreatment process; its purpose is to reduce the redundancy of input data, so that the whitened input data has the following properties: (i) low correlation between features; (ii) all features having the same variance, then the formula is as formula (4) .
In formula (4) the rotation matrix of  layer can be obtained, and the output of the hidden layer can be used as the input of the second self-encoder to continue training the second self-encoder.Then, the output ℎ () 2 of the second hidden layer of the self-encoder can be obtained.The output ℎ () 1 of the first hidden layer of the self-encoder is called a first-order feature, and the output ℎ () 2 of the second hidden layer of the self-encoder is called a second-order feature.In order to classify, the two-order feature ℎ ()  2 can be used as the input of Softmax regression.
Figure 2 shows the flowchart of the proposed method.And there are mainly three processes (supervised pretraining, rough image retrieval, and accurate image retrieval).The object of the first process is to transform the highdimensional feature vector into a low-dimensional compact two value codes through hash function.In the second procedure, we pick out M candidate images by calculating Hamming distance.In the third process, we calculate the Euclidean distance between the candidate image and the image to be retrieved and accurately extract K images from the M candidate images.
Figure 3 shows a block diagram of a self-encoding neural network.The stacking self-encoding network contains 3 hidden layers (feature layers).The input layer inputs the original data i into the first layer of the feature layer, the output result of the former layer serves as the input of the next layer, and the output of the third layer serves as the feature expression of the image.In our method, it is also used as the input of the hash classifier, and it is possible to use the characteristics of the STD neural network to classify the features.
By using the matrix representation of the binary codes vectors and the output of the 3th layer of the network, we use the gradient descent method to solve the neural network.
For a single sample (x, y), the cost function is as shown in For datasets containing m samples, the optimization cost function is formulated by the following formula: The first term (, ) represents the mean variance term.The second term aims to prevent the data from overfitting by reducing the magnitude of the weight. is a weight attenuation parameter.It is used to balance the relative importance of mean square deviation terms and weight attenuation terms.Our purpose is to minimize the quantization loss (, ) between the learned binary values and the real values of an input image according to parameters  and .

Hash Algorithm Retrieval.
The image retrieval method based on hash algorithm maps the high-dimensional content features of images into Hamming space (binary space) and generates a low-dimensional hash sequence to represent a picture.This method reduces the requirement of computer memory space for image retrieval system, improves the retrieval speed, and better adapts to the requirements of mass image retrieval.
Inspired by [6,8], we use a set of hash functions to hash data into different buckets.After we do some hash mapping on the original image feature data, we hope that the original two adjacent feature data can be hash into the same bucket with the same bucket number.And then, after hash mapping of all the data in the original feature set, we can get a hash table.These original feature data sets are scattered into hash table buckets, and the data belonging to the same bucket is probably adjacent to the original data.However, there is also  Output ℎ (3)  k from layer 3 is used as binary code with b k = f(ℎ (3)  k ) Figure 3: Self-learning network based on stack self-encoding network.The neurons labeled x i is the input of the neural network and "+ 1" are the offset nodes (intercept entries) of the neural network.The layer 0 is the input layer of neural network, and layer 3 is the output layer of neural network.The middle layers of layer 0 layer to layer 3 are the hidden layers of neural network.
a small probability in events; that is, the nonadjacent data is hash to the same barrel.Set the hash function as the follows: Here   is the projection vector and   is the corresponding intercept.The code value generated by formula ( 7) is {−1, 1}, and we use the following formula to convert it into two value codes: Given a sample point ∈   , we can compute a K-bit binary code  for  with formula (9).The hash function performs the mapping as ℎ  :   →   = {ℎ 1 () , ℎ 2 () , . . ., ℎ  ()} Then, for a given set of hash functions, we can map them to a set of corresponding binary codes by formula (10).
Here  = {  }  =1 ∈  × is the feature data matrix with points as columns.Such a binary encoding process can also be viewed as mapping the original data point to a binary valued space, namely, Hamming space.

Similarity Measure.
After obtaining the binary hash code of the image, it is necessary to measure similarity between the retrieved image and the library image in the Hamming space.The smaller the Hamming distance is, the closer distance between the two data is, and the degree of similarity is higher; otherwise, the two data similarity is lower.Here ⊕ is an XOR operation.The two sets of   and   represent the hash code of the search image feature and the image library is mapped through the hash function.The new image features learned by the stack self-encoding network are generated by the hash function.The storage structure of the image feature vectors is shown in Table 3.
As can be seen from Table 3, the hash code of the image is related to the image ID and the image name one by one.In the process of searching, the image feature vector is obtained through deep learning by a hash function, the original data is mapped into a new data space, and a corresponding hash code is obtained.The hash code is used to calculate the Hamming distance in the Hamming space as a measure of similarity between images.Finally, the storage structure of the image feature vector is used to find the corresponding image ID of the hash code, and the output retrieval result is output to the user.

Image Secondary Search.
In the first-level search phase, the features learned from the deep learning network are mapped into the Hamming space using the hash function.In the similarity measurement phase, the traditional Euclidean distance is abandoned.Measure the similarity between images by comparing the Hamming distance between the image features of the query image and the image of the library image.In order to further improve the accuracy of retrieval without affecting the real-time performance, we can retrieve the image by the second level retrieval.These steps are described in detail as follows: After one level retrieval, we choose the K images with the most similarity in the first-level retrieval result and then calculate the Euclidean distance between the original feature vector of the K images and the original feature vector of the query image.The results obtained as the similarity measure of the images and output the retrieval result that has been ranked from the high and low with the similarity distance.
Although the Hash algorithm maps the high-dimensional feature vectors of the image into a hash-coded form, the problem of "dimensional disasters" is solved, and the retrieval efficiency is greatly accelerated.However, when the similarity comparison is performed, the Hamming distances of the image features are simply compared using the results of the primary search, and occasionally undesirable results may still appear on the search results.If we want to increase the accuracy of the search, we must increase the hash code length.However, excessively long codes will increase the amount of calculations, increase the memory burden, and reduce the real-time nature of retrieval, failing to achieve the goal of reducing the size of data.In order to solve this problem, keep the retrieval efficiency, and further improve the retrieval accuracy, we propose a search strategy for secondary retrieval, the specific steps of which are as follows.
Step 1. Through the first-level search in the Hamming space, the similarity degree of the images is sorted, and the top K sorting images are selected.
Step 2. For the  images in Step 1, calculate the Euclidean distance one by one from its original image feature vector to the image feature vector of the query image.
Step 3. The Euclidean distance calculated in Step 2 is sorted.The smaller the calculated value is, the higher similarity between images is, and the similarity is sorted from high to low and output as the final search result.
In the second search, it is necessary to pay attention to the selection of the  value, although the larger the  value is, the better the search effect is, but accordingly, the longer the time is consumed.Therefore, it is necessary to combine various factors to select the appropriate  value.

Experimental and Performance Analysis
In this section, we thoroughly compare the proposed approach with the improved deep learning hash retrieval methods on several benchmark datasets.Through a series of experiments, the effectiveness and feasibility of the proposed algorithm are verified.CIFAR-10 dataset contains 10 object categories and each class consists of 6,000 images, resulting in a total of 60,000 images.The dataset is split into training and test sets which are averagely divided into 10 object classes.The Caltech 256 image library dataset contains 29,780 color images which are grouped intro 256 classes.
First, test the CIFAR-10 image dataset.There are a total of 50,000 training samples which are used for training on the deep learning network.The remaining 10,000 images are used as test samples.And then we randomly select 50 images from database as the query images.For the Hidden Image Retrieval algorithm based on deep learning mentioned in this paper, the image pixel data is directly used as input, while for other algorithms, the 512-dimensional GIST feature is used as the feature expression of the image.Note quantization all images into 32 * 32 sizes before experiment.
For the Caltech 256 image, a total of 256 classes are included, and each class contains at least 70 images.Therefore, 70 images of each class, a total of 17,920 images, are randomly selected and are used as training images.The remaining images are used as test samples.In addition, all of the images' size is set to 64 * 64 again when training.

Evaluation Metrics.
We measure the performance of compared methods using Precision-recall and Average-Retrieval Precision (ARP) curves.Precision is the ratio of the correct number of images m in the search result to the number k of all returned images.The formula is as follows: Recall is the ratio of the correct number of images m in the search results to the number g of images in the image library.The formula is as follows: Assume that the search result of the query image  is   and   means that the category is the same between the query image and the return image, then the accuracy rate for the image query result () can be defined by the following formula: Average-Retrieval Precision (ARP): the average value of all the images in the same class as the retrieval rate obtained from the retrieval image is defined as follows: Here   is the category index number of the image,  is the category index,  is the number of images whose category is   , and () is the category index number of the query image.

Performance Analysis.
In the proposed algorithm, IDLH, the length of the hash sequence and the depth of the hidden layer in the deep learning network are two key parameters.When the hash sequence length is small, different feature vectors can easily be mapped into the same hash sequence, so the retrieval accuracy is low.However, if the hash sequence is too long, a large storage space is required and a long time is consumed, which reduces the real-time performance.For the number of hidden layers, the number of layers in the hidden layer is too small, which is not conducive to learning strong image features.However, if the depth of the hidden layer is too large, the difficulty of training is increased.In order to verify the effectiveness and feasibility of our algorithm, we conducted the following experiments.
(1) Results on CIFAR-10 Dataset. Figure 4(a) shows the search results of the Average-Retrieval Precision using our proposed algorithm, IDLH, compared with the LSH algorithm [3] and other three deep learning algorithms, the DH algorithm [21], the DeepBit algorithm [40], and the UH-BDNN algorithm [41] on CIFAR-10 dataset with 8, 16, 32, 48, 64, 96, 128, and 256 bits.Figure 4(b) shows the Precision-recall curve under 48-bit encoding.It can be seen that the algorithm has a higher precision than the other hashing algorithms with the same recall rate.However, the advantage is not obvious, and the average accuracy is slightly higher than other hash algorithms.
In order to overcome the above defects, we use deep learning to perform hash mapping on image features, perform hash encoding of different bits on the same feature, calculate the Precision-recall of the search results under the condition of different coded bits, and determine the impact of the encoding length on the retrieval results.
As Figure 5 shows, with the increase in the number of coded bits, the Precision-recall is continuously increasing.With the increase in the number of coded bits, the image is better expressed.However, after the number of coded bits reaches 64, even if the number of coded bits increases, the average accuracy rate increases relatively slowly.Because the information of the tiny image is relatively simple, when the number of encoded bits reaches 64 bits, a relatively good image expression has been obtained, and the performance of the algorithm has basically stabilized.At this time, despite increase in the number of encoding bits, it is not very helpful to improve the accuracy rate.
In addition, we want to test the influence of the number of hidden layers in the deep learning network on the retrieval result by changing the number of hidden layers.
Figure 6 shows the effect of deep learning networks on experimental results in the case of different hidden layer numbers.
It can be seen that deeper networks do not have much improvement in performance, which is different from the expectation that more hidden layers will help learn stronger image features.Since the image library data used in the experiment is a tiny image library, relatively good image characteristics can be learned using a deep learning network with fewer layers.However, if the image library is replaced with a more colorful image, the deep neural network can acquire more detailed image features, and the deepening of the learning network will greatly help the study of image features.
(2) Results on Caltech 256 Image Data Set. Figure 7(a) shows the results of the Average-Retrieval Precision results when the number of coded bits is different.Compared with the black and white image library, the proposed algorithm embodies the advantage of image feature learning and leads the Average-Retrieval Precision to other hash retrieval algorithms.In Figure 7(b), we can also see that the algorithm proposed in this paper has a higher Precision-recall than other algorithms under the same recall rate, and it has better search performance.
As shown in Figure 8, as the number of coded bits increases, the precision rate increases with the same recall rate.This feature of the color image library is more pronounced than the black and white image library.Because the color image contains more information, more coding is needed to express it, and the increase of encoding helps to learn the features of the image.The experimental results also show that the deep learning network has learned more excellent image features.
In Figure 9, the precision rate is significantly improved by the increase in the number of hidden layers in the    color image library Caltech 256.This is because the information contained in a more colorful image is more complex.Adding a hidden layer can learn more details of the image and help improve the accuracy of the search.
Next, we tested the performance of secondary image retrieval.The value of k in the secondary search is 20, and the experimental results are shown in Figure 10.As can be seen from the results, secondary retrieval can effectively improve the retrieval accuracy when the number of coded bits is small.However, with the increase in the number of encoding bits, the results of the secondary search and the accuracy of the primary search are not much different.This is because the shorter the hash sequence is, the easier the feature vectors with different original features are mapped to the same hash code.In order to make up for the errors    caused by the short hash code, it is necessary to perform secondary search.When the number of encoding bits is small, a secondary retrieval method is used in IDLH, and the search accuracy rate can be improved at the expense of a small search speed.

Conclusion
With the rapid development of data storage and digital process, more and more digital information is transformed and transmitted over the Internet day by day, which brings people a series of security problems as well as convenience [42].The researches on digital image security, that is, image encryption, image data hiding, and image authentication, become more important than ever [38,39].The most essential problem of  image recognition is to extract robust features.The quality of feature extraction is directly related to the effect of recognition, so most of the previous work on image recognition is spent on artificial design features [43].In recent years, the emergence of deep learning technology has changed the status of artificial design classification characteristics.Deep learning technology simulates the mechanism of human visual system information classification processing, from the most primitive image pixels to lower edge features, then to the target components that are combined on the edge, and finally to the whole target, depth learning can be combined by layer by layer.The high-level feature is the combination of low level features.From low level to high level, features are more and more abstract and show semantics more and more.
From the underlying features to the combination of high-level features, it is the depth of learning that is done by itself.It does not require manual intervention.Compared with the characteristics of the artificial design, this combination of features can be closer to the semantic expression.
In terms of illegal image retrieval, the traditional recognition method should establish a recognition model for each type of recognition task.In the actual application, a recognition model needs a recognition server.If there are many identification tasks, the cost is too high.We used the deep neural network to recognize the illegal image, it only needs to collect the samples of every kind of illegal image and participate in the training of the deep neural network.Finally, a multiclassification recognition model is trained.When classifying unknown samples, deep neural network accounting calculates the probability that the image belongs to each class.
We all know that, in the image detection process, the accuracy and recall rate are mutually influential.Ideally, both must be high, but in general, the accuracy is high and the recall rate is low; the recall rate is high and the accuracy is low.For image retrieval, we need to improve the accuracy under the condition of guaranteeing the recall rate.For image disease surveillance and anti-illegal images, we need to enhance the recall under the condition of ensuring accuracy.Therefore, in different application scenarios, in order to achieve a balance between accuracy and recall, perhaps some game theory (such as Nash Equilibrium [44,45]) and penalty function [46][47][48] can provide related optimization solutions.
In this paper, we proposed an improved deep-learninghashing approach, IDLH, which optimized over two major image retrieval process.
(a) In the feature extraction process, the self-encoded network of the look-ahead type is trained by using unlabeled image data, and the expression of robust image features is learned.This unlabeled learning method does not require image library labeling and reduces the requirements for the image library.At the same time, it also takes advantage of the deep learning network's strong learning ability and obtains better image feature expression than ordinary algorithms.
(b) On the index structure, a secondary search is proposed, which further increases the accuracy of the search, at the expense of very little retrieval time.
Through experiments, the algorithm proposed in this paper is compared with other classic hashing algorithms on multiple evaluation indicators.Firstly, we tested the learning networks of different code lengths and depths in order to test their effect on the retrieval system and then tested the performance of the secondary search.Through the abovementioned series of experiments for different parameters, the effectiveness of the improved deep learning hash retrieval algorithm proposed in this paper is verified, and through the experimental data, the good retrieval results are proved.In addition, the proposed deep hashing training strategy can also be potentially applied to other hashing problems involving data similarity computation.

Figure 1 :
Figure 1: Deep learning-hash retrieval framework.IDLH consists of three main components (preprocessing layer, deep neural network layer, and hash layer).The object of the first layer is simply scaled, sample-by-sample mean-value reduction, and whitening.In the second component, we develop a deep neural network to obtain the feature expression function of the image.And the classifier weights and feature binary codes are simultaneously learned in the last component-hash layer.

4. 1 .
Database.Two mostly used databases in the recent deep learning hash works are taken into our evaluation.The two image libraries are derived from the CIFAR-10[11] core experimental image library dataset and the Caltech 256 image library dataset.
Precision-recall at 48 bits

Figure 4 :
Figure 4: Five kinds of algorithm retrieval performance comparison on CIFAR-10 dataset.

Figure 6 :
Figure 6: Precision-recall curves with different code different hidden layers.
Precision-recall at 48 bits

Figure 7 :
Figure 7: Five kinds of algorithm retrieval performance comparison on Caltech 256 set.

Figure 10 :
Figure 10: Average-Retrieval Precision with first and second retrieval.

Table 1 :
A huge number of parameters need to be learned and stored with different datasets.Considering a classification task with C different categories and D-dimensional feature representation, even the simplest linear models are comprised of D×C parameters.

Table 2 :
Summary of notations.

Table 3 :
Image library image storage structure.