Special Object Recognition Based on Sparse Representation in Multisource Data Fusion Samples

. Wireless sensor networks (WSNs) suﬀer from limited power and large amounts of redundant data. This paper describes a multisource data fusion method for WSNs that can be combined with the characteristics of a proﬁle detection system. First, principal component analysis is used to extract sample features and eliminate redundant information. Feature samples from diﬀerent sources are then fused using a method of superposition to reduce the amount of data transmitted by the network. Finally, a mathematical model is proposed. On the basis of this model, a novel method of special object recognition based on sparse representation is developed for multisource data fusion samples according to the distribution of nonzero coeﬃcients under an overcomplete dictionary. The experimental results from numerical simulations show that the proposed recognition method can eﬀectively identify special objects in the fusion samples, and the overall performance is better than that of traditional methods.


Introduction
In general, objects moving across borders or uninhabited regions could be either humans or animals. As there are significant differences between human and animal profiles, it is feasible to use certain features to recognize and monitor special objects (humans) [1][2][3]. However, previous studies are largely concerned with the acquisition of object profile samples and recognition using traditional methods such as k-nearest neighbors (KNN) and support vector machines (SVMs) [4][5][6]. ere has been no detailed discussion on the fusion of data or reducing the amount of data transmitted over the network. In this paper, we discuss these issues in detail based on previous work [7]. e main contribution of this paper is to construct a wireless monitoring network with a profile detection system as a network node. To reduce the amount of data transmitted in the network as much as possible, the network uses a sink node to fuse the feature samples sent by each network node; the fused data are then sent to the terminal for identification, allowing the determination of whether a special object (human) is passing through the monitoring area. is paper describes the fusion of multisource feature samples through a superposition approach, meaning that traditional recognition methods such as KNN, SVM, and sparse representation classification (SRC) [8][9][10] cannot accurately identify whether there is one or more class objects in the fused sample. To solve this problem, we propose a novel object recognition method based on sparse representation. Different from classical SRC, the method described in this paper uses the distribution of nonzero coefficients in the sparse representation to identify special objects in the fusion sample. e experimental results verify the effectiveness of this method.

Related Work
In our previous work, we designed the profile detection system shown in Figure 1. In the system, the signal sensing unit is composed of photoelectric sensors installed at even intervals on a vertical fixed bracket with a length of 2 m. When a moving object passes through the sensor's field of view, the profile feature information of the object is captured by the sensing unit [1,2,7]. is paper mainly studies the profile classification of three kinds of moving objects: humans, squatting humans, and animals (namely, dogs). In practice, when objects enter the sensor's field of view, their position, speed, attitude, and other conditions are different. To obtain a better recognition effect, samples of various angles, speeds, and attitudes entering the field of view are collected for training, as shown in

Sample Feature Extraction and Data Fusion
3.1. Feature Extraction. According to Zha et al. [7], the profile detection system of the network monitoring node in the wireless monitoring network expresses an object profile sample P as an S × L matrix. e number of columns of each sample matrix is different, and there is a large amount of redundant information. Sending the sample matrix signal directly through the network wastes network resources and is inconvenient for data processing because of the different dimensions of the matrix formed by P. To solve these problems, we use principal component analysis (PCA) [11][12][13] to pretreat the sample data. e specific process is as follows: (1) Input: sample matrix P ∈ R S×L .
(2) Calculate the covariance: (3) Determine eigenvalues and eigenvectors: namely, the eigenvalues Λ � λ 1 λ 2 · · · λ L and the corresponding eigenvectors U � μ 1 μ 2 . . . μ L . (4) Construct transformation matrix: select the eigenvectors corresponding to the k(k ≤ L) largest eigenvalues to construct the transformation matrix, namely, (5) Dimension reduction: (6) Output: matrix vectorization: After pretreatment, the vector y is taken as the feature sample of the object and sent to the sink node for data fusion.

Multisource Data Fusion Model.
Suppose that the wireless sensor network (WSN) contains N network monitoring nodes and that the jth network node acquires the feature sample of the object as y j ∈ R c (1 ≤ j ≤ N). e sink node receives the feature sample of each network node and performs data fusion. e network topology is shown in Figure 3. To reduce the computational complexity of the data fusion process, we adopt the method of superposition to achieve fusion. us, the fusion sample can be expressed as follows: Considering the signal transmission power, channel fading, noise interference, and other factors associated with each network node, equation (6) can be rewritten as follows: where r ∈ R c and the parameters ρ j and α j (1 ≤ j ≤ N) are the signal transmission power and channel link gain of the jth network node, respectively and n is the additive white Gaussian noise. When the dimension of the feature sample of each network node is large and the amount of data to be transmitted needs to be further reduced, the feature sample can be projected into a low-dimensional space through the projection matrix Ψ ∈ R d×c (d < c). e signal model is shown in Figure 4. e mathematical model of the fusion samples can be written as follows:

Sparsity Analysis and Construction of Overcomplete Dictionary
To analyze the sparsity of the fusion samples, it is assumed that r is composed of three different feature samples (i.e., y 1 , y 2 , and y 3 ), namely, Furthermore, suppose that x i (1 ≤ i ≤ 3) are sparse representation coefficients of eigenvector y i under the overcomplete dictionary A i . en, Combined with equation (10), equation (9) can be rewritten as follows: where It can be seen from equation (11) that the fusion sample r can be sparsely represented under the dictionary A a , and the sparse coefficients can be obtained by solving the following l 1 -minimization problem [14][15][16]: where the parameter ε is the error tolerance. Similarly, when the fusion sample r is composed of k feature samples, it can be sparsely represented, and the overcomplete dictionary is To represent the fusion sample sparsely, we use the training samples to construct an overcomplete dictionary directly [7,[17][18][19]. Assuming that there are T classes of training samples, the number of training samples in each class is N 1 , N 2 , . . . , N T , and the ith (1 ≤ i ≤ N j ) training sample in the jth (1 ≤ j ≤ T) class is expressed as P j,i ∈ R S×L i , and the specific process of constructing the dictionary is as follows: (1) Input: training samples P j,i ∈ R S×L i .
(2) Pretreatment of training samples: All training samples P j,i ∈ R S×L i are pretreated, and the feature vector φ j,i ∈ R c (c � S × k) of the pretreated sample is used as a dictionary atom. (3) Output: construct overcomplete dictionary: where, n � N 1 + N 2 + · · · + N T . rough the above process, we can obtain an overcomplete dictionary.

Special Object Recognition Method
According to the above sparsity analysis, when there is one class of feature samples in the fusion sample, the main nonzero coefficients in the sparse coefficient vector obtained by l 1 -minimization are distributed on the corresponding class of atoms, whereas the coefficients for other classes of atoms are zero or very small. If the fusion sample contains multiple classes, the main nonzero coefficients in the sparse coefficient vector are distributed on these classes. Based on this feature, we propose a special object recognition method for multisource data fusion samples. e method is illustrated in Figure 5, and the specific steps are as follows: (1) Input: dictionary A a � [A 1 , A 2 , . . . , A T ] ∈ R c×n for T classes and fusion sample r ∈ R c . (2) Sparse representation: where ε is the error tolerance. (3) Calculate coefficient l 1 -norm of each class: where δ j is the characteristic function that selects the coefficients associated with the jth class. (4) Multiclass discriminant rule: where τ ∈ [0, 1]. (5) Output: based on the results of step 4, check whether the special object is included.

Experimental Simulation and Results Analysis
Experiments are conducted based on the profile detection system, which uses 16 E3F-R2NK photoelectric sensors to construct a signal sensing unit. e effective distance of the sensors is 2 m. In the actual environment, the profile detection system mainly collects three kinds of object profile samples for training and testing: humans, squatting humans, and animals (dogs). e numbers of training samples and test samples are listed in Table 1. It is assumed that each network node can detect at most one object in a certain period of time. In the following experiments, the fusion samples are composed of humans and dogs, and we consider the humans as the special objects. Under different experimental conditions, the proposed method can judge whether there is a special object (human) in the fusion sample and considers this a correct recognition.
For the convenience of analysis and comparison, it is assumed that the signal transmission power of each network node is the same, and the channel link gain remains constant at 1. e performance of the proposed method is compared with that of the traditional nearest-neighbor classifier (NN) [20] and the KNN, SVM, and SRC methods. e experiments are conducted on a Windows 7 PC running MATLAB (R2014a) with a 3.0 GHz CPU and 8 GB memory.

Experiment 1: Verifying the Sparsity of Fusion Samples.
It is assumed that, within a certain period of time, two network nodes in the monitoring area detect objects passing by, among which one node is a single person and one node is a dog. After each node extracts the sample feature information, the sink node performs data fusion. e sparsity of the fused feature sample is analyzed, and the distribution of sparse representation coefficients is shown in Figure 6.
As can be seen from Figure 6, the coefficients of the fusion samples are still sparse and distributed on the atomic terms of the two classes of human and animal. e fusion samples do not contain images that include squatting humans, so the coefficients for the linear representation of a squatting human are close to zero. ese experimental results show that the following: (1) multisource data fusion samples can be sparsely represented under the overcomplete dictionary; (2) according to the distribution of the main nonzero coefficients in the coefficient vector, the combination of different classes in the fusion sample can be distinguished, and the special object (human) can be identified.

Relationship between the Number of Animals and Recognition Rate.
Assuming that there is one human, scenarios including 1-5 animals are considered (see Table 2). e experimental simulation results are shown in Figure 7.
As can be seen from Figure 7, the correct recognition rates achieved by the traditional NN, KNN, SVM, and SRC are relatively low. As the number of animals increases, the traditional methods struggle to recognize the special object correctly. e proposed method outperforms these traditional methods and exhibits better anti-interference ability.

Relationship between Number of Humans and Recognition Rate.
For the case of only one animal, experiments with 1-5 humans are considered (see Table 3). e experimental simulation results are shown in Figure 8.
From Figure 8, we can see that an increase in the number of humans enhances the recognition rate of various methods. However, the recognition rate of the method proposed in this paper is consistently 100%, which is obviously superior to that of traditional methods.

Experiment 3: Verification of Relationship between SNR and Recognition
Rate. For WSNs, reducing the transmission power of the signal can effectively extend the network lifetime. In this experiment, the number of humans and animals is known, and we analyze the impact of different SNRs on the recognition performance. e experimental condition settings are listed in Table 4, and the simulation results are shown in Figure 9.
It can be seen from Figure 9 that changes in the SNR have little effect on the recognition rates of the various methods.

Experiment 4: Verification of Relationship between
Number of Animals and False Alarm Rate. In border regions or special monitoring areas, there may be long periods      without humans passing through, during which only animals will be detected. erefore, it is very important to correctly judge that no humans are present, thus reducing the false alarm rate of the system. e experimental conditions are listed in Table 5, and the simulation results are shown in Figure 10.
As can be seen from Figure 10, when the SNR is 15 dB or 20 dB, the false alarm rate of the proposed method is slightly higher than that of traditional methods. However, as the number of animals increases, the animal class information in the fusion sample increases, which reduces the possibility of false alarms. e experimental results show that an increase in the number of animals will not increase the false alarm rate, but will actually help to reduce the false alarm rate.
In summary, the traditional recognition methods mainly focus on the test samples of a single class. When the samples are composed of multiple classes, the other classes in the fusion sample produce interference that affects their correct recognition performance. e proposed method is based on sparse representation theory. When the samples are composed of multiple classes, we can effectively separate the fusion samples in the sparse domain. According to the coefficient distribution, we can then judge the class combination in the fusion sample. e experimental results verify the validity of this method from many aspects of recognition tasks.

Conclusions
According to the characteristics of the profile detection system, combined with WSN and sparse representation theory, we have proposed a method of multisource data processing and an associated mathematical model. Based on this model, a novel special object recognition method has been developed. Compared with traditional methods, the method described in this paper achieves better performance in many aspects of recognition. Moreover, in practical applications, the system parameters (such as the transmission power of the network nodes) can be adjusted to satisfy the actual requirements.

Data Availability
e data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e authors declare that there are no conflicts of interest regarding the publication of this paper.