Big Data Fusion Method Based on Internet of Things Collection

With the development of information collection technology, the data that people need to deal with are also increasing, which brings problems such as isomerization of data types, poor data quality, and fast data generation speed. At present, as an important method of data fusion technology, data fusion method based on deep learning has become an effective way of data fusion under the background of big data, which has important research significance. There is a problem with heterogeneous data types between time series data and text data, and it is difficult to fuse them effectively by traditional data fusion methods. In order to make full use of the information contained in text data and improve the accuracy of time series prediction, this paper proposes a data fusion model based on FC-SAE. In this model, GloVe and CNN are used to extract the features of text data, FC neural network is used to extract the potential features of time series data, and then, the SEA model is used to fuse the data, which fully discovers the relationship between data and greatly improves the prediction accuracy.


Introduction
Now, the ecological environment has been seriously affected, and the integration of ecological data is conducive to solving ecological and environmental problems. In this paper, a rotating forest algorithm is proposed to solve the problem of data fusion [1]. e model proposed in this paper solves the problem of data fusion well. Using the nonlinear mapping ability of neural network, the parallel processing of largescale data is realized, and the fusion of multisource heterogeneous data is completed [2]. e effectiveness and usability of the convergence phase are of great significance for developers to improve the convergence [3]. is paper proposes a deep learning data fusion model to solve the traffic jam problem. e applicability of the model in traffic problems is obtained [4]. is paper proposes a data fusion method to improve the level of intelligent retrieval. e research shows that this method has obvious advantages in information recall and recall rate [5]. is paper studies how big data fusion mode method can effectively use big data to evaluate the relationship between them, and the effective evaluation rate is as high as 80% [6]. Deep learning is an advanced abstraction of modeling data, which is a branch of machine learning [7]. Deep learning uses a backpropagation algorithm to command machines to change their internal entries, thus discovering complex structures in large data sets [8]. is paper shows how to solve the research problems by training the deep network of learning features and how to share the representation learning patterns and evaluate them on a specific task [9]. In this paper, through a series of experimental studies, it is proved that deep learning can solve the problem of face recognition well. Face recognition tasks increase the differences between people by extracting verification features from different identities, while face verification tasks reduce the differences between people by combining verification features extracted from the same identity, which are necessary for face recognition [10].
rough a large number of systematic experiments, this paper shows why traditional methods cannot explain the generalization ability of large neural networks [11]. is paper briefly describes the application research of deep learning in each field. e most advanced technology at present is summarized, and the future research direction is given [12]. Although deep learning is widely used, it cannot capture the uncertainty of the model. Although the Bayesian model can reason the uncertainty of the model, its computational cost is very high [13]. e specificity of DNA binding to protein can be determined by deep learning technology. Experiments show that this algorithm is more excellent than other algorithms [14]. In this paper, an algorithm is created by deep learning, which can be applied to the field of medical imaging [15]. When the deep learning model uses data collection, it can obtain the complex underlying structure of the data system, especially for large data sets and high-dimensional data sets. When the features are complex, the data set collects more information and can retain more feature values.

Weighted Average Method.
is method directly processes the original information collected by the sensing nodes. Its fusion idea is to get the weight coefficient of each sensing node according to certain algorithm rules and then weigh the redundant data of each sensing node according to the weight, and the result is to calculate the weighted average value of the data collected by sensing nodes, which makes the data more accurate and concise. e advantages of this fusion method are easy and intuitive, but the main disadvantages are as follows: there is no deletion of redundant information, which will lead to an increase in computing energy consumption.

Bayesian
Reasoning. Bayesian estimation is a common uncertainty reasoning method, and it is also a statistical fusion method. It usually uses the detection area and prior knowledge to determine the target in the detection area. It can fuse nodes and information, and then get the hypothetical posterior probability.

Kalman Filtering Method.
Most of these fusion methods appear in dynamic topology, which can fuse perceptual data and remove redundant data in real time. is method uses the linear combination to get the final fusion result, which is an estimate with low accuracy. When necessary, it also needs to introduce the concept of adaptive.

Neural Network Method.
Neural network structure is a network model of general fusion methods, and fuzzy reasoning is generally adopted as the main idea of data fusion. It firstly establishes a fuzzy reasoning mechanism based on the sample data of the sensing nodes and then fuses the data, so it has good fault tolerance and robustness.

Fuzzy Logic Reasoning.
is method can deal with uncertain information reasonably. In addition, the information processing method of this method is close to human thinking, so it often appears in higher-level fusion operations. It also has many disadvantages, such as subjective description of data.
Weighted average method: easy to understand, good real time, suitable for changeable scenes. Kalman filtering method: It is suitable for linear system, so that the system does not need a lot of data storage and calculation, and can be processed in real time. Bayesian inference method: concise, medium computation, and good effect of related events. Template method: good effect in non-real-time environment. Fuzzy theory: It can solve the problem of information or decision conflict and realize the information fusion between subjective and objective.

Fusion Model Framework Based on FC-SAE
FC-SAE model mainly includes text feature extraction, time series modeling, and heterogeneous data feature fusion. Text feature extraction module mainly preprocesses text data, mines semantic relations through the GloVe word embedding algorithm, and extracts features at different abstract levels by convolution neural network; neural network is used to model time series. Heterogeneous data feature fusion module uses the encoder of sparse automatic encoder to fuse text features and time series features and extract the fused features.

Text Feature Extraction.
When the text data are converted into text sequence, it enters the word embedding layer. e essence of word embedding is to make words and vectors correspond one by one. To get the relationship in embedding space, we only need to calculate the distance between vectors. In FC-SAE model, GloVe word embedding model is used for word embedding operation. e basic idea of this model is to decompose the global vocabulary matrix and train the word vectorization model with window. e training process of GloVe word embedding model mainly includes constructing word co-occurrence matrix and training word vector.
GloVe model constructs a word co-occurrence matrix X. Suppose the corpus is "sheisthemostlovelyperson," and there are six words in the corpus: she, is, the, most, lovely, person. e window information is shown in Table 1. Take window 3 as an example, there are the following formulas: X the,most � X the,most + 1.
e co-occurrence matrix X can be obtained according to the above method. According to the co-occurrence matrix, the probability of occurrence of the word J in the context of the word I can be obtained, and the calculation is as shown in formula (3) 2 Computational Intelligence and Neuroscience In the formula, X i � k X ik . e relation of the words I, J, and K can be expressed as When the value of radio i,j,k is small, the correlation between J and K is high, while the correlation between I and K is low. When the value of radio i,j,k tends to 1, while the word I is related to the word K and the word J is also related to the word K, or while the word I is not related to the word K and the word J is not related to the word K. erefore, radio i,j,k can better measure the correlation between words.
e purpose of GloVe model is to construct a function F(v i , v j , v k ) by using word vectors v i , v j , and v k , so that the value of this function is as close as possible to (P ik /P jk ). e model loss function is shown in formula (5) where N is the number of words, and the co-occurrence matrix X is the N × N matrix; X ij denotes the number of times the word I and the word J appear in the same window; v T i denotes the transposition of the word vector corresponding to the word I; v j denotes the word vector corresponding to the word J when the word J is the central word of the context; b i and b j are bias terms; and F is a custom weight function. e common form of weight function is given by where x max � 100, α � 0.75, and word vector and word vector model can be obtained by minimizing loss function. We use convolution neural network to further extract the features of text data. e word vector matrix X[i : j] obtains a feature map C by convolution operation through a convolution layer and obtains a maximum value C in the feature map through a maximum pooling layer, which represents the most important feature. e output feature Z text of the word vector matrix is obtained under the action of a plurality of convolution kernels, and the process is shown in formulas (7)- (10): where n is the number of feature mappings and m is the number of convolution kernels. is model can obtain the variable length mode of input text with arbitrary length.

Time Series Modeling Based on Full Connection Layer.
Deep learning methods commonly used in time series prediction include RNN, LSTM network, and fully connected network. LSTM is the most commonly used deep learning algorithm for time series prediction because it can well obtain the medium-and long-term dependence of time series data. Detrending data can largely eliminate the need for long-term dependencies, which makes LSTM inferior to FC. erefore, this paper uses FC to model time series. e calculation steps of the fully connected network output are as follows: First calculate the weighted sum, as shown in the following formula: In the above formula, W ab represents the connection weight of nodes a to b, θ b is the offset term of the b hidden node, and X a is input. e output of each node is calculated according to formulas (12)- (14): where W bk is the connection weight between the b-th and kth output nodes, and θ k is the offset term of the k-th output node.
In FC-SAE data fusion model, the form of delayed observation is used as the fixed input vector, as shown in following formula: where Z ts is the time series feature, and T � (t 1 , t 2 , . . . t L ) and L are the lag observation lengths.

Feature Fusion of Heterogeneous Data.
At present, in the research of data fusion based on deep learning, most researchers use feature mosaic and full connection layer to fuse the extracted features. Although this fusion method is very simple, it cannot fully mine the relationship between data. FC-SAE model takes the sparse automatic encoder as the fusion module, which can fully mine the relationship between time series data and text data while retaining the maximum amount of information, so as to obtain the shared feature representation of data. e sparse automatic encoder is a deep learning model without supervision and management. In FC-SAE model, firstly, data features and sequence features are combined to obtain x, and then, the combined features x are input into SAE fusion model, output through hidden layer and nonlinear mapping, represented by y, and finally mapped back to x′. e above is the encoding and decoding process, which is represented by In the above formula, H represents the activation function of the encoder, H represents the activation function of the decoder, W represents the weight matrix of the encoder, W represents the weight matrix of the decoder, B represents the bias term of the encoder, and B represents the bias term of the decoder. e loss function can be expressed by formulas (18)-(20): where L is the number of hidden layer nodes, n is the number of data samples, and k is the input vector dimension. λ and β are given coefficients, which control the weighted coefficient regular term and the sparse regular term, respectively. ρ i is the average activation value of neurons in hidden layer, and ρ is the sparsity parameter.

Data Sets and Data
Preprocessing. e selection of time series data in this experiment is consistent with that in the previous chapter. Meteorological data are a public data set of the National Oceanic and Atmospheric Administration of the United States, which has 13 characteristics. e event text data are obtained from Barclays Center official website. e data set contains 751 event description text data, each of which includes title, time, and event description. e event presence data are generated from the event text data, and the value is 1 on the date when an event occurs and 0 on the date when no event occurs. Some data of the above experimental data set are shown in Tables 2-4. When dealing with taxi data, the original data are added first to get daily travel data. In order to reduce the impact on the data in the later period, the data are processed by trend, which makes the model pay more attention to the fluctuation of the data itself and improves the prediction effect. e preprocessing operation of meteorological data is to normalize the data. e preprocessing of text data is conventional text preprocessing, including HTML tag deletion, lowercase conversion, root, deletion of stop words, and prepositions.

Experimental
Design. FC-SAE model uses GloVe word embedding model and CNN neural network to extract features from text data in which CNN includes three convolution layers and three pooling layers, and ReLU function is used as activation function; a two-layer FC neural network with 100 neurons and 50 neurons is used to model the time series. Tanh function is used as activation function, and the input data are standardized before FC layer input. SAE is used as the fusion model, and Adam is used to optimize the model parameters.
is paper compares the nonlinear time series forecasting model with FC-SAE model and concludes that FC-SAE model has more advantages in forecasting. In addition, FC-SAE model is replaced by LSTM-SAE model, and its effectiveness is verified. en, the contribution of information from different sources to data fusion is evaluated.
is paper also analyzes the incremental experiments of FC-SAE model and LSTM-SAE model. Finally, in order to compare the impact of different fusion strategies on the prediction results, under the condition of complete model, the fusion model based on full connection layer is compared with the fusion model based on the sparse automatic encoder.

Analysis of Experimental Results.
In this paper, root mean squared error (RMSE), mean absolute error (MAE), and goodness-of-fit R2 are selected to evaluate the proposed model, as shown in formulas (21)-(23): where N represents the amount of data in the test set, y n represents the true value of the nth instance, and y n represents the predicted value of the nth instance. Both MAE and RMSE can reflect the deviation degree of the two values. e smaller the two values, the better the prediction effect, and RMSE is the most sensitive to outlier data. e fitting degree of the model can be measured by R, and the larger the value, the better the fitting effect of the model.

Comparison with Other Time Series Forecasting
Methods. Firstly, this paper compares the differences in MAE, RMSE, and R2 indexes among FC-SAE model and SVR model, GP model, LSTM neural network, and LSTM-SAE model, as shown in Table 3. e results show that LSTM is not the best choice for the time series prediction in this paper. And by adjusting the parameters of LSTM-SAE model, it cannot achieve the performance of FC-SAE model. e results also verify the conclusions of early studies on LSTM. Compared with SVR algorithm, GP algorithm, and LSTM algorithm, FC-SAE model has improved by 1.3%, 5.3%, and 6.6% in MAE index; 1.8%, 5.4%, and 4% in RMSE index; and 1.8%, 6.6%, and 7.7% in R2 index, respectively, which has obvious advantages. From the experimental results in Table 3, it can be seen that FC-SAE model and SVR model are better methods in this research field, and the time series prediction method based on data fusion has obvious advantages over the traditional time series prediction method, as shown in Figure 1 and Table 5.

Comparative Experiment of Multisource Heterogeneous Data Fusion.
In order to further analyze the influence of each data on the fusion effect, the incremental analysis method is selected for the experiment. Firstly, only time series is modeled, and then, meteorological data, event existence data, and event text description data are added in turn. e results are shown in Figures 2-4.
Looking at Figures 2-4, we can see that the best way to improve the prediction results of time series is to introduce meteorological data. From an empirical point of view, bad weather conditions will change people's travel mode, which is consistent with our experimental results. e event presence data (E) have the greatest influence on the prediction results. After introducing the event presence data, the MAE index and RMSE index of FC-SAE algorithm are reduced by 17.1% and 12.6%, respectively. e MEA algorithm decreased by 17%, and the RMSE algorithm decreased by 15.3%. Although the introduction of event existence data has significantly improved the prediction effect of FC-SAE algorithm and LSTM-SAE algorithm, it is impossible to judge the real impact of events on prediction results without further analysis of the information contained in event description. In this paper, the text description of events is introduced into time series prediction through GloVe word embedding model and convolution neural network. Experimental results show that the MAE index of FC-SAE algorithm decreases from 99.3 to 95.9, RMSE index also decreases from 140.3 to 135.8, and R2 increases from 0.599 to 0.626 after combining event text description data (TS) with other data. In the same way, the MAE and RMSE of LSTM-SAE     During the experiment, it is found that the advantage of using SAE model for fusion is that it can fully mine the correlation between time series data and text data, while retaining the maximum amount of information. e experimental results verify that SAE fusion model performs better in heterogeneous data fusion than full connection layer, as shown in Figure 5 and Table 6.
is chapter takes common time series data and text data as data sources, establishes multisource heterogeneous data fusion model based on FC-SAE, studies the fusion of time series data and text data, and discusses the influence of event text data on time series prediction. By comparing FC-SAE data fusion model with other popular time series prediction methods, it is proved that FC-SAE data fusion model has better prediction accuracy; the comparative experiment of multisource heterogeneous data fusion proves that the error of time series prediction results can be effectively reduced by multisource heterogeneous data fusion, and different data have different influences on prediction results. In addition, the comparison experiments of fusion methods show the Computational Intelligence and Neuroscience influence of different fusion strategies on the prediction results, and the results show that SAE fusion model has better fusion performance than FC fusion model.

Conclusion
is paper presents a multisource heterogeneous data fusion model based on FC-SAE. is paper presents a data fusion model based on FC-SAE to solve the data fusion processing of time series data and text data. In this model, FC neural network is used to extract the features of time series, GloVe and CNN are used to extract the features of text data, and finally, SAE model is used to fuse the features of data. is model solves the problem that data fusion is difficult caused by data type differences. By analyzing the experimental results, we find that FC-SAE model can make the time series achieve the best accuracy in heterogeneous data fusion. Moreover, through the analysis of incremental experiments, we can know the influence of different data on the fusion results in data fusion. Finally, by comparing different fusion strategies, it is proved that the fusion effect of SAE model is better than that of full connection layer fusion.

Data Availability
e experimental data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e authors declare that they have no conflicts of interest regarding this work.