Situation Element Extraction Based on Fuzzy Rough Set and Combination Classifier

Generalized network security situation awareness technology is divided into three processes: situation element extraction, situation understanding, and situation prediction. Situation element extraction is the most critical step in the whole process, and its extraction quality will directly affect the accuracy of situation understanding and prediction. In view of the shortcomings of current situation element extraction methods, this study makes an in-depth study on the network security situation element extraction algorithm and proposes a situation element extraction model based on the fuzzy rough set and combined classifier, which is used to improve the accuracy of situation elements acquisition, so as to provide a better data basis for situation understanding and prediction. In this study, the theory of fuzzy rough set is used to reduce the attributes of data without reducing the ability of data classification, which reduces the complexity of data; using the combination classifier theory and particle swarm optimization algorithm, a framework of situation element extraction is built, which can extract situation elements more accurately. The experimental results show that the network security situation element extraction framework proposed in this study can effectively shorten the extraction time of situation elements and improve the accuracy of situation element acquisition under the premise of ensuring the ability of data classification, thus proving the effectiveness and feasibility of the situation element extraction framework proposed in this study.


Introduction
With the gradual expansion of the scope of the network, the amount of information generated on it is increasing. With the development of network, the security threats caused by the network are increasing and more complex. e network attacks against these security problems are also increasing and more complex. As a result, the network intrusion detection and defense system based on passive defense and other network security devices are tired of coping with such a large number of complex network attacks. Even though there are many passive security products on the network, they are like isolated islands, fighting on their own, and lacking necessary communication and cooperation [1] and cannot play a very good effect on complex, multi-step, and hidden attacks. Network security situation awareness technology can quickly and accurately identify the type of attack and predict the attack intention, so that network managers can respond in time to avoid greater losses. Because of the great advantages of network security situation awareness technology, it has been the focus of the majority of scholars since its birth. e extraction of network security situation elements is the basis of the perception process, and the accuracy of the whole perception process will be directly affected by its extraction accuracy.
Network security situational elements refer to elements or attributes that can reflect the current security state of the network; that is, the current network state can be inferred through the changes in these elements or attribute values, and situational element extraction is to use extraction technology to extract these situational elements from the multisource heterogeneous network security information generated by a large number of network security devices which is accurately extracted. e first step in network security situational awareness is to extract situational elements, and the quality of the extracted elements will directly affect the accuracy of situational understanding and prediction. Situational elements are extracted from information sources, and the extracted situational elements must meet the requirements of being accurate, comprehensive, easy to extract, and highly recognizable. According to the different sources of security data information, we can divide the network situation elements into three parts: basic network environment elements, security vulnerability elements, and attack information elements. Basic network environment elements refer to the elements that reflect the configuration of the network and various network access devices, network topology, etc.; security vulnerability elements refer to the elements that reflect the vulnerability of the network, that is, information that can reflect the inherent defects of the network or equipment. Element: attack information element refers to the element that reflects the type of attack the network is currently receiving. It can be seen from the above explanation that the basic network environment elements and security vulnerability elements are inherent information of the network itself. ey can only reflect the hidden dangers existing in the current network and possible attacks and cannot directly affect the change in network security status. Information elements can not only directly reflect the current network security situation but also reflect the network loopholes and hidden dangers. erefore, the extraction of attack information elements is the most important part of network situation element extraction. In the face of increasing network data, it is of far-reaching significance to study how to remove redundant information from it and extract comprehensive and accurate situation elements [2].
In the aspect of network security situation awareness, Endsley first put forward the concept of situation awareness in 1988 and established a three-layer awareness model [3]. In 1999, Bass introduced it into the field of network security and then established a JDL model based on data fusion [4]. Since then, the active network defense technology has developed rapidly in the field. Liu and others combined evidence theory with particle swarm optimization to build a hierarchical situation awareness model [5]. Leau et al. constructed a hierarchical evaluation mechanism based on entropy and a network situation awareness mechanism based on the Grey Verhulst-Kalman prediction [6]. Zhang et al. introduced the stochastic game theory into situation awareness and quantified the situation value through the game effectiveness of both sides in the cloud computing environment [7]. Chen et al. constructed a network situation awareness model using a variable step learning mechanism and simulated annealing method to optimize BP neural network [8]. Liu et al. used different weights to fuse data sources, quantified situation value with the hierarchical method, and realized automatic control of network security situation awareness by applying cognitive awareness for the first time [9]. Li and Duan optimized the hidden Markov model with the crowd search method and realized effective situation evaluation [10]. He and Zhu combined the GRU neural network and particle swarm optimization method to mine the time relationship of data, used the attention mechanism to measure the situation value of current network, and constructed a neural network situation awareness model with the attention mechanism [11]. Samuel effectively combines a decision tree and artificial neural network to build an enhanced network situation awareness model [12]. e authors have also done some in-depth research in the field of network security situational awareness and achieved a series of relevant results [13][14][15][16][17].
In the aspect of fuzzy rough set theory, in 1990, Dubois and Prade first proposed the concepts of rough fuzzy set and fuzzy rough set [18] and published a monograph on fuzzy rough set in 1992 [19]. Since then, fuzzy rough set theory has become the main attribute reduction method. In 2011, Chen et al. introduced the Gaussian kernel function into fuzzy rough set theory, described the similarity of fuzzy attributes with the Gaussian kernel, and constructed an attribute selection algorithm based on the difference matrix [20]. In 2019, Wang et al. added a distance measurement index in fuzzy rough set theory and used variable distance to measure the importance of attributes, so that the attribute reduction was completed with the variable distance measure function [21]. Senthilnayaki et al. introduced the maximum correlation algorithm to fuzzy rough set theory for attribute reduction and then used the improved k-nearest neighbor algorithm to extract situation elements [22], which effectively reduced the false alarm rate of attacks. In 2020, Ghoutkhar and Nehi have effectively integrated the traditional rough set theory and the regional value fuzzy set theory and put forward a general method of selecting fuzzy rough set attributes and a new development direction [23]. In the same year, Ni and others designed a dynamic attribute reduction algorithm, which added an incremental mechanism on the basis of fuzzy rough set theory, which can dynamically and quickly update the attributes based on the existing attribute selection results [24].
In terms of combined classifiers, Hansen's and Salamon's paper published in 1990 shows that different classifiers can provide complementary information for sample classification [25]; that is to say, using multiple classifiers for fusion classification can improve the performance of classification system, which lays a foundation for the development of combined classifiers. In 1998, Kittler and others deeply studied the combined classifier, gave the specific framework of the model for fusion classification, and defined two fusion methods based on the sum rule and product rule [26]. Since then, the theory of combined classifier has gradually developed and become a research hotspot in various fields. In 2002, Ma carried out in-depth research on combined classifiers and proposed two classifier set selection methods based on minimum dependency and genetic algorithm [27]. en, the results of classifier set generation were fused with a neural network to improve the accuracy of system classification. In 2003, Giacinto and others constructed a combined classifier fusion model to detect network intrusion [28]. In 2008, Wang and others proposed the hierarchical model of situation awareness and applied the 2 Computational Intelligence and Neuroscience combination classifier fusion method [29] at each level, which achieved good results in situation understanding and prediction. In 2016, Mi et al. applied the weighted vector to the fusion layer of the combined classifier and then used the k-nearest neighbor algorithm to adjust the misclassified samples [30], thus improving the classification performance. In 2019, Mo and Ma constructed a new value integral algorithm [31] based on the value integral algorithm of fuzzy sets and applied the algorithm to the fusion layer of combined classifiers, which greatly improved the recognition ability of combined classifiers. In 2020, TTN and others used the interval to represent each category of data, trained the combined classifier, and then used PSO algorithm to find the optimal parameters of fusion [32], which improved the accuracy of fusion recognition. In 2021, Wang et al. constructed a regularized combined classifier based on the random gradient descent algorithm and then determined the fusion weight [33] using the geometric structure of the data set to prevent the overfitting of the combined classifier. e author analyzes and summarizes the shortcomings of existing research methods. Aiming at some problems faced in the process of situation element extraction, this study focuses on the fuzzy rough set theory and combined classifier theory. In the process of attribute reduction, the fuzzy rough set theory is introduced and effectively improved, and some new measurement indexes and methods are added, and the reduced attributes are made more in line with the actual situation; in the process of extracting situation elements, the combined classifier theory is introduced to make up for the low accuracy of extracting elements by a single classifier. Finally, the improved compression factor particle swarm optimization method based on random vector is used to train the BP network as the fusion layer to make the extracted situation elements more accurate and provide a good data basis for situation understanding and prediction and improve the accuracy of situation understanding and prediction. It can be seen from the results that the situation element extraction framework designed in this study not only improves the accuracy of situation element acquisition but also reduces the operation time, which verifies the advantages of this framework.

Design of Attribute Reduction Method
If you want to use the classical rough set to deal with continuous data, you must discretize it, but some original data information will be lost. e traditional fuzzy rough set can overcome the above problems, but it only reduces attributes according to the dependence of decision attributes on conditional attributes and does not consider the redundancy of conditional attributes, the maximum membership of samples belonging to real categories, and the influence of noise samples on classification results. Aiming at the shortcomings of traditional fuzzy rough sets, this study proposes a method of attribute secondary reduction based on conditional attribute similarity and maximum similarity criterion, which can delete redundant conditional attributes while ensuring the breadth of classification information. To reduce the influence of noise samples and ensure that the samples belong to the real category of membership, this study proposes an approximate calculation method based on k-order weighted average in the nearest neighbor region, which makes the attribute reduction more accurate and improves the accuracy of situation element extraction.

Attribute Secondary Reduction Method.
To reduce the similarity between the condition attributes in the attribute subset after reduction and ensure the information breadth of the subset, a similarity measure method of condition attributes and an attribute selection method based on the maximum similarity criterion are proposed to delete redundant condition features, so as to achieve the goal of attribute secondary reduction.

Similarity of Conditional Attributes.
For the similarity measurement between samples and the contribution of conditional attributes to decision attributes, experts and scholars have proposed some classical measurement methods, but there is no unified measurement method for similarity of conditional attributes. is study draws on the experience of similarity between samples and the contribution of conditional attributes to decision attributes, and the similarity measurement methods between discrete and continuous attributes are proposed.
For discrete attributes, we can measure the importance of conditional attributes to decision-making attributes by contributing to classification, which is defined as follows.
Definition 1. Let the decision system be S � S � U, C, D, V, f , and given the discrete characteristic cεC, then the importance of characteristic c to D can be expressed as follows: where | · | is the square sum of the number of objects in each class divided by feature set , that is, the classification degree of features. Based on the above idea, we can define the similarity between discrete conditional features by different classification degrees of discrete conditional features. Definition 2. Let S � U, C, D, V, f be the decision system and C � c 1 , c 2 , c 3 , . . . , c m be the conditional feature set. Given discrete features c k , c l ∈ C(1 ≤ k, l ≤ m), the similarity of features c k and c l can be measured by the following formula: According to formula (2), if the classification degree of features c k and c l is the same, then | c k ∪ c l | � ���� � | c k | · ���� | c l |; that is, similarity s(c k , c l ) � 1; if the Computational Intelligence and Neuroscience classification degree of features c k and c l is different, then ���� | c l |; that is, similarity s(c k , c l ) < 1. Assuming that there are only two discrete conditional features of c 1 and c 2 in four samples, the optimal situation is that c 1 and c 2 cannot effectively classify the samples, but the combination can completely classify the samples. e four samples are (0, 0), (0, 1), (1, 0), and (1, 1); then, s(c 1 , c 2 ) � 1 2 + 1 2 + 1 2 + 1 2 / ������ 2 2 + 2 2 √ · ������ 2 2 + 2 2 √ � 0.5, and the similarity of the two features is the minimum.
When the conditional features are continuous features, the similarity between features cannot be measured by the classification degree of features. Because if we do not discretize the continuous features, we cannot describe the classification degree of the features by the values of the continuous features.
is study uses the experience of distance-based similarity measurement between samples for reference and finds that the similarity between continuous features can be effectively measured according to the distribution of eigenvalues, so a method of measuring the similarity between continuous features according to the distribution of features is proposed.
be the sample set, and C � c 1 , c 2 , c 3 , . . . , c m be the conditional feature set, and given continuous features c k , c l ∈ C(1 ≤ k, l ≤ m), then the similarity between features c k and c l can be expressed as follows: where x ik is the value of the kth feature of the ith sample in the sample set U, and a k is the average value of the kth feature in the conditional feature set; that is, a k � 1/n n i�1 x ik . According to formula (3), if the value distribution of the two features is the same, then n 2 ; that is, the similarity of the two features is s(c k , c l ) � 1; if the distribution is different, the similarity of the two features is s(c k , c l ) < 1.

Attribute Cyclic Secondary Reduction Based on Maximum Similarity Criterion.
After the matrix of attribute similarity is calculated according to the formula of conditional attribute similarity, we can see the correlation between attributes.
To reduce the secondary conditional attributes under the condition of ensuring the information breadth, we first cluster the attributes with high similarity by the direct clustering method and then select representative conditional attributes according to the maximum similarity criterion, so as to realize the secondary reduction of conditional attributes. Finally, the advantages and disadvantages of the secondary reduction attribute set are measured according to the attribute dependency index. us, it is decided whether the attribute reduction is needed again. Figure 1 shows the specific process of conditional attribute reduction.
(1) Conditional attribute clustering using the direct clustering method: the calculation process of direct clustering method is simple and easy to implement, which can save the clustering time of conditional attributes. erefore, this study uses this method to cluster similar conditional attributes. e specific process is introduced as follows: Definition 4. In universe U, a sequence of weighted values composed of a group of elements can be called a path: where R i (i � 1, 2, 3, . . .) is the ith path in the object set, x ik (k � 1, 2, 3, . . . , t) is the kth element in the ith path, and x i1 and x it are the starting point and ending point of the ith path, respectively. If two roads have the same starting point and ending point, they are considered to be equivalent. An element can appear many times in the path, but it cannot make the path form a loop. Each arrow in the path represents a step in the path, and c i(k−1) c ik (k � 1, 2, 3, . . . , t) on the arrow represents the cost or weight of each step.
Definition 5. e weight of the path is equal to the minimum weight of the step in the path: Definition 6. e decision system is S � U, C, D, V, f , C � c 1 , c 2 , c 3 , . . . , c m is the conditional feature set, and c k , c l ∈ C(1 ≤ k, l ≤ m) are given. If there is a path in the conditional feature similarity matrix with c k as the starting point (or endpoint) and the weight of c l endpoint (or start point) is not less than λ, then c k and c l belong to the same class under threshold λ. According to the weight definition of the path, the weight of all steps in the path is not less than λ, so all elements in the path belong to the same class under the threshold of λ.
According to Definition 6, the conditional feature set can be divided into several feature equivalence classes. However, the same feature attribute may belong to different feature equivalence classes at the same time. For those feature equivalence classes whose intersection is not an empty set, they should be combined into one class to get the final feature equivalence class.
(2) Secondary reduction in conditional attributes based on a maximum similarity criterion After clustering the condition attributes with high similarity, a representative condition attribute is selected from the attribute set of each cluster, and other condition attributes are deleted, to achieve a fast reduction in secondary redundant condition attributes. To complete this process, this study proposes a clustering conditional attribute reduction method based on maximum similarity criterion.
Definition 7. S � U, C, D, V, f is set as the decision system, C is set as the set of conditional attributes, and the similarity matrix of conditional attributes determined by the similarity measurement formula of conditional attributes is SM. en, according to SM and threshold λ, the clustering result of conditional attributes is describes the conditional attributes contained in a certain category after attribute clustering, en, the representative attributes of each clustered category can be expressed as follows: Among them, arg max( vm 1 , vm 2 , vm 3 , . . . , vm s ) is the condition attribute index corresponding to the maximum similarity sum of attributes in the set, and vm j is the similarity sum of the jth condition attribute and other condition attributes in the ith attribute cluster C i : If there are two or more conditional attributes in C i to maximize the total similarity of attributes, then the corresponding conditional attribute with the minimum standard deviation of attribute similarity as the representative attribute is selected, to ensure that the selected conditional attributes represent the corresponding clustering subset of similar attributes to the maximum extent, to reduce the information loss caused by eliminating redundant attributes. e analysis shows that the number of similar conditional attribute subsets after clustering does not decrease with respect to the clustering threshold, because the reduction process is to select a representative attribute from each similar conditional attribute subset, so the number of conditional attributes in the reduced representative attribute set does not decrease with respect to the clustering threshold; and the more attributes in the reduced representative attribute set, the more decision-making characteristics depend on it [34]. Taking advantage of this phenomenon, we can build a conditional attribute based on the related knowledge in the next section.

Attribute Final Reduction Method.
After the redundant conditional attributes are deleted to get the secondary reduction in the conditional attributes, we need to consider the contribution of the conditional attributes to the decision attributes. Some experts and scholars have conducted indepth research on the importance of conditional attributes to decision attributes as a measure to reduce redundant conditional attributes and put forward some feasible methods. ese methods are superior in some aspects, but there are also some shortcomings. In view of the shortcomings of some methods at present, this study redefines the upper and lower approximations of fuzzy rough set theory and proposes a feature selection method of fuzzy rough set based on k-order weighted average.
In Section 2, we have introduced the standard definition of upper and lower approximation of fuzzy rough sets. To facilitate calculation, some experts have given another equivalent definition. Definition 8. Let the decision system be S � U, C, D, V, f and U be divided into r distinct indiscernible relations X 1 , X 2 , X 3 , . . . , X r by D. Given the feature subset B⊆C, R B is a binary similar fuzzy relation on the object set U under the condition of B, and then, the lower and upper approximation sets of X i (1 ≤ i ≤ r) with respect to B are, respectively, expressed as follows: where B X i (x) is the minimum value of the inconsistency degree of sample x ∈ U relative to all samples y that do not belong to category X i . In other words, B X i (x) depends on the sample y that is most similar to x among all samples that do not belong to category X i . BX i (x) is the maximum value of consistency between sample x ∈ U and all samples y belonging to category X i . e above definition uses the fuzzy Computational Intelligence and Neuroscience neighborhood of the sample x to measure the importance of the sample to the decision category X i , which is similar to the classical rough set in thought.
Although the above definition can simplify the upper and lower approximation process of fuzzy rough sets and reduce the computational complexity, it does not consider the influence of noise samples on the upper and lower approximation process. If the noise sample y is exactly the sample of all samples not belonging to category X i and the most similar to x in the approximate extraction process under category X i , then the lower approximation of X i obtained at this time is not accurate.
For example, there are two categories of samples in a data set, and the sample distribution is shown in Figure 2. It can be seen that the sample y 1 is far away from its category 2. In this case, it can be considered as a noise sample. When formula (8) is used to calculate the degree that sample x belongs to category 1, noise sample y 1 will still be used for calculation, which will lead to an inaccurate calculation of the lower approximation of category 1. However, if y 2 is used to calculate the lower approximation of category 1, more accurate results will be obtained.
To solve the above problems, this study proposes an upper and lower approximation method based on k-order weighted average, which can effectively resist the influence of noise samples.
Definition 9. Let the decision system be S � U, C, D, V, f and U be divided into r distinct indiscernible relations X 1 , X 2 , X 3 , . . . , X r by D. Given the feature subset B⊆C, R B is a binary similar fuzzy relation on the object set U under the condition of B, and then, the lower and upper approximation sets of X i (1 ≤ i ≤ r) with respect to B are, respectively, expressed as follows.
Let y 1 , y 2 , y 3 , . . . , y k be k samples with the least degree of inconsistency with sample x that are not classified as class X i ; then, the lower approximate set of X i (1 ≤ i ≤ r) about B is expressed as follows: (10) Let z 1 , z 2 , z 3 , . . . , z k be the k samples, which belong to class X i and are most consistent with sample x, and then, the upper approximation set of X i (1 ≤ i ≤ r) with respect to B is expressed as follows: Firstly, k samples that meet the requirements are selected, then the dissimilarity (or similarity) of each sample is multiplied by the corresponding dissimilarity (or similarity) weight, and finally they are added as the final value. By using this method to calculate the upper and lower approximations of fuzzy rough sets, the idea of weighted average can be effectively used to resist the interference of noise. If the sample distribution of a certain data set is shown in Figure 2, the sample y 1 is the noise sample of category 2, and this sample happens to be the sample with the least similarity to sample x when calculating the lower approximation of category 1, then this sample must be the sample most similar to category 1 relative to other samples of category 2, that is, relative to other samples of category 2.
e similarity weight of the noise sample y 1 is the smallest, so the influence of the noise sample y 1 can be resisted when k qualified samples are selected and the lower approximation of category 1 is calculated by the weighted average method.
e approximate decision table under the fuzzy rough set obtained by formula (8) can correctly classify all samples in the example, but it needs to calculate the fuzzy neighborhood of 3 3 � 9 samples in the process of obtaining the approximation. If the number of samples in the data set is very large, the calculation time will increase explosively. To effectively reduce the time of calculating the upper and lower approximation of fuzzy rough sets and reduce the influence of noise samples, this study improves the k-order weighted average method and proposes a method of calculating the upper and lower approximation of close neighborhood based on order weighted average.
Given the feature subset B⊆C, R B is a binary similar fuzzy relation on the object set U under the condition of B, and then, the lower and upper approximation sets of X i (1 ≤ i ≤ r) with respect to B are, respectively, expressed as follows: 6 Computational Intelligence and Neuroscience Among them, y j (1 ≤ j ≤ k) represents the jth sample among the k samples that do not belong to category X i and are least inconsistent with sample x, and z j (1 ≤ j ≤ k) represents the jth sample among the k samples that belong to category X i and are most consistent with sample x.
Using the above formula to calculate the upper and lower approximations of fuzzy rough sets can ensure that the membership of samples in the category they do not belong to is zero, thus indirectly ensuring that the membership of samples in the real category is the largest. In this way, the method of directly constructing the upper and lower approximations can save most of the unnecessary calculation of the sample fuzzy neighborhood and greatly improve the speed of the upper and lower approximations of fuzzy rough sets. For example, the lower approximation decision table of the data set shown in Table 1 can be obtained using the upper and lower approximation calculation method of the nearest neighbor based on the k-order weighted average, which can not only make all samples correctly classified but also only calculate the fuzzy neighborhood of three samples, greatly reducing the calculation time.
Applying the upper and lower approximation calculation method based on k-order weighted average to the attribute reduction process of fuzzy rough sets, a heuristic attribute selection algorithm based on addition can be designed for the secondary reduction set of conditional attributes obtained in the previous section. e specific process is shown in Figure 3. Firstly, we take the secondary reduction set of conditional attributes as all the attributes to be reduced and set the target attribute set as an empty set. Secondly, a conditional attribute in the attribute set to be reduced is copied to the target attribute set, and the lower approximation of the fuzzy rough set is obtained according to the upper and lower approximation calculation method of the nearest neighbor based on k-order weighted average, so as to obtain the importance of the current target attribute set to the decision attributes, and then, the added conditional attributes need to be removed from the target attribute set. irdly, the selection of the next attribute in the attribute set is continued to be reduced and the corresponding operation until each attribute in the attribute set to be reduced is traversed is completed, and then, the condition attribute is selected that makes the target attribute set increase the most importance to the decision attribute to join the target attribute set and delete the attribute from the attribute set to be reduced. Finally, whether the target attribute set meets the requirements is judged. If not, the qualified attributes selected from the attribute set to be reduced to the target attribute set are added. Otherwise, the attribute selection process is terminated and the final target attribute set is obtained.

Description of Attribute Selection Algorithm.
Previously, we have introduced the secondary reduction in conditional attributes based on the similarity and maximum similarity criteria of conditional attributes and the final reduction process and related theories of attributes in the nearest neighbor region based on k-order weighted average. To effectively combine the two, reduce the dimension of network security data, and accurately extract the situation elements, this study designs the following attribute selection algorithm. Computational Intelligence and Neuroscience rough the above algorithm, the redundant information between condition attributes is considered, and the redundant condition attributes are deleted under the condition of ensuring the information breadth. At the same time, the membership degree of samples belonging to the real class is ensured to be the largest. It can reduce the influence of noise samples, reduce the misclassification probability of samples, and shorten the time of calculating the upper and lower approximation of fuzzy rough sets. e algorithm also does not need to solve the core attributes of conditional attribute set in attribute reduction, which reduces the time of feature selection to a certain extent.

Design of Situation Element Extraction Model Based on Combination Classifier
Although the traditional recognition system relying on a single classifier can achieve high recognition rate in some problems, it generally needs more samples to train, which increases the time and space cost of system training. When faced with complex problems, we cannot require a single classifier recognition system to have good recognition ability for all target categories, and the recognition system is prone to overfitting. A combined classifier recognition system can make full use of the complementary classification  information of different classifiers, to overcome the shortcomings of single classifier. e construction of combined classifier recognition system mainly consists of two parts, namely, the design of classifier group and the design of fusion layer. e design of these two parts has no sequence. We can design the classifier group or the fusion layer first, but the design of these two parts must be related to each other. e classification method adopted by the classifier group should match the fusion method adopted by the fusion layer, and the output of the classifier group can be used as the input of the fusion layer. Only when the two parts are designed reasonably and combined effectively, the performance of the combined classifier recognition system can be brought into full play.

Design of Classifier Group in the Model.
e traditional single classifier recognition system generally selects a suitable classifier to train and recognize the target according to the specific problems and expert experience. Even if multiple classifiers are used, the best one is selected by comparing these classifiers, and the remaining classifiers are discarded. Combined classifier recognition system needs to select a group of classifiers from many classifiers and effectively combine these classifiers to make full use of the complementary information between different classifiers, so as to improve the performance of the recognition system. e design of classifier group in combined classifier recognition system is relatively simple. Generally, several classifiers that meet the corresponding requirements are selected from many classifiers to construct a classifier group through some measurement indexes. For example, in reference, a method of using the genetic algorithm to select a group of classifiers is proposed. In this method, each classifier is constructed as a gene of the genetic algorithm, and then, the chromosome of the optimal gene combination is obtained through genetic evolution; thus, a subset of classifiers is selected in the classifier set. In this study, from the recognition rate, ROC curve, learning time, and other indicators, the performance of each classifier is comprehensively considered. Four classifiers are selected from 15 commonly used classifiers, and they are constructed into a classifier group. e specific selection process will be introduced in detail in the experimental part. In addition, it has been mentioned that to make the output of the classifier contain more information, the classifier with metric output form should be selected in the process of classifier group design, so the four classifiers selected in this study are classifiers with metric output form.
Using the classifier group for target recognition can make full use of the complementary information between different classifiers without demanding a certain classifier. In this way, we can not only use fewer samples to train the classifier group and prevent the classifier group from overfitting, thus effectively improving its generalization ability, but also because of the mutual restriction between different classifiers. e search space of the solution is enlarged, and the risk of getting the local optimal solution is reduced.

Design of Fusion Layer in the Model.
In Section 2, we have introduced six fusion methods commonly used in the fusion layer of the combined classifier recognition system. ese six fusion methods can achieve good results in solving simple Input: network security information decision system S � U, C, D, V, f , initial clustering threshold λ, and empty attribute set A. Output: attribute reduction set A that meets the requirements.
Step 1: according to the k-order weighted average method, the lower approximation of fuzzy rough set under condition feature set C is obtained, and then the dependence of decision feature D on C is obtained; Step 2: calculate the similarity matrix SM of condition attributes according to the similarity measurement formula of condition attributes; Step 3: complete the clustering of conditional attributes according to the direct clustering method and clustering threshold λ, and divide the set of conditional attributes into r similar subsets C 1 , C 2 , C 3 , . . . , C r ; Step 4: according to the maximum similarity criterion, select a representative condition attribute from each attribute similarity subset to form the secondary reduction attribute set B; Step 5: get the lower approximation of the fuzzy rough set under the attribute subset B according to the method of order weighted average in the next neighborhood, and then get the dependence degree of the decision attribute D on the attribute subset B; Step 6: judge whether D depends on B more than or equal to C. If it does not meet the requirements, increase the clustering threshold λ and return to Step 3 to continue. Otherwise, complete the secondary reduction of conditional attributes and get the attribute set; Step 7: take attribute subset B as the attribute set to be reduced, and select an unselected condition attribute c i from it to copy it to attribute subset A; Step 8: the lower approximation of fuzzy rough set under attribute subset A is obtained according to the k-order weighted average lower approximation calculation method in the nearest neighbor region, and then the dependence degree of decision attribute D on attribute subset A is obtained, and then attribute c i is deleted from attribute subset A; Step 9: repeat Step 7 and Step 8 until every element in B is traversed; Step 10: select the attribute c j that increases the dependence of decision attribute D on attribute subset A the most and add it to attribute subset A and delete it from attribute subset B; Step 11: check whether the dependency of D on A is greater than or equal to that of C after attribute c j is added to attribute subset A. If it does not meet the requirements, it will return to Step 7 to continue. Otherwise, it will complete the final reduction of conditional attributes and output the final attribute reduction set A. Computational Intelligence and Neuroscience fusion problems, but they are faced with the complex information fusion problems. e results obtained are not obvious, and the accuracy of fusion results is lower than that of single classifier in the classifier group. To make the classification results of the classifier group output more effective fusion, this study constructs a fusion method based on random vector to improve the compression factor particle population optimization BP network. Although the method is more complex than the six fusion methods, it can improve the accuracy of classification and recognition effectively.

Improved Compression Factor PSO Algorithm Based on
Random Vector. Although the traditional particle swarm optimization method can quickly get the approximate solution of the problem through iterative operation, which shows good problem optimization ability, when facing complex problems, its particles are prone to prematurity and cannot get the optimal solution of the problem, resulting in large solution error. To effectively improve the traditional PSO method, we mainly start from two aspects: (1) some advanced theories are integrated with the PSO method to study the improved PSO method. (2) Some good performance algorithms are combined with PSO to study various combinatorial optimization methods. Here are two classic improved PSO methods.
(1) PSO method with inertia weight: in order to improve the global optimization ability of the traditional PSO method, the weight with inertia was added to the particle swarm optimization method by Shi in 1998, which makes its performance better [35]. e specific form of the improved PSO method is as follows: where w is the inertia weight. When w � 1, this method is the traditional PSO method; when w is large, this method will search for the approximate solution globally to prevent premature particles, but it is easy to skip the optimal solution; and when w is small, the method will search for the optimal solution in the local solution space and accelerate the convergence speed, but it is easy to get the local optimal solution. Some experts and scholars have studied the value of w inertia weight and put forward two common methods.
① e method of inertia weight decreasing linearly is as follows: Among them, w max represents the maximum value that w can obtain, which is generally set as 0.9, w min represents the minimum value that w can obtain, which is generally set as 0.4, t represents the number of current iterations of the population, and t max represents the total number of iterations of the population. e above method can ensure that the value of w decreases gradually when the number of iterations increases. In the early stage of iteration, it can ensure to find the approximate solution globally and lock the position of the approximate solution as soon as possible, to prevent premature particles to a certain extent; in the later stage of iteration, it can ensure that the local optimal solution is searched carefully and the convergence speed is accelerated. ② e adaptive value method of inertia weight is as follows: where f is the fitness of the corresponding particles, f min and f avg are the minimum and mean fitness of all particles under the current iteration number. When the fitness of particles is better than the average of all particles, it means that the particle is near the optimal solution.
When w is small, the particle searches locally. On the contrary, it shows that it is far away from the optimal solution, and when w is larger, the particle carries out the global search. By using the fitness value of each particle to control the inertia weight of each particle, the difference between particles can be ensured, the optimal solution can be searched more comprehensively, and the premature of particles can be prevented to a certain extent.
(2) PSO algorithm with compression factor is as follows: where ϕ is the compression factor: Although PSO with inertia weight can improve convergence speed and reduce the risk of getting local solution of the problem, it will lose the ability to explore the new solution space due to too small inertia weight in the later iteration period. PSO method with compression factor can effectively avoid this problem and improve the local exploration ability of the method, the better constraint particle flying to the optimal solution.
Although the performance of PSO method with compression factor is good, it can solve many optimization problems, but when it faces high-dimensional or complicated optimization problems, the solution may be local optimal. To solve this kind of complex nonlinear problem efficiently, this study introduces a compression factor PSO method based on random vector.
If the solution space of a problem is n-dimensional, the iterative optimization process can be expressed as follows: where r k 1d and r k 2d represent two d-dimensional random vectors of the kth iteration, and each element of r k 1d and r k 2d belongs to [0, 1]. is compression factor PSO method based on random vector not only has the advantage of iterative optimization of compression factor PSO method but also makes the optimization information of the previous generation of particles affect the current speed of the particles randomly through the control of random vector, to increase the differences among particles and effectively suppress the method to get the local optimal solution, further enhance the ability of particle to explore new regions, and improve the convergence speed of the method.

Structure Description of Fusion Layer.
Research has shown that if the output form of each classifier in the classifier group is metric output form, we can use neural network as the fusion method of fusion layer to realize the fusion decision. Because BP neural network has a simple structure, it is relatively easy to implement and has good performance, so this study uses it as a fusion decision method. In the optimization of BP network, the derivative of error is generally used to update the parameters in the network by back propagation. e problem solution obtained by this method of updating the parameters by gradient descent is easy to be locally optimal, so it cannot well approximate the real solution of the problem [36]. To solve this problem, this study uses the compression coefficient PSO method based on random vector to optimize the parameters of BP neural network and realizes the fast solution of parameters under the condition of reducing the computational complexity.
Because the output result of each classifier in the classifier group is the class information of the input sample, the output result of the classifier group can be regarded as the extended features of the identified sample, but such extended features are obtained through the complex transformation of the classifier and cannot fully represent the information of the sample. In order to input more information into BP fusion neural network and improve the accuracy of fusion decision, this study inputs the original features of samples and the output results of classifier group to the fusion layer, and its specific structure is shown in Figure 4. e application process of fusion layer is mainly divided into the following stages: firstly, the improved compression factor PSO method based on random vector is used to iteratively optimize the relevant parameters of BP neural network, and the trained BP fusion neural network is obtained. en, the network security data and the corresponding results of the classifier group are input to the BP fusion network, and the BP fusion network classifies the input data to get the final fusion result. Finally, the fusion results are compared with the real value to judge the accuracy of network situation element extraction.

Extraction Process of Situation Elements Based on Combined Classifier.
e use of the network situation element extraction system based on the combination classifier is mainly divided into two stages: training and target recognition. In the process of training, we need to train each classifier in the classifier group with network security samples; then, the optimized classifier group is used to identify the target of the network security training samples; finally, the recognition results of the classifier group and the corresponding training samples are input into BP neural network. e parameters of BP neural network are optimized using the improved compression factor PSO algorithm based on random vector, and the optimized BP fusion network is obtained. In the process of classification recognition, the first step is to input the data set of network security test into the classifier group to get the target recognition result of the classifier group to the test set; then, the target recognition results obtained by the classifier group and the corresponding network security samples are input into the fusion layer to get the final fusion recognition results; finally, the fusion recognition results are evaluated as network situation elements to verify the accuracy of extraction. e specific process is shown in Figure 5.

Experimental Environment and Data.
e specific hardware environment for simulation is as follows: the CPU used is 2.60 GHz Intel Core-i5 3200M, the graphics card used is 2 GB NVIDIA GT 720M, the capacity is 12.0 GB RAM, and the capacity is 128 GB (solid state) + 500 GB (mechanical) hard disk. e specific programming environment for simulation is as follows: Win 7 (64 bit) desktop system, PyCharm of 2019.1, Python of 3.7, TensorFlow of 1.15.1, and Keras of 2.3.1. e experiment uses the authoritative and classic NSL-KDD data set, which is an improved version of KDD99 data set and corrects some defects of the original data set. e specific improvements of the data set are as follows: (1) the redundant samples in the original data set are deleted, which inhibits the overlearning of some samples to a certain extent and improves  is data structure solves the problem of uneven distribution of sample categories to a certain extent, which makes the recognition algorithm easier to perform, and the results are more effective. ere are 21 attack modes in the training sample set, and in addition to these attacks, there are 16 new attack modes in the test sample set [37]. ese attack modes can be divided into four types, and their specific distribution is shown in Table 1. e training sample set and test sample set are classified according to the attack type, and the specific classification is shown in Table 2.

Pretreatment of Experimental Data
① Character Data Digitization: in addition to numerical data, there are four kinds of character data (proto-col_type, service, flag, label) in each sample of experimental data. To unify the standard and meet the experimental requirements, the first three of the four character data are directly converted into numerical data; for example, the value of TCP in the proto-col_type condition attribute is assigned to 1, the value of UDP is assigned to 2, and the value of ICMP is assigned to 3; secondly, label data are classified according to attack types and then converted into numerical data directly; finally, the transformed label data are encoded by one hot. ② Standardization: we standardize the sample set to prevent the extreme value or outlier of the sample from affecting the experimental results. ③ Normalization: we normalize the sample set to unify the dimensions of the sample.

e Determination of Classifiers in Classifier Group.
To select several classifiers with good performance from some classifiers to construct a classifier group, 15 classical classifiers based on different principles are optimized with training samples; then, the test samples are input into 15 optimized classifiers, and the classifiers classify them, respectively; finally, four classifiers with better comprehensive performance are selected to construct a group of classifiers. First of all, it is necessary to select the classifier with high recognition rate for U2R and R2L attack types. Because the samples of U2R and R2L attack types in the training data set are relatively small, most classifiers have the problem of under learning. If the selected classifier has a low recognition rate for both types of attacks, the fusion layer will also get a low recognition rate when fusing the recognition results, which will affect the performance of the combined classifier system; secondly, to ensure the stability of the classifier in different states, we need to choose the classifier with relatively large AUC coefficient (the area formed by ROC curve and X axis is relatively large); finally, considering the performance of the classifier group, it needs less training time and relatively high total recognition rate to select the classifier.
rough the above process, the classifier training and recognition in Table 3, and the ROC curve and AUC coefficient of the classifier in Figure 6, this study selects MLP classifier, k-neighbor classifier, logistic regression, and random forest classifier from 15 classical classifiers to construct the classifier group.

Selection of Experimental Parameters.
In the process of secondary reduction in conditional attributes, the initial value of clustering threshold is 0.5 and the step size is 0.01, which increases gradually until the termination condition is satisfied. In the construction process of fusion layer, the particle number of the improved compression factor PSO method based on random vector is set to 50, the values of c 1 and c 2 are set to 2.05, the maximum number of iterations is set to 200, and the recognition rate of samples is taken as a function of particle fitness; the activation function in the network is set to sigmoid function, the number of nodes in the output layer is set to 5, and the number of nodes in the input layer and the hidden layer depends on the specific problem.

Experimental Results and Analysis.
We need to build the system according to the design process in Section 4. e previous construction of classifier group has been completed, so let us build the fusion layer.
To show the performance advantages of the improved compression factor PSO method based on random vector in the process of constructing fusion layer, this study compares the proposed improved PSO method with the other two PSO methods, and the results are shown in Table 4. All attributes of NSL-KDD data samples are used in the experiment, so it is Compared with the other two PSO methods, the approximate solution obtained using the improved compression factor PSO method proposed in this study is more close to the optimal solution, which improves the recognition rate and the effect of the system. e effectiveness of the improved PSO method is verified. e improved PSO method can optimize BP fusion neural network to achieve good results. To improve the performance of the network, we need to compare the network performance under different numbers of hidden layer nodes, so as to find the best hidden layer junction number. e experimental results show that, when the number of hidden layer nodes is 18, the sample recognition rate of the system is the highest, so the number of hidden layer nodes should be set to 18.
After constructing the target recognition system of combined classifier, to verify its effectiveness, this study designs three comparative experiments.
First, the recall rate of the target recognition system of the combined classifier is compared with that of the single classifier in the classifier group, and the results are shown in Table 5.
It can be seen that the combined classifier target recognition system is superior to the single classifier in the classifier group in both the total recognition rate and the recognition recall rate of each attack category, which proves that the combined classifier in this study is better than other single classifiers in the classifier group, thus proving the effectiveness of the combined classifier system in this study.  Second, the output of the classifier group is fused through the fusion layer designed in this study and the fusion method mentioned in Section 2.
e results are shown in Table 6.
It can be seen that although the six classical fusion methods are relatively simple in calculation and the recognition rate of some categories is higher than that of the fusion method in this study, the recognition recall rate of each attack category is lower than that of the fusion method in this study, which can also be proved from the average F1 value of macro (the average F1 value of each attack category). Moreover, the total recognition rate of the six classical fusion methods is lower than that of the single classifier of the classifier group, so in the face of complex problems, the classical fusion methods not only cannot well fuse the output of the classifier group but also can play a counterproductive role. Compared with the six classical fusion methods, the fusion method proposed in this study not only greatly improves the total recognition rate but also greatly improves the F1 value of each category; that is, it ensures a higher recognition rate and recall rate at the same time, which proves that the fusion method proposed in this study is effective.
ird, to further prove the effectiveness of the combined classifier system, this study compares it with several current classic neural networks. In the experiment, the number of iterations of each recognition system is set to 200, and the results are shown in Table 7.
It can be seen from the table that the above target recognition systems have fully studied the training set, but the recognition effect of the first four neural networks on the test set is not ideal; that is, the generalization rate is not high, and the recognition rate of the test set and the recall rate of each attack category have been significantly improved by the combined classifier system in this study, which proves the feasibility of the recognition system in this study.
e value of num_outbound_cmds attribute in NSL-KDD data set is all zero, which has no effect on situation element extraction, so it should be deleted before attribute reduction. e secondary attribute set obtained by the attribute secondary reduction method in this study is shown in Table 8, and the final attribute set obtained by the attribute final reduction method in this study is shown in Table 9.
To verify the effectiveness of the proposed attribute reduction method, three comparative experiments are designed.
First, the four classifiers in the classifier group are optimized using the training samples that are not reduced and the training samples that are reduced by the feature selection method in this study. en, the four classifiers are identified using the test samples that are not reduced and the test samples that are reduced by the method in this study.
It can be seen from the corresponding experimental results in Table 10 that the training time of the classifier is reduced by attribute reduction, and the recognition rate of three of the four classifiers has been improved to varying degrees, which proves that the attribute reduction method proposed in this study is feasible.
Secondly, the feature selection methods in this study are compared with those of literature [38] and document [39], and the results are shown in Table 11.
It can be seen that the performance of training and recognition using the data set reduced by the method of this study is better than the other two methods. It shows that the attribute reduction method is effective.
irdly, the combination classifier system is trained and recognized with the original data set and the data set is reduced by the attribute reduction method to improve the performance of the attribute reduction method. Because the feature of the sample becomes 12 after feature reduction, the node number of the input layer of BP network should be adjusted to 32, and the optimal node number of the hidden layer needs to be determined again. e experimental results show that the optimal node number is 13.
After the combined classifier system used before and after attribute reduction is constructed, the corresponding sample sets are used to train and identify them, respectively. e corresponding experimental results are given in Table 12.
It can be seen that although attribute reduction can reduce the recall rate of DoS attack recognition, it can be ignored. erefore, using the attribute reduction method in this study can further improve the performance of the combination classifier system and extract more accurate situation elements, which proves the effectiveness of the         attribute reduction method in this study and also proves that it is feasible to extract situation elements combined with the combination classifier.

Conclusion
e main contributions of this study are as follows: (1) e relevant theories of situation element extraction are summarized. e classical situation element extraction process is analyzed and improved, the corresponding algorithm is designed, and a situation element extraction framework based on fuzzy rough set and combined classifier is constructed.
(2) To delete redundant information between conditional attributes on the premise of ensuring information breadth and data classification ability, an attribute secondary reduction method based on conditional attribute similarity and maximum similarity criterion is proposed. Firstly, the similarity matrix of conditional attributes is obtained by this method, and then, the conditional attributes are clustered using the direct clustering method and reasonable threshold. Finally, the representative attributes are selected using the maximum similarity criterion to complete the secondary reduction in conditional attributes. (3) To ensure the maximum membership of the samples belonging to the real category, reduce the probability of misclassification, and reduce the influence of noise samples, an upper and lower approximation calculation method of the nearest neighbor domain based on k-order weighted average is proposed and applied to the fuzzy rough set theory. A heuristic reduction method for the secondary conditional attributes is designed to obtain the final reduced attribute set. (4) To establish a better combined classifier, this study selects 4 classifiers with good personality from 15 classical classifiers to construct a classifier group according to some measurement indexes of the classifier. en, the improved PSO method is used to optimize the BP neural network as the fusion layer of the combined classifier, so as to complete the construction of the model. (5) e attribute selection method and combined classifier model in this study are coded, and a variety of comparative experiments of fusion method, attribute selection method, and target recognition system are carried out on the authoritative NSL-KDD data set to verify the effectiveness and feasibility of the network situation element extraction model based on fuzzy rough set and combined classifier.
At the same time, this study also has some shortcomings: (1) e model is applied to a more real network environment, so as to further test the advantages and disadvantages of the situation element extraction method.
(2) In the process of attribute reduction, although some complex computation is avoided, it still needs a certain amount of computation. In the future research, we should further explore measures to improve the efficiency of attribute selection methods or make effective attempts on parallel attribute reduction. (3) In the process of building classifier group, we explore more reasonable construction methods, to make the selected classifier group better and further improve its performance. (4) e extracted network situation elements are effectively applied to situation understanding and prediction to realize its real value.
Data Availability e data sets used and analyzed during this study are available from the corresponding author on reasonable request.

Conflicts of Interest
e authors declare that they have no conflicts of interest.