Retraction Retracted: Mathematical Methods for Sensitive Information Mining Method of News Communication Platform Based on Big Data IOT Analysis

It is urgent to eﬀectively monitor the public opinion of the news communication platform. +e platform designed in this paper takes microblog public opinion as the research goal, uses MongoDB to build a distributed computing platform for sensitive information of news communication platform, establishes a corpus of sensitive event topics, introduces PageRank algorithm to deal with microblog social relations, obtains the characteristics of sensitive information of news communication platform, and carries out information screening, so as to accurately screen and mine the keywords in high impact information. To ensure the practical application eﬀect of sensitive information mining method of news communication platform based on big data analysis. Finally, the experiment proves that the sensitive information mining method of news communication platform based on big data analysis has the advantages of high timeliness and high accuracy, which fully meets the research requirements. +is is fully in line with the requirements of the study.

This article has been retracted by Hindawi following an investigation undertaken by the publisher [1]. This investigation has uncovered evidence of one or more of the following indicators of systematic manipulation of the publication process: (1) Discrepancies in scope (2) Discrepancies in the description of the research reported (3) Discrepancies between the availability of data and the research described (4) Inappropriate citations (5) Incoherent, meaningless and/or irrelevant content included in the article (6) Peer-review manipulation The presence of these indicators undermines our confidence in the integrity of the article's content and we cannot, therefore, vouch for its reliability. Please note that this notice is intended solely to alert readers that the content of this article is unreliable. We have not investigated whether authors were aware of or involved in the systematic manipulation of the publication process.
Wiley and Hindawi regrets that the usual quality checks did not identify these issues before publication and have since put additional measures in place to safeguard research integrity.
We wish to credit our own Research Integrity and Research Publishing teams and anonymous and named external researchers and research integrity experts for contributing to this investigation.
The corresponding author, as the representative of all authors, has been given the opportunity to register their agreement or disagreement to this retraction. We have kept a record of any response received.

Introduction
With the popularization of the Internet and the improvement of netizens' sense of social responsibility, the network public opinion broke out a huge vitality that cannot be ignored. It is the public's strong influence and tendentious views on some hot issues in real life. Generally speaking, sensitive information is mainly composed of four parts, which are sensitive words, the related words of sensitive words, the degree of correlation between them and the association rules between them [1]. At present, sensitive information mining technology mainly uses association analysis and cluster analysis to obtain sensitive information related to sensitive words. +e application range of correlation analysis technology is relatively wide, and the development speed is fast. Association analysis technology mainly includes two parts: association words and association rules [2]. Clustering analysis technology is mainly to find the text information of related topics, so as to realize the monitoring of topics and achieve the purpose of topic tracking [3]. For big data analysis, this paper aims to establish a big data platform in which IoT and smart devices can work together to collect the data. +erefore, the final objectives of this paper are to utilize the developed big data communication platform enable quick information collection and real-time feedback, to aggregate and analyze the data collected through repeaters, and to utilize structured databases in the form of big data. +e application process of this technology mainly includes three steps. +e first step is feature extraction, which mainly refers to filtering the information after the input of information, obtaining the feature vector of the sample, and finally obtaining a matrix; the second step is text clustering, which mainly refers to clustering the results of feature extraction, which can obtain a matrix reflecting all the features in the n-dimensional space. +e final step is to select the classification threshold, which mainly refers to the determination of the threshold after the clustering spectrum is obtained, and then the classification scheme can be directly obtained, in order to ensure the effectiveness of sensitive information mining method based on big data analysis.

Information Filtering Algorithm of News Communication
Platform. Network information is complex and diverse, and there are many kinds of good information to serve the public. At the same time, some reactionary, superstitious, violence, and other sensitive information pose a serious threat to social and public security [4]. +erefore, to carry out Internet public opinion information mining requires not only to identify the hot topics concerned by the public from the Internet public opinion information, but also to analyze whether the public's attitude towards an event is positive or negative. In addition, deep-seated public opinion information mining requires good control of negative information dissemination, timely discovery, and disposal of sensitive information, in order to prevent it from causing serious harm to the government, enterprises, individuals, and so on [5]. In the process of traditional topic detection, to judge the similarity between a report and a topic, we need to calculate the similarity between the report and each report in the topic cluster. When the size of topic cluster is large, the number of comparisons will increase exponentially, which will affect the processing speed [6]. To solve this problem, the center vector is used to represent a topic. At this time, we only need to calculate the similarity with the center vector, which improves the effectiveness of topic discovery. By analyzing the public opinion information such as network news and forum posts, the existing forms and the structural characteristics of network text information, this paper puts forward the idea of "text reconstruction", that is to say, the representative information of the topic is gathered together to form a "theme block", and the remaining part forms a "content block". News headlines contain a lot of classified information, which can let the public know the basic situation of the event in the briefest prompt and evaluation, and guide the public to further read [7]. It is a high generalization of the content of the web page, and has a high accuracy when used to classify news web pages [8]. Topic is the supporting point of Title Construction, and the title of the core event under the same topic is the same or similar to that of its related follow-up reports. Title Information has significant ability to distinguish topics in topic detection, but with the continuous development and change of events, the topic center drifts, and the titles of subsequent reports also change.
+e first paragraph of the news web page is a supplement to the title, which is a general description of the event, including the time, place, event, which people or units are involved and so on. It makes a great contribution to the classification [9]. "Universal" emphasizes the statistical characteristics of public opinion information, a single web page cannot be regarded as public opinion, many web pages about a topic and the participation of many Internet users can become network public opinion information [10,11]. In this sense, network public opinion is accompanied by many news pages, BBS/forums, blog, many netizens browse or comment on a certain topic. We can say that the topic spread by multiple media and concerned by multiple netizens is a hot topic.

Construction of Sensitive Information Database of News
Communication Platform. From the header of the file I record, you can find the first attribute address in the file i record body, followed by two flag words (flags), where the first bit represents the file deletion flag, and the second bit represents the directory 1 division flag (normal directory). When reading the disk information, make a suitable way to read the information according to the two flag bits. If it is a normal file or directory, continue the following reading operation [12]. Otherwise, you need to recover the file first, and then filter the recovered file to get the required file information, and then hand it to the following text information extraction module for processing.
As can be seen from the figure, a file may contain multiple attributes. A complete file needs all the attributes in the file records to be combined according to certain rules. +erefore, when reading the disk file information, all the attributes in the file records need to be read into the memory according to the order in the file records [13]. Each attribute has its specific content, that is, the operation of each attribute is different. +e serial number of the file i recorded in the main file table NFTstarts with 0, and the files i recorded from 0 to 16 belong to platform files, or metafiles, which are mainly used to store the metadata of the platform. +ese metafiles are transparent to users and are hidden files [14]. +e difference between 16 files and other files and directories is that they have a unique fixed address in the MFT table,  while other files and directories can be stored anywhere in  the table. +e content part starts with the attribute name, and then defines whether the attribute is resident or nonresident. If it is the former, then the attribute value is the content of the attribute; conversely, if it is a nonresident attribute, then the flow of the attribute will be stored in one or more runs. For the sake of simplicity, the storage running area is continuous on the logical cluster number [15]. A run table is stored after the file attribute name, through which the run date table can access the run table belonging to the attribute. +e purpose is to calculate the similarity between the text information extracted from the new web page and the existing text cluster, and taking into account the life cycle of public opinion information; its importance decreases with the loss of time [16]. +at is to say, the same keyword in different time intervals is likely to represent different meanings, so we add the calculation of time interval [17,18]. For example, after the new text information is calculated in the time interval, the smaller the similarity value is, the higher the possibility that it is a new event is, and the higher the score is. +e expression is shown in formula (1):

Mathematical Problems in Engineering
In the formula, x → is the new file information, c → 1 is the first cluster in the time interval, i is the number of files in the time interval, and k is the number of files added between the latest file collection time in cluster c → 1 and the arrival time of the new file x → . In the case of setting the closed value, as long as the score is greater than the set value, the new file is considered to be a new topic. Based on this, the data processing steps of the news communication platform are optimized as follows in Figure 1: +e design of network public opinion monitoring platform mainly includes three modules: text preprocessing module, sensitive information analysis module, and public opinion analysis module. +e text preprocessing module mainly includes two steps: Chinese word segmentation and information filtering. Among them, Chinese word segmentation mainly transforms the irregular key text obtained by the platform to form a sensitive word set, and then further processes the word set to obtain the corresponding associated word set. When using word segmentation tools for word segmentation, its speed is relatively fast, and has high efficiency. After inputting the original text information, through the process of Chinese word segmentation, filtering meaningless words, calculating word frequency, scoring feature items and so on, the feature vector of the sample is finally obtained, and the final output of this step is a matrix. Whether the feature selection is good or bad will greatly affect the later analysis. +rough text clustering, we can get the distance between these sample points which can reflect the n-dimensional space. +e output result of clustering algorithm operation is a clustering pedigree graph, which can generally reflect all the classification situations, or directly give a specific classification scheme, including a total of several categories, the specific sample points in each cluster, etc. After getting the cluster pedigree, we need to choose the appropriate threshold. After determining the ill value, the platform can directly see the classification scheme through the existing clustering pedigree.

Realization of Sensitive Information Mining in News
Communication Platform. Propensity analysis of network public opinion is essentially to distinguish the network text information and determine whether it belongs to the positive category or the negative category. +e main process of classifier construction is to use the word sequence check suffix tree representation model to calculate, get the similarity calculation results in the feature space, and use support vector machine algorithm to find the optimal classification hyperplane, so as to achieve the purpose of accurate judgment of network information public opinion tendency. Different from structured data, there are polysemy and polysemy in Chinese [19,20]. At the same time, the context understanding of sentences brings challenges to public opinion information monitoring, which needs the support of corresponding natural language processing technology. +e network public opinion monitoring based on the network information extraction and semantic analysis technology cannot understand the deeper semantics, can only stay in the stage of passive monitoring of network public opinion, and cannot realize the automatic identification of network hot spot information, or track the discovered public opinion information. With the development of natural language processing, data mining and other technologies, especially the wide application of search engine, we can efficiently organize the originally scattered information together through the analysis of its relevance [21,22]. Build a sensitive information knowledge thesaurus, through the analysis of users' concerns about sensitive words, for the detection documents, infer the relationship of sensitive information in the knowledge base, determine the query conditions of sensitive words, submit to the search engine, build the basic analysis set, carry out sensitivity analysis, get the sensitivity evaluation of sensitive words, and issue early warning according to the analysis results. +rough the analysis of Web usage records, web structure information and web content information, some quantitative indexes of public opinion are given for decision makers to use, as shown in Table 1.
+e basic task of information collection layer is to collect the rich and various public opinion information from the web pages with various data formats. It provides the required data for the public opinion information mining layer, and is the premise of public opinion deep mining. It is the main task of the Internet public opinion information mining layer to carry out in-depth mining of public opinion information, find the hot issues of public concern, analyze the attitude of the public, and deal with the sensitive information that constitutes harm. By analyzing the data provided by the  public opinion information collection layer, it can detect network topics, analyze people's attitudes, monitor networksensitive information, evaluate public opinion situation, etc., and provide objective basis for the public opinion information service layer to serve relevant departments.
+e sensitive information analysis module mainly includes three ways, namely, association analysis, cluster analysis, and feature extraction. Among them, the common algorithm of association rule mining is Apriori algorithm. +e application of this algorithm must first form all frequent item sets, and then form all credible association rules from these frequent item sets. +e most important feature of this algorithm is to start from the single item, and then filter it layer by layer, so as to get an effective item set, effectively avoiding the search for impossible items. +e public opinion analysis module mainly provides two functions, which are the discovery of public opinion hot spots and the tracking of network public opinion topics. Among them, the application of hot spot discovery can let users know the current hot topics in time, and comprehensively grasp the current network public opinion information. In the process of hot spot discovery, the public opinion monitoring platform mainly obtains information, frequently searched words, browsed web pages, forum replies, and other relevant information based on the keywords entered by users, then monitors the hot spots, and automatically identifies the "hot spot" information in the network, thus forming hot spot alarm [23][24][25]. Topic tracking in public opinion monitoring platform is mainly realized by topic tracking method, which mainly forms tracking expression expressed by query vector from training set, and then uses this tracking expression to judge the newly captured web page information, and finally obtains the information related to the current topic [26,27].

Experimental Analysis
If the threshold is set too low, it will lead to the separation of reports on the same topic. If the threshold is set too high, it will make the topic larger and introduce a lot of irrelevant reports. In the research of Chinese text orientation analysis, there is no open corpus at present https://www.drip.com 16000 comments were collected as experimental data, and emotion categories of the text were manually labeled, including 8000 positive category documents and 8000 negative category documents. A total of 9804 documents are divided into 20 categories. Among them, there are no more than 100 documents in 11 categories such as literature and education, and more than 1000 documents in 6 categories such as computer, environment, agriculture, economy, politics, and sports. Because the training process of genetic algorithm needs a large number of samples, we only select 6 categories with more than 1000 documents. At the same time, because the algorithm will eventually be applied to information filtering, the project team collected 276 and 192 documents of violence and pornography, respectively. As a result, there are 7947 documents in 8 categories. +e distribution of training documents is shown in Table 2: In order to set a reasonable threshold, ten experiments were carried out. In each experiment, firstly, 10 reports were randomly selected from 8 topics in data set 1 to form the original report set; secondly, the reports corresponding to the original report set were selected from data set 2 to form the reconstructed report set; +en, each report in the original report set and the theme block of each report in the reconstructed report set are segmented, feature selected, and weighted, and each topic in the two report sets is represented as a central vector; Finally, cosine similarity is used to calculate the similarity of each report to its topic and other topics in the report set. +is experiment uses eight topics to measure the performance of traditional single-pass clustering algorithm and hierarchical topic detection algorithm in topic detection. +e experiments were conducted five times and the performance was evaluated by the cost of testing. +rough the average measurement of five experiments, the two methods can identify the topic well under the similar threshold settings of different topics, but the detection costs are different, and the change curve is shown in Figure 2.
It can be seen from the figure that when the topic threshold is given, the detection cost of hierarchical topic detection algorithm is lower than that of traditional singlepass topic detection algorithm, which indicates that the former has better topic detection ability. With the different topic similarity threshold setting, the detection cost Mathematical Problems in Engineering fluctuates up and down. When the topic similarity threshold rises from 0.24 to 0.30, the detection cost decreases. When the topic similarity threshold is greater than 0.30, the false detection rate changes to miss detection trend, and the detection rate is improved. +erefore, when the similarity threshold is 0.30, the detection cost is the least and the topic detection performance is the best. In the experiment, the topic threshold BR is set as 0.30, and the subtopic threshold value is in the range of 0.4 to 0.6. A hierarchical subject detection algorithm is used to test the subject. +is method detects the topic and identifies five subtopics, and detects the same number of subtopic reports as the manual identification reports, as shown in Figure 3. By comparing the detection results of similarity mining effect of sensitive information, we can know the discovery that the number of reports of subtopics identified by the hierarchical topic detection algorithm is roughly the same as that of manually annotated subtopics, indicating that the method can distinguish molecular topics and present the hierarchical structure of topics to a certain extent and the accuracy of the proposed method in each category of test data 1.
+rough analysis, we can find that there are some similarities between the two categories with poor classification effect. For example, the political category often contains economic, environmental, agricultural, and other factors, resulting in the low accuracy. In the above experimental data in Table 3, the improved calculation method can achieve better results. However, we cannot rule out that the above experimental results are obtained on the basis of data 1, and there may be some overfitting problems. +erefore, the above second set of test data is used for further test, and the analysis data are as follows: Among the above experimental data, in terms of accuracy in Table 4, although the computer finance and closed test have a slight decline, there is little difference, while the sports class has a big gap. After analyzing the training documents and test documents, we can find that the sportsrelated documents in the original training documents belong to sports theory research, while the test documents come from the network, so there is a big difference between the distance. In view of the purpose of the research which is to apply to content-based information filtering, this experiment is designed to apply the above classifier to the test experiment of network-sensitive information filtering. In the experiment, test data 1 is divided into two categories, legal documents and illegal documents. +e illegal documents are composed of pornographic and violent documents in test data 1, while the legal documents are randomly selected from the other six categories. +e experimental data composition and test results are as follows.
In the experimental data shown in Table 5, in the two numbers before and after each table item, the former one uses the template generation method based on dynamic genetic algorithm to generate the template, but does not use the weight calculation method in this paper to calculate the filtering effect of the filtered information. +e latter one uses both the template generation method based on dynamic genetic algorithm and the template generation method based on this paper. +e weight calculation method calculates the filtering effect of the filtered information. Simultaneously interpreting the data in the above table with traditional methods, the proposed method is obviously better than the traditional method in terms of the accuracy of illegal information. Because the test data used by traditional methods are completely consistent, the proposed method has better filtering effect.

Conclusions
With the increase of the number of network users, the network environment has become more complex. +erefore, the establishment of network public opinion monitoring platform is particularly important. On the basis of sensitive information mining, relevant key technologies are studied, and the design method of network public opinion monitoring platform is proposed to effectively realize the network public opinion monitoring, realize the maintenance of social stability, and promote government departments to make decisions more democratic and scientific.

Data Availability
+e data used to support the findings of the study can be obtained from the corresponding author upon request.

Conflicts of Interest
+e authors declare that they have no conflicts of interest.