Key Technologies of Media Data In-Depth Analysis System Based on Artificial Intelligence-Based Big Data

At present, big data related technologies are developing rapidly, andmajor companies provide big data analysis services. However, the big data analysis system formed by the combination method cannot sense each other and lacks cooperation, resulting in a certain amount of waste of resources in the big data analysis system. In order to find the key technology of the data analysis system and conduct in-depth analysis of the media data, this paper proposes a scheduling algorithm based on artificial intelligence (AI) to implement task scheduling and logical data block migration. By analyzing the experimental results, we know that the performance of LAS (Logistic-Block Affinity Scheduler) is improved by 23.97%, 16.11%, and 10.56%, respectively, compared with the other three algorithms. Based on real newmedia data, this article analyzes the content of media data and user behavior in depth through big data analysis methods. Compared with other methods, the algorithm model in this paper optimizes the accuracy of hot topic extraction, which has important implications for media data mining. In addition, the analysis results of the emotional characteristics, audience characteristics, and hot topic communication characteristics obtained by the research also have practical value. *is method improves the recall rate and F value by 5% and 4.7%, respectively, and the overall F value of emotional judgment is about 88.9%.


Introduction
In the field of big data, many excellent products have been tested. rough the combination of these products, a variety of big data analysis systems can be formed. However, in the existing big data analysis system, the parallel processing layer and the data storage layer lack cooperation, which cannot guarantee the locality of tasks, make the system load unbalanced, and ultimately make the system resource utilization rate low. In the existing big data analysis system, the parallel processing layer and the data storage layer lack cooperation. e main reason is that, under the background of the big data environment, there is a large amount of data expansion and the collision of the system's CPU in data processing capabilities. Data processing is full, but a large amount of data is loaded, which affects the processing of the tasks currently being processed, which leads to unbalanced system load.
Artificial intelligence is the study of making computers to simulate certain human thinking processes and intelligent behaviors (such as learning, reasoning, thinking, planning, etc.). It mainly includes the principles of computer realization of intelligence, manufacturing computers similar to human brain intelligence, and making computers achieve higher level applications. e deep data analysis system is a combination of computer, communication, and control. Liu et al. believe that the hydrostatic system is considered an indispensable support structure in heavy-duty machine tools. e calculation and analysis of hydrostatic bearings are always laborious [1]. Li et al. believe that the high-pressure hydrogenation heat exchanger is an important equipment in the refinery, but it has exposed the leakage problem caused by ammonium salt corrosion. erefore, it is very important to evaluate the operating status of the hydrogenation heat exchanger. In order to improve the traditional method of evaluating the operation status of the hydrogenation heat exchanger, they proposed a new method for evaluating the operation status of the hydrogenation heat exchanger based on big data. Aiming at the noise data that is ubiquitous in the industry, they proposed an automatic noise interval detection algorithm. Aiming at the problem that sensor parameters have huge and uncorrelated dimensions, a key parameter detection algorithm based on Pearson correlation coefficient is proposed. Finally, a system-based health scoring algorithm based on PCA (principal component analysis) is proposed to help field operators evaluate the health of heat exchanger hydrogenation [2]. Noguchi et al. believe that as 2020 approaches, Japan's tourism industry is expected to increase. In every possible situation, people need access to stable transportation and communication services and other social infrastructure. NTT is researching and developing solutions to meet these requirements by using big data technology. In their article, they introduced one such advanced, high-performance big data technology and described the results of a field test (Fukuoka test) of services provided to tourists visiting Japan [3]. e calculation and analysis of hydrostatic bearings are always laborious. e diagnostic evaluation system can evaluate the raw material and power consumption of different product companies to avoid the loss caused by the mismatch of raw materials and processes in the production process.
Although there are many solutions for big data analysis systems, the current big data analysis systems also face many challenges. One of them is about the problem of system load balancing. A large-scale big data analysis system is composed of dozens or even hundreds of servers. It is very important to provide efficient data analysis services.
In recent years, artificial intelligence (AI) has become the key to growth in developed countries such as Europe and the United States and developing countries such as China and India [4]. e fields of neuroscience and artificial intelligence have a long and intertwined history [5]. In [6], the author investigates the current state of AI applications in healthcare and discusses its future. e main disease areas using AI tools include cancer, neurology, and cardiology. In [7], the author focuses on decentralized event detection, where sensor nodes use artificial intelligence, data fusion, and distributed pattern recognition performed locally in WSN to collaboratively detect events. In [8], the author emphasizes the most basic features of revolutionary technologies in the 5G era and discusses the relationship between AI and candidate technologies in 5G cellular networks. In [9], the author developed the theory of AI replacement work, which stipulates four intelligent methods required for service tasks: mechanical, analytical, intuitive, and empathetic. In [10], the author provides practical case studies and resource links for AI educators, as well as specific suggestions on how to integrate AI ethics into general AI courses and how to teach independent AI ethics courses. In [11], the author introduced a new artificial intelligence (AI) model for mapping flood sensitivity. e results show that the model is superior to all these models and that the model can be used for sustainable management of flood-prone areas [12]. In [13,14], the authors evaluated the use of artificial intelligence platforms on mobile devices to measure drug compliance in stroke patients during anticoagulation therapy. In [15,16], the author gives a comprehensive review of the application of Al technology in improving the performance of optical communication systems and networks and also reviews applications related to optical network control and management. In [17,18], the author explores the phenomenon of using artificial intelligence in higher education teaching to predict the future nature of higher education in a world where artificial intelligence is part of our university architecture.
Based on the analysis of the big data in-depth analysis system, this paper proposes a scheduling algorithm based on artificial intelligence (AI) that uses task scheduling and logical data block migration as implementation methods and verifies the algorithm through experiment and analysis verification [19,20]. Based on real new media data, this article analyzes the content of media data and user behavior in depth through big data analysis methods. e algorithm model in this paper optimizes the accuracy of hot topic extraction and has important implications for media data mining. In addition, the analysis results of the emotional characteristics, audience characteristics, and hot topic communication characteristics obtained by the research also have practical value.

In-Depth Analysis Method of Media Data
Based on Artificial Intelligence Big data analysis is usually a complex process that contains semistructured and unstructured data for user behavior analysis and decision support [21]. Big data is almost everywhere. As an important asset for the survival of enterprises, it is very necessary to save a large amount of production data. But traditional techniques cannot handle such large data sets, so artificial intelligence is often used to process such data. With the help of ML (Machine Learning) and AI (Artificial Intelligence), complex analysis tasks completed on big data are faster than humans imagine. Artificial intelligence has such excellent working ability in data analysis, which is the main reason why artificial intelligence and big data are now inseparable. Artificial intelligence, machine learning, and deep learning learn from each data input and use these inputs to generate new rules for future business analysis [22].

Hadoop Database.
Hadoop Database is a columnoriented distributed database with high reliability, high performance, and horizontal scalability, all relying on HDFS; HBase can also provide data support for multiple parallel processing frameworks, such as MapReduce and Spark. A data unit (Cell) in HBase contains a data record; each data unit is unique and is identified by a combination of RowKey, Column Family, Column Qualifier, and Timestamp. e row key is the unique identifier of each row of data, and the column family is the division of data attributes. e column modifier represents the description of the column data. e column family and the column modifier together form the data column, and the time stamp represents the data insertion time [23]. Each cell uniquely identifies the data stored in the cell through a key consisting of a row key, column family, column qualifier, and time stamp and performs CRUD operations on the data [24]. In order to describe the relationship between the work content of each component and each component in the big data analysis system more clearly, this paper uses the data model to model the entire system. For big data analysis systems, this paper proposes a discrete-time model [25]. Assume that the entire big data analysis system consists of n virtual machines, and the set of virtual machines is V � v 1 , v 2 , . . . , v n . e data analysis task in the system is submitted to the system to run at different times and then is divided into multiple subtasks to be submitted to different working nodes to run. Here, a working node refers to a virtual machine set and a virtual machine in V. At time t, the set of m tasks running in the big data analysis system is denoted as J t � j 1 , j 2 , j 3 , . . . , j m . For any task j i ∈ J t in the set, S i � s i,1 , s i,2 , . . . , s i,pt is used to denote the p i stages of the task. e storage model of the big data analysis system chooses the HBase storage structure as the modeling target. e entire database consists of k data tables, D � t 1 , t 2 , t 3 , . . . , t k }. Let T i � l i,1 , l i,2 , . . . , l i,qi denote the q i logical data blocks that make up the T i table, where logical data block L i � p i,1 , p i,2 , . . . , p i,wi contains w i physical data blocks [26].

JLQ Algorithm Design (Join Local Queue).
e task of the JLQ (Join Local Queue) scheduling algorithm is to distribute subtasks to the corresponding executor for execution, in order to simplify the model complexity. In the JLQ model, it is assumed that each worker node is configured as a single CPU; that is, only one subtask can be executed at a time. For a big data analysis system with n task execution nodes, JLQ generates a set consisting of n queues, with Q i representing the i-th queue in the set, corresponding to the subtask set assigned to worker node S i . At time t, vector Q(t) � Q 1 (t), Q 2 (t), . . . , Q n (t) represents all subtasks in the system at the current state. Vector Q(t) updates internal data when two subtasks are issued and a subtask has been completed [27].
In this paper, f i (t) represents the execution status of node i's current task, as shown in the above formula. When the value of f i (t) is −1, it means that node i is in idle state at time t and can perform tasks. If node i is performing a task, the value of f i (t) j indicates that the current node is executing a task from the j-th queue.

LAS Algorithm Design (Logistic-Block Affinity
Scheduler). LAS (Logistic-Block Affinity Scheduler) scheduling algorithm is an independent module in the big data analysis system [28]. e main task is to determine the hot spots of the nodes by detecting the status of each working node in the system and, based on the results of the determination, analyze whether the migration of logical data blocks is required to ensure the locality of the program or reduce the possibility of hot spots. In the LAS algorithm, set is used to represent the load state of the n working nodes in the current system at time, where h i (t) represents the load state of node i at time t. e value of h i (t) is as shown below. When the value of h i (t) is 0, it indicates that the node is in a light load state, and the value of h i (t) 1 represents that the working node i is in a hot state [29].
In order to clearly represent the relationship between subtasks and logical data blocks in the big data analysis system, set there are a total of m logical data blocks in the distributed storage system. U i (t) in the model represents the logical data block t with the number i, where U i (t) � w i,1 (t) + w i,2 (t) + · · · + w i,n (t) and n represents the total number of working nodes in the cluster, each addend in U i (t), and w i,j (t) represents the number of logical data blocks i used in working node j at time t.

In-Depth Analysis Model of Media Data Based on Emotion Judgment and Data Characteristics.
e data analysis model means how to construct a structure in the data analysis process for specific implementation and establishment, including what data to collect, what content to extract as the basis for analysis, and which analysis methods, algorithms, and statistical items need to be more appropriate. e results of the analysis now make the data association or change trend clear at a glance, thereby more clearly showing the data development trend and logical association; you can also use the report summary method to present the data items that need to be highlighted, such as statistical sum, mean, proportion, ranking, best value, etc., which clearly show the association, statistics, and changes of data. ey add color to data, quickly convey information, show the relationship between data more accurately, and highlight key points.

Model
Mobile Information Systems e symbol "−i" here means the count variable value after excluding the word whose index number is i in the user corpus. e symbol n e k d,−i indicates the number of subject terms k generated from the user document d e in addition to the current allocation. Symbol n e wk indicates the number of times the subject word w is generated from the subject k. Next, sample the hidden variable l e of the user sentiment index according to the following formula: e symbol indicates the total number of occurrences of all words in user document N e k d . After a series of Gibbs sampling, the values of user document-topic distribution, user topic-word distribution, and user topic-emotional distribution can be calculated approximately according to the obtained parameters. e specific calculation method is shown in the following three formulas:

Emotional
Score Acquisition. e model can be used to calculate the sentiment value of each sentiment word (between −1 and 1). e closer the sentiment value of the word is to 1, the more likely it is to be a positive vocabulary; otherwise it is more likely to be a negative vocabulary. According to this model, the sentiment score for each topic is shown in the formula Q(s) and Q(w) are the sentiment scores of a specific word. Q(s) represents the sentiment word score, and Q(w) represents the text word score. E represents the overall emotional orientation of a topic.

Acquisition of Media Data Sets.
is article uses two parts of data. e collected data mainly comes from media data, which is mainly used to measure the proportion of nearby activities. en there is the friend information data of social media users. is part of the data will become a data source for measuring the connection between the target user and the local people. By using the two parts of the data, we can collect the information we need very well. e media data set contains the following attributes: tweet number, tweet text, time stamp, name and number of the location, and longitude and latitude of the location. For Twitter users, the data set contains the following attributes: user ID, user name, user description, user location, number of friends, and number of tweets. e data set contains 23.45 million tweets and a total of 819,000 users. e earliest tweet in the data set was sent on July 11, 2018, and the last tweet was posted on December 1, 2019. erefore, the total time interval of the case study data set used is 505 days. Table 1 shows the data source characteristics of the original data set and the sample data set. e sample data set contains 304,000 tweets totaling 10,000 users, and about 12% of the tweets have coordinates. And the average number of tweets, proportion of coordinated tweets, and time interval of each user in the sample data set are basically consistent with the original data set.

Experimental Environment of Deep Data Analysis System.
In this paper, the distributed environment composed of Spark and HBase is used as the big data analysis system used in the experiment.
e entire system is deployed on the Alibaba Cloud ECS server. A master node is used in conjunction with 4 slave nodes. e LAS module is placed on the master node to run, and there is no separate virtual machine setup. e virtual machine hardware settings are shown in Table 2 and the software version is shown in Table 3 throughout the experiment.
Each section in the big data analysis system uses CentOS 6.5 as the virtual machine operating system, as shown in Table 2. e CPU model of the physical machine where the master node and the slave node are located is the same. e five nodes all use 4-core CPU and 8 GB memory configuration. e difference is that, in addition to the 40 GB hard disk used by the system, the master node mounts an additional 50 GB hard disk, and each slave node mounts an additional 200 GB hard disk for environment deployment and experimental data storage. e software environment used in the experiment is shown in Table 3. e entire system adopts the stable version combination of Hadoop 2.6.0, HBase 1.0.3, and Spark 1.4.0 as the big data analysis system used for experimental testing, and the basic operating environment JDK version is 1.8.0.

Evaluation Index of Data Distribution.
In order to compare the performance of various algorithms in the same environment in the experiment, this paper presents an important data distribution evaluation index: distribution deviation. e distribution deviation is expressed by λ, and the value is in the interval [0, 1], which represents the balance between the distribution of all logical data blocks in the database and the Spark data analysis task.
In order to calculate the value of λ, we first need to introduce three important matrices: A, B, and D. e following formula is of matrix A expression, where n represents n working nodes in the system, and k represents k tables in HBase. e temporary expression C i, j, l indicates whether a logical data block l in HBase belongs to table j and is distributed on worker node i.
e matrix B expression is shown in the following formula. e k value has the same meaning as the matrix A, and m represents m data analysis programs in the system. e temporary expression E i, j, l indicates whether the lth stage of the data analysis task in the system needs to access the data table i, as shown in the formula. erefore, each element b i, j in the matrix B represents the data access relationship between the table numbered i and the table numbered j in the system. e larger the value of b i, j , the closer the data dependence.
Matrix D is the product of matrix A and matrix B; see the formula below. e matrix D represents the connection between the data analysis program in the cluster and the data distribution on each node.

Algorithm Evaluation Index.
During the experiment, three scheduling algorithms were selected for comparative experiments: HFS, JSQ-MAX Weight, and LTF. In the experimental analysis process, in order to more intuitively complete the performance comparison between the comparison algorithm and the logical data block affinity scheduling algorithm, this article uses query execution time (QET) as an important comparison condition for the comparison experiment. QET represents the program execution delay, that is, the time from the start of sending a data request to the acquisition of all data, as shown in the following formula. QET other represents the query time of the comparison algorithm, and QET LAS represents the query time of the affinity scheduling algorithm. If Imp(other) is greater than 0, it means that the efficiency of the logical data block affinity scheduling algorithm is higher than that of the comparison algorithm. Otherwise, it means that the efficiency is lower than the comparison algorithm.

Simple SQL Query.
is article selects Q12 as a comparative use case for simple SQL query operations. Q12 query description is as follows: given the year, month, and category of media data, find out all users who have searched for media data of the specified category on the web site for that month and have used this type of media data in the following three months.    Figure 1. It can be seen from the comparison and analysis in Figure 1 that, under simple SQL query operations such as Q12, the distribution deviation λ has a certain impact on the operation time and the execution time of the algorithm. In the three comparison algorithms, the data query time becomes longer with the increase of the distribution deviation λ, indicating that the original distribution state of the data will affect the execution efficiency of the three comparison algorithms. e execution time of the proposed algorithm is not affected by the distribution deviation λ and is shorter than the execution time of the three algorithms. It proves that, in a simple SQL query program, the performance of the proposed algorithm is significantly improved.

Distributed SQL Query and Analysis.
Choose Q2 as a comparative example of distributed SQL query and analysis. Q2's query description is as follows: given a type of media data, find the products that are often browsed online with the media data and ranked in the top 30 page views, and the user session timeout is set to 60 min. e query time of Q2 in each algorithm is shown in Figure 2. As can be seen from Figure 2, among various algorithms, the JSQ-MaxWeight algorithm generally performs the longest in distributed SQL query and analysis operations, and there are large data fluctuations. e HFS algorithm and LTF algorithm gradually increase with the distribution deviation λ, and the query time is similar. ere are still cases where the value of the distribution deviation λ becomes larger and the algorithm execution time becomes longer. e experimental results of the logical data block affinity scheduling algorithm are relatively stable; that is, when the distribution deviation λ value is 0, the data query and processing time are similar to the other three algorithms. However, as the distribution deviation λ becomes larger, the polyline of the execution time before and after is relatively smooth. e execution time of each data distribution is the shortest of the four algorithms.

Data Analysis Realized by Custom
Function. Q1 is a typical data analysis implemented by a custom function. e query about Q1 is described as follows: Find out 100 types of media data that are often browsed together in the specified media data. e execution time of Q1 in each algorithm is shown in Figure 3. It can be seen from the analysis in Figure 3 that the peak value appears in the line chart during the change of the distribution deviation λ in the three comparison algorithms. When the distribution deviation λ value is 0.5 to 0.6, the program execution time of the three comparison algorithms becomes significantly longer, and the results of the JSQ-MaxWeight algorithm fluctuate greatly. Among the four algorithms, the data affinity scheduling algorithm is relatively stable, is less affected by the data distribution, and has the best improvement effect.

Comparison of Media Data Testing and Adaptability Analysis Results under the Background of Big Data Based on Artificial Intelligence.
e media data set uses the data format and collection time as the table division scheme, which represents all the data collected on May 29, 2019. In fact, the tasks X1, X2, and X3 performed by the data set are all the same data analysis program but are only used for data analysis performed on different dates, and the data accessed is irrelevant to each other. e test results of media data under the background of    big data based on artificial intelligence are shown in Figure 4, respectively, the test results of the media data set and the comparison results of the polyline. rough calculation, the performance of the LAS algorithm is 23.97% higher than the HSF algorithm, 16.11% higher than the JSQ-MAX algorithm, and 10.56% higher than the LTF algorithm. It can be seen from the chart analysis that in the big data analysis system for multitask execution, various algorithms fluctuate to a certain extent as the distribution deviation λ changes from 0 to 1. In general, the larger the value of the distribution deviation λ, the longer the program execution time. When the distribution deviation λ is 0, the effect of the logical data block affinity scheduling algorithm and the LTF algorithm is almost the same, but as the distribution deviation λ becomes larger, the logical data block affinity scheduling algorithm is smaller than the LTF algorithm in execution time. e LTF algorithm is better than the HFS algorithm and the JSQ-MaxWeight algorithm. e HFS algorithm program has the longest execution time and the affinity scheduling algorithm program has the shortest execution time. e scheduling results are the best among all algorithms, and the performance of the big data analysis system is fully utilized. e test results of the media data in-depth analysis system are shown in Figure 5. rough the above analysis, it can be seen that the logical data block affinity scheduling algorithm has a good performance advantage in a large data analysis system in which multiple data query operations and multiple applications coexist. When the distribution deviation λ value is 0, the logical data block affinity scheduling algorithm and the optimal algorithm among the three comparison algorithms have little difference in execution time. However, with the increase of the distribution deviation λ, the execution time of the three comparison algorithms obviously increased or fluctuated. e data affinity scheduling algorithm has a small change interval during the execution of a single application, and there is a certain degree of increase when multiple applications are parallel, but the relative change is not large. When the distribution deviation λ value is 0, it means that each table in HBase satisfies load balancing in units of tables; when the distribution deviation λ value is 1, it means that all tables corresponding to the task are on the same node. From the above analysis, it can be seen that the effect of the data affinity scheduling algorithm when the data is in the best distribution state is not much different from the Spark default scheduling algorithm HFS. But it is applicable to the situation where the data distribution is relatively uneven.

In-Depth Analysis Results of Media Data under the
Background of Big Data Based on Artificial Intelligence 4.3.1. Emotional Judgment. Figure 6 compares the results of this model and the basic method. After calculation, in the judgment of the emotional characteristics of new media content, the accuracy rate of the method in this paper is 87.7% for positive emotion judgment, 91.1% for negative emotion judgment, and 88.1% for neutral emotion judgment. e overall accuracy rate of emotional judgment is about 89.0%, and the accuracy rate has increased by 4%. e judgment recall rate for positive emotions is 88.1%, the recall rate for negative emotion judgments is 88.4%, the recall rate for neutral emotion judgments is 89.8%, the overall recall rate for emotion judgments is about 88.8%, and the recall rate is increased by 5%. e F value for positive emotion judgment is 88.0%, the F value for negative emotion judgment is 89.7%, the F value for neutral emotion judgment is 88.9%, and the overall F value for emotion judgment is about 88.9%, an increase of 4.7%. After testing, our model based on emotional judgment and data characteristics has better judgment results than other research results, which improves the accuracy of judgment, recall rate, and F value, and can meet the relevant requirements of the application. Figure 7 reflects the distribution ratio of sentiment tendency of the article headlines of today's Toutiao platform and Yidian information platform. From Figure 7, we can see that nearly 34% of the article titles of a little information platform show negative emotions, nearly 46% show neutral emotions, and nearly 20% show positive emotions. Today's Toutiao platform's article titles have 32% negative emotions, 44% neutral emotions, and 24% positive emotions. It can be seen from this that although the sentiment span of the article headlines of today's Toutiao platform is large, the sentiment distribution is more scattered. But overall, the headlines of today's headlines are more positive.  Figure 8 reflects the calculation results of the richness of the content of each new media platform. We arranged the calculation results of the articles on the two platforms in descending order, using the article serial number as the horizontal axis and the content richness value as the vertical axis. When the content richness calculation result is greater than 1, the content richness is higher than the average level; when the content richness calculation result is less than 1, the content richness is lower than the average level. It can be seen from Figure 8 that the content richness of today's headlines and little information platform articles conforms to the average distribution. About 37% of the article content belongs to the relatively rich content, and the remaining 63% of the article content belongs to the relatively rich content. At the same time, comparing the distribution of the two curves in the figure, we can find that the content richness of today's Toutiao platform articles will be slightly higher than a little information. Figure 9 reflects the results of user engagement calculations for each new media platform content. We arranged the calculation results of user engagement of the two platforms in descending order, using the article serial number as the horizontal axis and the user engagement value as the vertical axis. When the user participation calculation result is greater than 1, the user participation degree of the content is higher than the average level; when the user participation calculation result is less than 1, the user participation degree of the content is lower than the average level.
We can see from Figure 9 that the user participation of today's headlines and little information platform articles conforms to the long-tail distribution. About 20% of the content of the article belongs to the content with high user       engagement, and they tend to attract 80% of user reviews and user interactions. e remaining 80% of the content of the article belongs to the content with low user participation, and they tend to divide the remaining 20% of the comment volume and interaction level. e conclusion of this analysis is in line with the law of practice. e more popular reviews and interactive articles will attract more people to comment on the interaction, and the topic is easy to ferment. At the same time, comparing the distribution of the two curves in the figure, we can find that the user participation of Toutiao platform today is closer to that of a little information platform.

Conclusions
Based on the analysis of the big data in-depth analysis system, this paper proposes a scheduling algorithm based on artificial intelligence (AI) that uses task scheduling and logical data block migration as implementation methods and verifies the algorithm through experiment and analysis verification. By conducting multiple sets of experiments in the existing big data set, the experimental results are analyzed. e performance of the LAS algorithm is 23.97% higher than the HSF algorithm, 16.11% higher than the JSQ-MAX algorithm, and 10.56% higher than the LTF algorithm. Obtain the scheduling algorithm to improve the execution time of the program in single-task execution and multitask parallelization, respectively, and find the applicable scenario of the scheduling algorithm by analyzing the results during the operation of multiple programs.
Based on real new media data, this article analyzes the content of media data and user behavior in depth through big data analysis methods. Compared with sentiment analysis based on the basic sentiment dictionary, the analysis accuracy is optimized. Compared with other methods, the algorithm model in this paper optimizes the accuracy of hot topic extraction, which has important implications for media data mining. In addition, the analysis results of the emotional characteristics, audience characteristics, and hot topic communication characteristics obtained by the research also have practical value. is method improves the recall rate and F value by 5% and 4.7%, respectively, and the overall F value of emotional judgment is about 88.9%.

Data Availability
No data were used to support this study.

Conflicts of Interest
e author states that this article has no conflicts of interest.