M-DA: A Multifeature Text Data-Augmentation Model for Improving Accuracy of Chinese Sentiment Analysis

A neural network based on a word or character embedding is a mainstream model framework in text sentiment analysis and has achieved good results. However, there is a lack of learning about POS-Tagging and Sequence-Tagging . In this research, we propose a multifeature text data-augmentation model (M-DA) with a multiple-input single-output network structure to overcome this problem of Chinese text sentiment analysis. First, this paper sequentially obtains various sequences of Chinese text, including word sequence, pos sequence, char sequence, char_pos sequence, and char_4tag sequence, we use char_pos and the char_4tag to construct a new sequence ( 4tag_pos ) and then use 4tag_pos to mark the characters to obtain the reconstructed characters sequence (char_4tag_pos), so as to achieve the purpose of text enhancement. Then, the Word2Vec method is used to train the initial reconstruction of the character embedding. Finally, the BiLSTM network is used to capture the long-term dependence between the sequences, and the dropout technology and attention are used to improve the accuracy. In the course of the experiment, we also realized that it is better to use the original sequence and the sequence after text enhancement technology as the input of the BiLSTM network. Therefore, our proposed model also discusses the concatenate or dot method to fuse multiple sequences as the ﬁnal embedding. Multigroup comparison experiments are conducted on the data set, and the results show that the proposed M-DA model is superior to the traditional deep learning technology in terms of accuracy, recall rate, f-measure, and accuracy, and the relative time cost is small.


Introduction
Natural Language Processing (NLP) has become an important direction in the field of artificial intelligence, which promotes the continuous development and breakthrough of language intelligence. Text sentiment analysis is the extraction of opinions and tendencies contained in texts with the characteristics of subjective consciousness [1,2]. It is an important research direction in the field of natural language processing and is widely used in various industries [3]. With the continuous development of the Internet, Weibo, Zhihu, Douban, Tianya Forum, Jingdong, Meituan, Eleme, Taobao, etc., the many APPs and online comment platforms continue to increase. e Internet has deeply affected all aspects of life, and the number of Internet users has reached a high limit. While using the Internet to get information, individuals are also contributing to the creation of that information. Speeches, declarations, and other materials are being shared, and online information is exploding. rough text emotion analysis technology, users' views or emotional tendencies are mined from a large amount of user data [4]. Consumers can help themselves make purchase decisions according to the emotional tendencies of other users' comments.
Enterprise managers can understand the market demand through the emotional tendency of user comments, to update and improve the products in time. Government personnel can analyze public opinion based on users' opinions in popular events on social media, to correctly guide the spread of emotions of netizens, effectively control the trend of event development, or provide support for formulating related policies. erefore, text sentiment analysis has a high research value. A neural network based on a word or character embedding is a mainstream model framework in text sentiment analysis and has achieved good results. However, there is a lack of knowledge of POS-Tagging, Sequence-Tagging, etc. In this research, we propose a multifeature text data-augmentation model (M-DA) with a multiple-input single-output network structure to overcome the problem of Chinese text sentiment analysis. e rest of the paper is organized as follows: Section 1 provides the related work, Section 2 provides the text dataaugmentation steps and algorithm, Section 3 provides the language preprocessing. e BiLSTM model based on multifeature text data-augmentation is described in Section 4. Experimental analyses and results are explained in Section 5. e conclusion is given in S.

Related Work
Text sentiment analysis methods can be divided into two categories, based on sentiment dictionaries and machine learning methods. e method based on the sentiment dictionary is to first use the existing sentiment words in the dictionary to match the words in the sentence and then calculate the sentiment words in the sentence to obtain the comprehensive sentiment tendency of the sentence. Kamps et al. [5] used the WordNet dictionary to conduct word-level sentiment analysis. Hiroshi Kanayama et al. [6] proposed a fully automatic dictionary expansion method for domainoriented sentiment analysis. However, this method does not consider the connection between words in the text and lacks semantic information, and the classification effect is too dependent on the quality of the dictionary. e successful application of machine learning in text sentiment analysis has effectively promoted the research and development of text sentiment analysis. Traditional machine learning methods require artificially labeled training sets and artificially designed features for emotional feature extraction and then use text classifiers for classification. Commonly used classifiers include naive Bayes, maximum entropy, support vector machines, etc. [7,8]. Due to the performance of the classifier, the reliance on the number and quality of manually labeled training sets leads to excessive influence of human factors and huge human engineering. Later, methods based on deep learning appeared, and deep learning is an important branch of machine learning. e recurrent neural network [9] is a popular model, and most studies use it as the method's basic module. Zhang et al. [10] used distributed word representation technology and RNN network for sentiment classification. Sequential deep learning model LSTM network [11] solves the problem of gradient disappearance or gradient dispersion in the RNN network [12]. BiLSTM [13] is composed of two LSTMs with opposite front and back directions to obtain contextual features. Zhou et al. [14] performed Chinese sentiment analysis by combining Word2Vec and Stacked Bi-LSTM model. In 2017, the Google machine translation team completely abandoned network structures such as RNN and CNN and only used the attention mechanism [15] for machine translation tasks and achieved good results. Luong et al. [16] proposed global and local attention mechanisms, which promoted the application of attention-based models in the field of NLP. Kokkinos et al. [17] proposed self-attention and applied self-attention to sentiment analysis tasks. Part of the research adds the attention mechanism screening feature at the end of the recurrent neural network. Fei et al. [18] combined the Bi-LSTM model with self-attention to form the SA-BiLSTM method. Long et al. [19] studied the sentiment analysis of Chinese text in social media by combining the BiLSTM network with the multihead attention mechanism (MHAT). It shows that adding attention mechanism screening features can effectively improve the classification effect.
Text representation has been successfully applied to many downstream natural language processing (NLP) tasks as input features, which has a direct impact on the effect of deep learning models [20]. In 1986, Rumelhart et al. [21] first proposed the distributed representation of words used in deep neural language models. Bengio et al. [22] first used neural networks to build language models. Mikolov et al. [23] proposed the Word2Vec [24] technology based on the Log-Bilinear model [25] in 2013, which promoted the rapid development of word vectors. Despite the success and popularity of word embedding, most existing methods use each word as the smallest unit and ignore the morphological information of the word. When optimizing the cost function related to rare words and their context, the rare words cannot be represented well. To solve this problem, recently, Wieting et al. [26] proposed Charagram embedding, in which words or sentences are represented by character n-gram count vectors.
is is an easy way to learn character-based combination models to embed text sequences. Sun et al. [27] proposed two new models to build better word representations by modeling external context and internal morphemes in a joint prediction method, called BEING and SEING.
ese two models can also be extended to learn phrase representations based on distributed morphology theory. Rei et al. [28] proposed a new architecture for combining alternative word representations. In the framework of sequence labeling, both character-level and word-level embedding are used. Rezaeinia et al. [29] showed that the improved word vector (IWV) is very effective for sentiment analysis. Rahimi et al. [30] proposed two new unsupervised models that integrate word polarity information and word cooccurrence into more tailored sentiment analysis features. Word polarity and cooccurrence are clustered together in the form of tension and tension factorization to generate word embeddings. Unlike other alphabet writing systems, the target of Chinese is three different levels of granularity of radicals, characters, and words. Yu et al. [31] proposed a method of joint embedding of Chinese words and their characters and radical subcharacter components and quantitatively evaluated the quality evaluation and word analogy tasks of word embedding learned by the model on word similarity. Peng et al. [32] inspired by aggressive hierarchical embedding [33] designed two fusion mechanisms to merge three granularities and achieved good results on Chinese sentiment analysis tasks. In addition, Wu et al. [34] proposed using dictionary embedding and polarity reversal for sentiment analysis.
A Chinese character can be a word or part of a multisyllable word. However, in the above method, the vector calculation that is used to represent the document does not consider the position or part of speech of the words that are composed. So in this work, we propose a multifeature text data-augmentation model (M-DA). First, this paper sequentially obtains various sequences of Chinese text, including word sequence (word), part-of-speech sequence (pos), character sequence (char), word part-of-speech sequence (char_pos), and character position ;sequence (char_4tag), and then uses character position. e part-ofspeech sequence constructs a new sequence (4tag_pos) and then uses 4tag_pos to mark the characters to obtain the reconstructed characters sequence (char_4tag_pos), to achieve the purpose of text enhancement. For example, after "十分" is divided into characters, "十" will be marked as the first character B_m of the quantifier, and "分" will be marked as the last character E_m of the quantifier. en, the Word2Vec method is used to train the initial reconstruction of the word embedding. Finally, the BiLSTM network is used to capture the long-term dependence between the sequences, and the dropout technology and attention mechanism are used to improve the accuracy. In the course of the experiment, we also realized that it is better to use the original sequence and the sequence after text enhancement technology as the input of the BiLSTM network. erefore, our proposed model also discusses the concatenate or dot method to fuse multiple sequences as the final embedding, to get our final multifeature text data-augmentation model (M-DA).

Text Data-Augmentation
Text Data-Augmentation is also called data-augmentation, which means that the value of limited data is equivalent to more data without a substantial increase in data. Data enhancement technology is already a standard configuration in the image field, and data enhancement is achieved through techniques such as image flipping, rotation, mirroring, and Gaussian white noise. In the field of NLP, Wei et al. [35] introduced NLP data enhancement technology, proposed an EDA model, and showed that the data-augmentation can prevent overfitting and improve the generalization ability of the model. is paper considers that the knowledge of text POS-Tagging and Sequence-Tagging is usually very useful, so this paper proposes to use the position and part-of-speech sequence of the characters in the text to reconstruct the original character sequence, to achieve the purpose of data enhancement. Before obtaining the position and part-ofspeech sequence (4tag_pos) of a word, it is necessary to obtain the word sequence (word), part-of-speech sequence (pos), character sequence (char), word part-of-speech sequence (char_pos), and character position sequence (char_4tag). Finally, use 4tag_pos to reconstruct char as a new text sequence (char_4tag_pos).
Example: "It's really good. Breakfast makes me very satisfied." e text data-augmentation process is shown in Table 1.
e process of text data-augmentation is shown in Table 1. First, the posseg method in jieba is used to obtain POS at the same time as word segmentation. en the words are divided into characters, and in the same way, the POS is iterated one by one to obtain the POS of the character (char_pos). Regarding text Sequence-Tagging, 2tag oversimplifies the model and cannot obtain sufficient information; 6tag makes the model complicated. It is good if the training sample is large enough, but the training set is too small to obtain accurate information through 6tag [36]. erefore, this paper uses 4tag for marking. 'B' represents the beginning position of the word, 'M' represents the middle position of the word,' E′ represents the end position of the word, and 'S' represents a single character. By iterating the words, mark the characters one by one with the mark 4tag to generate a new text sequence (char_4tag). en, the char_pos and char_4tag are merged to obtain the position and POS information corresponding to the character (4tag_pos). Finally, we used 4tag_pos to mark the characters to get the reconstructed text (Char_4tag_pos). e position-POS vector corresponding to the character will always follow the character, which can strengthen the semantic logic of the model. e enhanced dependencies in the text are shown in Figure 1. e quantifier "十分" (very) can highlight the main morphological features of things. e "B" marked character "十" is the beginning of the word, and the "分" marked character B belongs to the part of the word and the end position.

Language Preprocessing
Google's open-source tool Word2Vec converts text strings into numeric vectors, calculates the distance between words, and groups similar words according to their meanings. e CBOW model uses surrounding words to predict the center word, to use the prediction result of the center word, and to use the Gradient Descent method to continuously adjust the vector of the surrounding words. After the training is completed, each word will be used as the central word, and the word vectors of the surrounding words are adjusted so that the word vectors of all words in the entire text are obtained.
is paper uses the CBOW model based on Negative Sampling to construct the above seven sequence vectors. Take, for example, the vectorization of reconstructed text. Given the context of the reconstructed character w, w needs to be predicted. erefore, for a given Context(w), the reconstructed character w is a positive sample, and other reconstructed characters are negative samples. Assuming that a negative sample subset NEG(w) ≠ ∅ about Context(w) has been selected, and ∀w ∈ D, define L w (w) ∈ 1，0 { } to represent the label of the word w.
e label of positive samples is 1, and the label of negative samples is 0.
For a given positive sample (Context(w), w), the optimization goal of this model is to maximize the posterior probability of a given text.

Scientific Programming
In equations (1)-(3), x w represents the sum of the vectors of the reconstructed characters in Context(w), and θ u ϵR m represents an auxiliary vector corresponding to the word u, which is the parameter to be trained.

The BiLSTM Model Based on Multifeature Text Data-Augmentation
BiLSTM network is one of the commonly used neural networks in text tasks. is paper combines the proposed multifeature text data-augmentation model (M-DA) with the BiLSTM network to complete Chinese sentiment analysis tasks. e M-DA-BiLSTM model is a multiple-input single-output network structure. e schematic diagram of the model is shown in Figure 2.
For a given text sentence, use the reconstructed text method to obtain three types of text, such as word, char, Char_4tag_pos, and input the three types of text into the Word2Vec model for training and obtain the word vector w i , w i ∈ R n×k , and the character vector c i , c i ∈ R n′×k′ , and reconstruct the character vector ′ is the vector dimension (vec_dim), and the obtained dictionary index is used to represent the sentence S.
ere are 3 ways of representation: 5.1. Input Layer. Instantiate 3 Keras tensors; the shape values are all set to 128, which means that the input will be a onedimensional vector with 128 elements in this dimension.

Embedding Layer.
In this paper, we use the pretrained Word2Vec model for word embedding. e preprocessed data set provides a unique and meaningful sequence of words; each word has a unique ID. e embedding layer uses Word2Vec pretraining weights to initialize embedding weights and introduces external semantic information, which is often very helpful to the model. e input of an embedding layer should be a series of integer sequences, and all integers in the sequence will be replaced by corresponding columns in the corresponding word vector matrix, which is its word vector. For example, the word vector represents the sentence S as shown in Figure 3.
After passing through the dropout layer, change the mode of unified learning of weights and unified update of parameters in the previous network. In each training iteration, some parameters in the network are learned.

Concatenate Layer. Connect the layers of the input list.
It takes a list of tensors as input, all shapes are the same except for the connected axis, and it returns a tensor connected as all inputs. e sentence S output at time t is shown in Figure 4.

BiLSTM Layer.
It is a combination of forwarding LSTM and backward LSTM. e memory unit in LSTM is controlled by the forget gate f t , the memory gate i t , the temporary memory state c t , the current memory state c t , and the output gate o t , and they are calculated by the hidden state h t−1 at the last moment and the current input s t . See equations (5)- (7). In the calculation, the first hidden state needs to be used, but it does not exist and is generally set to a 0 vector in reality.   Scientific Programming 5 W, U is the weight matrix, b is the bias vector, and tanh, logistic are the activation functions.
At time t, the hidden state output by the forward LSTM is h t → , and the hidden state output by the backward LSTM is h t ← ; then the hidden state h t output by the BiLSTM is merged by the output of the forward and backward LSTM, as shown in

Attention.
is paper uses attention to express the correlation between the words in the text sentence and the output result. First, generate the target attention weight v t , then use the softmax function for probabilism of the attention weight to generate a probability vector p t . Finally, the generated attention weight is assigned to the corresponding hidden layer state h t . Different weights are applied to the state at each moment to ensure that the problem of information redundancy is solved while retaining valid information.
5.6. Dense Layer. e parameter is 1, and the activation function is the sigmoid function. h w (x i ) is the probability that the predicted input x i belongs to category 1, as shown in In the above equations, x i is the sample feature vector, y i ∈ [0, 1] is the true label of the sample, and w is a trainable parameter. e model uses a logarithmic loss function to update the parameter weight matrix w to complete the experience minimization, as shown in e summary function is used to print the M-DA-BiLSTM model, as shown in Figure 5.

Experiments
e experimental data is a commonly used and public data set in this field, [37], the data set is Chinese shopping review text with binary sentiment tags, and the tag∈[0,1], where negative emotions are 0 and positive emotions are 1. Data evaluation objects include multiple types, including hotels, milk, books, and mobile phones. e division of the experimental data set is shown in Table 2.
In the experiment section, we will discuss experimental settings, evaluation indicators, comparison methods, and result analysis in detail.

Experimental Settings.
e experimental environment configuration data are shown in Table 3.
Parameter settings will directly affect the classification effect of subsequent models. e specific parameter settings are shown in Table 4.

Evaluation Indicators.
e four model evaluation indicators of accuracy, precision, recall, and F1 are commonly used standards for NLP model evaluation. Accuracy is an evaluation of the correct classification ability of the model. e higher the accuracy, the better the classification ability of the model. Precision is accuracy, and recall is an evaluation of recall. F1 is a weighted average of precision and recall and is a comprehensive evaluation index. e larger the value, the better the model, as shown in equations (12)- (15).
In the above formula, its specific meaning is shown in Table 5. TP represents positive emotions labeled by the model, which are positive emotions. FP is the positive emotion labeled by the model, which is negative. TN is a negative emotion labeled by the model, which is a negative emotion. FN represents the negative emotion labeled by the model, which is positive. 6 Scientific Programming

Comparison Methods.
e experiment uses the deep learning mainstream network BiLSTM and the classical language pretraining technology Word2Vec as the baseline model and sets the following series of comparative experiments.
(1) Word-BiLSTM [14]: e mainstream method BiLSTM network is used for sentiment analysis, and the model uses Word2Vec technology to train word vectors.
(2) Word-BiLSTM-attention [15,18]: Compared with the model (1), the attention mechanism is conducive to the selection of important features. (3) Char-BiLSTM-attention [26,28]: For characters, the language pretraining models are used to convert characters into vector representations, which is also a popular method. (4) Char_4tag_pos-BiLSTM-attention: e reconstructed text (Char_4tag_pos) proposed in this paper was used as input.       (5), the results of Char_4-tag_pos vector and word vector using the Dot method are input into the BiLSTM network. (7) (Char: 4tag_pos: Word)-BiLSTM-attention: It is a network structure with three inputs and a single output. e inputs are Char, 4tag_pos, and word vector, and then the concatenate method is used to merge the three, so that the overall model can obtain more text representation information. (8) (Char_4tag_pos: Word: Char)-BiLSTM-attention: Compared with model (7), the inputs are the Char_4tag_pos, Char, and word vectors, and then the concatenate method is used to merge the three.
Observe that the Char_4tag_pos vector is better than the 4tag_pos vector. is model is also the final model (M-DA-BiLSTM) proposed in this paper.
In this work, we propose a BiLSTM model (M-DA-BiLSTM) enhanced based on multifeature text data. In addition, the model uses dropout technology and attention mechanism to improve accuracy. e model uses text 4tag sequence and pos sequence to reconstruct char to generate a new sequence (Char_4tag_pos) to achieve the purpose of text data enhancement for char. Here, we need to consider the following three situations. e first is whether the text data enhancement method proposed in this paper has a positive effect when applied to the experimental data set. So we set up experimental groups 1, 2, 3, and 4, send the Char_4tag_pos vector directly to the network, and compare it with the mainstream model that uses word vector or char vector as input. e second is whether the effect of using the enhanced text data sequence and the original sequence as input together is better than a single input. And what is the difference between different fusion methods? So we set up experimental groups 5 and 6, using concatenate or dot method to fuse Char_4tag_pos vector and word vector as input.
ird, what impact will more inputs and more features have on the model? And how do different subelements affect the model? So we set up experimental groups 7 and 8, using concatenate to fuse 4tag_pos, word and char and compare and fuse Char_4tag_pos, word, and char.

Result Analysis.
On the test set, the four model evaluation indicators of accuracy, Precision, recall, and F-1 are used to evaluate the model after the training. e results are shown in Table 6, unit (%).
In Table 6, e Word-BiLSTM model is the mainstream model of sentiment analysis [14]. On its basis, e Word-BiLSTM-attention model uses the attention mechanism proposed in [15] to optimize the model. e Chinese have character-level granularity, which is different from English.
e Char-BiLSTM-attention model uses the character-level vector proposed in [26] as the input of the neural network BiLSTM. By comparing with 3 groups of popular models, the Char_4tag_pos-BiLSTM-attention model has scores of 91.13% and 91.07% on the two comprehensive indicators accuracy and F1, which is better than the control group. e experiment verifies the feasibility and effectiveness of using the reconstructed word vector as the input of the BiLSTM network. e models in Table 7 are all network structures with multiple inputs and a single output so that the model can obtain more textual information. e first two groups of models are the dual-input structure of Char_4tag_pos and word vector. ey use two different methods to merge the two types of vectors. By comparing these two different merging methods, we can know that the concatenate method is 91.53% better than the dot method on the accuracy, and the dot method is 91.43% better than the concatenate method on F1. erefore, Figure 6 shows the time cost of all comparison models. e abscissa (x-axis) is the epochs of iterations when training the model, and the ordinate (y-axis) is the time of each iteration, in seconds (s). It can be seen from Figure 6 that model 1 without attention has the smallest time cost, followed by single-input models 4, 2, and 3, then dual-input models 5, and three-input models 8, 7. e dual-input model based on the dot method has the greatest time cost. Comprehensive analysis shows that the scores of models 5 and 6 are the same inaccuracy. However, the time cost of the dot method is about three times that of concatenate method, so concatenate method is more practical than the dot method.
Compared with model 4, models 5 and 6 are better than the single-input structure of reconstructed word vector inaccuracy, which shows that the dual input of Char_4-tag_pos and word vector can further optimize the model. Based on this, Char_4tag_pos is composed of the basic elements char and 4tag_pos. So, we discussed the three inputs of char, 4tag_pos, and word vector and used the merge method concatenate, which is less expensive in time.
By comparing model 7 with models 5 and 6, we can know that such a three-input model 7 achieves scores of 92.35% and 92.25% on the two comprehensive indicators accuracy and F1, and the results are better. e average time is only increased by about 20 s/epoch compared to the dualinput model 5, and it is much shorter than that of model 6.
Model 8 is the final model proposed in this paper. It is a three-input single-output network structure. e model scores 92.35%, 93.87%, and 92.25% on accuracy, precision, and F-1. e training model takes an average of 102 s/epoch. Compared with model 7, the reconstructed character vector is replaced with the 4tag_pos vector as an input item, and the concatenate merging method can be used to further optimize the model under the premise that the time cost is as small as possible. Figure 7 shows the distribution of the accuracy rate (val_acc) and the change of the loss rate (val_loss) of the 8 groups of comparative models on the validation set. Figure 7(a) uses a box plot to show the distribution of val_acc values during the entire model training process. e abscissa is the comparison model number, which has the same meaning as the label in Figure 6, and the ordinate is the val_acc value. e biggest advantage is that it is not affected by outliers and can describe the discrete distribution of data in a relatively stable way. From Figure 7(a), model 4 is compared with the first three groups of models. In the single-input model, the reconstructed word vector contains the highest value as the input. Comparing models 5 and 6 with the first four groups, the median of the dual-input model is higher, and model 5 contains the highest value. Model 8 contains the highest value of val_acc and the highest median value, with a concentrated value distribution. e output result is the probability that the predicted sample is 1. For the (i) e meaning of bold in "model" column: just to make the model name more visually prominent. (ii) e meaning of the bold values provided in the "Accuracy\Precision\Recall\F1" columns: just to highlight the highest value in each column.   convenience of statistics, the output result greater than 0.5 is classified as 1, and the rest are 0. In Figure 8(a), TP represents the correct number of positive samples predicted by the model, and TN represents the correct number of negative samples predicted by the model. e higher the value, the better. In Figure 8 In summary, for the data set used in this paper, we verify the proposed M-DA-BiLSTM model. Here, we considered the following three situations: In the first, we set up experimental groups 1, 2, 3, and 4, send the Char_4tag_pos vector directly to the network, and compare it with the mainstream model that uses word vector or char vector as input. It can be seen that the Char_4tag_pos vector is more appropriate on the experimental data in this paper, which indicates that the text data enhancement method proposed in this paper has a positive effect when applied to the experimental data set.
In the second, based on the results of experimental groups 1, 2, 3, and 4, we set up experimental groups 5 and 6, using concatenate or dot to fuse the Char_4tag_pos vector and the word vector as input. It can be seen that the effect of using the text data-enhanced sequence and the original sequence as the input is better than a single input, but the accuracy of the two fusion methods is equivalent, so we have analyzed the time cost, and the results show that the time cost of the concatenate method is small.
In the third, based on the results of experimental groups 5 and 6, we set up experimental groups 7 and 8, using concatenate to fuse 4tag_pos, word, and char, and compare and fuse Char_4tag_pos, word, and char. It can be seen that three-input is better than two-input, and Char_4tag_pos is more suitable as a subelement. However, the time cost of three inputs is nearly twice that of dual inputs. erefore, training models with more inputs and more features will require higher and higher experimental hardware and software configuration.

Conclusion
In this work, we proposed a multifeature text data-augmentation model (M-DA). First, this work sequentially gathers several Chinese text sequences, including word, pos, char, char pos, and char 4tag (which we employ), and char pos produces a new text sequence (4tag pos). To achieve the goal of text enhancement, utilize 4tag pos to mark the characters and get the reconstructed characters sequence (char 4tag pos). en, the Word2Vec method is used to train the initial reconstruction of the characters embedding. Finally, the BiLSTM network is used to capture the long-term dependence between the sequences, and the dropout technology and attention are used to improve the accuracy. In the course of the experiment, we realized that it is better to use the original sequence and the sequence after text enhancement technology as the input of the BiLSTM network. erefore, the proposed model compares the experimental results of the concatenate or dot method and chooses to use concatenate. e method fuses multiple sequences as the final embedding, thereby further is paper focuses on binary polarity detection in sentence-level sentiment analysis. In the future, we recommend investigating the effectiveness of our proposed M-DA-BiLSTM for other sentiment analysis tasks (such as aspect level or multiple sentiment analysis).

Data Availability
e data used to support the findings of this study are included in the article.

Conflicts of Interest
e authors declare that they have no competing interests.