Research on the Guidance of Youth Labor Education Based on the “Combination of Education and Production Labor” Program Based on the Deep Learning Model

At present, there is a lack of research on Marx's idea of “combining education and productive labor” and its guiding significance for youth labor education, and no effective teaching model has been formed. In response to this problem, this study proposes a semi-supervised deep learning model based on u-wordMixup (SD-uwM). When there is a shortage of labeled samples, semi-supervised learning uses a large number of unlabeled samples to solve the problem of labeling bottlenecks. However, since the unlabeled samples and labeled samples come from different fields, there may be quality problems in the unlabeled samples, which makes the generalization ability of the model worse., resulting in a decrease in classification accuracy. The model uses the u-wordMixup method to perform data augmentation on unlabeled samples. Under the constraints of supervised cross-entropy and unsupervised consistency loss, it can improve the quality of unlabeled samples and reduce overfitting. The comparative experimental results on the AGNews, THUCNews, and 20Newsgroups data sets show that the proposed method can improve the generalization ability of the model and also effectively improve the time performance. The study found that the SD-uwM model uses the u-wordMixup method to enhance the unlabeled samples and combines the idea of the Mean Teacher model, which can significantly improve the text classification performance. The SD-uwM model can improve the generalization ability and time performance of the model, respectively, 86.4 ± 1.3 and 90.5 ± 1.3. Therefore, the use of SD-uwM in Marx's program is of great practical significance for the guidance process of youth labor education.


Introduction
From the standpoint of the proletariat, Marx opposed the exploitation and oppression of workers by the bourgeoisie, combined with the actual development of the modern machinery industry to construct a theoretical system combining productive labor and education, and demonstrated its inevitability in the process of social development [1].
is theory also had a profound impact on Marx's educational thought and provided important guidance on how the proletariat should carry out educational activities. In our country, it is very necessary to excavate the connotation of Marx's thought of "combining education with productive labor," which has theoretical value and practical significance. is thought of Marx can guide our country to educate and cultivate laborers with professional quality and promote the improvement of productivity, thereby promoting social development [2]. Since the 18th National Congress of the Communist Party of China, General Secretary Xi Jinping has delivered many important speeches on the topic of labor based on Marx's discussion on labor. He highly recognized the important position of labor in socialist countries, and also clarified the importance of cultivating young people's correct labor values and loving physical labor in the new era. General Secretary Xi Jinping clearly stated at the 2018 National Education Conference: "We must educate and guide students to advocate and respect labor." [3] In today's society, we must attach importance to labor education and give full play to the educating function of physical labor and use Marxist labor Ideology arm the minds of young people, guide them to love labor, and train young people into laborers who meet the needs of social development in practice, so that they can better serve my country's modernization drive. At the same time, the "Opinions on Comprehensively Strengthening Labor Education in Colleges and Elementary Schools in the New Era" issued by my country pointed out that labor education is an indispensable part and an important part of education in the education system of socialist countries. We must pay attention to the unique educational value of labor and strengthen modern labor education [4]. rough classroom teaching, labor practice, etc., effective measures are taken to change the attitude of young people towards labor and to curb the phenomena of extravagance, lack of progress, and greed for pleasure in the campus. In this context, it is more necessary to study Marx's thought of "combining education and productive labor" and its guiding significance for youth labor education. e famous Russian educator Ushinsky believes that education is an important hand to guide students to respect and love labor, and it is also an indispensable part of cultivating students' labor habits [5,6] Lenin attached great importance to the combination of education and productive labor. He inherited and developed Marx's idea and pointed out that the ideal realization of the future society cannot be separated from the combination of education and productive labor of the new generation. In a socialist country, labor courses should be an indispensable course in school education. Students can master labor skills by participating in voluntary labor and combine practice and theory to help themselves better master scientific knowledge and become socialist laborers [7]. Makarenko believes that labor has an important impact on a person's future development. Individual labor can improve one's living standard and enhance happiness; in collective labor, people help each other and can establish good interpersonal relationships [8]. In the book "Dedicate the whole heart to the child," Sukhomlinsky said that the beauty of human beings is the brightest in labor, so an important task of education is to make the child's surroundings full of the natural world and through labor a world created and built. Labor has the function of aesthetic education, and schools should make labor a need for students' spiritual life and become a powerful educational force on this basis. Lenin, Makarenko, and Sukhomlinsky all stood from the perspective of the proletariat and believed that the ultimate goal of labor education was to train laborers for a communist society. e discussion of this topic by British scholars can be traced back to the early utopian socialist omas More who required children to learn agricultural knowledge in school, go to work in the fields, and required every youth to learn at least one kind of handicraft. Owen also mentioned that people will form their own characters during labor, so labor education should be paid attention to in childhood [9]. Locke believes that labor can make people physically and mentally healthy and overcome many vices. Students can choose one of crafts, gardening, or agriculture to practice [10]. German educator Comenius said in " e Great Teaching eory" that when realizing the ideal of education through labor, attention should be paid to adapting to the natural order, adapting to human nature and age characteristics. Pestalozzi, a Swiss educator, was the first educator to put the idea of combining education and labor into practice in the history of Western education. He believed that this method could coordinate and develop people's abilities in all aspects and advocated the integration of education and labor [11].
rough the consistency training framework [12], based on the wordMixup data enhancement method, this paper proposes a u-wordMixup (unlabeled sample word mixture) data enhancement method for unlabeled samples. Under the constraint of loss, the enhancement generates high-quality additional training samples and reduces overfitting. Based on the u-wordMixup method, a new semi-supervised deep learning model (SD-uwM) is proposed using the Mean Teacher model for consistent training. e u-wordMixup data augmentation method enhances unlabeled samples with the goal of reducing unsupervised consistency loss, constrains the quality of unlabeled training samples, and reduces model overfitting.
e objective loss function combines supervised cross-entropy loss and unsupervised consistency loss and uses the MeanTeacher method for consistency training to improve the generalization ability of the model. Optimize the key words of the "Combination of Education and Productive Labor" program to achieve the optimal application of youth education.

wordMixup Data Enhancement. wordMixup is a data enhancement method for labeled samples.
e idea is to interpolate the word embedding vectors of two samples to generate a new sample word embedding matrix as an enhanced sample [13]. Given a pair of labeled samples (x i , y i ) and (x j , y j ), perform word embedding to get (x i , y i ) and (x j , y j ), where x i ∈ RN × d, x j ∈ RN × d is the word embedding matrix of text x i and x j , N is the number of words, d is the word vector dimension, and y i and y j are the corresponding class labels. en perform interpolation according to formulas (1) and (2) to obtain a new sample (x ij , y ij ), where x ij is the word embedding matrix of the enhanced sample, and y ij is its class label.
Among them, x k i and x k j represent the word vector of the kth word in the text x i and x j , respectively, λ ∈ [0, 1] is the interpolation weight factor, and x k ij is the word vector of the kth word of the new sample generated by interpolation. Perform word vector interpolation on each word in (x i , y i ) and (x j , y j ) one-to-one to obtain the embedding matrix x ij of the new sample, where y ij is the class label of x ij , and (x ij , y ij ) is the enhanced additional training samples. e wordMixup method achieves good results in supervised text classification. But unlabeled in semi-supervised learning has no labels, how to generate pseudo labels for its 2 Computational Intelligence and Neuroscience interpolation? To this end, based on wordMixup, this paper proposes an improved data enhancement method u-word-Mixup for unlabeled samples, which will be introduced in the following chapters [14].

SD-uwM Semi-Supervised Deep Learning Model
e framework of the semi-supervised deep learning model (SD-uwM) is shown in Figure 1 [15]. It uses the idea of the Mean Teacher model to construct a teacher model T and a student model S, using labeled samples and unlabeled samples, based on supervised crossentropy loss and unsupervised consistency loss objective functions for semi-supervised deep learning [16].
As shown in Figure 1, DL � (xli, yli) Nl i � 1 represents the labeled training sample set, DU � (xui ) { }Nui � 1 represents the unlabeled training sample set, Nl represents the labeled sample set number, and Nu represents the number of unlabeled samples. T is the teacher model, S is the student model, and T and S have the same network structure. LS stands for supervised loss, LUS stands for unsupervised loss, and L stands for consistency objective loss function [17].
SD-uwM trains and learns simultaneously on labeled samples and unlabeled samples based on a consistent target loss function L [18]. As shown in the left half of Figure 1, the student model S computes a supervised cross-entropy loss LS on DL. At the same time, as shown in the right half of Figure 1, the student model S uses the u-wordMixup method to enhance the unlabeled samples. According to the prediction of the enhanced samples by the student model S and the prediction of the unlabeled samples by the teacher model T, the calculation is unsupervised and consistent. e performance loss LUS, LS and LUS together constitute the consistency target loss function L of the SD-uwM model. After many iterations, the SD-uwM model is trained to output the parameters of the student model S as the parameters of the final classification model [19].

u-wordMixup Data Enhancement.
Using the Mean Teacher model idea and consistency training [20], the target loss function L of SD-uwM takes into account both supervised cross-entropy loss and unsupervised consistency loss and is defined as shown in where LS is the supervised cross-entropy loss on the labeled samples DL, LUS is the unsupervised consistency loss on the unlabeled samples DU, and β is the scale coefficient. LS is the supervised loss of the student model S on the labeled sample set DL, which is calculated as Among them, y l i is the true label of the labeled sample x l i , θ′ represents the parameter of the student model S, and p θ ′ (x l i ) represents the predicted pseudo-label of the sample x l i by the student model S, that is, y l i .
Based on wordMixup, a u-wordMixup data augmentation method for unlabeled samples is proposed as part of the SD-uwM model [21]. Different from the wordMixup method, the interpolation operation object of the u-word-Mixup method has no real class label. e u-wordMixup method is shown in Figure 2 [22].
As shown in Figure 2, x u i and x u j represent two unlabeled samples, where x u i ∈ RN × d, x u j ∈ RN × d, N is the number of words, and d is the word vector dimension. e feature interpolation of the word vector is performed for each word in x u i and x u j one-to-one, and the embedding matrix x u ij of a new unlabeled sample is obtained as an additional training sample [23]. e teacher model T predicts the unlabeled samples x u i and x u j to generate pseudo-labels y u i and y u j and performs pseudo-label interpolation ony u i and y u j to obtain y u ij , which is used as the pseudo-label of the enhanced sample x u ij . en the student model S predicts the enhanced sample x u ij to get the predicted label y u ij . Among them, the calculation of feature interpolation and pseudo-label interpolation are as follows: Among them, qλ (x u i , x u j ) represents the u-wordMixup data enhancement transformation, x u ki and x u kj represent the word vector of the kth word of the samples x u i and x u j , respectively, λ ∈ [0, 1] is the interpolation weight factor, and x u kij is the interpolation. e word vector of the kth word of the generated augmented sample. y u i is the predicted pseudo-label of x u i by the teacher model T, y u j is the predicted pseudo-label of x u i by the teacher model T, and y u ij is the pseudo-label generated by interpolation, that is, the pseudo-label of x u ij . Based on consistent training, the pseudo-label predicted by the student model S for the enhanced unlabeled sample x u ij is y u ij , which should be as consistent as possible with the pseudo-label y u ij generated by interpolation, that is, y u ij ≈ y u ij , and ideally the two are equal. erefore, the unsupervised consistency loss LUS is computed as where θ denotes the parameters of the teacher model T, θ′ denotes the parameters of the student model S, and θ is the moving average of θ′. p θ (x u i ) represents the predicted pseudo-label of the sample x u i by the teacher model T, p θ (x u j ) represents the predicted pseudo-label of the sample x u j by the teacher model T, and p θ′ (x u ij ) represents the prediction of the newly generated sample x u ij by the student model S Pseudo tags. λ ∈ [0, 1] is the interpolation weight factor, and qλ(x u i , x u j ) represents the u-wordMixup data augmentation transformation [24].

SD-uwM Model Application.
e algorithm description of SD-uwM is shown in Algorithm 1 [26]. e objective function L of the algorithm takes into account the supervised cross-entropy loss LS and the unsupervised consistency loss LUS, and the constraint enhancement generates unlabeled training samples [27]. In each iteration, according to the objective function L, the parameter θ′ of the student model S is optimized. After many iterations, the parameter θ′ of the optimal student model S is finally obtained.

Analysis of Experimental Results.
e comparison method in the experiment is as follows: (1) SD-uwM: the semi-supervised deep learning model based on u-wordMixup data augmentation proposed in this paper (2) wM-SL: a supervised text classification method based on wordMixup data augmentation method in the literature (3) SL: supervised text classification method without data augmentation (4) Mean Teacher: semi-supervised method applied to image classification in the literature, modified for text classification tasks

Comparison of Target Loss between SD-uwM and Mean
Teacher. In order to verify the effectiveness of the u-wordMixup method, a comparative experiment was conducted on SD-uwM and Mean Teacher models, and the training loss changes are shown in Figure 3. As can be seen from Figure 3, compared to the Mean Teacher model, the SD-uwM model using the u-word-Mixup data augmentation method has lower training loss.
is is because the SD-uwM model objective loss function is more realistic, in which LUS is combined with the u-wordMixup method, aiming to reduce the unsupervised consistency loss, which can improve the quality of unlabeled samples, thereby improving the performance of the model.  Table 1.

Classification Comparison between SD-uwM Model and Other Methods
As can be seen from Table 1, on the three data sets, regardless of whether the network structure is LSTM or TextCNN, the classification accuracy of the SD-uwM model is better than that of the SL, wM-SL, and Mean Teacher models, up to 90.4. e SL model is a supervised learning method, which requires a large number of labeled samples to achieve better performance. e wM-SL model only enhances the labeled samples, and the Mean Teacher model does not use the u-wordMixup method to enhance the samples.
e SD-uwM model uses the u-wordMixup method to enhance the data of unlabeled samples and uses the unsupervised consistency loss constraint to improve the generalization ability of the model. It can be seen from Figures 4 and 5 that with the increase of the number of iterations, although the indicators of the SL, wM-SL, Mean Teacher, and SD-uwM models generally show an upward trend and converge to a certain upper limit, the SD-uwM model classification.
e performance is significantly better than SL, wM-SL, and Mean Teacher models. As shown in Figure 5(a), using LSTM on AGNews, compared with Mean Teacher, wM-SL, and SL, the Macro-F1 of SD-uwM reaches 90.3%, which is increased by 8%, 9.9%, and 14.5%, respectively. is is because SD-uwM uses the u-wordMixup method to target unsupervised consistency loss for unlabeled sample enhancement, which can reduce overfitting and improve classification performance.
As can be seen from the above Figure 5, in order to verify the influence of unlabeled samples on the SD-uwM model, the number of labeled samples Nl � 300 is fixed on AGNews and THUCNews, the number of labeled samples Nl � 200 is fixed on 20Newsgroups, and the number of unlabeled samples is constantly increasing. Compare the classification results of SD-uwM model with SL, wM-SL, and Mean Teacher models. It can be seen that the indicators of SD-uwM and Mean Teacher show an upward trend with the increase of unlabeled samples, but the classification results of SD-uwM are significantly better than that of Mean Teacher, wM-SL, and SL. Using LSTM on THUCNews, compared with Mean Teacher, wM-SL, and SL, the Macro-F1 of SD-uwM has reached 91.4%, an increase of 5.3%, 8.2%, and 13.9%, respectively. It can be seen that since the SD-uwM model uses the u-wordMixup method to enhance the unlabeled samples and combines the idea of the Mean Teacher model, it can improve the text classification performance.

Time Performance Analysis of the SD-uwM Model.
When selecting unlabeled samples, the usual semi-supervised learning algorithm needs to calculate the similarity matrix between the unlabeled samples and the labeled samples, which will increase the time complexity. e SD-uwM model in this paper is random sampling, and there is no need to calculate the similarity between the two samples. e temporal performance comparison between SD-uwM model and typical semi-supervised learning method Cotraining is shown in Table 2.
As can be seen from Table 2, the classification accuracy of the SD-uwM model is 86.4% ± 1.3 and 90.5% ± 1.3 in the two different data sets, respectively, and the classification accuracy of the SD-uwM model is significantly higher than that of the co-training model. e classification accuracy was 83.3% ± 1.2 and 88.4% ± 1.2, respectively. At the same time, the time performance of SD-uwM model is significantly better than that of co-training, in which SD-uwM model time is basically maintained at 0.01-0.02 s, while co-training model has exceeded 30 s. e reason is that when selecting unlabeled training samples, the SD-uwM model is random sampling, and the

Conclusion
It is extremely necessary to study Marx's thought of "combining education with productive labor" and its guiding significance to youth labor education. is paper proposes a u-wordMixup method for the data augmentation of unlabeled samples and combines the consistent training framework and the Mean Teacher method to propose a semi-supervised deep learning model SD-uwM. e model uses the u-wordMixup method to enhance the data of unlabeled samples and takes into account the supervised cross-entropy loss and unsupervised consistency loss to construct a new objective function, so as to realize the teaching optimization of the combination of education and production labor. e findings of the study show that the experimental days are as follows: (1) On the three data sets, regardless of whether the network structure is LSTM or TextCNN, the classification accuracy of the SD-uwM model is better than that of the SL, wM-SL, and Mean Teacher models

Conflicts of Interest
e authors declare that they have no conflicts of interest.   Computational Intelligence and Neuroscience