Neural Network-Oriented Big Data Model for Yoga Movement Recognition

The use of computer vision for target detection and recognition has been an interesting and challenging area of research for the past three decades. Professional athletes and sports enthusiasts in general can be trained with appropriate systems for corrective training and assistive training. Such a need has motivated researchers to combine artificial intelligence with the field of sports to conduct research. In this paper, we propose a Mask Region-Convolutional Neural Network (MR-CNN)- based method for yoga movement recognition based on the image task of yoga movement recognition. The improved MR-CNN model is based on the framework and structure of the region-convolutional network, which proposes a certain number of candidate regions for the image by feature extraction and classifies them, then outputs these regions as detected bounding boxes, and does mask prediction for the candidate regions using segmentation branches. The improved MR-CNN model uses an improved deep residual network as the backbone network for feature extraction, bilinear interpolation of the extracted candidate regions using Region of Interest (RoI) Align, followed by target classification and detection, and segmentation of the image using the segmentation branch. The model improves the convolution part in the segmentation branch by replacing the original standard convolution with a depth-separable convolution to improve the network efficiency. Experimentally constructed polygon-labeled datasets are simulated using the algorithm. The deepening of the network and the use of depth-separable network improve the accuracy of detection while maintaining the reliability of the network and validate the effectiveness of the improved MR-CNN.


Introduction
Gigabytes of images are generated every day in the Internet, which contain a huge amount of information. People need to process these images in order to browse and retrieve them effectively. Image retrieval has become very active in related research areas since the 1970s. e advancement of image retrieval is also inseparable from the development of database management systems and the effective promotion of computer vision as a field. Most of the information received by humans comes from vision, and vision-based image retrieval is particularly important. Among the components of image retrieval, human action behavior classification is a very important part, and an in-depth investigation into human action image classification can improve the efficiency of retrieval containing human action images [1][2][3][4]. Research related to image-based human action recognition in the field of pattern recognition has become a cutting-edge research topic. Human action recognition focuses on recognizing the action or behavior of a person from a single image. Given the limited sources of information and the complex background of images collected from the web and the presence of a large number of still images on the web, action recognition based on still images requires the development of robust and effective methods to better understand the web images for image retrieval or search. At the same time action recognition has great utility and broad application prospects [5][6][7][8]. Action recognition systems can be subsequently applied to exercise data evaluation, intelligent training assistance, etc., for example, calculating exercise exertion, training scoring, and virtual coach teaching [9][10][11][12].
Yoga, as a convenient and fashionable form of fitness, can relieve anxiety and regulate mental state and enhance body immunity [13,14]. When people learn yoga, the mainstream learning method is still to enroll in offline yoga training courses, but the practice time and venue are not fully free, and the learning content and progress are usually mastered by the teacher, which brings many inconveniences to the workplace people's exercise in [15]. At the same time, the current level of instructors varies, and the professional quality of teachers is more difficult to guarantee when full understanding is not conducted, so many people choose to find resources from the Internet for self-learning. is way is very convenient, but the accuracy and scientific nature of the exercise cannot be guaranteed, and unreasonable exercise habits and wrong postures can cause physical injury, which is contrary to the original purpose of exercising [16][17][18].
In order to solve the problem, this paper is divided into the following five chapters: Section 1 briefly introduces the research background of this paper, the current status of the research, and the structure of the paper; Section 2 briefly introduces the research progress and shortcomings of yoga movement recognition and also elaborates and describes the significance and main contents of this paper. Section 3 specifically introduces the MR-CNN-based yoga action recognition network. e feature pyramid technique is used in the feature extraction part to improve the performance of the network in multiscale target detection. Regions of interest are extracted using a region candidate network for target classification and detection after passing RoI Align, while images are segmented using mask branches. Finally, the improved deep backbone network and the improved mask branching network are applied to the yoga action recognition network to improve the accuracy of target detection. Section 4 firstly introduces the evaluation indexes of the adopted recognition performance and then validates and evaluates the recognition effect of the MR-CNN-based yoga action recognition model. e experimental results show that the scheme proposed in this paper has high accuracy for yoga movement recognition. Section 5 briefly summarizes the research and practical contents of this paper and describes the shortcomings of the research contents and the outlook of future research work.

Related Works
Yoga action recognition belongs to a type of action in human action recognition, which is mainly applied in the field of sports and can also be extended to other types of action recognition [19]. Human action recognition is jointly implemented by many different and intersecting disciplines, such as machine learning, artificial intelligence, sensor technology, and computer vision. ere are different ways of acquiring human movements, according to which they can be classified as, first, human movement recognition based on wearable sensors and, second, human movement recognition based on vision [20].
Wearable sensors use sensors fixed to key locations on the human body to capture motion data and analyze the computational data to recognize the actions performed by the human body. is type of action recognition system has more comprehensive data analysis capability, but requires higher equipment and expertise, and the acquisition of parameters is more inconvenient. Obviously vision-based action recognition has a higher universal applicability. Meanwhile video and image have become the main carriers of information due to the development of network technology, and the activities of human society are mainly recorded in video and image. Whether from the perspective of entertainment, sports, surveillance, and security, the study of human action recognition in images is of great significance. Human action recognition is mainly used to analyze and understand human actions by processing and analyzing image data or image sequences. e researcher divided human motion recognition into four stages: initialization system, bone extraction, pose estimation, and pose recognition [21]. e researcher designed a pose recognition model as based on the hidden Markov model; this model models human motion in a cascade form, and the main feature is the use of expectation maximization algorithm in order to ensure reliability and accuracy [22]. Some researchers designed a yoga pose recognition system for self-training and used a star algorithm to extract the human star skeletal point vector to detect the ongoing yoga poses [23]. e researcher designed an electronic yoga teaching system based on somatic devices, using the Hausdorff algorithm and others to assess the similarity of poses and determine the names [24].
A variety of different features are widely used in yoga action recognition, including location, contour, and timespace features, which are specifically classified as static, dynamic, and spatiotemporal features [25][26][27]. Static features are color, size, contour, edge, object shape, and depth from which behavioral details and contour states are extracted. Dynamic features are direction, velocity, trajectory, etc., from which motion patterns are obtained. Spatiotemporal features are used for video and image sequences to extract 3D data models such as spatial and spatiotemporal cubes.
Although many scholars have achieved certain research results, the accuracy of yoga action recognition still needs to be improved as it is easy to lose the preliminary feature information in the process of action recognition. In order to further improve the accuracy of yoga action recognition, the significance of this paper is to improve the accuracy of yoga action recognition and fully exploit the existing resources to further promote the development of intelligent sports industry. In view of the fact that the research on yoga movement recognition technology is in the development stage, this paper carries out the research on yoga movement recognition technology based on MR-CNN.
is paper provides theoretical support for yoga movement recognition, which is of great practical significance to the field of intelligent sports and has long-term application prospects for the development of intelligent software and hardware.
Action classification and localization belong to two tasks of target detection, and the most representative algorithm in target detection is faster region candidate convolutional network (Faster R-CNN). In order to achieve further yoga action segmentation tasks, this paper extends the Faster R-CNN and selects the MR-CNN algorithm as the detection and segmentation network for the study. Considering that MR-CNN achieves action classification and localization with the addition of mask branch for segmentation of yoga actions, this chapter uses MR-CNN to implement the detection and segmentation task, improves the feature extraction backbone network, uses the improved depth residual network with the addition of batch normalization layer instead of the traditional depth residual network, improves the mask branch, and uses the depth-separable convolution instead of the traditional convolution. Compared with the traditional convolutional neural network, the residual network adds shortcut connections and adds a batch normalization layer to the residual block to improve it. e superimposition of the residual block constitutes an improved deep residual network. e improved deep residual network can achieve a smaller average error after multiple trainings and has a higher detection accuracy.

Faster Region Candidate Convolutional Neural Networks.
Currently, target detection can be classified into two types of neural network algorithms based on regression and candidate region. e neural network algorithms based on regression have a high computational efficiency and perform rule and dense sampling using features such as aspect ratio, scale, and position to detect targets. However, the detection accuracy is lower compared to the neural network algorithms based on candidate region. One of the main reasons is due to the category imbalance problem. is paper focuses on a target detection algorithm based on a two-stage detector. In the first stage of the two-stage approach is the generation of candidate target frames, and in the second stage further classification and regression are performed. e optimal performance is obtained for the two-stage approach on several challenging datasets such as PASCAL VOC and MS COCO. e network framework for faster region candidate convolutional neural contains two parts, the extracted candidate frame part and the target detection part. It can be divided into four parts: first, using the image classification model as the backbone network to extract the image features; second, inputting the image features to the region candidate network to obtain the candidate regions; third, inputting the results obtained from the previous two steps, i.e., image features and candidate regions, to the RoI Pooling layer to obtain the integrated candidate region features; fourth, predicting the bounding box of the object and the class of the object based on the candidate region features.
Region proposal network (RPN) uses a convolutional network to construct candidate regions without considering categories by sliding convolution over the features. e network uses classifiers with only two categories: the category with targets and the category without targets. RPN performs ka RoI prediction for each point in the feature map. erefore, RPN outputs 4 × kone RoI coordinate and 2 × kone target score for each pixel location.
A set of anchor points is generated for each point on the processed convolutional feature map, and the anchor points generated on the feature map also need to be mapped to the size of the input image in the end. e feature extraction process includes only convolution and pooling layers, so the final dimension of the feature map is proportional to the original image. If the size of the image isw × h, then the final feature map is compressed to a size of w/rand h/r, where ris the subsampling rate. If an anchor point is defined at each spatial location on the feature map, the anchor points of the final image will be separated by rone pixel, and Figure 1 shows the concrete framework of the network implementation.
Network specific implementations are as follows: (1) Firstly, the features of the original image are extracted using a convolutional neural network, which is a commonly used image classification backbone.
Formation of feature maps is as follows: (2) In the feature map generated by the sliding scan, the sliding window size is n × n, and the lower dimensional feature vector obtained in the next step is derived from the convolutional layer mapping the sliding window position. e sliding window is designed as3 × 3. Although n � 3 looks small, each rectangular window frame is perceptible to a large extent considering the small size of the higher-level feature map itself. After mapping to the low-dimensional feature vector using Rectified Linear Unit (ReLU), and considering the kkinds of possible anchor frames for each sliding window position without considering beyond the bounding anchor frame, nine candidate regions will be predicted with feature map size W × H, and the region proposal will yield W × H × kone.
(3) After the region suggestion, there are two fully connected layers, regression layer and classification layer, in which the input is a low-dimensional feature vector, and the regression layer serves to generate the bounding box corresponding tok, k, (x, y, w, h), to ensure that the candidate box does not exceed the image boundary, to crop the part that exceeds it to be close to the edge and to determine whether the candidate region is the background part or the foreground part and score it. In order to accurately represent the coordinates of the candidate regions, the number of results in the window regression layer is 4k, and the number of results in the classification layer is 2k, indicating the probability of the candidate kregions being background and foreground, respectively. e goal of the task is to unify the bounding box P and the ground truth G. e idea is not to learn the specific coordinates of G, but to learn the scale of the deformation Computational Intelligence and Neuroscience performed during the transformation of the bounding box. e idea of the transformation is to move the coordinates of the bounding box positioning (x, y) and to scale the size of the bounding box according to the ratio (w, h).
Suppose the original predicted bounding box is P(x, y, w, h), and the calibrated bounding box isG(x, y, w, h), where x, y, w, h denote the coordinates of the center point of the bounding box and the width and height, respectively. In order that the regression window obtained after the mapping transformation of the bounding boxPcan be closer to the real window G, learning the transformation relation of the bounding box is the main goal. at is, given (P x , P y , P w , P h ), find a mapping f, such that P x , P y , P w , P h � G x , G y , G w , G h ≈ G x , G y , G w , G h . (1) Perform the translation first: (Δx, Δy), Δx � P w d x (P), Δy � P h d y (P), G y � P h d y (P) + P y .
en, perform scaling: erefore, it can be seen from the above formula that the four transformation parameters to be learned are d x (P), d y (P), d w (P), and d h (P). P is not the true value G but the predicted G value after the transformation of the four parameters; then, the objective function can be expressed as d * (P) � w T * Φ 5 (P), Φ 5 (P) is the feature vector of the input Proposal,w T * is the parameter to be learned, where * denotesx, y, w, h , andd * (P) is the calculated predicted value. According to the distance relationship between the predicted value and the true value, the loss function is obtained as e function optimization objective is In the research of action recognition, the depth and step length of the network usually restrict each other. Common solutions to this problem include image pyramids and feature layering. e multiscale training and testing of image pyramids are time consuming and computationally intensive, making it difficult to apply in practice. Feature layering, i.e., each layer predicts the detection results for the corresponding scan resolution separately, allows different feature layers to learn the same semantic information. However, since in convolutional neural networks different layers correspond to different semantic features at their respective levels, shallow networks with high resolution learn more detailed features, and deeper networks with low resolution learn more semantic features. e feature pyramid network (FPN) improves on this problem by introducing feature maps for each resolution into the latter one scaled by twice the resolution to do the summation operation. Since this method only adds additional cross-layer connections to the original network, it adds almost no additional time and computation in practical applications. e network structure is characterized by the ability to fuse the features of each layer and strengthen the semantic information while enhancing the spatial information, and the network structure is shown in Figure 2.
e left-hand model of the FPN structure is the bottomup part, with bottom-up paths for feature extraction, using a skeleton network for computation. e model on the right is the top-down part, using nearest-neighbor upsampling for upsampling starting from the highest layer instead of the deconvolution operation, which is simpler to implement and can effectively reduce the training parameters. e horizontal arrow is a lateral connection that fuses the result obtained by upsampling with the feature map generated from the bottom-up.

MR-CNN Framework and Structure.
e framework in this paper is extended from the fast region-convolutional network by adding a semantic segmentation branch to perform predictive segmentation for each candidate region, while being parallel to the existing backbone used for classification and bounding box regression, and the overall instance segmentation framework is structured as in Figure 3 where the segmentation task is mainly implemented by the extended branch, which is a full convolutional network acting on each candidate region to perform prediction at the pixel level segmentation regions. While common instance segmentation systems perform classification on top of segmentation completion, MR-CNN is implemented in parallel with classification and segmentation.

Computational Intelligence and Neuroscience
RoI pooling is not aligned pixel by pixel, which has little effect on the bounding box but has a significant impact on the accuracy of the mask.
After using the RPN network candidate window, the predicted targets are processed afterwards. Since the predicted regions vary in size and resolution, a uniform quantization operation is required before the extracted features are input to the fully connected layer. Since the network has an FPN part as well as a segmentation image task, the traditional ROI Pooling layer is not suitable, so MR-CNN uses the ROI Align layer for its optimization. e specific steps of traditional ROI Pooling are as follows: (1) Based on the input image, the ROI is mapped to the corresponding position of the feature map, and a rounding operation is performed during the calculation, i.e., the first quantization. (2) e mapped region is divided into parts of the same size, the number of parts divided is the same as the dimensionality of the output, and the forensic operation is performed when the size of each part of the region is calculated, i.e., the second quantization. (3) Maximum pooling operation is performed for each part.
After the above steps, which mainly process the boxes with different sizes, the resultant output feature map is of fixed size, which can achieve ROI, and convolution feature map size does not affect the output feature map size and can improve the processing speed. However, it can be seen that after two quantizations to round the floating-point numbers, the candidate regions originally mapped on the feature map will produce deviations, which will have an impact on the regression localization in the later layers. In order to solve the error caused by the quantization operation, the optimized candidate window processing eliminates the quantization operation and uses bilinear interpolation when obtaining the values on the pixel points of the floating-point coordinates as follows: (1) Iterate over all candidate regions to ensure that the floating-point boundaries are not quantized. In the target segmentation task, a L mask component is added to calculate the cross-entropy of the predicted target segmented image compared to the conventional detection network. Since the overall network is in multitask learning mode, the loss function is as follows: where L class represents the classification loss, L bbox is the target detection regression loss, and L mask is the segmentation loss.

Classified Losses.
During the training process, the target detection network generates an mregion recommendation window. Letp be the probability of correct classification, and the network is used to achieve the multiclassification task, so the cross-entropy function is usually chosen as the loss function, and the formula is as follows:

Target Detection Coordinate Regression Loss.
For the prediction of edges, which is a regression problem, a squared loss function, or L2 loss, can usually be chosen. e L2 loss function has a high penalty at large errors. erefore, a slightly more moderate absolute loss function (L1 loss) is used, which grows linearly with error rather than squared. However, this function does not have a derivative at the 0 point, so it may affect the sword collection. A common solution is to segment the function, using a squared function near the 0 point to make it smoother. is segmentation function is called the smooth L1 loss function, or SmoothL1Loss.  Computational Intelligence and Neuroscience

Segmentation of Losses.
After the target detection network passes through the RPN network and obtains m regional recommendation windows, the target branch will output m a × a matrices, and the matrix elements are the probability values of [0, 1]. e logarithmic loss function is applied to measure the target segmentation results. e single-pixel point loss is as follows.
e candidate window segmentation image matrix has a dimension a × a, and the overall loss is as follows: p(y | x)).

(11)
a × a is the image matrix dimension, and the overall loss is as follows: To further simplify the network parameters and improve the segmentation accuracy, a separable convolutional neural network is used to replace the normal convolutional structure in the MR-CNN algorithm for mask separation. While the conventional convolution considers the region first and then the channel, the deep separable convolution considers both the channel and the region, which also reduces the required parameters of the network.
A separable convolution is composed of a depth convolution and a point-by-point convolution compared to a conventional convolution. A deep convolution is a set of two-dimensional convolution kernels that perform spatial convolution on each input channel. One channel is responsible for one convolution kernel, so a deep convolution operation requires fewer parameters. Deep convolution learns each channel instead of one convolution kernel corresponding to all channels, allowing for a richer feature set. Point-by-point convolution uses 1 × 1 convolution window operation on the feature map obtained in the previous step to map the output to a specified number of channels. e size of the convolution kernel of the standard convolution is set as D k × D k × M × N, the input feature map is set asF in × F in × M , and the output is set asF out × F out × N, where F in and F out represent the input and output feature map size, respectively, and M, N the number of input and output channels, respectively, and the size D k of the convolution kernel. e required computation size for the standard convolution kernel is Using the depth convolution combined with 1 × 1 the convolution instead of the traditional convolution, the computation of a single depth convolution is From the above two computational quantities, it can be seen that a single depth convolution is very efficient and less computationally intensive, and it needs to be followed by a 1 × 1 convolution for the linear combination of the output channels. e computational effort of the depth-separable convolution is e ratio of the computational effort of the deeply separable convolution to the conventional convolution for equal number of channels and convolution kernels is Deeply separable convolution parameters are fewer and less computationally intensive, which improves the network performance.
erefore, it is combined with the mask branch in MR-CNN.

Experimental Results and Analysis
e images of yoga pose in the dataset were sourced from the web and downloaded from the web using a search engine. Due to the lack of a relevant dataset of yoga poses, the downloaded images were manually annotated. e images also contain poses captured from different camera views. Each category of the dataset has an average of 100 images. e images include not only yoga poses with clean backgrounds, but also yoga poses with different backgrounds such as forest, grass, and indoor. e dataset contains 200 images, and 80% of the total number of images are used as the training set and 20% as the test set. e order of the anchors is important in the construction of the dataset, and it is necessary to ensure that the same order is used in the training and prediction phases and to match the order of convolutional execution. For the FPN network, the anchors had to be sorted to make it easy to match the anchors with the output of the convolutional layer that predicts the anchor scores and displacements. First, they are sorted by pyramid level, all anchors in the first level, followed by all second levels, and so on. is makes it easier to separate anchor points by level. Within each level, the anchors are sorted in the order of feature map processing. Typically, the convolutional layer processes the feature maps starting from the top left and moving right row by row. Figure 4 shows the number of anchor points and the size of the feature map corresponding to each level.

Network Performance Metrics.
e performance of network classification recognition is mainly evaluated using a combination of accuracy, precision, recall, and F1 values. e examples in this paper are classified into two types, i.e., positive and negative, and the calculation formulas are 6 Computational Intelligence and Neuroscience e precision rate is denoted by p and the recall rate is denoted by R. TP denotes the number of cases that are actually positive and correctly identified as positive, FP denotes the number of cases that are actually negative but incorrectly identified as positive, and FN denotes the number of cases that are actually positive but incorrectly identified as negative. TN denotes the number of cases that are actually negative and correctly identified as negative.
e evaluation metrics used in the target detection task to evaluate the performance of the network are mainly two indices: Mean Average Precision (MAP) and Mean Intersection over Union (MIoU). MAP determines the accuracy of the network in predicting the class of objects in the box, and MIoU determines whether the detection box predicted by the network overlaps with the manual. e MAP determines how accurately the network predicts the object category in the box, and the MIoU determines whether the detection box predicted by the network overlaps with the manually marked box. e target detection sets a threshold to determine whether the target is correctly predicted, and the threshold affects the accuracy and recall rate. e AP curve is obtained by averaging the APs of different categories within the interval [0, 1], and the larger the MAP value, the more accurate the prediction. Complete overlap is the ideal result of network prediction, when the value of intersectionto-merge ratio is 1. e calculation formula for the cross-merge ratio is C in the formula that represents the labeled target area, and G represents the candidate area for network detection. At this point, the larger the value of IOU, the more accurate the detection frame of the framed target and the better the network performance. Figure 5 shows the comparison of the average error of MR-CNN with the classical residual network during the training process. e figure shows that the average error decreases rapidly during the first few iterations, but the MR-CNN is able to achieve a much smaller average error during the subsequent training process. Figure 6 shows the comparison of the average error of the MR-CNN with the classical residual network during the test. e figure shows that after 20 iterations, the improved residual network has a lower average error and performs better during the test. e performance of MR-CNN and the traditional deep residual network for yoga action classification is shown in Figure 7. It can be found that the improved residual network has higher recognition performance and better feature extraction performance and can accurately identify the corresponding actions in the images. e use of a more lightweight depth-separable network in the mask segmentation branch of the improved MR-CNN speeds up the network and improves the accuracy rate. e improved MR-CNN designed in this paper is used to conduct yoga action recognition and detection experiments.

Analysis of Results.
During the experiments, test images are fed into the trained network, prediction calculations are performed, and the results of the images output from the network are compared. e respective detection results using MR-CNN and the improved network with the training dataset are shown in Figure 8.
e improved network has a stronger learning capability, which can segment the target region more accurately, improve segmentation accuracy, and better fit the predicted segmented edges to the target edges.  Computational Intelligence and Neuroscience e variation of the loss function during the training process also reflects the performance of the network at the same time. Figures 9 and 10 show the loss profiles of the network during the training process of 30 epoch. It can be seen that the overall loss performance of the improved network is better and the loss function converges faster.

Summary and Outlook
is paper describes the current situation and problems in the field of image detection and segmentation of yoga poses, and the research designs a target detection and segmentation network. Based on the theoretical basis of convolutional neural network and combined with current related networks, a recognition detection network based on improved depth residual network is proposed, and a network for optimal segmentation of images is proposed in combination with the recognition detection network to realize the yoga action recognition and detection segmentation and candidate region processing network. Experiments show that the proposed scheme in this paper can effectively recognize yoga poses. Some problems have also been discovered during the experiment in this article. For example, the network is slow to decline during the training process, and the network weight parameters may fail to find the global optimal solution. More parameter optimization strategies can be studied in the future to optimize network training. e related field still faces     many problems, such that the application of the former MR-CNN mainly focuses on two-dimensional images, and the three-dimensional features of the target can complement the flat image features in complex environments. Follow-up research should explore object detection in combination with 3D technology in order to achieve efficient recognition detection.

Data Availability
e data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e author declares no conflicts of interest.