Indian Classical Dance Classification with Adaboost Multiclass Classifier on Multifeature Fusion

Extracting and recognizing complex human movements from unconstraint online video sequence is an interesting task. In this paper the complicated problem from the class is approached using unconstraint video sequences belonging to Indian classical dance forms. A new segmentation model is developed using discrete wavelet transform and local binary pattern (LBP) features for segmentation. A 2D point cloud is created from the local human shape changes in subsequent video frames. The classifier is fed with 5 types of features calculated from Zernike moments, Hu moments, shape signature, LBP features, and Haar features. We also exploremultiple feature fusionmodels with early fusion during segmentation stage and late fusion after segmentation for improving the classification process. The extracted features input the Adaboost multiclass classifier with labels from the corresponding song (tala). We test the classifier on online dance videos and on an Indian classical dance dataset prepared in our lab. The algorithms were tested for accuracy and correctness in identifying the dance postures.


Introduction
Automatic human action recognition is a complicated problem for computer vision scientists, which involves mining and categorizing spatial patterns of human poses in videos.Human action is defined as a temporal variation of human body in a video sequence, which can be any action such as dancing, running, jumping, or simply walking.Automation encompasses mining the video sequences with computer algorithms for identifying similarities between actions in the unknown query dataset with that of the known dataset.Last decade has seen a jump in online video creation and the need for algorithms that can search within the video sequence for a specific human pose or object of interest.The problem is to extract, identify a human pose, and classify into labels based on trained human signature action models [1].The objective of this work is to extract the signature of Indian classical dance poses from both online and offline videos given a specific dance pose sequence as input.
However, the constraints are video resolution, frame rate, background lighting, scene change rate, and blurring to name a few.The analysis on online content is a complicated process as the most of the users end up uploading the videos with poor quality, which shows all the constraints as a hindrance in automation of video object segmentation and classification.Dance video sequences online are having a far many constraints for smooth extraction of human dance signatures.Automatic dance motion extraction is complicated due to complex poses and actions performed at different speeds in sink to music or vocal sounds.Figure 1 shows a set of online and offline (lab captured) Indian classical dance videos for testing the proposed algorithm.
Indian classical dance forms are a set of complex body signatures produced from rotation, bending, and twisting of fingers, hands, and body along with their motion trajectory and spatial location.There are 8 different classical Indian dance forms; Bharatanatyam, Kathakali, Kathak, Kuchipudi, Odissi, Sattriya, Manipuri, and Mohiniyattam [2][3][4].Extracting these complex movements from online videos and classification requires a complex set of algorithms working in sequence.We propose to use silhouette detection and background elimination, human object extraction, local texture with shape reference model, and 2D point cloud to represent the dancer pose.Five features are calculated that represent the exact shape of the dancer in the video sequence.For recognition, a multiclass multilabel Adaboost algorithm is proposed to classify query dance video based on the dance dataset.
The rest of the paper is organized into literature survey on the proposed techniques, theoretical background on the proposed models, and experimental results.The proposed model is compared with SVM and Graph Matching (GM) classifier already proposed by us in our previous work.

Literature Survey
Local information of the human in the video is the popular features for action segmentation and classification in recent times.This section focuses on giving a current trend in human action recognition and how it is used in recent works for classifying dance performances.The human action recognition is subdivided into video object extraction, feature representation, and pattern classification [5,6].Based on these models, numerous visual illustrations have been proposed for discriminating human action based on shape templates in space-time [5], shape matching [6], interest points in 2D space time models [7], and representations using motion trajectories [8].Impressively, dense trajectory based methods [9] have shown good results for action recognition by tracking sampled points through optical flow fields.Optical flow fields are based on preconditioned on brightness and object motion in a video [10].The algorithms assume uniform brightness variations and object motions in consecutive frames which produce excellent results under minimum constrained video recording.Minimum constrained video recordings are having uniform brightness, less blurring, fixed camera angle, and high contrast between object and background.Finding such a video happens in a movie or in a lab setup.Hence, these approaches need to be robust in estimating human actions, which is still an open-ended problem on real time videos.Data driven methods with multiple feature fusion [11] with artificial intelligence models [12] are currently being explored with the increase in computing power.
In this work, human action recognition on Indian Classical Dance [13] videos is performed on recordings from both offline (controlled recording) and online (Live Performances, YouTube) data.Indian classical dance forms are practised from 5000 years worldwide.However, it is difficult for a dance lover to fully hold the content of the performance as it is made up of hand poses, body poses, leg movements, hands with respect to face and torso, and finally facial expressions.All these movements should synchronize in precision with both vocal song and the corresponding music for various instruments.Apart from these complications, the dancer wears complicated dresses with nice makeup and at times during performance the backgrounds are changing  depending on the story which truly makes this an openended problem.Mohanty et al. [14] highlight the difficulties in using state-of-the-art pose estimation algorithms such as skeleton estimation [15] and pose estimation [16] which fail to track the dancers moves in both offline and online videos.The author in [14] proposes using deep learning based convolutional neural networks (CNNs) and shows they perform well in estimating the correct pose of dancers on both 3D Kinect dance poses and online videos.CNNs require large training data for a specific class of inputs which makes them computationally slow for a video datasets that change for every 2 frames.In real time, there is no Kinect [17] like effects and hence 2D video analysis must be refined in accuracy for identifying poses of Indian dance forms.Samanta et al. [18] used histogram of oriented optical flow (HOOF) features with sparse representations.Support vector machine classifier (SVM) classifies the Indian classical dance poses from KTH dataset with an accuracy of 86.67%.In our previous work [19], we approached the same problem with SVM classifier on dance videos and found that only multiclass SVMs should be considered.Moreover, optical flow on online videos suffers at lot due to inconsistencies during capture and sharing process.
In [20], Samanta and Chanda proposed a video descriptor manifold on ICD YouTube videos and KTH dataset with nonlinear SVM classifier and recorded recognition accuracies in the range 70 to 95%.The other works proposed also used SVM classifier with simple image processing models on images of dancers for pose estimation which can be found in [21][22][23].In [24], authors used Kinect sensor to capture leg poses in Indian classical dance forms and classification is initiated with SVM classifier for a set of 40 poses.Kinect sensor produces skeleton data of the human body pose and fails to reproduce data related to fingers which are important in classifying a dance pose.
The objective is to select features that represent a sign and are easily distinguishable in closely related sign words and are computationally efficient.The attributes for a selfie sign language recognizer chosen are shape signature [25] for hand and head shapes, Hu moments [26] for hand orientations, hand-head distance, and hand position vectors for tracking.The chosen attributes perfectly characterize a sign in Indian sign language.
Classifying at faster rate on a huge dataset is a complicated problem.Adaboost [27] classifier is fast and efficient algorithm for large datasets [28].Inspired by [29,30], the feature matrix is labelled and inputted to Adaboost classifier for training and testing.The performance indicators are recall-precision curves and execution time on mobile and are recorded to check the robustness of the algorithm and feasibility of implementing more efficiently.
In this paper, we propose an multiclass multilabel Adaboost (MCMLA) based classification problem on multidimensional feature vector.We show that this can be used to match large unconstrained dance features which are automatically extracted from video datasets.The feature representation of video objects depends on the efficiency of video segmentation algorithms.As illustrated in Figure 2, the proposed Adaboost can effectively recover the query video frames from the dance dataset, by shape-texture observation model defined by discrete wavelet transform (DWT) and local binary patterns (LBP).
In summary, our MCMLA algorithm on online and offline Indian classical dance videos combines the representational flexibility and trivial computations.We perform experiments on two different datasets of Indian classical dance Bharatanatyam and Kuchipudi created from online downloads and offline controlled lab capture.The proposed method is compared with other GM models which are outperformed by a considerable margin in speed.

Proposed Methodology
The proposed algorithm framework is shown in Figure 2.An Indian Classical Dance (ICD) video library is created combining online and offline videos.Dancer identification, dancer extraction, local shape feature extraction, and classifier are the modules of the system.Further feature fusion concept from [31] is also explored in this wok using 5 feature types, Zernike moments, Hu moments, shape signature, LBP features, and Haar features.Adaboost algorithm explores the relativity between the query dance sequence and known dataset.

Dancer Identification.
Most of the dance videos are poorly illuminated or fully brightened with too much background information during capture.Commercial video cameras have a frame rate of 30 fps and dance movements are sometimes faster and at time slower which makes the object blurry.The objective is to extract moving dancer and segment it for further processing.This helps to prevent the algorithm from constantly upgrading the background information and modelling the object characteristics in real time.The dancer identification module is based on one of the silhouette extraction methods proposed in [32].A significant indication in determining dancers motion for extraction lies in the temporal changes in the dancer's silhouette during performance.To avoid background modelling and foreground extraction models, we propose using the following procedure.
The dance video sequence V(x, y, t) ⊂ R + , with (x, y) ⊂ Z + , gives pixel location and t ⊂ Z + is the frame number.Each frame in V is having RGB planes and is of size N × M × 3.This part of the module is only for motion segmentation and object extraction; color can be discarded.RGB is converted to gray scale and contrast enhanced to improve the frame quality.The frame V t at t is mean filtered with mask defined by m(x, y) with The size of m is updated based on the frame size N×M for faster computations, where the object area is small compared to the background area.The ⊗ operator is linear convolution and the averaged frame is of the same size as the input frame.The next step applies a Gaussian filter of  mean and  variance on the input frame V t : The size of the Gaussian mask is determined by the input video frame.Euclidian distance metric S t (x, y) between V t m and V t g gives the saliency map of the moving pixels in the frame The second order normed distance map is shown in Figure 3 which identifies the dancer's silhouette.However, to extract the dancer, a mask of this silhouette is used to determine the connected components in the object.Figure 3(d) shows the silhouette mask and connected component output is in Figure 3(e).
The centroid of the mask is mapped on the frame to crop out the moving dancer in the frame.The method is effective in all lighting conditions putting constraints on the input video frame size in selecting the masks used for mean and Gaussian filters.The boxed and extracted dancer from the video sequence is shown in Figures 3(g) and 3(h), respectively.The extracted dancer is free from background variations in the video sequence.If a portion of background still appears at this stage it can be nullified during the matching phase.Applying feature extraction on the extracted dancer allows for lesser computations as the background is almost eliminated and leads to good matching accuracy.

Feature Extraction. Which features can help recognize
and classify dance correctly is the question.From a dancer's perspective, to identify a dance type, body posture, hand shapes, and their movements in space are the vital features.Feature extraction phase explores the methodology in extracting these features.There are many shape descriptors available in literature for characterizing shape features [33].Lighting, frame inconsistency, contrast, blurring, and frame size are some of the critical factors that affect feature extraction algorithms.In addition, the dancer velocity during performance has instincts for a faster shape extractor.

Haar Wavelet Features: Global Shape Descriptor.
For removing video frame noise during capture and to extract local shape information, we propose a hybrid algorithm with Discrete wavelet transform (DWT) [34] and Local Binary Patterns (LBP) [35].The objective at this stage is to represent moving dancers shape with a set of wavelet coefficients.Here we propose using Haar wavelet at level 1.At level 1, Haar wavelet decomposes the video frame V t into 4 subbands.Figure 4 shows the 4 subbands at 2 levels.At 1st level we have 4 subbands and at 2nd level have 8 subbands.In the 1st level, the three subbands represent the shape information at three different orientations: Vertical v, Horizontal h, and Diagonal d.Combining the three subbands and averaging the wavelet coefficients normalizes the large values.
The averaged shape harr wavelet coefficients W t S , along with {h, v, d} subband coefficients are reconstructed to spatial domain.Figure 4 shows the reconstructed spatial domain frame producing the exact hand shapes.These shape features can be used as nodes and a graph can be constructed for recognition.However, background noise is still a major concern at this stage.Local pixel information becomes vital in selection of nodes that exactly represent the graph.

Thresholding.
Apply threshold on the reconstructed ICD video frame V t r as The binarized video frame B t is To extract the nodes for the graph, local pixel patterns provide exact shape representation.

Local Binary Patterns and Local Shape Models.
LBP compares each pixel in a predefined neighbourhood to summarize the local structure of the image.For an image pixel B t (x, y) ∈ R + , (x, y) gives the pixel position in the intensity image.The neighbourhood of a pixel can vary from 3 pixels with radius r = 1 or a neighbourhood of 12 pixels with r = 2.5.The value of pixels using LBP code for a centre pixel (x c , y c ) is given by where g c is binary value of centre pixel at (x c ,y c ) and g p is binary value around the neighbourhood of g c .The value of P gives the number pixels in the neighbourhood of g c .The local shape descriptor L t S of the human dancers pose projects maximum number of points on to graph.

Multifeatures: Zernike, Hu
Moments, Shape Signature, LBP, and Haar. Figure 5 shows the extracted dancer represented with LBP features and Haar wavelet features.Local shape features in Figure 5(a) are used to construct a graph.Given a motion frame in a ICD video sequence V t successfully extracted local shape features L t S and transformed into a binary shape matrix B t S of ones and zeros using (5).A sparse representation of B t S eliminates all zeros and retains only ones and their locations in M t S (x, y, w), where x, y are shape point locations and w is shape feature weight vector.Figures 5(d) and 5(e) show a sparse representation for both wavelet reconstructed LBP (WR LBP) and only Harr wavelet features (HWF), respectively.The points on the motion object are formed by extracting the location of the pixel and its feature value determines the shape of the dance pose.From these feature point locations and values a graph is constructed in this work.
Haar and LBP features are enough to label the dancers in the frames and put them under the classifier.This is early fusion of features at the segmentation stage.The fusion operator is principle component analysis.PCA of wavelet features and LBP features is concatenated using the following expression: However, 3 more features are proposed in this work that can effectively represent shape features of the dancer.Dancers in Indian classical dance videos have a large motion vector field and Haar and LBP depend on variations in lighting, camera movement, and background.To counterbalance camera movements, we propose using Zernike Moments (ZM) to represent dancer in each frame on the 2D point cloud extracted from WR LBP vectors.
Dancer body orientations provide rotation invariant feature of a dance movement in a dancing space.However, these movements are incorrectly classified if there are unavoidable sudden camera vibrations.Moments M pq project 2D points in dance segments B t s (x, y) on to basis x p y q which results in a piecewise continuous linear function in the spatial plane.Moments and moment functions are used as pattern shape features in number of applications [36][37][38].Geometric moments are usually defined as where M pq is the (p + q)th order moment of B t s > 0. However, geometric moments do not exhibit any invariance properties such as translation, rotation, and scaling.Teague [39] proposed ZM to recover image from moments representing the image in terms of orthogonal polynomials.ZM are used in pattern recognition applications [40,41].
ZM project the 2D point cloud B t s (x, y) to a set of orthogonal polynomials, called Zernike polynomials defined as a complete set where R nm () are real valued radial polynomials defined in [39] and  is moment magnitude and  is angle.The orthogonality of ZM is modelled as where " * " indicates complex conjugate and  nn  should satisfy ZM of order n with repetition m for continuous feature set B t s (x, y) per frame t over a unit disk is defined as If  0 is the rotational angle, with original ZM and rotated ZM as A nm and A R nm , respectively, we have where |A nm | and Θ nm represent magnitude and phase, respectively.In ( 14) the magnitude remains constant while the image rotates, whereas the phase changes with image rotation.Hence, most of the applications use ZM magnitude as a feature vector for pattern classification.In this work, we propose ZM magnitude on 2D shape point cloud as invariant feature representing the small variations in camera movement that occur in online video capture of the dance performance.To represent changes in the dancer dimensions which happen nonlinearly, we propose using a nonlinear function defined over geometric moments.Hu moments in [42] are nonorthogonal centralized moments that are scale, translation, and rotation invariant.Human dancers come in all shapes and sizes and the features describing them may change with dancer.Modelling invariance in dancer shape features from the 2D point segments of Figure 5(e) or Figure 5(j) is done with Hu moments.Hu moments are derived for 2D normalized central moments    using algebraic invariants: These 7 moments are calculated for every extracted dance shape in each frame.
Every pixel is represented with a shape feature [43] or a descriptor [44] in each frame to model dancer shape for classification.However, to reduce the computations on the feature vector during recognition, we propose using shape signatures in [25] to represent dancer shapes.In [25], shapes are represented on a multiscale model based on integral kernels.A shape descriptor or signature is an integral invariant at various scales forming a shape signature function on dancer shapes    (, ) where   is a Gaussian kernel in 2D with  > 0 and zero mean.Also, Integrating shape feature values over shape 2D spatial domain results in a shape signature value normalized by area of the shape to achieve scale invariance.The range of shape signature is [0, 1] based on value of scale .For low values of scale, it is 0 and for large scales it approaches 1.

Feature Matrix Construction.
From the 5 features, a complete feature matrix is constructed per dance frame.Early fusion is performed with Haar and LBP features with PCA at the end of dancer segmentation and are labelled with dance vocal words.This feature matrix is named as EFDF (early fused dance features).It is a 2D feature matrix of size 256 × 256 × 2. Max pooling is performed to reduce the feature set to 1 × 256 × 2 per frame.Further, late fusion model is introduced with more robust set of features in the Zernike moments, Hu moments, and shape signatures (SS) on 2D shape point cloud of the dancer in the video frame.These late features are mixed with Haar and LBP features to create a multifeature matrix per frame.We calculated 5 ZM, 7 Hu Moments, 1 SS, max pooled, and thresholded LBP and Haar features limited to 30 features per frame.The final feature matrix in late fusion strategy is 5 × 7 × 1 × 30 × 30 per frame.These features are carefully labelled with vocal words representing the dance form in the video frame.For a 25-frame dance word "swami ra ra," we have 73 × 25 feature matrix.This feature matrix or a set of matrices are inputted to MCMLAB classifier.
where  = ||.For  = 1, . . ., , we select a weak classifier   :  ×  → [−1, 1] with distribution   , to maximize the absolute value of We choose the biasing value   as and update the distribution function as where   is normalization factor to keep the distribution as probability density function.The final output strong classifier is For the multiclass problem  1 , . . .,   , we use the real valued 2D Look-Up Table model in [41], which is defined as where ()  = ( ∈   | frame ∈ sign) and  ,  (, ) = {1,  ∈ ; 0, otherwise}.From this weak hypothesis, through training a strong hypothesis is generated to recognize sign labels   = (  ).

Experimentation and Results
A set of 4 experiments are initiated to test the robustness of the proposed multifeature fusion with Adaboost classifier.Our Indian classical dance datasets consist of performances on "Bharatanatyam" and "Kuchipudi" from online YouTube videos and offline dance videos in controlled environment at KL University, cams department studio.We have created 4 dance videos from 5 dancers for 2 songs in two different dance styles.Similar online YouTube downloaded dance performances are also collected.Matching frames in each dataset are shown in Figure 6.
In  We use three performance evaluators for validating the results.They are precision-recall curves, percentage recognition rate, and the computation time per sign.For a strong hypothesis H resulted from Adaboost training and testing for an input feature f i on a trained distribution D with Z i = H(f i ) predicted labels.The following metrics in [46] are Recall (H, Exp-I uses input videos from dance data set captured in the controlled environment.The dancer identification, feature extraction, and graph representation for the dancer are shown in Figure 7. Saliency maps from average and Gaussian distance metric create Silhouette, which identifies the dancer in the video frame.The dimensions of the bounding box extract the dancer.Then the dancers features are extracted with segmentation as early features.Haar wavelet at level 1 is averaged in high frequency components to remove background and IDWT is performed to recover the global shapes on the dancer.Applying LBP on the resulting IDWT dance frame captures local shape information.At this stage, we perform Adaboost multiclass multilabel classification with the PCA based Haar and LBP feature fusion.
Early fused Haar-LBP features of offline dance video of a dancer are used to train the Adaboost classifier.For the same dance video shot with slight variations is provided as query dance video for same set of labels.The resulting confusion matrix from early fusion of features on same training and testing set is shown in Figure 8.
The normalized values in the confusion matrix are calculated using (26).We did plot the confusion matrix for the song sequence "Siva Shamboo" with first 28 word labels.The average recognition on total 116 labels from two dances in "kuchipudi" from a dancer is 0.96.From other dancer's videos, the Adaboost classifier averaged around 0.955.Exp-1 is repeated with all the same parameters with late fusion with Zernike moments, Hu Moments, and liner shape signature along with Haar and LBP features.Late feature matrix is a 73 × 25 matrix per label.For our 116-label dance sequence we have a 73 × 25 × 116 dimension feature vector.Figure 9 shows the results of the classifier in the form of confusion matrix (showing 28 labels) from (26).
The results show an average of 0.99 for all dance videos in the dataset.Clearly, multiple features and late fusion have increased the ability of the classifier to recognize dance poses correctly.False matching is much less in this experiment as the datasets used for training and testing are similar, that is, same dancer and same performance.
In Exp-2, the performance is the same, meaning same labels for training but the dancer performing the dance will be different.However, testing with a different query video having different dancer results in lower recognition rates compared to the previous experiment.Here also, the simulations are performed for early feature fusion and late fusion.Confusion matrices from the two simulations are shown in Figures 10  and 11, respectively.The average recognition dropped due to noncoherence between dance moves with respect to body   shapes, pose shapes, and movement speeds between the two dances.
Early feature fusion in exp-2 provided a recognition 0.81 averaged over a set of 4 video samples of the same song.However, late feature fusion with more versatile feature representing a dance frame produced a recognition rate of 0.91.Exp-3 trains the Adaboost classifier with online dance video content and tests with the same set for early fusion and late feature fusion.The initial shape extraction from online dance video content is shown in Figure 12.   0.12 0.12 0.12 0.12 0.13 0.17 0.17 0.19 0.14 0.24 0.24 0.14 0.17    The confusion matrices for early fusion and late fusion are presented in Figures 13 and 14 for exp-3, online dance videos, respectively.
The recognition rate decreased for online dance videos as these are unconstrained in real time capture.Problems such as vibrant video backgrounding, poor or over lighting, and motion blurring in online videos made the feature extraction complicated.However, the proposed algorithm with early feature fusion has resulted in an average recognition rate of 0.84 for one test video under these circumstances.The average recognition rate for 4 test samples in exp-3 is 0.83.The problems in online videos are effectively handled by increasing the number of features to represent a dance pose.Figure 14 gives the confusion matrix with late fusion features.The average recognition rate is 0.93 for 4 test videos.
Similarly, we obtained average recognition of 0.68 with early feature fusion and 0.82 with late fusion, respectively.The confusion matrices are shown in Figures 15 and 16.
To summarize, the results obtained with the proposed early fusion and late feature fusion on offline and online dance videos are presented in Table 1.The dataset consists of 4 online and offline videos.For each video, we computed the average recognition rate obtained by taking mean of all recognition rates per dance pose.
From Table 1, the performance of the Adaboost classifier on different online and offline dance videos can be recorded.This data can be interpreted to understand the ability of features to uniquely model dancers pose in a real time dance video to measure the dancers' performance.Harr and LBP together model global and local dance shapes, respectively.However, these features have constraints on motion blurring, scale variations, brightness, and contrast variation in the video frames.Apart from these image variations, we have camera vibrations, dancer shape variations, and video backgrounds to restrict these feature vectors from uniquely modelling a dancer in the video frame.To compensate for these variations, additional features in the form of shape signatures, Zernike moments, and Hu moments are proposed to represent a dance pose.Shape signature models the dancer pose uniquely with an integral kernel having Gaussian characteristics.The shape signature is calculated on the shape curve on a set of frames which matches similar dance pose at a different location in a video sequence.Similarly, ZM are linear orthogonal representation of video data that handle dancer's movements in spatial domain.Hu     features and the same number is used for testing.The average recognition rates were 0.84 for HOG, 0.82 for SIFT, and 0.8 for SURF in Exp-1 where same training and query dance video is used.But for different dancer videos the recognition dropped to 0.67, 0.65, and 0.59 in case of exp-2 for HOG, SIFT, and SURF, respectively.Similar results were seen in Exp-3 and Exp-4 for online dance videos.However, in Exp-4, for different dance video sets for training and testing have drastically reduced the recognition rate of the classifier by 50%.The drop-in classifier performance can be attributed to the poor feature extraction due a large variation in the video frames even though they have same dance pose.
HOG, SIFT, and SURF features are extracted from the original gray scale video frame.To improve their performance the algorithms are applied on the extracted dancer from our Haar-LBP sparse segmentation module.On the segmented and extracted dancer, the average recognition rate from HOG is 0.89, SOFT is 0.91, and SURF is 0.82 for exp-1.In exp-2 these values were again reduced by a factor of 18%.Exp-3 and Exp-4 showed an improvement of 30% increase in average recognition rate for all three feature extraction models.However, after consecutive testing and measurements these state-of-the art features reported less overall average recognition rates compared to the proposed late fusion features that form a complete set to model a dance video sequence.
The set of 5 late fused features with 73 attributes per frame performed well with MCMLAB classifier compared to early fusion and other feature extraction models; it is time to test for different classifiers.We applied the same late fusion features to adaptive graph matching (AGM) and Support Vector machine classifier.In AGM, each node is modelled with 73 features and the distance between the features is models as edge.For each dance pose video, a oneto-one matching on both node and edge feature is calculated.Recognition rate is calculated using (24).Multiclass SVM is the most trusted classifier for character recognition.Hence we apply the late fused features to a SVM classifier and each dance pose is measured.
Three classifier are compared here with late fusion features and the other Adaboost with early fusion features taking it to 4 classifiers.All these classifiers are compared for recognition rate and efficiency.We plot average recognition for the 4 classification models averaged across 4 offline and online videos in Figure 17.
From Figure 17, the Adaboost with late feature fusion with 5 features outperforms the AGM and SVM.However, AGM comes close to Adaboost classifier and sometimes is better than MCMLAB on late fusion features for dance pose recognition.Nevertheless, AGM is far slower than MCMLAB algorithm.SVM came last in the comparison due to inefficiency in defining initial support vectors for classification from the fused feature vectors.MCMLAB with early fusion features is better than SVM classifier.MCMLAB is better, if the example set is uniquely defined by the feature set and the proposed features are the best choice for Indian classical dance pose recognition.
Equations ( 25) and ( 26) are used to calculate precision and recall with late and early fusion on Adaboost for 2 offline and 2 online video sets.Similarly, the same datasets are used for AGM and SVM classifiers.One video will be used for same feature train-test model and other is a different train-test model.The average precision and recall values are plotted as a graph in Figure 18.
The ability to recall precisely the same label as the query video as a performance measure is plotted in Figure 18.It shows the same result as above, making MCMLAB as the  Figure 19 shows the percentage recognition obtained from methods in [20,21,23,24] along with the proposed early and late feature fusion models.The recognition percentage is averaged over the entire dataset.The plots in Figure 19 highlight the use of multiple features for various representations of moving objects in a dance video for accurate classification and recall.

Conclusion
Indian classical dance classification is a complex problem for machine vision research.The features representing the dancer should focus on the entire human body shapes.Hand and leg shape segmentation is a critical part of a ICD.In this work, we proposed a fully automated ICD consisting of dancer identification, extraction, segmentation, and feature representation and classification.Saliency based dancer identification and extraction helps in reducing the image space.Wavelet reconstructed local binary patterns are used for feature representation preserving local shape content of hands and legs.Two fusion models are proposed for feature fusion.Early fusion at the segmentation stage with PCA based Haar wavelet and LBP is used and late fusion using the Zernike moments, Hu moments, and shape signatures is used with Haar and LBP is proposed.Multiclass multilabel Adaboost on features of early fusion and late fusion between two sets of dance video data is the classifier.Multiple experiments on online and offline ICD video data are tested.Dance video data is labelled as per the vocal song sequence.The early and late features and classifiers performance tests show that the proposed late fusion features and multiclass multilabel Adaboost classifier give better classification accuracy and seed compared to AGM and SVM.More action features can be added for representing dancer more realistically by elimination backgrounds and blurring artefacts to improve the efficiency of the classifier.

Figure 1 :
Figure 1: Online and offline dance datasets used in this work and the video constraints.

Figure 2 :
Figure 2: Flow diagram of the proposed process for Indian classical dance recognition.(a) Training datasets, (b) detected dance object, (c) extracted, (d) DWT and LBP features, (e) feature points, (f) query dance video, (g) query feature points, and (h) multifeature extraction and classification.

Figure 5 :
Figure 5: (a) Original dancer in online video frame, (b) LBP features from reconstructed and thresholded wavelet coefficients, (c) Haar wavelet features, (d) sparse coded wavelet reconstructed LBP features, and (e) sparse coded wavelet features.(f)-(j) Same as (a)-(e) for offline lab captured video frame.

3. 5 .
Dance Classifier: Adaboost Multiclass Multilabel.Boosting based classifications[29,30,45] find very precise hypothesis from a set of weak hypotheses.Here hypothesis is a classification rule.Set of weak hypotheses are simple rules that generate a predictable classification.Let  = [( 1 ,  1 ), ( 2 ,  2 ), ( 3 ,  3 ), . . ., ( V ,  V )] be a set of training examples at an instance   on th frame in feature space  with labels   on label space .The algorithm accepts the training samples  along with some class distribution  = {1, . . ., } ∈ R represented as weak learners.On the input, the weak learner computes a weak hypothesis .Generally,  :  → R. The interpretation for classification is based on sign{()} = {+1, −1} → {  } for a binary classifier.|(  )| gives prediction confidence.The key to boosting is to use the weak learner to produce a very precise prediction rule by repeatedly addressing the weak learner on different distribution of training examples.In this work, a multiclass version of Adaboost is used having a set of strings as class labels.The problem is modelled as given  and size of final strong classifier .The Adaboost initializes the distribution function as  1 (, ) = 1 V ∀ = 1, . . ., V,  = 1, . . ., , exp-1 and exp-3, we use offline and online dataset of same dancer video for training and testing with early fusion and late fusion of multiple features 28 words in the dance sequence.Each mudra pose is coordinated with vocal manually by labelling each set.Variations in number of frames per label are nullified and normalized to 15 Key frames per dance pose across all video data.Exp-2 and Exp-4 are conducted based on different training and test

Figure 6 :
Figure 6: (a) Online Indian classical dance datasets from YouTube.(b) Offline dance video dataset created at KL University in a controlled lab environment.

Figure 8 :
Figure 8: Exp-1 early fusion confusion matrix with same offline dancer in training and test video.

Figure 10 :
Figure 10: Exp-2 early fusion confusion matrix with different offline dancer in training and test video.

Figure 11 :
Figure 11: Exp-2 late fusion confusion matrix with different offline dancer in training and test video.

Figure 13 :
Figure 13: Exp-3 early fusion confusion matrix with same online dancer in training and test video.

Figure 15 :
Figure 15: Exp-4 early fusion confusion matrix with different online dancer in training and test video.

Figure 16 :
Figure 16: Exp-4 late fusion confusion matrix with different online dancer in training and test video.

Figure 17 :
Figure 17: Comparison of classifiers for dance pose recognition.

Figure 19 :
Figure 19: Compassions with other models of ICD with the proposed early and late fusion models with Adaboost classifier.
Figure 9: Exp-1 late fusion confusion matrix with same offline dancer in training and test video.
moments handle camera vibrations and other nonlinear movements of the dancer in the video frames.Early fusion and late fusion concepts are introduced to understand which set of features can correctly and uniquely model a dance pose irrespective of the constraints during video capture.Simulations show that late fusion and more number of features are necessary for training Adaboost classifier.Same dancer videos resulted in better recognition rates compared to different dancer in both offline and online dance videos.We also tested the classifier with HOG (Histogram of oriented Gradients), SIFT (Scale Invariant Feature Transform), and SURF (Speeded Up Robust Features) with MCMSAB classifier.Training vector is made of 50 best Figure 14: Exp-3 late fusion confusion matrix with same online dancer in training and test video.

Table 1 :
Summary of mean recognition rates for 4 offline and online dance performances for the song "Siva Shamboo" in kuchipudi dance.