The Study of Scene Classification in the Multisensor Remote Sensing Image Fusion

We propose a scene classification method for speeding up the multisensor remote sensing image fusion by using the singular value decomposition of quaternion matrix and the kernel principal component analysis (KPCA) to extract features. At first, images are segmented to patches by a regular grid, and for each patch, we extract color features by using quaternion singular value decomposition (QSVD) method, and the grey features are extracted by Gabor filter and then by using orientation histogram to describe the grey information. After that, we combine the color features and the orientation histogram together with the same weight to obtain the descriptor for each patch. All the patch descriptors are clustered to get visual words for each category. Then we apply KPCA to the visual words to get the subspaces of the category. The descriptors of a test image then are projected to the subspaces of all categories to get the projection length to all categories for the test image. Finally, support vector machine (SVM) with linear kernel function is used to get the scene classification performance.We experiment with three classification situations on OT8 dataset and compare our method with the typical scene classification method, probabilistic latent semantic analysis (pLSA), and the results confirm the feasibility of our method.


Introduction
Image fusion refers to fusing two or more images which are in the same scene into a composite image [1].The composite image contains the significant information of these various source images.Then we can obtain a more comprehensive and clear description of the scene.Image fusion information is usually obtained from the multisensor remote sensing images.Information fusion of multisensor remote sensing images can collect a wealth of information, and remote sensing is able to improve the means of the global scope dynamic observation data.The categories of ground targets and the scene environment are different, and the remote sensing images obtained from each sensor are not necessarily the same object or the same scene.We need to classify these remote sensing images into various categories, such as city, forest, open country, mountain, and coast.By that way, it will be conducive to speed up the image fusion in the same scene, which is more conducive to improve the important influence of the multisensor remote sensing images in many monitoring analytical applications, such as city planning, forest monitoring, land, and use monitoring [2], and flood monitoring.
Scene classification is an important part of the multisensor remote sensing image interpretation.Conventional scene classification methods mainly use the images' low level features (global color or texture histograms, the power spectrum, etc.) to classify only a small number of scene categories [3,4].Szummer and Picard proposed indoor and outdoor classification method.They used color, texture, and frequency features of local regions in their method and got the 90.3% correct classification [3].Paek and Chang used the HSV color histogram and the edge direction histogram to represent each block of the image in the color and texture features and got the similar result with Szummer and Picard [4].However, the classification accuracy of the method based on low-level features is not satisfactory.
In recent years, many authors proposed bag-of-visual words methods based on local patches without topological information for scene classification, which have been applied to cases where there are a larger number of scene categories [5][6][7][8][9].Fei-Fei and Perona proposed a scene classification method based on bag of keypoints to classify thirteen scene categories [5].Bosch et al. also presented a method based on pLSA to classify a larger number of scene categories and got good classification performance [6].These methods used local patches without topological information.However, they both [5,6] reported that the classification result is better when local parts are selected from evenly sampled grid which were named dense patches in [6].Therefore, in this paper we use dense patches from which we extract features for classification.Dense patches are constructed by segmenting images by regular grid.The visual words vocabulary can be obtained by clustering the features of patches with the -means algorithm.Thus, an image can be viewed as a collection of "visual words." The feature vector is developed from patches.In [6], they used color patches and extracted SIFT features in each color channel, respectively.But the dimension of the feature vector they got is very high, and the method spent a lot of time in feature extraction, clustering, and quantification.In order to reduce the time needed for classification, we should reduce the dimension of the feature vector.
To answer this demand, in this paper we use the quaternion singular value decomposition (QSVD) method to extract the color features of the image.Because the RGB channels are treated as an integrated whole in the quaternion model of a color image, the three colors do not have to be operated in each color channel independently.In this way, we take into account the direct interchannel effects in the three color channels and reduce the dimension of the feature vector.The singular value vector of a color image is robust to shift variations, scale variations, and rotation variations [10].Therefore, the singular value decomposition of the quaternion model of a color image is effective for scene classification.We obtain a quaternion matrix from each patch of an input image.
Simultaneously, we also obtain the orientation histogram of multiscale Gabor feature from each patch.In recent years, it is reported that the orientation histogram of Gabor features with rough topological information which is developed from dense patches is effective for scene classification [7,8].We develop the orientation histogram from multiscale Gabor features because Gabor features which extract frequency and orientation information are better than the simple gradient [8].Then we combine the color features obtained from the quaternion model and the orientation histogram to form the final feature vector computed from each patch.Although the most famous approach for scene classification is bag-of-visual words method, superior accuracy of KPCA of visual words to bag-of-visual words was reported in recent years [8,11].However, the computational cost of KPCA becomes high when the number of training samples increases.To reduce the computational cost, we must decrease the number of training samples.To solve the problem, we use the bag-of-visual words method to extract the representative feature vectors of each category.Then we applied KPCA to the set of visual words of each category.The uncorrelated mean projection length to the subspace of each category is used as input features for classification.Finally, we use SVM to classify the input features.The mean projection length in subspace is used as features for support vector machine.The proposed method is evaluated using OT8 dataset [12].We compare our method with the recent popular scene classification method pLSA in [6].
The rest of paper is organized as follows.In Section 2, the methods of context information extraction is presented.In Section 3 briefly introduces similarities with subspaces and explains how the information specialized for each category can be extracted.The SVM and devices on kernel are introduced in Section 4. Section 5 shows the experimental results and evaluates the results.Conclusion and future works are described in Section 6.

Context Information in Patch Space
For scene images, dense patches are obtained from a regular  ×  grid, and here we use 16 × 16 grid.The visual words are developed by clustering descriptors of patches that are computed from the training images.Accordingly, all the patches are represented by the clustering centers called the visual words obtained by -means algorithm.In this paper, for each patch we extract color features, and grey features, respectively, and use weights to integrate them as the entire features for classification.

Singular Value Decomposition of Quaternion Matrices.
Color in images processing has been widely used recently.For color images, the color of each pixel is usually expressed in RGB color, and each element of the image matrix saves three values of RGB color.The traditional methods of color images processing usually separate the color images into RGB three channels images first and then use a variety of image processing methods to process images.But as a result of inherent relationship among RGB components, image separation leads to the distortion of color image.Recently, the effectiveness of QSVD in color images processing was reported [13][14][15][16].In this paper, we develop the color features by using QSVD method.This is because the algorithm uses a quaternion to represent the RGB values of a pixel in color images.In this way we make the RGB color channels as an integrated whole, and take into account the direct interchannel effects in the three color channels.
First, we explain the quaternion briefly.The quaternion [13] is defined as where   ,   ,   ,   are all real and , ,  are orthogonal imaginary operators, which satisfy the following identities: is the real part, and   ,   ,    are the imaginary parts.
A quaternion with no real part (  = 0) is said to be pure quaternion.
If we set   in (1) as 0 and use   =   (, ),   =   (, ), and   =   (, ) to represent the RGB values of a pixel in color image, then an × color image can be represented as a pure quaternion image by the following function  () (, ): where   (, ),   (, ), and   (, ), respectively, are red, green and blue components of the pixel at position (, ) in the image.Thus, the color image is a matrix of size  ×  over the quaternion field (restricted to pure quaternion).So a color image can be represented as a quaternion matrix.
In this paper, quaternion is represented as a matrix , Consider an  ×  quaternion matrix  () ,  () ∈  × and  () =   +    +    +   , where   ∈  × ( = , , , ), then the equivalent real matrix of  () is The definition of the equivalent real matrix of a quaternion matrix  () denotes the relationship between the real matrix and the quaternion matrix.In addition, we can use the eigenvalues and eigenvectors of the equivalent real matrix to compute the eigenvalues and eigenvectors of a quaternion matrix [14].The relationship between the singular value decomposition of a quaternion matrix and the singular value decomposition of the equivalent real matrix can be found in [14].
Next, we consider the QSVD of a color image.The proof of existence and meaning of the SVD of a quaternion matrix can be found in [14].Any matrix  () ∈  × admits a singular value decomposition given as where  is the conjugate-transposition operator,  () ∈  × , and  () ∈  × are two unitary quaternion matrices.Besides,  ()   () =  × and  ()   () =  × , where  is the identify matrix.These matrices contain the left and right quaternion singular vectors of  () .Λ  is a real diagonal matrix, where  is the rank of  () .The values on the diagonal of Λ  ,   (with 1 ≤  ≤ ), are the singular values of the quaternion matrix, arranged in decreasing magnitude order along the diagonal.The method for obtaining the decomposition of a quaternion matrix can be found in [14], which is based on the equivalent real matrix.Now, we explain how to extract color features from the QSVD results of a patch.In this paper, the color features are developed from the two quaternion unitary matrices of QSVD.The most important information is contained in the two quaternion unitary matrices of QSVD [14].Like the singular value decomposition (SVD) of a grey image, the SVD of a color image also can be represented as follows: Here, ],  () and V () are the th column vectors of  () and  () , respectively.  is the diagonal element of the real matrix Λ, and  is the rank of  () .Each result of  () V  () is referred as the th weighted eigenimage [14].From the original image and its eigenimages, we can find that most image information, that is, energy, color, and structure, is concentrated on the foreside of the eigenimages.Figure 1 illustrates an image and a part of its eigenimages.From the figure we can discover that the first three eigenimages possess most energy of the image [14].Then we use the first three eigenimages to reconstruct the original image, and the reconstructed image is shown in Figure 1.
From Figure 1, we can observe that the first several eigenimages contain the average information about color and structure of the original image (low frequency information).So in this paper, we use the first eigenimage as the color feature for classification.We can get the average information of the original image, not the detail information.In this way, we can avoid the problem of over fitting and improve the generalization ability of the proposed method.
In the experiments, we use evenly sampled 16 × 16 grid, and for each grid (each patch) we compute its equivalent real matrix, then obtain the result of QSVD of the real matrix, and then compute the first eigenimage as the color feature.

Orientation Histogram Extraction Based on Gabor Features.
Recently, orientation histogram has been widely used in pattern recognition.Here, we develop the orientation histogram from Gabor features [8].A distinct advantage of Gabor features is their optimality in time and frequency, or space and spatial frequency in two dimensions, providing the smallest possible pieces of information about time-frequency events.
In this paper, we define Gabor filters as where  = (, ),  =  V exp() = ( V cos(),  V sin())  ,  V =  max / V ,  =  ⋅ /8,  = √ 2, and  =  [7]. is the frequency of a sinusoidal plane wave,  is the anticlockwise rotation of the Gaussian envelope and the sinusoid, and  and V represent the orientation and the scale of the Gabor filter respectively.In the experiments of this essay, we use Gabor filters of eight different orientations ( = {0, 1, . . ., 7}) with three frequency levels (V = {0, 1, 2}).The size of Gabor filters of three different frequency levels is set to 9 × 9, 13 × 13, and 17 × 17 pixels, respectively [7]. Figure 2 shows the output (norm of real and imaginary parts) of the Gabor filters with various , V.   Now, we introduce how to develop the orientation histogram from the output of Gabor filters.In this paper, the orientation histogram is developed from a regular 16 × 16 grid.Figure 3 lists an orientation histogram of the original image.First, for each scale, we extracted Gabor features (real and imaginary parts) of eight orientations from each grid of the input image.We compute the norm of real and imaginary parts for each pixel.Then, the orientation histogram with eight bins at each grid is developed by voting the output value of the maximum orientation at each pixel to the orientation bin [8].In order to reduce the impact of light changes, we normalized the histogram of each patch.This process is repeated at each scale parameter independently.So we finally get an orientation histogram with 24 bins (24 = 3 scales × 8 orientation bins) from each grid.
We fuse the orientation histogram and the first eigenimage (the color feature vector) by the same weight for each patch to form the descriptor of each patch.

Subspaces Similarities
The proposed method uses KPCA to model the similarities with subspaces and extract features specialized for each category.KPCA can achieve the transformation from input space to feature space through the nonlinear mapping Φ and then perform linear PCA on the mapped data, and so it has a strong nonlinear processing ability.However, the computational cost of KPCA becomes high when the number of training samples increases.To reduce the computational cost, we must decrease the number of training samples.To solve this problem, we use clustering algorithm.The cluster centers are selected from the descriptors of each category by using k-means method, and the cluster centers are used to construct the category-specific subspace by KPCA.The cluster centers are called visual words vividly, so KPCA was applied to the visual words of each category.Since the visual words of each category include various kinds of local parts, the distribution becomes nonlinear.Thus, the nonlinearity of KPCA is appropriate for modeling them.

Introduction of KPCA.
In the following, we introduce KPCA briefly [11].Suppose we have a set of  samples in an -dimensional space  1 ,  2 , . . .,   ∈   , and  is mapped into high dimensional space by nonlinear mapping Φ().
By applying standard linear PCA in high dimensional space, nonlinear principal components are obtained.Nonlinear Thus, the covariance matrix in high dimensional space is computed by Therefore, the PCA in feature space is to solve the eigenvalues  and the eigenvectors  for KPCA in equation  = .Where  is defined as: where   is the coefficient.Define the  ×  matrix  ( is the kernel function) as Then we obtain the following eigenvalue problem: By solving the formula (10), we can obtain the required eigenvalues, eigenvectors, and .
In this paper, the th descriptor   is projected to the subspace of each category, and the square of projection value to each axis is computed.The projection length represents the similarity with an axis in subspace of a certain category.The projection length    of the descriptor   to th principal component axis of the subspace of category  is computed as where  is the solution of KPCA,  is the number of visual words,   is the th visual word of the category , and  is the kernel function.This equation shows that the projection length is computed by the weighted sum of similarities with the ensemble of visual words because kernel function measures the similarity with visual words [7,11].Here we use the normalized polynomial kernel as the kernel function in KPCA [11].

Procedure of KPCA of Visual
Words.First, we get 256 descriptors from each training image of certain category.Then, by using -means clustering method, we obtain the  visual words.Next, KPCA is applied to  visual words.Note that each category has one subspace.Finally, we compute the mean projection length for each subspace of the certain category.In this paper, the mean projection length is used to integrate all descriptors to be robust to the order of the descriptors [11].In other words, th projection length in the subspace of category  is computed as The mean projection length to the subspace is used as the features for classification.
Let us consider the number of visual words ( in the -means vector quantization) and the number of subspace axes ( the number of principal components in KPCA).We investigate the variation of classification performance with change in the number of visual words and subspace axes for the case of the 4 natural categories of OT8 dataset.We carry out experiments with the original images in OT8 dataset.In our experiments, datasets are split randomly into two separate sets of images, half for training and half for testing.We take 100 random images from the training set to form a validation set for finding the optimal parameters, and the rest of the training images are used to compute the visual words.The visual words are learnt from about 30 random training images of each category.This protocol is the same as the conventional method [6].
We find the optimal parameters ( and ) over the validation set by using the cross-validation algorithm.Figure 4(a) shows the relationship between the classification performance and the number of visual words . Figure 4(b) shows the relationship between the classification performance and the number of subspace axes .To get the optimal parameters, we repeat the experiment ten times with varying random selection of the training and test sets and building the visual words vocabulary afresh each time.The mean classification rate of ten runs is used as a final result.
From Figure 4, we can see that the classification accuracy reaches the maximum value when the number of visual words  is 1500 and the number of subspace axes  is 250.Then the optimal parameters are fixed and subsequent experiments use the optimal parameters.

Classifications by SVM
We adopt SVM to the mean projection length to the subspaces of all categories.First, we explain the support vector machine (SVM) [8,17] briefly.The SVM algorithm seeks the optimal hyperplane which maximizes the distance between hyperplane and nearest sample from it.SVM is basically a hyperplane classifier () = ⟨, ⟩ +  aimed at solving the two-class problem.For nonlinearly separable data, a nonlinear mapping function Φ:   → ,  → Φ() is used to map them into a higher dimension feature space where the hyperplane classifier can be applied.When the training set (sample and its label) is denoted as  = ((  ,   ), . . ., (  ,   )), the nonlinear SVM using kernel is defined to be where SV shows the support vectors and   is the label of training sample   , and (  , ) is a kernel function, for example, a polynomial kernel, a linear kernel, a radial basis function (RBF) kernel, and so forth.However, since categoryspecific nonlinear processing is done by KPCA, SVM with linear kernel is enough to categorize many scene categories.Moreover, the computational cost of linear SVM is low because the decision function is computed by inner product.

Proposed Method for Scenes Classification
Most of the remote sensing images have a clear scene, such as desert, farmland, forest, and coast.So the multisensor remote sensing images are usually treated as scene images.This paper is trying to recognize the scenes of 8 different categories: coast, forest, highway, inside city, mountain, open country, street, and tall building, and the classifier is trained to identify an input image to the right scene category.The detailed work flow of the proposed method is shown in Figure 5.
The quaternion method and the orientation histogram of Gabor features are used to extract color features and grey features, respectively.Then they are integrated to form the descriptor of a patch.Next, we use the -means clustering algorithm to obtain the visual words of a certain category.After that, KPCA is applied to the visual words to get the subspace of the category.The descriptors of a test image then project to the subspaces of all categories to get the projection length to all categories of the test image.Finally, mean projection length is computed by (14), and is inputted into SVM classifier.

Evaluation Classification Results
In this paper, we consider three classification situations like the experiments in [6]: classification into eight categories, and also classification within the two subsets of natural (four categories) and man-made (four categories) images.Then we compare our method with the typical scene classification method pLSA in [6].
We repeat the experiment six times with varying random selection of the training and test sets and building the visual words vocabulary afresh each time.The mean classification rate of six runs is used as a final result.All parameters are fixed with the number of visual words  = 1500 and the number of subspace axes  = 250.All the experiments are carried by using Matlab on the same desktop computer with AMD Athlon(tm) II X2 3.0 G CPU.
Figure 6 illustrates the results of comparing our classification method with the typical scene classification methods in [6].We compare our classification results for four natural categories with the pLSA methods [4]. Figure 6(a) shows the result of this comparison.We only compare with the performance when classifying the four natural categories using unnormalized images and with nonoverlapping 11 × 11 patch in [6], since this situation in [6] is the closest to our experimental conditions in this paper.In our experiments, we use unnormalized images with nonoverlapping 16 × 16 patch.From Figure 6(a), we can get that the classification rate of our method is about 1% better than the performance described in [6].From [6], we can find that the size of patch is larger and that the classification accuracy is lower.In this paper, we use 16 × 16 patch larger than the 11 × 11 patch in [6], and so if we use 16 × 16 patch to extract features by using the method in [6], we can get more poor results.In [6], they use color patches, and they consider the three color components HSV and obtain a 112 × 3 = 363 dimensional vector.However, in our experiments, we have a 3 × 8 = 24 dimensional grey feature vector and a 16 × 16 = 256 dimensional color feature vector.So the total dimension of the descriptor for each patch is 24 + 256 = 280.As we all know, the computational cost will increase when the feature dimensions increase.Our method is superior to pLSA in computational cost and computation speed.To sum up, our method outperforms the method proposed in [6] in this situation.
In the following, we investigate the impact of color information for scene classification.Figure 6(b) shows the results when classifying the images of scenes with color patches and grey patches.The performance when using color patches is nearly 5%-6% better than that when using grey patches.Color information is an important factor in scene classification.Here, the color patch means fusing QSVD features and Gabor features; the grey patch means only Gabor features.
The comparative results of our method and pLSA proposed in [6] are illustrated in Table 1, where we can get that our method outperforms the pLSA method in classification accuracy and feature dimensions by using grey patches.For color patches, in [6] Bosch et al. eventually adopted the Color SIFT four Concentric Circles (C4CC) patch.The feature dimension of the C4CC patch is 3 × 4 × 128 = 1536, so the total dimension of features which are used to build the visual words vocabulary with the OT8 dataset is 1536 × 240 × 961 (1536 is the feature dimensions 240 = 30 × 8, where 30 is the number of images for building visual words for each category  and 8 is the number of categories; 961 is the number of patches in each image).When using single-precision floating-point integer to represent each feature dimension, it will totally take more than 2.5 G of memory to store the matrix used to build visual words.Therefore, this method will spend a lot of time in feature extraction, clustering, and quantification.The average processing time for each image is about 7.22 seconds.As a comparison, the feature dimension in our method is 24 + 256 = 280, and the total dimension of features used to build the visual words vocabulary with the OT8 dataset is 280 × 240 × 256.When using single-precision floating-point integer to represent each feature dimension, it only takes less than 0.13 G of memory to store the matrix used to build visual words.And the average processing time for each image in our method is about 3.06 seconds.Our method has an advantage over the method proposed in [6], regarding the run time, computation speed, and memory requirements.
Even if [6] has achieved higher classification accuracy than that of our method by using color patches, our method is superior to theirs in processing time, feature dimensions and hardware requirements.In some real-time emergency response systems, we need to consider the processing time firstly, then consider the hardware requirements, and finally consider the accuracy.In real-time emergency response systems about scene classification, accuracy is not the ultimate goal, quickly and accurately locating the target is the final demand.So the feature dimensions should not be too high.By our method, the feature dimension is lower than Bosch's method and processing time is also faster than [6].Thus our method can be used in real-time emergency response systems about scene classification.

Conclusions
In this paper, we propose a scene classification method for color scene images.We use quaternion matrices to represent the color images and extract color feature by quaternion singular value decomposition.In this way, we take into account the direct interchannel effects in the RGB channels and reduce the dimension of the feature vector.To get the optimal texture information in time domain and frequency domain, we use Gabor filter to extract grey features.We also find the optimal parameters for getting the best classification accuracy by cross-validation experiments.From the evaluation results, we can observe that our method can classify the input image to the right category fastly and accurately.This work can be further extended to classify scene images in some real-time emergency response systems.In our method, for building the visual words vocabulary, we adopt the k-means algorithm, but the calculation of this algorithm is high.In the future work, we can consider the improvements in clustering and quantitative technologies to improve the classification performance.We also will look for better ways to improve the classification rate.

Figure 2 (
Figure 2 shows the output (norm of real and imaginary parts) of the Gabor filters with various , V.Figure 2(a) shows the original image.Figure 2(b) is the Gabor output when Gabor filters with  = 4, V = 0 are applied to Figure 2(a).Figure 2(c) is the output of the Gabor filters with  = 4, V = 1.Figure 2(d) is the output of the Gabor filters with  = 4, V = 2. From the figure, we can see that Gabor filter with small V being sensitive to fine edges, so the output image (b) is clearer than (c) and (d).Now, we introduce how to develop the orientation histogram from the output of Gabor filters.In this paper, the orientation histogram is developed from a regular 16 × 16 grid.Figure3lists an orientation histogram of the original image.First, for each scale, we extracted Gabor features (real and imaginary parts) of eight orientations from each grid of the input image.We compute the norm of real and imaginary parts for each pixel.Then, the orientation histogram with eight bins at each grid is developed by voting the output value of the maximum orientation at each pixel to the orientation bin[8].In order to reduce the impact of light changes, we normalized the histogram of each patch.This process is repeated at each scale parameter independently.So we finally get an orientation histogram with 24 bins (24 = 3 scales × 8 orientation bins) from each grid.

Figure 2 (
Figure 2 shows the output (norm of real and imaginary parts) of the Gabor filters with various , V.Figure 2(a) shows the original image.Figure 2(b) is the Gabor output when Gabor filters with  = 4, V = 0 are applied to Figure 2(a).Figure 2(c) is the output of the Gabor filters with  = 4, V = 1.Figure 2(d) is the output of the Gabor filters with  = 4, V = 2. From the figure, we can see that Gabor filter with small V being sensitive to fine edges, so the output image (b) is clearer than (c) and (d).Now, we introduce how to develop the orientation histogram from the output of Gabor filters.In this paper, the orientation histogram is developed from a regular 16 × 16 grid.Figure3lists an orientation histogram of the original image.First, for each scale, we extracted Gabor features (real and imaginary parts) of eight orientations from each grid of the input image.We compute the norm of real and imaginary parts for each pixel.Then, the orientation histogram with eight bins at each grid is developed by voting the output value of the maximum orientation at each pixel to the orientation bin[8].In order to reduce the impact of light changes, we normalized the histogram of each patch.This process is repeated at each scale parameter independently.So we finally get an orientation histogram with 24 bins (24 = 3 scales × 8 orientation bins) from each grid.

Figure 2 (
Figure 2 shows the output (norm of real and imaginary parts) of the Gabor filters with various , V.Figure 2(a) shows the original image.Figure 2(b) is the Gabor output when Gabor filters with  = 4, V = 0 are applied to Figure 2(a).Figure 2(c) is the output of the Gabor filters with  = 4, V = 1.Figure 2(d) is the output of the Gabor filters with  = 4, V = 2. From the figure, we can see that Gabor filter with small V being sensitive to fine edges, so the output image (b) is clearer than (c) and (d).Now, we introduce how to develop the orientation histogram from the output of Gabor filters.In this paper, the orientation histogram is developed from a regular 16 × 16 grid.Figure3lists an orientation histogram of the original image.First, for each scale, we extracted Gabor features (real and imaginary parts) of eight orientations from each grid of the input image.We compute the norm of real and imaginary parts for each pixel.Then, the orientation histogram with eight bins at each grid is developed by voting the output value of the maximum orientation at each pixel to the orientation bin[8].In order to reduce the impact of light changes, we normalized the histogram of each patch.This process is repeated at each scale parameter independently.So we finally get an orientation histogram with 24 bins (24 = 3 scales × 8 orientation bins) from each grid.

Figure 3 :
Figure 3: (a) Original image; (b) the orientation histogram with eight different orientations and three scales of the original image (the different colors represent different scales).

Figure 4 :
Figure 4: (a) Varying number of visual words  and fixing number of subspace axes  = 250; (b) varying number of subspace axes  and fixing number of visual words  = 1500.

Figure 5 :Figure 6 :
Figure 5: Detailed description of the proposed method for scenes classification.

Table 1 :
Comparative results of our method and pLSA.