Rotation Covariant Image Processing for Biomedical Applications

With the advent of novel biomedical 3D image acquisition techniques, the efficient and reliable analysis of volumetric images has become more and more important. The amount of data is enormous and demands an automated processing. The applications are manifold, ranging from image enhancement, image reconstruction, and image description to object/feature detection and high-level contextual feature extraction. In most scenarios, it is expected that geometric transformations alter the output in a mathematically well-defined manner. In this paper we emphasis on 3D translations and rotations. Many algorithms rely on intensity or low-order tensorial-like descriptions to fulfill this demand. This paper proposes a general mathematical framework based on mathematical concepts and theories transferred from mathematical physics and harmonic analysis into the domain of image analysis and pattern recognition. Based on two basic operations, spherical tensor differentiation and spherical tensor multiplication, we show how to design a variety of 3D image processing methods in an efficient way. The framework has already been applied to several biomedical applications ranging from feature and object detection tasks to image enhancement and image restoration techniques. In this paper, the proposed methods are applied on a variety of different 3D data modalities stemming from medical and biological sciences.


Introduction
The analysis of three-dimensional images has gained more and more importance in recent years. Particular in the medical and biological sciences, new acquisition techniques lead to an enormous amount of 3D data calling for automated analysis. In this paper, we show how the harmonic analysis of the 3D rotation group offers a convenient and computationally efficient framework for rotation covariant image processing and analysis. Most of the state-of-the-art techniques rely on "low"-order features such as intensities, gradients of intensities, or second order tensors like the Hessian matrix or the structure tensor [1]. For example, consider a lesion detection/segmentation problem in a 1 -weighted magnetic resonance image. A typical procedure for solving such a task would rely on a local image feature extraction step such as the computation of a Laplacian-or a Gaussian-pyramid. Once the feature images are computed, a healthy group of volunteers is used to determine the distribution of such features for subjects in a healthy condition.
From such a distribution, we can estimate the probabilities for the absence or presence of lesions in a voxel-byvoxel manner. Instead of solely using 0-order features, such as the Laplacian-pyramid, higher order tensor fields can be used to derive further scalar valued quantities. Such features can be the smoothed intensity gradient magnitudes (1-order features), or the eigenstructures of a Hessian matrix field or a structure tensor field (2-order features). However, due to their mathematical and computational complexity, features of order three or even higher order are rarely used. This paper proposes a unified framework that can cope with highorder features in a systematic way. The proposed framework is based on the harmonic, irreducible representations of the 3D rotation group. This guarantees the most sparse tensor representation. Consequently, in comparison with ordinary Cartesian tensor analysis, the algorithms and the handling are operationally clearer and more efficient.
Given a Cartesian tensor 1 ,..., of rank , such a tensor can be the result from a simple projection onto moment functions or from a differentiation process (for instance, a gradient is a tensor of order 1, and a Hessian matrix is a tensor of oder 2). A tensor can be considered as a feature describing an object in a rotation covariant way; that is, if the original object is rotated by a rotation matrix R, the tensor rotates in the following manner: ( ) 1 ,..., = ∑ 1 ,..., where denotes an element of the 3D rotation group. A tensor rotation is a common operation in many applications, for instance, steering a local image descriptor (a tensor) with respect to some data dependent reference frame. From a computationally point of view a Cartesian tensor rotation is quite inconvenient. Typically, there are symmetries with respect to index permutations (for instance, the Hessian matrix is a symmetric tensor). These symmetries have to be taken into account to provide an efficient computation. Another problem is that the tensor rotation matrix 1 , 1 ⋅ ⋅ ⋅ , is "full"; that is, all elements 1 ,..., mix under rotations. Spherical tensor analysis, where tensors appear in their irreducible representations, solves these problems, and, even more, it offers further advantages regarding tensor operations. Suppose that we aim at extracting rotation invariant features. Given a Cartesian tensor , for Cartesian tensors, the basic operation is tensor contraction. The tensor is contracted down to a rank 0 tensor by repeatedly combining two indexes (with the Kronecker delta ), or three indexes (with the -tensor ). This can be done in several ways, for example, linearly ∑ ⋅, ,⋅, ,⋅ , quadratically ∑ ⋅, ,⋅ ⋅, ,⋅ , or even cubicly ∑ ⋅, ,⋅ ⋅, ,⋅ ⋅, ,⋅ . It is possible to combine different tensors as well. A problem that occurs is ambiguities; in the presence of symmetries, some ways might end up in the same result, some may not. In contrast, the spherical tensor analysis offers a systematic way of performing such operations. The Kronecker and products are replaced by one single spherical product which allows for multiplying spherical tensors of arbitrary rank. The even parity products are related to the Kronecker product, the odd parity products to products.
In this paper, we want to review the basics of spherical tensor analysis and how it can be applied to image processing problems. In Section 2, we introduce the basic concepts such as the notion of a spherical tensors. We define the spherical product and introduce its properties. We also show how spherical tensors are related to ordinary Cartesian tensors. In Section 3, the so-called spherical tensor derivative operators (shortly spherical derivatives) are introduced. The spherical derivative operators are able to connect spherical tensor fields of different rank. We discuss several properties and derive their representation in polar coordinates. We focus on two types of basis systems evolving from the spherical differentiation process: the Gauss-Laguerre functions and the spherical Gabor-functions. Both are known to be very important in pattern analysis. The differential relationships of these functions offer an efficient way to compute projections onto these type of functions. In Section 4, expansions in terms of tensorial harmonics are discussed, which are just the straight-forward generalization of ordinary scalar-valued spherical harmonic expansions. Finally, in Section 5, several biomedical applications are reviewed and discussed.

Spherical Tensor Analysis
Let D be the unitary irreducible representation of a ∈ (3) of order with ∈ N. They are also known as the Wigner D-matrices (e.g., see [3]). The representation D acts on a vector space which is represented by C 2 +1 . We write the elements of in bold face, for example, u ∈ , and write the 2 + 1 components in unbolt face ∈ C where = − , . . . , . For the transposition of a vector/matrix, we write u ; the joint complex conjugation and transposition is denoted by u ⊤ = u . In this terms, the unitarity of D is expressed by the formula (D ) ⊤ D = I.
Note that we treat the space as a real vector space of dimensions 2 + 1, although the components of u may be complex. This means that the space is only closed under weighted superpositions with real numbers. As a consequence of this, we always have that the components are interrelated by = (−1) − . From a computational point of view, this is an important issue. Although the vectors are elements of C 2 +1 , we just have to store just 2 + 1 real numbers.
We denote the standard basis of C 2 +1 by e , where the th component of e is . In contrast, the standard basis of is written as c = ((1 + i)/2)e + (−1) ((1 − i)/2)e − . We denote the corresponding "imaginary" space by i ; that is, elements of i can be written as ik where k ∈ . So, elements w ∈ i fulfill = (−1) +1 − . Hence, we can write the space C 2 +1 as the direct sum of the two spaces C 2 +1 = ⊕ i . The standard coordinate vector r = ( , , ) ∈ R 3 has a natural relation to elements u ∈ 1 by Note that S is an unitary coordinate transformation. The representation D 1 is directly related to the real-valued rotation matrix U ∈ (3) ⊂ R 3×3 by D 1 = SU S ⊤ .
is called a spherical tensor field of rank if it transforms with respect to rotations as for all ∈ (3). The space of all spherical tensor fields of rank is denoted by T .

Spherical Tensor
Coupling. Now, we define a family of bilinear forms that connect tensors of different ranks.
Computational and Mathematical Methods in Medicine 3 Definition 2. For every ≥ 0, we define a family of bilinear forms of type where 1 , 2 ∈ N has to be chosen according to the triangle inequality | 1 − 2 | ≤ ≤ 1 + 2 . It is defined by where ⟨ | 1 1 , 2 2 ⟩ are the Clebsch-Gordan coefficients.
The characterizing property of these products is that they respect the rotations of the arguments. Proposition 3. Let k ∈ 1 and w ∈ 2 ; then, for any ∈ (3), holds.
Proof. The components of the left-hand side look as First one has to insert the identity by using orthogonality relation (B.1) with respect to 1 and 2 . Then, we can use relation (C.2) and the definition of ∘ to prove the assertion.
Proposition 4. If 1 + 2 + is even, then ∘ is symmetric, otherwise antisymmetric. The spaces are closed for the symmetric product, and for the antisymmetric product this is not the case. Consider where k ∈ 1 and w ∈ 2 . Proof. The proposition is proved by the symmetry properties of the Clebsch-Gordan coefficients (B.6). To show the closure property, consider Hence, we have for even + 1 + 2 the "realness" condition complying to and for odd + 1 + 2 the "imaginariness" condition for i , which prove the statements.
We will later see that the symmetric product plays an important role, in particular, because we can normalize it in an special way such that it shows a more gentle behavior with respect to the spherical harmonics.
Definition 5. For every ≥ 0 with | 1 − 2 | ≤ ≤ 1 + 2 and even + 1 + 2 , we define a family of symmetric bilinear forms by For the special case = 0, the arguments have to be of the same rank due to the triangle inequality. Actually in this case, the symmetric product coincides with the standard inner product where is the rank of k and w.
The introduced product can also be used to combine tensor fields of different rank by point-wise multiplication. Proposition 6. Let k ∈ T 1 and w ∈ T 2 and chosen such that | 1 − 2 | ≤ ≤ 1 + 2 ; then, is in T , that is, a tensor field of rank .
In fact, there is another way to combine two tensor fields: by convolution. The advantage of the convolution is that the evolving product also is covariant with respect to translation; that is, the product is covariant to 3D Euclidean motion.

Proposition 7.
Let v ∈ T 1 and w ∈ T 2 and chosen such that | 1 − 2 | ≤ ≤ 1 + 2 ; then, is in T , that is, a tensor field of rank .
Given a translation , the following two relations hold: Further important properties of the products are their associativity rules.

Spherical and Solid
Harmonics. Due to their special properties, the spherical harmonics (see, Appendix A for definition) play the central role in spherical tensor analysis. One of the most important ones is that each Y , interpreted as a tensor field of rank , is a fix-point with respect to rotations; that is, Consequently, The Y form an orthogonal and complete basis of the functions defined on the 2-sphere. Hence, any real squareintegrable scalar field ∈ T 0 can be written as A band-limited spherical harmonic representation of two images is illustrated in Figure 1.
The expansion coefficients of the rotated function ( )(r) = (U ⊤ r) are simply D a ( ), which can be concluded from the fix-point property. In the following, we always use Racah's normalization (also known as semi-Schmidt normalization); that is, where the integral ranges over a sphere using the standard measure. With this, the coupling of two spherical harmonics gives, again, a spherical harmonic From a computational perspective, this property can be used to efficiently compute higher order harmonics for lower ones.
Besides the spherical harmonics, the so-called solid harmonics, often appear in the context of harmonic analysis of the 3D rotation group. They are the homogeneous solutions of the Laplace-equation and are just related by R := Y , and they are homogeneous polynomials of degree ; that is, R ( r) = R (r).

Relation to Cartesian
Tensors. The correspondence of spherical and Cartesian tensors of rank 0 is trivial. For rank 1, it is just the matrix S that connects the real-valued vector r ∈ R 3 with the spherical coordinate vector u = Sr ∈ 1 . For rank 2, the consideration gets more intricate. Consider a realvalued Cartesian rank-2 tensor T ∈ R 3×3 and the following unique decomposition: where ∈ R, T anti is an antisymmetric matrix, and T sym a traceless symmetric matrix. In fact, this decomposition follows the same manner as the spherical tensor decomposition. A rank 0 spherical tensor corresponds to the identity matrix in Cartesian notation, while the rank 1 spherical tensor to a antisymmetric 3 × 3 matrix or, equivalently, to a vector. And finally, the rank 2 spherical tensor corresponds to a traceless, symmetric matrix. So, let us consider the spherical decomposition. For convenience, let T = STS ⊤ ; then, the components of the corresponding spherical tensors b with = 0, 1, 2 are where b 0 corresponds to , b 1 to T anti and b 2 to T sym . Explicitly, the relation to T is ) .
The inverse of this "Cartesian to spherical"-transformation is Note that for arbitrary ranked Cartesian tensor, the relations are not that trivial.

Spherical Derivatives
This section proposes the concepts of differentiation in the context of spherical tensor analysis. First, we will introduce Computational and Mathematical Methods in Medicine 5 Original < 1 < 4 < 5 < 9 < 15 < 32 < 32 < 2 < 3 < 5 < 7 < 9 < 16 < 32 Figure 1: A spherical harmonic decomposition of images can be seen as some kind of frequency decomposition. A band limited expansion of a volumetric images is illustrated. We see that lower frequency components (right-hand side) are roughly representing the important characteristics of the objects. However, higher frequency components are necessary to represent the details. For the expansion here, we use a Fourier-like basis for representing the images in radial direction. Here, ℓ represents the order of the spherical harmonics and the number of radial frequency components taken into account. The image shows an isosurface rendering together with the centered , , and -slice. The interested reader is referred to [2]. the spherical derivative operator which connects spherical tensor fields of different ranks by differentiation. The basic idea is simple; formally replace the coordinates r = ( , , ) appearing within the solid harmonics R by the gradient operator ( , , ).
In fact there are much more rotation covariant differential operators than the two defined previously. Given a tensor field f, any field of the form g = R (∇)• ℓ f, which we obtain via differentiation, is a spherical tensor field, too.
But the up-and down-derivatives are from a computational point very attractive, because, as shown earlier, they allow an iterative computation of higher order differentials, which is computationally much more efficient than the direct way. For further discussion on the spherical tensor derivative operator, consider the spherical derivatives in the Fourier domain, where they act by point-wise •-multiplications with a solid harmonic i Y 1 (k) = iR 1 (k) = iSk where = |k| is the frequency magnitude.
Proposition 10 (Fourier representation). Letf(k) be the Fourier transformation of some f ∈ T ℓ and∇ representations of the spherical derivative in the Fourier domain that are implicitly defined by(∇f) =∇f; then, Proof. See [4].
Both statements are direct consequences of the Fourier correspondences for the ordinary partial derivatives. For scalar fields, we can generalize this statement also for higher orders.
In the Fourier domain, these multiple derivatives act by Using this one can show that Proof. See [5].
We want to emphasize that both statements only hold for scalar-valued fields, and generalizations to tensor-valued do not hold in general due to the nontrivial associativity rules.
Proposition 12 (product rule). Let f ∈ T ℓ and ℎ ∈ T 0 ; then, one has the product rules It is well known that convolutions commute with differentiation, and actually there are generalized commutation rules for spherical tensor fields.
Proof. Both assertions are founded by the associativity of the spherical product. Consider the first statement in the Fourier domain by using (28) and then apply the associativity given in (17) as follows: where we abbreviated R 1 = R 1 (ik). A repeated application of this proves the first assertion. For the second statement, it is similar but using the associativity as given in (15).
This proposition shows again the importance of the up-and down-derivatives. For general derivative operators R (∇)• ℓ f, the previous commutations rules do not hold. The previous convolution property is of particular importance for the efficient covariant processing of 3D images. The major motivation is to compute convolutions with the spherical harmonic basis in an efficient way. Suppose that the goal is to compute where is some arbitrary scalar image. In fact, as we will show in the next section, one can show that ∇ − 2 /2 = (−1) R − 2 /2 . Together with the convolution theorem, we get which enables us to compute the convolution by an repeated application of the spherical derivatives, which is computationally much cheaper than a direct convolution (even by the use of the Fast Fourier Transform).

Spherical Derivatives in Polar Representation.
To get a better understanding of what happens during the differentiation via spherical derivatives, we consider their properties in polar representations.

Lemma 14.
Given a spherical tensor field f ∈ T whose angular and radial component are separable such that f (r) = Y (r) ( ), where : R → C denotes the function representing the radial component of f , then the spherical upand down-derivatives of f can be computed by respectively.

Gauss-Laguerre Functions.
Previously, we already stated that ∇ − 2 /2 = (−1) R − 2 /2 holds; in fact, there is a more general statement involving the so-called Laguerre polynomials. This offers the possibility to compute convolutions with the evolving functions in an iterative and efficient way. We denote by the associated Laguerre polynomial of order (F.1). We further denote by the spherical tensor valued polynomials L ∈ T − . These polynomials are widely known as Laguerre Gaussian-type functions in the field of theoretical chemistry (e.g., see [8] or [9]). In the image processing community, these functions are known as generic neighborhood operators [10] and are used, for example, for key-point detection [11].

Gabor Functions.
Gabor functions, that is, Gaussianwindowed plane waves, play an important role in image processing due to the fact that the different frequency components of signals can be studied locally. This information is, for example, used for tracking [12] or feature extraction [6]. Thus, it is of particular interest to provide efficient methods to apply Gabor filters. One way is to explicitly represent a finite number of Gabor kernels, each representing a certain orientation of the plane-wave [13]. The problem is that the orientation space must be discretized. However, representing Gabor functions in terms of spherical derivatives offers a way to compute Gabor filter responses for the whole range of possible orientations. First, note that applying spherical derivatives on a plane wave gives a quite neat result as Following the proof from Section 3.1, a similar result holds for the spherical Bessel function, which constitutes the radial part in the harmonic expansion of the plane wave as In the following, we show that there exists a very similar way to represent the Gaussian windowed wave in terms of the derivatives of the Gaussian windowed Bessel functions. Let where ( ) ∈ R are real-valued weighting factors.

Tensorial Harmonic Expansions
In most image processing applications, the data to be processed is of scalar nature; that is, for each voxel, we observe one single intensity value. But there are actually acquisition techniques, where the measurement itself is already a tensorial quantity. For example, in diffusion weighted magnet resonance imaging (DW-MRI), rank 2 tensors are common. Or, in phase contrast MRI velocity, vectors are measured. Thus, there is a great interest to represent these measurement in an appropriate way. In [14], we proposed to expand a spherical tensor field f ∈ T ℓ of rank ℓ as follows: where a ( ) ∈ T + are expansion coefficients. For ℓ = 0, the expansion coincides with the ordinary scalar spherical harmonic expansion. We can observe properties very similar to the ordinary SH expansion; that is, A rotation of the tensor field affects the expansion coefficients a to be multiplied from the left with D + . So, the previous expansion shows the same, very convenient, rotation behavior like an SH expansion, which can be used, for example, to extract invariant local descriptors in a simple way. And in fact, the previous representation is orthogonal and complete. By setting a ( ) = ∑ = + =−( + ) ( )e + , we can identify the functional basis Z as Proposition 17 (tensorial harmonics). The functions Z : 2 → ℓ provide a complete and orthogonal basis of the angular part of T ℓ , that is; where , = 1 2ℓ + 1 (2 + 1) (2 ( + ) + 1) .
The functions Z are called the tensorial harmonics.

Symmetric Tensor Fields.
In this section, we discuss the properties of expansion coefficients of specific tensor fields, expanded in terms of tensorial harmonics. We show that symmetries in a tensor field are simplifying the tensorial harmonic expansion coefficients. This is similar to the ordinary spherical harmonic expansion. For example, the point symmetry (r) = (−r) of a scalar fields leads to vanishing spherical harmonic coefficients for odd . In the following, we consider similar symmetries for tensorial harmonics.
The rotation symmetry of a spherical tensor field f ∈ T ℓ around the -axis is expressed algebraically by the fact that f = f for all rotation around the -axis. Such fields can easily be obtained by averaging a general tensor field f over all these rotations as It is well known that the representation D of such a rotation is diagonal; namely, , = i . Hence, the expansion coefficients of f vanish for all ̸ = 0. Thus, we can write any rotation symmetric tensor field as We call such a rotation symmetric field torsion-free if f = f , where ∈ (3) is a reflection with respect to the -plane (or -plane). The action of such a reflection on spherical tensors is given by , = (−1) (− ) . Similar to the rotational symmetry, we can obtain such fields by averaging over the symmetry operation as Note that the mirroring operation for a spherical harmonic is just a complex conjugation; that is, Y (U r) = Y (r). The consequence for (53) is that all terms where the + ℓ are odd vanish. The reason for that is mainly Proposition 4 because with its help we can show that holds. Finally, consider the reflection symmetry with respect to the -plane. This symmetry is particularly important for fields of even rank. The symmetry is algebraically expressed by f = f where ∈ (3) is a reflection with respect to the -plane, whose action on spherical tensors is given by , = (−1) . Averaging over this symmetry operation has the consequence that expansion terms with odd are vanishing. For odd rank tensor fields, the reflection symmetry is not imperative. But there is typically an antisymmetry of the form f = −f . This antisymmetry lets the expansion terms vanish with even index .

Applications
In the context of rotation covariant image processing, the applications of the proposed framework are manifold. The mathematical representation might appear unfamiliar, but the provided tools can be used quite easily. Basically, there are two types of operations: differentiation by spherical tensor derivatives and multiplication by spherical tensor products. The spherical derivatives can be used in two ways. On the one hand, the up-derivatives can be used to "create" new tensor fields out of existing fields by incorporating neighborhood relations. This can be regarded as a simple and efficient way to compute local meaningful image descriptors in a covariant way. On the other hand, the down-derivatives can be used to gather information from a local point neighborhood and form a lower ranked tensor field via superposition. Due to the tensorial nature, the information is able to interfere in a destructive or constructive way. The spherical products are the basic nonlinear ingredient in the framework. They can be used to combine tensor fields in a nonlinear, covariant manner.
Several principles in the image processing and pattern recognition [15][16][17] literature are based on the following principle: compute, in a first step, local descriptors at several image locations, make some inference based on this knowledge, and cast this information back by combining evidence from several locations. In fact, our framework is ideally suited to adopt this principle. First, local descriptors are densely computed by differentiation for all image locations. Then, the information is combined by using spherical products in a nonlinear and nontrivial way. Finally, we use again the spherical derivative to form neighborhood descriptors. The descriptors are then used for object or feature detection.
In the following, we give examples of the proposed framework in several application domains.

Implementation.
For implementing the discrete spherical derivatives, we propose to utilize central differences of 4th order accuracy for computing the partial derivatives (see Figure 2(b)). We observed that this scheme is a good tradeoff between computational complexity and accuracy. We experienced that the standard Laplace operator (considering a six voxel neighborhood) is numerically very unstable (even if double precision numbers are used!). Therefore, we propose the usage of the scheme depicted in Figure 2(a) which performed significantly better regarding numerical stability in our experiments. This is illustrated in Figure 3. As an example, we show the expansion images obtained via the proposed schemes together with the images obtained via a standard scheme. For comparison, we also show explicitly computed expansion images. The example illustrates that the ordinary Laplace operator leads to strong artifacts after a few number of applications.

Tensor
Voting. The Tensor Voting framework was originally proposed by [15] and has found several application in low-level vision in 2D and 3D. For example, it is used for perceptual grouping and extraction of line, curves, and surfaces. The key idea is to make unreliable measurements more robust by incorporating neighborhood information in a consistent and coherent manner. Following [4], the key expression that has to be computed is where V n : R 3 → ℓ is the voting field, : R 3 → R a scalar valued feature image giving evidence for the occurrence of the feature, and n : R 3 → R 3 the orientation of the feature of interest. In the following, we restrict ourselves to axial symmetric voting fields. Therefore, let be a axial symmetric function, where the -axis is the symmetry axis. Then, the voting field is where n is a rotation such that the -axis is mapped onto the axis defined by the normalized vector n. In [4], we have shown that (56) simplifies to where are combined tensor-valued evidence images and is the harmonic expansion of the voting field V r steered in -direction. The coefficients ( ) can be obtained by a projection on the tensorial harmonics Due to the symmetry of V r , only Z 0 are involved. Further information concerning a practical point of view can be found in [14].

Nonlinear Covariant Filters.
In the following, we briefly show how to design trainable rotation covariant image filters which can be used for rotation invariant object or landmark detection. The idea is that expansion coefficients of a spherically expanded voting function are learned in a data driven way. The filter is mainly based on two steps. Rotation covariant image descriptors are densely computed in a voxelby-voxel manner. Then, a weighted superposition of these image descriptors is used to form expansion coefficients of a spherical voting function. The expansion coefficients are formed such that each voting function votes for the presence or absence of landmarks or objects. The weights are found by a least square fit to a given training data set. For a fast implementation, we propose to use voting functions based on an expansion of spherical functions having a differential relationship in terms of spherical derivatives. In [18,19], we used a spherical superposition of Gaussian windowed solid harmonics for representing the voting function. However, we are not restricted to them. For instance, we also can use the spherical plane-wave expansion leading to a voting function that is not only highly adaptable in angular direction, but also highly adaptable in radial direction, too; see the paper by [20]. The Fourier like voting function can be written as where V ( ) ∈ C 2 +1 are the expansion coefficients of the filter and B are spherical Fourier basis functions known as Bessel functions (see (43)). The filter response is a saliency map representing the evidence for the presence or absence of objects. The saliency map is computed by collecting all contributions (votes) utilizing simple scalar valued convolutions. The explicit expression of the filter is For implementation we use a band-limited expansion (up to order ∈ N) and only take a small set of frequencies ( 0 , ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ , ∈ R) into account. We further make use of Gabor waves (see Theorem 16) to gain a filter that adapts and votes locally. In this case, the filter simplifies to Trainable filters based on the Gabor waves have shown superior performance over the standard harmonic filters [20].   Figure 4 shows some qualitative results of an experiment where we detect the pores of airborne-pollen. The database contains 3D recordings of airborne-pollen acquired via a confocal laser scanning microscope. In Figure 4(a), we see the training image. The three porates are marked by red circles. In Figure 4(b), we exemplary show three datasets belonging to the test set together with the maximum intensity projection of the filter response.

Voxel-Wise Classification.
Especially in the field of biomedical imaging, the third dimension becomes more and more important due to the fact that organism can be studied in their natural constellation. Objects and organism can be located in any number at any position and, much more challenging, in any orientation. The third dimension does not only lead to larger datasets, but also the interrelation of neighboring intensity values becomes more complex. With a fast voxel-wise transformation of volumetric images into the harmonic domain, we are capable to compute rotation invariant image descriptors in an analytical way. In [6,7], we used a fast Gabor transform to locally analyze images by decomposing local image patches into basic frequency components. For the experiments, we used confocal recordings of Arabidopsis root tips. We exemplary aimed at detecting differentiated cells located in the root cap. They morphologically differ from the other cells by their nonroundish shape. For this experiment, two datasets were used: one dataset for training and one dataset for evaluation. All cells (about 3600 in each root) were manually labeled by an expert. We transformed the Gabor expansion coefficients into invariant features utilizing the spherical tensor product; we combine the expansion coefficients corresponding to the same angular frequency, but not necessarily the same radial frequencies, whereas where ( 1 , 2 ) ∈ C are the rotation invariant image descriptors. It is worth mentioning that the combination of the same expansion coefficient coincides with the powerspectrum; namely, ( ) = (a ( )• 0 a ( )) = ‖a ( )‖ 2 . In Figure 5(a), we depict the center slice of the training data together with the training samples. Based on the rotation invariant image descriptors representing the training samples, an SVM classifier is trained. We used the SVM to classify test-set in a voxel-by-voxel manner (Figures 5(b) and 5(c)). We classed each voxel into root-cap cell or non-root-cap cell. For further details regarding the experiment, we refer to [6,7].

DTI Processing.
Diffusion weighted magnetic resonance imaging (DWI) plays a substantial role in neuroscience and clinical applications. One field of interest is the investigation of the neuronal fiber architecture located in the brain white matter connecting different regions in the brain. The fibers themselves cannot be recorded directly. However, the data is usually recorded using the high angular resolution diffusion imaging (HARDI) technique [21], a specific kind of diffusion tensor imaging (DTI) technique. The resulting signal is an angular dependent, volumetric image. From such an image representation, the fiber architecture can be estimated (e.g., see [22]). Due to the angular dependency of HARDI signals, spherical harmonics are a common tool for signal representation. Therefore, in the context of DTI, there exist several applications worth considering spherical tensor algebra.

Tissue Classification.
For the analysis of the fiber structure, a preprocessing step that identifies the brain white matter within the image is required. For group studies, the parcellation of the human brain into anatomical regions is of great interest. Preliminary results have been published in conference papers [23,24]. We utilize the fact that the given recordings are tensor valued. We first transform the local measurements into the spherical harmonic domain (e.g., see [25]). Based on these rotation covariant image representations, we compute voxelwise rotation invariant image features. This is done by first comprising the voxels surrounding using the spherical down derivative operators. This can be seen as some kind of Taylor expansion of the given data. Then, we compute rotation invariant image features by computing the power spectrum of the resulting expansion coefficients. We finally use a random forest classifier [26] to learn the appearance of different kinds of brain regions and tissue types based on labeled training images. Such a parcellation might be for example, gray brain matter, white brain matter, and background signal. Qualitative results showing the resulting decisions of the random forest on an unclassified image are shown for the gray matter/white matter scenario in Figure 7. . In (c) we depict an isosurface rendering of the classified root. Further details concerning the experiment can be found in [6].
Ground truth regions used in experiment 1

Prediction in data 1
Ground truth regions used in experiment 2 Prediction in data 1 Figure 6: The ground truth regions that we used to train and evaluate our algorithm shown together with our algorithm's regions prediction. We can clearly see that our predictions are much more consistent with the data. Figure 7: Isosurface showing the predictions for dataset 3 using GND and a random forest (RF) classifier. The classifier can distinguish between background, brain white matter (green), and gray matter (red).

13
Background Gray matter White matter Figure 8: The confidence of the classifier represents the probability that a certain voxel belongs either to the background class, gray matter class, or the white matter class. The probability is represented by the intensity. A final decision is made by decision by majority (as shown in Figure 7). Furthermore, the votes for a certain class can be used as a kind of evidence value in further processing steps. Examples for the three classes background, white matter, and gray matter are depicted in Figure 8.
In Figures 9 and 10, we show the probability map of different kinds of brain regions that have been detected within unlabeled test images via a random forest classifier. Figure 6 shows final predictions for one of the test sets.

Unique Point-Landmark Detection.
Group studies often require the coregistration of images or partial image structures of different individuals. In such applications, the detection of characteristic landmarks is often an indispensable prerequisite.
Similar to [27], where features are used to find correspondences in scalar valued MR contrasts, we used tensorbased features in [28] offering a unique signature of a voxel's surrounding in tensor-valued HARDI signals. Thanks to these features, a large number of corresponding points can be reliably found in images of different individuals using a linear classifier. The features are computed in three steps. (1) We first entirely fit the HARDI signal to spherical harmonics.
(2) The resulting fields are then efficiently expanded in terms of tensorial harmonics (Section 4) via tensor derivatives (see Section 3). (3) We obtain new covariant feature images which we use to form a trainable filter (see Section 5.3). The filter is used for the landmark detection task.
Second-order features which are sufficient for most applications are not providing enough information to solve the detection task in a human brain; they are invariant against reflection about an axis. Hence, they cannot distinguish the left and the right hemisphere. It is known that the spherical triple-correlation [29] yields complete rotation invariant features. Hence, they must solve this issue. Based on this idea we designed new 3rd order rotation invariant differential features fitting into our framework that are variant with respect to reflections about an axis. These features are additionally included in the harmonic filter framework. The triple product is given by where b 1 ∈ C 2 1 +1 , b 2 ∈ C 2 2 +1 , b 3 ∈ C 2 3 +1 are the local tensorial harmonics expansion coefficients. A proof can be found in [28].
The resulting filter has shown very promising results on a training set of 7 and a test set of 14 images. For the experiment, we placed about 20000 landmarks within the brain gray and white matter in an equidistant manner. For each dataset, the computation of the features and the detection of of all landmarks took about 5 minutes. We show some detection results in Figures 11, 12, 13, and 14.

A. Spherical Harmonic Functions
The Schmidt seminormalized spherical harmonics : where are the with associated Legendre polynomials of order [30]. The spherical harmonics build a complete orthogonal basis for functions on the 2-sphere, whereas

D. Spherical Bessel Functions
The spherical Bessel functions : R ≥0 → R are related to the Bessel functions of the kind ] (e.g., see [30]) For the spherical Bessel functions, we have the following differential relations [30]:

E. Plane Wave
Using the addition theorem of the spherical harmonics, we can express the spherical expansion of the plane wave (e.g., see [3, page 136]) in terms of the tensor product • 0 leading to where are the Legendre polynomials [30] of order and Y = ( − , . . . , ) the semi-Schmidt normalized spherical harmonics written as vector.

F. Associated Laguerre Polynomials
The associated Laguerre polynomials [30] are defined by The following 3-point-rule [30] is used in this work: We further need the the following differential equation [30]: For positive integers , we have the following relation between the Gamma function and the double factorial [32,33]: