DDNS Discrete Dynamics in Nature and Society 1607-887X 1026-0226 Hindawi Publishing Corporation 10.1155/2015/649423 649423 Research Article A Novel Iterative Method for Polar Decomposition and Matrix Sign Function http://orcid.org/0000-0002-6905-8951 Soleymani F. 1 Stanimirović Predrag S. 2 Stojanović Igor 3 Bischi Gian I. 1 Department of Mathematics Islamic Azad University, Zahedan Branch Zahedan Iran iauzah.ac.ir 2 Faculty of Sciences and Mathematics University of Niš Visegradska 33 18000 Niš Serbia ni.ac.rs 3 Faculty of Computer Science Goce Delčev University Goce Delčev 89 2000 Štip Macedonia ugd.edu.mk 2015 11102015 2015 03 04 2015 21 06 2015 11102015 2015 Copyright © 2015 F. Soleymani et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

We define and investigate a globally convergent iterative method possessing sixth order of convergence which is intended to calculate the polar decomposition and the matrix sign function. Some analysis of stability and computational complexity are brought forward. The behaviors of the proposed algorithms are illustrated by numerical experiments.

1. Introduction

Recently, a large number of papers intended for calculating simple and multiple zeros of the nonlinear equation f(x)=0 have been published. There are several attempts to overcome the difficulties in designing new iterative methods for solving f(x)=0 or in the application of such iterative methods in computing matrix functions (see [1, 2]). Our current work discusses an application of a high order iterative method for solving nonlinear scalar equations in calculating the polar decomposition and the matrix sign function.

We now restate some preliminaries about the polar decomposition and the matrix sign.

Let Cm×n denote the linear space of all m×n complex matrices. The polar decomposition of a complex matrix ACm×n is defined as(1)A=UH,UU=Ir,rankU=r=rankA,wherein H is a Hermitian positive semidefinite matrix of the order n and UCm×n is a subunitary matrix (partial isometry). Here, the inequality mn is assumed. A matrix U is subunitary if Ux2=x2 for any xR(U)=N(U), where R(X) and N(X) denote the linear space spanned by columns of the matrix X (range of X) and the null space of X, respectively. Note that if rank(A)=n, then UU=In, and U is an orthonormal Stiefel matrix.

It is assumed that the singular value decomposition (SVD) of A has the following form:(2)A=PΣ0Q,wherein PCm×m and QCn×n are unitary matrices and(3)Σ=diagσ1,,σn,σ1σn0.The Hermitian factor H is always unique and it can be expressed as (AA)1/2, and the unitary factor is unique if A is nonsingular. Note that, here, the exponent 1/2 denotes the principal square root: the one whose eigenvalues lie in the right half-plane.

The polar decomposition may be easily constructed from an SVD of the given matrix A. However, the SVD is a substantial calculation that displays much more about the structure of A than does the polar decomposition. Constructing the polar decomposition from the SVD destroys this extra information and wastes the arithmetic work used to compute it. It is intuitively more appealing to use the polar decomposition as a preliminary step in the computation of the SVD.

On the other hand, a related issue to the polar decomposition is the matrix sign decomposition, which is defined for a matrix ACn×n having no pure imaginary eigenvalues. The most concise definition of the matrix sign decomposition is given in (4)A=SN=AA2-1/2A21/2.Here, S=sign(A) is the matrix sign function, introduced by Roberts . We herewith remark that, in this work, whenever we write about the computation of a matrix sign function, we mean a square matrix with no eigenvalues on the imaginary axis.

Now we briefly review some of the most important iterative methods for computing the matrix polar decomposition.

Among many iterations available for computing the polar decomposition, the most practically useful is the scaled Newton iteration  as well as the recently proposed dynamically weighted Halley iteration . The method of Newton introduced for computing the polar decomposition (via unitary polar factor) is defined in  by the iterative scheme(5)Uk+1=12Uk+Uk-,for the square nonsingular cases and by the following alternative for general rectangular cases :(6)Uk+1=12Uk+Uk.Note that U stands for the Moore-Penrose generalized inverse, Uk-=Uk-1, and Uk=Uk.

The cubically convergent method of Halley(7)Uk+1=Uk3I+UkUkI+3UkUk-1has been developed in  for computing the unitary polar factor.

The particular formula (7) is also applicable to singular or rectangular matrices. This scheme has further been developed adaptively in . In fact the dynamically weighted Halley method (DWH) for computing unitary polar factor has been introduced as follows:(8)Uk+1=UkakI+bkUkUkI+ckUkUk-1.The parameters ak, bk, and ck are dynamically chosen to accelerate the convergence. They are computed by(9)ak=hlk,bk=ak-124,ck=ak+bk-1,where(10)hl=1+γ+128-4γ+82-l2l21+γ,γ=41-l2l41/3.In (9), lk is a lower bound for the smallest singular value of Uk. Fortunately, once l0σmin(U0) is obtained, then effective and sharp bounds can be attained at no cost from the following recurrence:(11)lk=lk-1ak-1+bk-1lk-121+ck-1lk-12,k1.

An initial matrix U0 must be employed in a matrix fixed-point type method so as to arrive at the convergence phase. Such a sharp initial approximation for the unitary factor can be expressed as(12)U0=1αA,whereas α>0 is an estimate of A2 (a safe choice of which is AF).

The rest of this paper unfolds the contents in what follows. Section 2 derives a new iterative scheme for solving nonlinear equations. Additionally, by applying the illustration via basins of attraction, we find a scheme with a global convergence. The global convergence of the constructed solver is verified analytically. In Section 3, we generalize the proposed nonlinear equation solver into the iterative method for finding the polar decomposition. Furthermore, we prove that the new scheme is convergent. The computational complexity is discussed in Section 4. In Section 5, we define an extension of the proposed method in numerical computation of the matrix sign function and show its asymptotic stability. Section 6 is devoted to the application of the contributed methods in solving two numerical examples (one in double precision arithmetic and the other in a high precision computing environment). Finally, Section 7 draws a conclusion of this paper.

2. Chebyshev-Halley Type Iteration and Its Extension

Many of the iterative methods for computing matrix functions can be deduced by applying a nonlinear equation solver to a special mapping. For example, applying Newton’s method to the mapping(13)FUUU-I=0,in which I is the identity matrix of the appropriate size, could result in iterates (5) (note that the equivalent form of (13) for the matrix sign is U2-I=0). This reveals a close relation between matrix functions and iterative root-finding methods .

Gutiérrez and Hernández in  developed a Chebyshev-Halley type scheme in Banach space for finding simple zeros of f(x)=0, which can be written in what follows:(14)xk+1=xk-1+12Lxk1-ηLxkfxkfxk,wherein ηR, L(xk)=f(xk)f(xk)/f(xk)2, and the convergence order is cubic.

If we decide to apply (14) for solving (13), we will obtain a family of at least cubically convergent schemes in finding the polar decomposition. But an important barrier occurred and it would be the nonglobal convergence of some of the schemes.

On the other hand, Iannazzo in  showed that the matrix convergence is governed by the scalar convergence for the so-called pure matrix iterations. We remark that this is true for the matrix sign function when the scalars are the eigenvalues, while for polar decomposition it is true when scalars are singular values. This is also well illustrated in the textbook .

Hence, we should find a member from (14), so that the derived method is new and also possesses the global convergence. Toward this goal, we employ the theory of basins of attraction for (14) so as to solve the quadratic polynomial x2-1=0 in a square [-2,2]×[-2,2] of the complex plane whereas the maximal number of iterations is fixed to 100 and the stopping criterion is fxk10-4.

This is done in Figures 13 for different values of η. In Figures 1(a) and 2(a), the convergence in the process of solving x2-1=0 is local, and therefore convergence of the matrix iterations could happen only for very sharp initial matrices. Moreover, the method in Figure 3(b) behaves chaotically and also includes divergence points. The divergence points are included in the black areas and, accordingly, should be outside of interest.

Basins of attraction for different methods of family (14) for f(x)=x2-1 shaded according to the number of iterations.

η = 0

η = 1 / 2

Basins of attraction for different methods of family (14) for f(x)=x2-1 shaded according to the number of iterations.

η = 1 / 3

η = 3 / 4

Basins of attraction for different methods of family (14) for f(x)=x2-1 shaded according to the number of iterations.

η = 3 / 2

η = 5 / 2

We remark that η has been chosen randomly and experimentally. Clearly, the investigation of the other cases can be used in future studies.

Among the six tested methods extracted by different particular values of η and illustrated in Figures 13, only the methods illustrated in Figures 1(b), 2(b), and 3(a) have a global convergence. A deeper verification reveals that the methods used in Figures 1(b) and 3(a) are in fact Halley’s method and its reciprocal for x2-1=0, respectively. Since those are not new, the only novel and useful method which has global convergence and remains from the six considered methods of (14) is the iterative method used in Figure 2(b). It can be written as follows:(15)xk+1=1+12xk2+3xk46xk+10xk3.

To increase the order of convergence, let us now compose the method of Newton with iterations (14) corresponding to η=3/4 in order to derive a useful high order globally convergent iterative scheme for solving f(x)=0. This intention leads to(16)yk=xk-fxkfxk,xk+1=yk-1+12Lyk1-3/4Lykfykfyk,wherein L(yk)=f(yk)f(yk)/f(yk)2.

Theorem 1.

Let αD be a simple zero of a sufficiently differentiable function f:DRR for an open interval D, which contains x0 as an initial approximation of α. Then the iterative expression (16) without memory has a sixth order of convergence.

Proof.

To save the space and in order not to deviate from the main topic, we here exclude the proof. In addition, the steps of the proof are similar to those taken in .

The iterative method (16) reaches the sixth-order convergence using five functional evaluations and thus achieves the efficiency index 61/51.430, which is higher than that of Newton; that is, 21/21.414. We remark that the cost of one function evaluation and its first and second derivatives are assumed to be unity. Furthermore, the application of (16) in solving the polynomial equation g(x)x2-1=0 possesses the global convergence (except for the points lying on the imaginary axis). This is illustrated analytically in what follows.

In terms of the fractal theory, it is necessary to find the global basins of attraction for a zero z:(17)SzzC:Υkzz,  as  k,where Υk(x)=Υ(Υ((Υ(x)))) is the k-fold composition ΥΥ of the iteration function Υ. Here using (16), we have (in its reciprocal form)(18)xk+1=Ixk=20xk+108xk3+108xk5+20xk73+60xk2+130xk4+60xk6+3xk8,k=0,1,.

To check the global convergence of (16) in the case of the quadratic polynomial g(z)=z2-1, with the zeros ±1, we start from(19)Bz=20z+108z3+108z5+20z73+60z2+130z4+60z6+3z8and find(20)Bz+1Bz-1=λU+1U-16,wherein λ=-3+U(2+3U)/3+U(-2+3U).

Let S denote the boundary of the set S. One of basic notions in the fractal theory connected to iterative processes and convergence of an iterative function f is a Julia set for the proposed operator Υ. Thus, when k, we obtain the following:

If z+1/z-1<1, then Bk(z)+1/Bk(z)-10, and Bk(z)-1.

If z+1/z-1>1, then Bk(z)+1/Bk(z)-10, and Bk(z)+1.

Furthermore, the basins of attraction S(-1) and S(1) in the case of the operator B are the half-planes on either side in relation to the line z=0 (the imaginary axis). Since ±1 are attractive fixed points of B, the Julia set J(B) is the boundary of the basins of attraction S(-1) and S(1); that is,(21)JB=S-1=S1=γi:γR.

The Julia set J(B) is just the line z=0 for (19), and thus the new sixth-order method (18) is globally convergent. Therefore, the presented method has global behavior, even outside the square [-2,2]×[-2,2] which is considered in Figure 1.

In addition, we remark that a globally convergent sixth-order method can be easily constructed by composing standard Halley and Newton methods.

The main contribution of this work lies in the next two sections at which we extend (16) into the iterative methods for computing the polar decomposition and the matrix sign function.

3. Extension to Polar Decomposition

This section is devoted to the extension of (16) to solve the matrix equation (13). This application enables us to obtain a fast sixth-order iterative matrix method for constructing the polar decomposition. Note that the improvements in hardware and software have been ultimately indispensable, since higher order methods produce approximations of great accuracy and require complicated convergence analysis, which is feasible only by symbolic computation. Subsequently, in this paper, we use Mathematica  to illustrate the speed of convergence.

To this goal, an application of (16) on (13) in conjunction with further simplifications produces the following reciprocal form as the rational iteration:(22)Uk+1=Uk20I+108Yk+108Zk+20Wk·3I+60Yk+130Zk+60Wk+3Tk-1,where (23)Yk=UkUk,Zk=YkYk,Wk=ZkYk,Tk=WkYk,and U0 is given by (12).

Now, we have a novel iterative fixed-point type method for finding the polar decomposition via calculating the unitary matrix U.

Theorem 2.

Assume that ACm×n is an arbitrary matrix. Then, the matrix iterates {Uk}k=0k= of (22) converge to U using U0=A.

Proof.

To prove the statement, we make use of the SVD of A in the form A=PΣQ, where Σ=Σr000, r=rank(A) and the zero blocks in Σ may be absent. Note that the steps of the proof are similar to those which recently have been taken in . Define(24)Dk=PUkQ.Subsequently, from (22), we have(25)D0=Σ,Dk+1=20Dk+108Dk3+108Dk5+20Dk7·3I+60Dk2+130Dk4+60Dk6+3Dk8-1.Since D0Rm×n has diagonal and zero elements, it follows by the induction that the sequence Dkk=0 is defined by(26)Dk=diagdik000,dik>0.Accordingly, (25) represents r uncoupled scalar iterations(27)di0=σi,1ir,dik+1=20dik+108dik3+108dik5+20dik7·3+60dik2+130dik4+60dik6+3dik8-1.Simple manipulations yield the relation (28)dik+1-1dik+1+1=-3+20dik-60dik2+108dik3-130dik4+108dik5-60dik6+20dik7-3dik83+20dik+60dik2+108dik3+130dik4+108dik5+60dik6+20dik7+3dik8.Since σi is positive, (28) holds for each i. It follows that di(k+1)-1/di(k+1)+10 as k; that is to say,(29)DkIr000.Therefore, as k, UkUrVr=U. The proof is complete.

Theorem 3.

Let ACm×n be an arbitrary matrix. Then, method (22) has sixth order of convergence to find the unitary polar factor of A.

Proof.

The proposed scheme (22) transforms the singular values of Uk according to(30)σik+1=20σik+108σik3+108σik5+20σik7·3+60σik2+130σik4+60σik6+3σik8-1,1ir,and leaves the singular vectors invariant. From (30), it is enough to show that the convergence of the singular values to unity has sixth order of convergence for k1:(31)σik+1-1σik+1+1=--1+σik63+σik-2+3σik1+σik63+σik2+3σik.Now, we attain(32)σik+1-1σik+1+13+σik-2+3σik3+σik2+3σikσik-1σik+16.This reveals the sixth order of convergence for the new method (22). The proof is ended.

Note that, in 1991, Kenney and Laub  have proposed a family of rational iterative methods for the matrix sign, based on Padé approximation. Their principal Padé iterations are convergent globally. Thus, we have convergent methods of arbitrary orders for the matrix sign (subsequently for the polar decomposition). However, here we tried to propose another new and useful method for this purpose.

We now end this section by recalling an important approach for speeding up the convergence speed of (22). The sixth-order convergence for iteration (22) ensures rapid convergence in the final stages of the iterates. The speed of convergence can be slow at the beginning of the process, so it is necessary to scale the matrix Uk before each cycle. An important scaling approach was derived in  in the Frobenius norm as comes next:(33)θk=UkFUkF1/2.

See  for more details. We remark that numerical experiments show improvements in the speed of convergence applying (33).

Remark 4.

The new scheme can be expressed in the following accelerated form:(34)Computeθk  by  33,k=0,1,,Mk=3I+60θk2Yk+130θk4Zk+60θk6Wk+3θk8Tk,Uk+1=θkUk20I+108θk2Yk+108θk4Zk+20θk6Wk·Mk-1.

4. Computational Complexity

In this section, we evaluate the computational efficiency of (22). To compare the behavior of different matrix methods for finding U, we recall the definition of efficiency index:(35)EI=p1/C,wherein C and p denote the computational cost and the convergence order per cycle. Here, in order to have a fair comparison and since there are matrix-matrix multiplications (denoted by mmm) and matrix inversion(s) per computing cycles of (6), (7), and (22), we extend (35) as follows:(36)CEI=p1/sm+c,so as to be able to incorporate all the existing factors of an algorithm into the definition of the computational efficiency index. In (36), s, m, and c denote the whole number of iterations required by a method to converge to U, the number of mmm per cycle, and the considered cost of matrix inversion per cycle, respectively.

On the other hand, it is assumed that the cost of mmm is 1 unit and subsequently the cost of one matrix inversion is 1.2 and the cost for computing the Moore-Penrose inverse is 1.4. We considered these weights empirically after testing many random matrices on a typical PC. We also neglect the cost of additions, subtractions, and so forth, because their costs are negligible in contrast to the cost of mmm and matrix inversion.

Now, the approximated CEI for the studied methods would be (37)CEI621/s10+11.4,CEI731/s23+11.2,CEI2261/s36+11.2,wherein s1, s2, and s3 are the whole number of iterations required by (6), (7), and (22) to converge, respectively, in the same environment.

Finally, we assume that s3=s and thus the number of iterations for (6) and (7) would roughly be 5/2s-1 and 5/3s-1, respectively, since they have the second and the third orders of convergence in contrast to the sixth order for (22). We have obtained these factors empirically via solving many numerical examples.

The results of comparisons have now been drawn and illustrated in the bar chart of Figure 4. We here remark that NMP, HMP, and PMP stand for Newton’s method for polar decomposition, Halley’s method for polar decomposition, and proposed method for polar decomposition. Such naming will be used from now on. In this figure, one may easily observe that when a higher number of iterations are required (occasionally for large scale matrices), then the new method performs much better, because it requires smaller number of matrix inversions. In Figure 4, for example, s=4 means that the number of steps required for the convergence of (22) is 4 and subsequently around s1=9 and s2=6 steps are required for (6) and (7) to converge. Note that we plotted the logarithm of CEI in Figure 4 to show the distinctions obviously.

Comparison of computational efficiency indices for different matrix methods in finding the unitary polar factor.

The only difficulty in our high order method is that if, in one iteration, it produces results of a lower accuracy (a low tolerance, e.g., ζ=10-1), then an expensive further cycle should be carried out to reach the tolerance in double precision (e.g., 10-8). In our proposed algorithm (ALP, i.e., Algorithm for polar decomposition) hereby, we first apply (22) to arrive at the convergence phase much rapidly and accelerate the beginning of the process. And next, we apply the simple method (6) for the last step only. This switching idea has been illustrated in Algorithm 1.

<bold>Algorithm 1: </bold>The hybrid algorithm for computing polar decomposition.

( 1 ) U0 is given

(2) use (22) until Uk+1-Uk/Ukζ

(3) If the final stop termination Uk+1-Uk/Ukϵ has already been occurred, go to (5), else

(4) do (6) to fulfill the stop termination Uk+1-Uk/Ukϵ

(5) end for

5. Extension for Matrix Sign

Herein, we show an application of the proposed iterative method (16) in finding the matrix sign function. In fact, there is a tight relationship between the matrix sign and the polar decomposition. As introduced in Section 1, the sign of a nonsingular square matrix is an important matrix function with potential applications in different branches of Mathematics; see [16, 17].

The iterative rule(38)Xk+1=12Xk+Xk-1,which is also known as the Newton method, is the most common and well-known way for finding the sign of a square nonsingular matrix. It converges quadratically when X0=A has been chosen as the initial matrix.

A general family of matrix iterations for finding the matrix sign function S was discussed thoroughly in . An example from this class of methods is the method of Halley which is a modification of (7) and is defined by(39)Xk+1=Xk3I+Xk2I+3Xk2-1.

Accordingly, the new scheme (22) can simply be used for finding the matrix sign function S with a little modification as comes next:(40)Xk+1=Xk20I+108Xk2+108Xk4+20Xk6·3I+60Xk2+130Xk4+60Xk6+3Xk8-1.

Iteration (40) is rational. We investigate the stability of (40) for finding S in a neighborhood of the solution. In fact, we analyze how a small perturbation at the kth iterate is amplified or damped along the iterates.

Lemma 5.

Under the same conditions as in Theorem 3, the sequence {Xk}k=0k= generated by (40) is asymptotically stable.

Proof.

If X0 is a function of A, then the iterates from (40) are all functions of A and hence commute with A. Let Δk be a numerical perturbation introduced at the kth iterate of (40). Next, one has(41)X~k=Xk+Δk.Here, we perform a first-order error analysis, that is, formally using approximations (Δk)i0, since (Δk)i, i2, is close to zero (matrix). This formal approximation is meaningful if Δk is sufficiently small. We have(42)X~k+1=20X~k+108X~k3+108X~k5+20X~k73I+60X~k2+130X~k4+60X~k6+3X~k8-1=20Xk+Δk+108Xk+Δk3+108Xk+Δk5+20Xk+Δk7×3I+60Xk+Δk2+130Xk+Δk4+60Xk+Δk6+3Xk+Δk8-1256S+640Δk+384SΔkS256I+512SΔk+512ΔkS-1S+52Δk+32SΔkSI-2SΔk-2ΔkSS+12Δk-12SΔkS,wherein the identities(43)B+C-1B-1-B-1CB-1,were used (for any nonsingular matrix B and the matrix C). Note that, after some algebraic manipulations and the approximation Δk+1=X~k+1-Xk+1=X~k+1-S, one can verify (assuming Xksign(A)=S for enough large k)(44)Δk+112Δk-12SΔkS.We also used the equalities S2=I and S-1=S, relative to the matrix sign function. We can now conclude that the perturbation at the iterate k+1 is bounded; that is,(45)Δk+112k+1Δ0-SΔ0S.Therefore, the sequence {Xk}k=0k= generated by (40) is asymptotically stable. This ends the proof.

It is now not difficult to also show that (40) has a sixth order of convergence and reads in the following error inequality:(46)Xk+1-SMk-1S-13I+SXk-2I+3SXk·Xk-S6,where Mk=3I+60Xk2+130Xk4+60Xk6+3Xk8.

Here we make a comment that we have not given any discussions on backward stability for the polar decomposition in the sense of measuring A-UHF/AF (this is a much stronger notion than the stability).

An acceleration via scaling, similar to that of (34), is applicable to (40). See  for an interesting choice of the scaling parameter. In Figure 5, we have drawn the basins of attractions for (16) in order to solve the complex polynomial of different orders. Although Figure 5(a) shows a global convergence of our method for the matrix sign function (also theoretically shown at the end of Section 2), the application of the proposed method in computation of the matrix sector function needs a deeper care.

Basins of attraction (16) shaded according to the number of iterations.

x 2 - 1 = 0

x 3 - 1 = 0

A similar implementation for (40) is advised and could be found in Algorithm 2 (we denote it by ALS, i.e., Algorithm for Sign).

<bold>Algorithm 2: </bold>The hybrid algorithm for computing matrix sign function.

( 1 ) X0 is given

(2) Use (40) until Xk+1-Xk/Xk10-1

(3) If the final stop termination Xk+1-Xk/Xkϵ has already been occurred, go to (5), else

(4) do (38) to fulfill the stop termination Xk+1-Xk/Xkϵ

(5) end for

6. Numerical Experiments

We have tested the contributed method (22), denoted by PMP, and (40), denoted by PMS, using the programming package Mathematica 8 with the machine precision ϵ=2.22045×10-16. Apart from this scheme, several iterative methods such as (6), denoted by NMP, and (7), denoted by HMP, which require one (pseudo)inverse per cycle, have been tested. We also use the algorithms ALP, ALS, NMS, and HMS. Note that a sixth-order method from the Padé family  for the matrix sign is defined as follows:(47)Xk+1=Xk6I+20Xk2+6Xk4·I+15Xk2+15Xk4+Xk6-1.We denote this iteration by SMS (sixth-order method for sign), that is, [2/3]-Padé approximant of the Padé family. Although the computational complexity of this method is lower than that of PMS, the PMS produces results of higher accuracies in high precision computing as will be observed in Example 2.

The considered stopping termination in performed numerical experiments is(48)Rk+1=Uk+1-UkUkϵ,wherein ϵ is the tolerance. There is a similar stopping termination once we used it to find the matrix sign function.

Example 1.

In this experiment, we study the behavior of different methods for finding the unitary polar factor of the complex rectangular 400×200 matrix which is randomly generated with the uniform distribution using the code

SeedRandom; m = 400; n = 200;

A = RandomComplex[-1 - I, 1 + I,

m, n];

The results of the comparison are arranged together in Table 1 applying the tolerance ϵ=10-6. It could easily be observed that a clear reduction in the number of iterations and the CPU time for finding the polar decomposition is obtained using ALP.

Results of comparisons in Example 1 with U0=A for polar decomposition.

Methods
NMP HMP PMP ALP
IT 9 6 4 3 + 1
R k + 1 2.93539 × 1 0 - 9 7.6991 × 1 0 - 9 7.63809 × 1 0 - 10 7.63809 × 1 0 - 10
Time 2.7812500 1.1250000 1.4218750 1.2750000
F k + 1 3.60456 × 1 0 - 14 1.05716 × 1 0 - 14 8.2024 × 1 0 - 15 3.52843 × 1 0 - 14

A considerable increase of the accuracy of approximations produced by the proposed method could be observed from numerical results. To check the accuracy of the numerical results, we have computed and reported Fk=UkUk-IF for the last iteration in Table 1.

An important application of higher order methods is in high precision computing. In the next test we considered an academical example, with 128-digit floating point. We do this consideration purposely so as to check the convergence behavior of different methods by the following definition for computational order of convergence (COC) using (48). The COC can be approximated by(49)COC=lnRk+1/RklnRk/Rk-1,at which the last three approximations for the polar decomposition or the matrix sign function are used.

In this experiment, we study the behavior of different methods for finding the matrix sign function of the following 5×5 matrix: (50)A=1078775658610975910.The results of comparison are carried out in Figure 6 and Table 2 applying the tolerance ϵ=10-20. It could easily be observed that there is a clear reduction in the number of iterations ensured by the application of PMS. To check the accuracy of the results we have computed and reported Ek=Xk2-I/Xk2.

Results of comparisons in Example 2 for the matrix sign.

Methods
NMS HMS SMS PMS
IT 13 9 6 6
R k + 1 2.13346 × 1 0 - 36 3.9478 × 1 0 - 58 3.83821 × 1 0 - 69 2.63194 × 1 0 - 95
COC 2.0 3.0 6.03717 6.0543
E k + 1 1.65001 × 1 0 - 72 0 0 0

Convergence history of different methods using  logplot  in Example 2 with high precision computing.

Example 3.

In this experiment, we compare the behavior of different accelerated methods via scaling defined by (33) for finding the unitary polar factors. We now compare the Newton method (NMP), accelerated Newton method (ANMP), PMP, and the accelerated proposed method for polar decomposition (34) denoted by APMP. We used the following six complex rectangular 310×300 matrices (with uniform distribution):

m = 310; n = 300; number = 6; SeedRandom;

Table[A[l] = RandomComplex [−10−10 I,

10+10 I, m, n];, l, number];

The results of comparison are carried out in Tables 3 and 4 applying the tolerance ϵ=10-10 in double precision. It could easily be observed that there is a clear reduction in the number of iterations, ensured by the acceleration via scaling (33).

Results of comparisons for Example 3 in terms of number of iterations.

Matrices Methods
NMP PMP ANMP APMP
# 1 11 5 9 4
# 2 11 5 9 4
# 3 11 5 9 4
# 4 11 5 9 4
# 5 11 5 9 4
# 6 12 5 9 4

Results of comparisons for Example 3 in terms of elapsed time (s).

Matrices Methods
NMP PMP ANMP APMP
# 1 5.1718750 3.0781250 4.6875000 4.2031250
# 2 5.2031250 3.1093750 4.6718750 4.2031250
# 3 5.1718750 3.0781250 4.6718750 4.1875000
# 4 5.1718750 3.0781250 4.7031250 4.1718750
# 5 5.1562500 3.0781250 4.6718750 4.2031250
# 6 5.5937500 3.0781250 4.6875000 4.1718750

It can be observed that taking into account every presented method (e.g., NMP) the number of iterations for all of matrices is the same for that method. It would be more interesting to make a comment that there is an upper bound for the maximal number of iterations for each method in computing the matrix sign and the polar decomposition in double precision arithmetic, as discussed fully in .

Example 4.

And finally to report the number of iterations when the input matrix is ill-conditioned, we take into consideration the Hilbert matrix of dimension 10 whose condition number is 3.53534×1013 (in l) . The unitary polar factor is known to be the unit matrix I in this case. Applying the tolerance 10-10, we found that NMP, HMP, and PMP require 49, 31, and 19 cycles, respectively. This again shows the superiority of the proposed approach.

7. Concluding Remarks

The polar decomposition is an important theoretical and computational tool, known because of its approximative properties, its sensitivity to perturbations, and its computation.

In this paper, we have developed a high order method for solving nonlinear scalar equations and then extend it to the iterative method applicable in the computation of the matrix polar decomposition. It has been shown that the convergence of the method is global and its convergence rate is six.

It has further been discussed how the proposed method could be applied in finding the matrix sign function. The presented scheme possesses asymptotic stability and it is very useful in a high precision computing environment. Some numerical tests have been provided to show the performance of the new methods.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The second author gratefully acknowledges support from the Research Project 174013 of the Serbian Ministry of Science. The second and third authors gratefully acknowledge support from the Project “Applying Direct Methods for Digital Image Restoring” of the Goce Delc˘ev University.

Gander W. On Halley's iteration method The American Mathematical Monthly 1985 92 2 131 134 10.2307/2322644 MR777559 ZBL0574.65041 Haghani F. K. Soleymani F. On a fourth-order matrix method for computing polar decomposition Computational & Applied Mathematics 2015 34 1 389 399 10.1007/s40314-014-0124-0 MR3326265 Higham N. J. Functions of Matrices: Theory and Computation 2008 Philadelphia, Pa, USA SIAM 10.1137/1.9780898717778 MR2396439 Roberts J. D. Linear model reduction and solution of the algebraic Riccati equation by use of the sign function International Journal of Control 1980 32 4 677 687 10.1080/00207178008922881 MR595962 ZBL0463.93050 2-s2.0-0019069037 Higham N. J. Computing the polar decomposition-with applications SIAM Journal on Scientific and Statistical Computing 1986 7 1160 1174 Nakatsukasa Y. Bai Z. Gygi F. Optimizing Halley's iteration for computing the matrix polar decomposition SIAM Journal on Matrix Analysis and Applications 2010 31 5 2700 2720 10.1137/090774999 MR2740628 2-s2.0-79251504591 Du K. The iterative methods for computing the polar decomposition of rank-deficient matrix Applied Mathematics and Computation 2005 162 1 95 102 10.1016/j.amc.2003.12.083 MR2107262 2-s2.0-10444243360 Gander W. Algorithms for the polar decomposition SIAM Journal on Scientific and Statistical Computing 1990 11 6 1102 1115 10.1137/0911062 MR1068499 Gutiérrez J. M. Hernández M. A. A family of Chebyshev-Halley type methods in Banach spaces Bulletin of the Australian Mathematical Society 1997 55 1 113 130 10.1017/s0004972700030586 MR1428535 2-s2.0-0031064164 Iannazzo B. A family of rational iterations and its application to the computation of the matrix pth root SIAM Journal on Matrix Analysis and Applications 2008 30 4 1445 1462 10.1137/070694351 Soleymani F. Shateyi S. Salmani H. Computing simple roots by an optimal sixteenth-order class Journal of Applied Mathematics 2012 2012 13 958020 10.1155/2012/958020 MR2997247 Hoste J. Mathematica Demystified 2009 New York, NY, USA McGraw-Hill Kenney C. Laub A. J. Rational iterative methods for the matrix sign function SIAM Journal on Matrix Analysis and Applications 1991 12 2 273 291 10.1137/0612020 MR1089159 Dubrulle A. A. Frobenius Iteration for the Matrix Polar Decomposition 1994 Palo Alto, Calif, USA Hewlett-Packard Company Higham N. J. The matrix sign decomposition and its relation to the polar decomposition Linear Algebra and Its Applications 1994 212-213 3 20 10.1016/0024-3795(94)90393-X Soheili A. R. Toutounian F. Soleymani F. A fast convergent numerical method for matrix sign function with application in SDEs Journal of Computational and Applied Mathematics 2015 282 167 178 10.1016/j.cam.2014.12.041 MR3313098 2-s2.0-84921805649 Soleymani F. Stanimirović P. S. Shateyi S. Haghani F. K. Approximating the matrix sign function using a novel iterative method Abstract and Applied Analysis 2014 2014 9 105301 10.1155/2014/105301 MR3240521 Byers R. Xu H. A new scaling for Newton's iteration for the polar decomposition and its backward stability SIAM Journal on Matrix Analysis and Applications 2008 30 2 822 843 10.1137/070699895 MR2443974 ZBL1170.65019 2-s2.0-67649206981