The Identification of Convex Function on Riemannian Manifold

The necessary and sufficient condition of convex function is significant in nonlinear convex programming. This paper presents the identification of convex function on Riemannian manifold by use of Penot generalized directional derivative and the Clarke generalized gradient. This paper also presents a method for judging whether a point is the global minimum point in the inequality constraints. Our objective here is to extend the content and proof the necessary and sufficient condition of convex function to Riemannian manifolds.


Introduction
Manifold is the space with the local property of Euclidean space.We often judge the warp of the space by measuring its space.The standard of the measurement is called metric.Metric is an intrinsic property and the space with metric is called Riemannian space.Riemann manifold is a differential manifold which has Riemannian metric.Manifold learning attempts to obtain the intrinsic structure of non-linearly distributed data, which can be used in non-linear dimensionality reduction (NLDR).We can find the meaningful low dimensional structure hidden in the high-dimensional observation data by the non-linear dimension reduction of high data space [1].Expanding the optimization based on linear space to the nonlinear space (such as Riemannian manifold) is a hot topic in this research field [2].Convex function is a kind of special function that has a broad application, especially has a wide range of applications in the areas of optimization and mathematical programming, cybernetics, and other fields.Since the useful data is usually not a linear combination of the features in the real world that the nonlinear convex programming problem has become a new research subject [3][4][5].In order to have a more indepth study of nonsmooth and its related problems, several important tools for analyzing nonsmooth problem have been extended from European space to Riemannian Manifold.
In 2004, Ledyacv and Zhu established the concept and algorithm of subdifferential of nonsmooth function in Riemannian manifold and proved that the solution of Hamilton-Jacobi equation defined on Riemannian manifold is unique.They also discussed the concept and variational principle of approximate subdifferential in Riemannian manifold [6,7].In 2005, Azagra and others discussed the J. Ferrera subdifferential which was defined on finite dimensional Riemannian manifold and relevant limit subdifferential and analyzed the issues of differential inclusion which were defined on differential manifold by using the concept of generalized derivative [8,9].Liu and others expanded (MP) problem from linear space to differential manifold and Fritz John necessary optimality conditions on Riemann manifold was given [10].In 2008, Gang and Sanyang solved the nondifferentiable multiobjective optimization problems with equality and inequality constraints from Euclidean space to Riemannian manifolds and derived the Fritz-John necessary conditions with generalized gradient formula for weak Pareto optimal solutions from Ekeland variational principle [11].Wang et al. established the semilocal convergence of sequences which was generated by the Gauss-Newton method (with quasiregular initial points) for the convex composite function ℎ ∘  on Riemannian manifold [12] by using the majorizing function technique.Bento et al. presented a steepest descent method with Armijo's rule for multicriteria optimization in 2 Mathematical Problems in Engineering the Riemannian context.With assuming quasiconvexity of the multicriteria function and nonnegative curvature of the Riemannian manifold, they proved full convergence of the sequence to a critical Pareto point [13].
Convex programming is of great significance in the study of nonlinear programming theory.Based on the above research results, we will parallel extend the identification of convex function and the optimality conditions of constraint problems from linear space to Riemannian manifold in order to solve the convex programming problem in Riemannian manifold.This paper is set out as follows.In Section 2, we briefly review some preliminary concepts.In Section 3, we introduce the identification of convex function on Riemannian manifold.We make some concluding remarks and suggest future research in this area in Section 4.

Preliminaries
Let   be an -dimensional Euclidean space; the th coordinate of point  ∈   is denoted by ()  ; that is ( )  is the th coordinate function of   .Definition 1 (see [14]).Let  be a Hausdorff topological space.If every point  of  has an open neighborhood  ⊂ , such that an open subset of  and -dimensional Euclidean space is homeomorphous, then  is called an dimensional topological manifold.
Then, () is a function which meets the local Lipschitz condition in the neighborhood of  0 .
Definition 3 (see [10]).Let function  :  →  be a meet local Lipschitz condition in the neighborhood of the point  ∈  and (, ) be a coordinate plot which contains .The Clarke generalized directional derivative of  at  along the direction of V ∈    is denoted by where  → 0.
Definition 4 (see [10]).Let () be a function denoted in Riemannian manifold (, ) and meet local Lipschitz condition in the neighborhood of .Then, the element of the subset of cotangent space is said to be Clarke generalized gradient in which  is at .

The norm of generalized gradient is denoted by
Definition 5 (see [15]).Let  :  →  meet local Lipschitz condition in the neighborhood of  ∈  and let (, ) be a coordinate plot which contains .We define Penot generalized directional derivative of  at  along the direction of V ∈    as follows: If (; V) = (; V), we define (; V) as Penot generalized directional derivative of  at  along the direction of V ∈   .
2 (; V; ), we define  2 (; V; ) to represent second-order Penot generalized directional derivative of  at  along the direction of V ∈   .

The Identification of Convex Function on Riemannian Manifold
Proof (necessary condition).For ∀ ∈ [0, 1], there exists Then, according to the first-order Taylor expansion, Since  is geodesic convex function in , then Then, from two simultaneous equations of ( 9) and ( 10), we get that is It follows from  → 0 + that (()/) → 0; hence, From the known conditions it follows that  ∘  −1 () Hence, Then, from (15) it follows that Since  =  + (1 − ), then Thus,  is geodesic convex function in .
Theorem 7 (first-order necessary and sufficient condition of strictly convex function).Let  be an -dimensional Riemannian manifold and  be a geodesic convex set.Let  ⊂ ,  :  →  meet local Lipschitz condition in , and then the necessary and sufficient condition that  is geodesic strictly convex function in  is as follows.For ∀,  ∈ , there exists Proof (necessary condition).In a similar way with proof of Theorem 6, we get Sufficient Condition.Since  is a strictly convex function, then  is a convex function.For any two different points ,  in which ,  ∈ , there exists  = (1/2) + (1/2), and then  ∈ .
From Theorem 6 it follows that Since  is a strictly convex function, then Then, it follows from ( 20) and ( 21) that that is, Theorem 8 (second-order necessary and sufficient condition of convex function).Let  be an -dimensional Riemannian manifold and  be a geodesic convex set.Let  ⊂ ,  :  →  meet local Lipschitz condition in  and let (, ) be a coordinate plot which contains , and second-order Penot generalized directional derivative of  at  along the direction of V ∈    exists.Then, the necessary and sufficient condition that  is geodesic convex function in  is as follows.For any  ∈ , there exists  2 (; V, ) ≥ 0.
Since  is a geodesic convex set, then  ∈ .From the known conditions it follows that Hence, so that Therefore,  is geodesic convex function in  according to Theorem 7.
Sufficient Condition.Since  is a geodesic convex set, thus for any  ∈  and any given nonzero vector , there exists a sufficiently small positive number  and thus  +  ∈ ; it follows from Taylor expansion that Since  is geodesic convex function in , using Theorem 7, we have Let  → 0 + ; then, Thus, That is, for any  ∈ , there exists  2 (; V, ) ≥ 0.
In particular, if   meets local Lipschitz condition in the neighborhood of  * ∈ , where  ∉ , then Theorem 11 is called Kuhn-Tucker theorem or K-T theorem for short.The necessary condition of the minimum point is called K-T condition and the point which meets the K-T condition is called K-T point.

Theorem 12 (sufficient condition of the minimum point).
Let the minimization problem (33) be a convex programming problem and let  * be a K-T point.Then,  * is the global minimum point.

Conclusions
In this paper, we introduce the classical Clarke generalized directional derivative and generalized gradient.Then, we extend the identification of convex function and prove the theorem of the first-order necessary and sufficient condition of convex function and strictly convex function to Riemannian manifolds by use of Penot generalized directional derivative.We have also given the second-order necessary and sufficient condition of convex function and strictly convex function and the Kuhn-Tucker theorem and sufficient condition of the minimum point of the inequality constrained optimization problems.In this way, we can do more research about optimization method on differential manifold.In the future, the theorem and the application of the equality constrained optimization problems and the inequality constrained optimization problems will be explored.
Theorem 9 (second-order necessary and sufficient condition of strictly convex function).Let  be an -dimensional Riemannian manifold and  be a geodesic convex set.Let  ⊂ ,  :  →  meet local Lipschitz condition in  and let (, ) be a coordinate plot which contains  and second-order Penot generalized directional derivative of  at  along the direction of V ∈    exists.Then, the necessary and sufficient condition that  is geodesic strictly convex function in  is as follows.For any  ∈ , there exists  2 (; V, ) > 0.  → ,  :  →   are geodesic convex functions in ,  is an open subset of , (, ) is a coordinate plot which contains , and the feasible region of the problem is denoted by  := { ∈  |   () ≤ 0,  ∈ } .Let  be an -dimensional Riemannian manifold and let  * be the optimal solution of the problem.If  :  → ,   :  → ,  ∈  meet local Lipschitz condition in the neighborhood of  * ∈  and   is continuous in  * , where  ∉ , or the vector group ∇(  ∘  −1 )(( * )) is linearly independent, where  ∈ , then there exists constant   ( ∈ ), such that