634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED...

12
634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian Distance Related Problems on Connected Locally Symmetric Manifolds Ricardo Ferreira, Member, IEEE, João Xavier, Member, IEEE, João P. Costeira, and Victor Barroso, Senior Member, IEEE Abstract—The squared distance function is one of the standard functions on which an optimization algorithm is commonly run, whether it is used directly or chained with other functions. Illus- trative examples include center of mass computation, implemen- tation of k-means algorithm and robot positioning. This function can have a simple expression (as in the Euclidean case), or it might not even have a closed form expression. Nonetheless, when used in an optimization problem formulated on non-Euclidean manifolds, the appropriate (intrinsic) version must be used and depending on the algorithm, its gradient and/or Hessian must be computed. For many commonly used manifolds a way to compute the intrinsic dis- tance is available as well as its gradient, the Hessian however is usually a much more involved process, rendering Newton methods unusable on many standard manifolds. This article presents a way of computing the Hessian on connected locally-symmetric spaces on which standard Riemannian operations are known (exponential map, logarithm map and curvature). Although not a requirement for the result, describing the manifold as naturally reductive ho- mogeneous spaces, a special class of manifolds, provides a way of computing these functions. The main example focused in this ar- ticle is centroid computation of a nite constellation of points on connected locally symmetric manifolds since it is directly formu- lated as an intrinsic squared distance optimization problem. Sim- ulation results shown here conrm the quadratic convergence rate of a Newton algorithm on commonly used manifolds such as the sphere, special orthogonal group, special Euclidean group, sym- metric positive denite matrices, Grassmann manifold and projec- tive space. Index Terms—Riemannian distance, Hessian computation, opti- mization. I. INTRODUCTION AND MOTIVATION T HE motivation behind the computation of the Hessian of the squared distance function is usually to use this im- portant object in intrinsic Newton-like optimization methods. The advantages of these methods when compared to gradient Manuscript received October 16, 2012; revised January 22, 2013 and April 19, 2013; accepted April 20, 2013. Date of publication May 06, 2013; date of current version July 15, 2013. This work was supported by the FCT project PEst-OE/EEI/LA0009/2011, partially funded by FCT Project Poeticon++ FP7- ICT-288382, FCT project Printart PTDC/EEA-CRO/098822/2008, and FCT Grant CMU-PT/SIA/0026/2009. The guest editor coordinating the review of this manuscript and approving it for publication was Dr. Nicolas Le Bihan. The authors are with the Institute for Systems and Robotics, Instituto Superior Técnico, Lisbon 1600-011, Portugal (e-mail: [email protected]; jxavier@isr. ist.utl.pt; [email protected]; [email protected]). Color versions of one or more of the gures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identier 10.1109/JSTSP.2013.2261799 methods are well known, especially when high precision is re- quired since its quadratic convergence rate is guaranteed to out- perform any gradient-based algorithm when enough iterations are run. Several authors have approached the problem of im- plementing intrinsic Newton algorithms on smooth manifolds. For example [1], [2] and [3] discuss several implementations of Newton algorithms on manifolds and applications can be found in robotics [4], signal processing [5], image processing [6], etc. Several approaches to the optimization of cost functions in- volving intrinsic squared distance exist, most of them relying on gradient methods, although there are a few exceptions where a Newton method is used. Hüper and Manton [7] have developed a Newton method for this cost function on the special orthog- onal group and in [8] a Newton method which operates on an approximation of the intrinsic distance function on the Grass- mann manifold. These manifolds, and others commonly used in engineering, are a subset of naturally reductive homogeneous spaces (NRHS) [9], [10]. This class of manifolds are interesting since the required operations for implementing Newton-like al- gorithms are easily obtained. Please note though that there is no known general relation between connected locally symmetric spaces (required for the results presented in this article) and nat- urally reductive homogeneous spaces (a worthy example is the Stiefel manifold which is NRHS but not locally symmetric). All the examples presented in this article: the Grassmann manifold , projective space , sphere , positive denite matrices , the special orthogonal group and the special Euclidean group belong to the intersec- tion of locally symmetric and naturally reductive homogeneous spaces. The proposed application is to solve optimization problems where the cost function depends on the squared distance func- tion using a Newton method on manifolds. In particular we pro- vide examples for the problems of computing the centroid of a constellation of points, MAP estimation and clustering using the k-means algorithm. This article does not focus on providing state of the art implementations for these examples, but rather demonstrate how the Hessian of the intrinsic squared distance function might be used in the context of these applications. Particularly in the context of centroid computation, several au- thors have proposed suitable algorithms and applications, for example Moakher [11] uses centroid computation on for smoothing experimental observations obtained with a signi- cant amount of noise in the study of plate tectonics and se- 1932-4553/$31.00 © 2013 IEEE

Transcript of 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED...

Page 1: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

Newton Algorithms for Riemannian DistanceRelated Problems on Connected Locally

Symmetric ManifoldsRicardo Ferreira, Member, IEEE, João Xavier, Member, IEEE, João P. Costeira, and

Victor Barroso, Senior Member, IEEE

Abstract—The squared distance function is one of the standardfunctions on which an optimization algorithm is commonly run,whether it is used directly or chained with other functions. Illus-trative examples include center of mass computation, implemen-tation of k-means algorithm and robot positioning. This functioncan have a simple expression (as in the Euclidean case), or it mightnot even have a closed form expression. Nonetheless, when used inan optimization problem formulated on non-Euclidean manifolds,the appropriate (intrinsic) version must be used and depending onthe algorithm, its gradient and/or Hessian must be computed. Formany commonly usedmanifolds a way to compute the intrinsic dis-tance is available as well as its gradient, the Hessian however isusually a much more involved process, rendering Newton methodsunusable on many standard manifolds. This article presents a wayof computing the Hessian on connected locally-symmetric spaceson which standard Riemannian operations are known (exponentialmap, logarithm map and curvature). Although not a requirementfor the result, describing the manifold as naturally reductive ho-mogeneous spaces, a special class of manifolds, provides a way ofcomputing these functions. The main example focused in this ar-ticle is centroid computation of a finite constellation of points onconnected locally symmetric manifolds since it is directly formu-lated as an intrinsic squared distance optimization problem. Sim-ulation results shown here confirm the quadratic convergence rateof a Newton algorithm on commonly used manifolds such as thesphere, special orthogonal group, special Euclidean group, sym-metric positive definite matrices, Grassmann manifold and projec-tive space.

Index Terms—Riemannian distance, Hessian computation, opti-mization.

I. INTRODUCTION AND MOTIVATION

T HE motivation behind the computation of the Hessian ofthe squared distance function is usually to use this im-

portant object in intrinsic Newton-like optimization methods.The advantages of these methods when compared to gradient

Manuscript received October 16, 2012; revised January 22, 2013 and April19, 2013; accepted April 20, 2013. Date of publication May 06, 2013; date ofcurrent version July 15, 2013. This work was supported by the FCT projectPEst-OE/EEI/LA0009/2011, partially funded by FCT Project Poeticon++ FP7-ICT-288382, FCT project Printart PTDC/EEA-CRO/098822/2008, and FCTGrant CMU-PT/SIA/0026/2009. The guest editor coordinating the review ofthis manuscript and approving it for publication was Dr. Nicolas Le Bihan.The authors are with the Institute for Systems and Robotics, Instituto Superior

Técnico, Lisbon 1600-011, Portugal (e-mail: [email protected]; [email protected]; [email protected]; [email protected]).Color versions of one or more of the figures in this paper are available online

at http://ieeexplore.ieee.org.Digital Object Identifier 10.1109/JSTSP.2013.2261799

methods are well known, especially when high precision is re-quired since its quadratic convergence rate is guaranteed to out-perform any gradient-based algorithm when enough iterationsare run. Several authors have approached the problem of im-plementing intrinsic Newton algorithms on smooth manifolds.For example [1], [2] and [3] discuss several implementations ofNewton algorithms on manifolds and applications can be foundin robotics [4], signal processing [5], image processing [6], etc.Several approaches to the optimization of cost functions in-

volving intrinsic squared distance exist, most of them relying ongradient methods, although there are a few exceptions where aNewton method is used. Hüper and Manton [7] have developeda Newton method for this cost function on the special orthog-onal group and in [8] a Newton method which operates on anapproximation of the intrinsic distance function on the Grass-mann manifold. These manifolds, and others commonly used inengineering, are a subset of naturally reductive homogeneousspaces (NRHS) [9], [10]. This class of manifolds are interestingsince the required operations for implementing Newton-like al-gorithms are easily obtained. Please note though that there is noknown general relation between connected locally symmetricspaces (required for the results presented in this article) and nat-urally reductive homogeneous spaces (a worthy example is theStiefel manifold which is NRHS but not locally symmetric). Allthe examples presented in this article: the Grassmann manifold

, projective space , sphere , positive definitematrices , the special orthogonal groupand the special Euclidean group belong to the intersec-tion of locally symmetric and naturally reductive homogeneousspaces.The proposed application is to solve optimization problems

where the cost function depends on the squared distance func-tion using a Newton method on manifolds. In particular we pro-vide examples for the problems of computing the centroid ofa constellation of points, MAP estimation and clustering usingthe k-means algorithm. This article does not focus on providingstate of the art implementations for these examples, but ratherdemonstrate how the Hessian of the intrinsic squared distancefunction might be used in the context of these applications.Particularly in the context of centroid computation, several au-thors have proposed suitable algorithms and applications, forexample Moakher [11] uses centroid computation on forsmoothing experimental observations obtained with a signifi-cant amount of noise in the study of plate tectonics and se-

1932-4553/$31.00 © 2013 IEEE

Page 2: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 635

quence-dependent continuum modeling of DNA; Manton [12]confirms the need of centroid computation algorithms on mani-folds (particularly compact Lie groups); Pennec [13] uses posi-tive definite symmetric matrices as covariance matrices for sta-tistical characterization, also subject to smoothing; Fletcher [14]uses the computation of centroids for analyzing shapes in med-ical imaging. In [15] a detailed analysis of the centroid compu-tation problem is presented as a special case of a more generalproblem, along with a Newton algorithm to solve it.This article follows from two conference papers [16] and [17]

where a method for computing the Hessian of the intrinsic Rie-mannian squared distance function on a connected locally-sym-metric manifold were presented without proof. The present ar-ticle is entirely self-contained with respect to the previous andconsolidates them in both clarity and detail. It is important tostate that most of the results required in the proof are availablein the literature, and this article’s role is mostly to chain them ina comprehensive way and provide a ready to use result requiringminimal knowledge of the underlying details.

II. RIEMANNIAN MANIFOLDS AND NOTATION

For a given smooth dimensional manifold [10], [18],[19], denote its tangent space at a point by . Thedisjoint union of all these tangent spaces is called the tangentbundle of and is denoted as . The set of real valuedfunctions on is . If and are smooth manifolds,given a smooth map between them, its push-for-ward is defined as the application such that forany tangent vector and any functionthe equality holds. If ,exterior differentiation is denoted by (here is seen as a de-gree 0 differential form).Additionally, the manifold is equipped with a non-degen-

erate, positive and symmetric 2-tensor field , called a Rie-mannian metric, providing the tangent space at each pointwith an inner product . The notation

for will be used exten-sively.The Riemannian exponential map is defined on any Rie-

mannian manifold and sends a vector to a point onthe manifold. If is the unique unit speed geodesic such and that

and , then .In general is only defined on a neighborhood of the originin . However, complete spaces, defined as those where the

map has domain are very interesting in view of man-ifold optimization. On a sufficiently small open neighborhoodthis map is a diffeomorphism and the image of a ball centeredat the origin contained within this neighborhood is known asa geodesic ball (or normal ball). Its inverse function known asthe logarithm, when defined, returns such that,

, and . Although the computationof these maps may be very involved, many manifolds used inengineering have already been widely studied and these mapsare usually available in the literature (see Section III-A for asimple way to compute them for the special class of naturallyreductive homogeneous spaces). The length of a smooth curve

is defined as . The

intrinsic distance between two points belonging to the sameconnected component of is defined as the infimum of thelength of all smooth curves joining and .On a Riemannian manifold there is a canonical way of identi-

fying nearby tangent spaces called the Levi-Civita connection,here denoted by . Once a connection is established, the cur-vature endomorphism is defined as

. Here are any vector fields ex-tending and is the Lie bracket. The op-erator is independent of the extension chosen.The gradient vector is then defined as the

unique tangent vector that satisfiesfor any . The Hessian is defined as the sym-metric 2-form such thatfor any . Note that once an orthonormalbasis for the tangent space is fixed, any tan-gent vector has a canonical expansion with respect to thisbasis and inner product given by , where

. These coefficients can be collected in acolumn matrix , easily inputed to acomputer. The hat will denote a coordinate representation fora given object on the manifold. Similarly the Hessian withrespect to this basis can be described as the matrix such that

for any withcoordinate representation in this basis .

III. NEWTON’S METHOD

A. Unconstrained Optimization

Gradient descent methods (familiarity with basic optimiza-tion techniques is assumed, see for example [20] or [21] fordetailed reference) are undoubtedly among the easiest to im-plement on smooth cost functions, as is the case of the squareddistance function on . Unfortunately their linear convergencerate might be prohibitively expensive on applications whereprecision is required. Newton’s method, when applicable,trades a little in implementation simplicity to gain greatly inconvergence speed, guaranteeing quadratic convergence ratewhen close enough to the optimum.Suppose a function is to be minimized (assumeis convex for simplicity). One way of interpreting Newton’s

method is to describe it as a sequence of second order Taylorexpansions and minimizations. Let

where is a matrix representation of the Hessian func-tion. In the gradient vector field is easily computed as

and the Hessian matrix has thefamiliar form where . Hereis a second order polynomial in attaining a minimum when

. The idea is that will be closer tothe point which minimizes .

B. Manifold Optimization

When the constraint set is a known manifold though, theprevious description still applies with only slight re-interpreta-tions (see [2], [3] and [22] for some generalizations). A search

Page 3: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

636 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

direction is generated as the solution of the system. If a basis for the tangent space is chosen,

then the former is written as

(1)

where is a matrix representation of the Hessian of the costfunction (with respect to the chosen basis for the tangent space)and is the representation of the gradientalso in the chosen basis. See Section II for a description of theseintrinsic objects and Section III-A for a way of computing themin certain spaces.As stated in the previous section, once a Newton direction has

been obtained, it should be checked if it’s a descent direction (itsdot product with the gradient vector should be negative). If thisis not verified, a fallback to the gradient descent direction shouldbe used. Once a direction has been obtained a step in that direc-tion must be taken. Although on a manifold it is not possible toadd a vector to a point directly, a canonical way of doing it isavailable through the Riemannian exponential map which sendsa vector to a point on the manifold as describedin Section II. So the update equation ,can be used to obtain a better estimate. Here is again a stepsize given by Armijo’s rule. The complete algorithm is now de-scribed, with only slight modifications relative to the case:

Manifold Newton Algorithm

Input: function to be minimized.

Output: which minimizes within tolerances.

1: choose and tolerance . Set .

2: loop

3: .

4: if set and return.

5: compute Newton direction as the solution of (1).

6: if set .

7: .

8: . Please note that due to finiteprecision limitations, after a few iterations the result should beenforced to lie on the manifold.

9: and re-run the loop.

10: end loop

IV. HESSIAN OF THE RIEMANNIAN SQUAREDDISTANCE FUNCTION

In [16] the following theorem was introduced without proof,and later updated in [17] still without proof. The proof is pre-sented as an appendix to this article.Theorem IV.1: Consider to be a connected locally-sym-

metric n-dimensional Riemannian manifold with curvature en-domorphism . Let be a geodesic ball centered atand the function returning the intrinsic

(geodesic) distance to . Let denote theunit speed geodesic connecting to a point , where

, and let be its velocity vector at . Definethe function , and con-sider any . Then

(2)

where is an orthonormal basis which diago-nalizes the linear operator ,

with eigenvalues , this means .Also,

Here the and signs denote parallel and orthogonal compo-nents of the vector with respect to the velocity vector of , i.e.

, , and .For practical applications though, presenting the Hessian in

matrix notation greatly improves its readability and comprehen-sion. Hence in [17] a second theoremwas presented also withoutproof which is included in the appendix as well.Collorary IV.2: Under the same assumptions as above, con-

sider an orthonormal basis. Ifis a vector, let the notation denote the column vector de-scribing the decomposition of with respect to the basis

, i.e. , let be the matrix with entriesand consider the eigenvalue decompo-

sition . Here will be used to describe the i’thdiagonal element of . Then the Hessian matrix (a represen-tation for the bilinear Hessian tensor on the finite dimensionaltangent space with respect to the fixed basis) is given by

where is diagonal with elements given by. Hence .

In spaces of constant curvature (such as the sphere and) with sectional curvature , computing the Hessian

has almost zero cost. Due to the symmetries of the curvaturetensor, whenever or are parallel to .Hence, matrix , which is the matrix representation on thegiven basis for the bilinear operator , has a nulleigenvalue with eigenvector . Since the sectional curvatureis by definition equal to and is constant, equal towhenever is not parallel to , then

hence using the Rayleigh quotient, the eigenvalues of areconstant and equal to . So an eigenvalue decomposition is

Page 4: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 637

where is any orthonormal complement of . It follows thenfrom the last theorem that the Hessian is given by

This removes the need for the numerical computation of matrixand its eigenvalue decomposition, significantly speeding the

computation of the Hessian matrix.

A. Algorithm

The complete algorithm is presented in both situations, whenthe space is not known to be of constant curvature:

Hessian of Riemannian squared distance function

Input: an orthonormal base , andthe Riemannian curvature tensor.

Output: the Hessian matrix of the Riemannian squareddistance function .

1: Build matrix .

2: Compute its eigenvalue decomposition .

3: Assemble diagonal matrix with elements .

4: .

or when it is known to be of constant curvature:

Hessian of Riemannian squared distance function (spacesof constant curvature)

Input: an orthonormal base , andthe Riemannian curvature tensor.

Output: the Hessian matrix of the Riemannian squareddistance function .

1: Represent in the given basis.

2: .

V. MANIFOLD APPLICATIONS

A. Centroid Computation

Let be a connected manifold anda constellation of points. Let be the

function that returns the intrinsic distance of any two points onthe manifold and define a cost function as

(3)

The set of solutions to the optimization problemis defined as the Fréchet mean set of the

constellation and each member will be called a centroid of. Depending on the manifold , the centroid might not be

unique, for example if the sphere is considered with a constel-lation consisting of two antipodal points, all the equator pointsare centroids. The set of points at which the function (3) attainsa local minimum is called the Karcher mean set and is denotedas . The objective is to find a centroid for the givenconstellation (which in the applications of interest should beunique), but the possibility of convergence to a local minimum

is not dealt with. Conditions for uniqueness of Karcher-Fréchetmeans usually involves the concept of manifold injectivityradius and the diameter of the constellation. Please see [12],[15], [23], [24] for the explicit treatment of these points.Using linearity of the gradient and the Hessian operators

(meaning in particular that if thenand ), the cost

function in (3) allows for the decomposition

(4)

where the fact that the gradient of the squared Riemannian dis-tance function is the symmetric of the Riemannian map isused (as stated in [25] as a corollary to Gauss’s lemma).The algorithm for centroid computation is then

Centroid computation

Input: A constellation with pointssufficiently close (see text for details) and an initial estimate

Output: An element of the Karcher mean set of theconstellation

1: Apply Newton’s algorithm as described in Section III tofunction where at each step the Hessian and gradient iscomputed as follows:

2: for each point in the constellation do

3:

4: (as described in Section IV)

5: end for

6:

7:

B. K-Means Algorithm

The implementation of a K-means algorithm is straightfor-ward once a working centroid computing algorithm is available.The algorithm is as follows:

k-means algorithm

Input: An dimensional manifold where the centroid iscomputable, a cloud of points and the number of desiredclasses.

Output: centroids of each class.

1: Choose randomly as initial estimate forthe centroids.

2: for each point in do

3: Compute distance to each centroid: .

4: Label point as belonging to set , where .

5: end for

6: Recompute the centroids .

7: If the centroids did not change position (or a maximumnumber of iterations reached), return.

Page 5: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

638 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

C. MAP Estimator

Consider a freely moving agent in whose position is rep-resented as a point , seen as the rigid transformation thattransforms points in the world referential, taking them to thelocal (agent) referential. Keeping the experiment simple, con-sider that the it observes several known landmarks in the world

. Hence, in the local referential, the agentobserves the points . If the agent is considered to be at witha certain uncertainty, it is possible to build a prior knowledgeprobability density function aswhere is a normalizing constant and describes an isotropiclevel of uncertainty. Notice that all directions are treated equallywhich is usually not the case. Please note that by the identity

a slightly more useful prior may bebuilt weighting differently translations from rotations. Withsimplicity in mind, assume that this description is useful.Assume also that the sensors are not perfect and the obser-vations obey the following Gaussian probability distribution

where, again is anormalizing constant and is a matrix encoding the uncertaintyof the sensor. If the observations are considered to be indepen-dent, the MAP estimator of the position is given by

Using the usual trick of applying the logarithm and discardingconstants, the former problem is equivalent to

This is formulated as an optimization problem on . Thegradient of each term is readily available and the Hessian of thefirst terms can be obtained using standard techniques (see thechapter of Riemannian Embeddings on any Riemannian geom-etry book, specifically the part about the second fundamentalform). The result presented in this paper allows for the Hes-sian of the last term to be obtained as well, thus allowing for aNewton algorithm to be implemented.

VI. RESULTS

This section holds experimental results for the main appli-cation of centroid computation. Fig. 1 compares the results ofapplying a Newton algorithm and a standard gradient algorithmwhen computing the centroid of a constellation on 6 differentmanifolds. The 20-point constellations were generated using aradius of except for the Grassman where the radius usedwas . The results presented in logarithmic scale clearly showthe quadratic convergence rate of Newton’s method and thelinear convergence rate of the gradient method. All examplesshow a plateau at due to finite precision issues. Note thatthe projective space manifold is a special caseof the Grassman, hence the previous expressions are applicable.Fig. 2 shows the results obtained for an implementation of the

MAP estimator described in SectionV-C. As described, an agent

is navigating in a world with 5 randomly placed landmarks. Inthis experiment and was used. The gradientmethod is clearly outperformed by the 5 iterations taken by theNewton method to attain the required precision.As a final note concerning absolute time to convergence, none

of the results of the Newton method shown in Fig. 1 are compet-itive with the gradient alternative. Although more iterations arerequired, these are simpler and thus convergence to the requiredprecision is faster. This no longer holds for an implementationof the MAP estimator where the gradient method shows ex-tremely slow convergence whereas Newton’s method still con-verges typically in less than 10 iterations and outperforms thegradient implementation. The choice of method does not havea simple answer in engineering applications and in the end theactual cost function, manifold and desired precision play crucialroles. Also, as mentioned in the discussion after the statement ofcorollary IV.2 for certain manifolds, e.g. , computation ofthe Hessian is almost trivial providing an additional advantageto Newton’s method.

VII. CONCLUSIONS

This article describes a simple algorithm to obtain the Hes-sian of the intrinsic squared distance function on connected lo-cally-symmetric manifolds on which it is known how to com-pute basic Riemannian differential operations. Results are pre-sented for centroid computation on the commonly used mani-folds , , , , , and . This is bynomeans an exhaustive list, and the result is valid for other man-ifolds fitting the requisites (for example the hyperbolic plane).Besides the main application, simple examples of MAP esti-mation and k-means clustering are also provided, extending therange of applications besides centroid computation.

APPENDIX ANATURALLY REDUCTIVE HOMOGENEOUS SPACES

Although this section is not critical for presenting the mainresult in this article, it does show that the method presented isviable by providing a recipe for obtaining the required data ina vast class of manifolds used in engineering. Note that a basicunderstanding of Lie group theory is assumed. Naturally Re-ductive Homogeneous Spaces, henceforth denoted by NRHS,(see for example [9] and [10]), are important since they canlead to closed formula solutions for the Riemannian exponen-tial maps, logarithm maps and curvature endomorphisms, ex-actly what’s needed to implement the Hessian algorithm pre-sented. A space with this property is defined as a coset manifold

, where is a Lie group (with Lie algebra ) anda closed subgroup (with Lie sub-algebra ), furnished

with a -invariant metric such that there exists an in-variant subspace that is complementary to . Notethat but is usually not a Lie sub-algebra since itis usually not closed under the Lie bracket operation. Further-more, the property

for

needs to hold. Here the subscript denotes projection on thissubspace. Henceforth, for spaces with this property, will becalled a Lie subspace for .

Page 6: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 639

Fig. 1. Simulation results for centroid computation. Except for the Grassman manifold, whose constellations were built with a radius of , all constellationswere built with a radius of .

A. NRHS Construction for a Particular Riemannian Manifold

When facedwith an optimization problem on a particular Rie-mannian manifold , it is not usually known whether or not itadmits an NRHS structure. Since many useful manifolds in en-gineering admit such structures, the process of identifying it willbe described here.First it is necessary to describe as a coset manifold

were the symbol states that the two sides are diffeo-morphic. Here, a proposition stated in [9] solves the problem,stating that all that needs to be done is to find a Lie groupwhich acts transitively on :

Theorem A.1: Let be a transitive action andlet be its isotropy subgroup at a point . Then there isa natural map which is a diffeomorphism. Inparticular, the projection , is a submersion.Furthermore, this action must be an isometry as stated in

the definition of an NRHS space, which means that for any, and the property

, must hold, where denotes the push forwardof the translation by .Examples:1) acts on the unit sphere

(seen as a Riemannian subspace) as the restriction

Page 7: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

640 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

Fig. 2. MAP position estimate results using 5 observed landmarks withand .

of the usual action of on . This actionis transitive. The isotropy subgroup ofconsists of the subgroup

Hence, ignoring the natural diffeomorphism yields. To verify that this action is G-invariant

let and . Then:

2) As a trivial example, acts transitively on itself(seen as a Riemannian submanifold of with theEuclidean inner product) through group multiplication.The isotropy subgroup at any point is the trivial subgroup

(where is the group identity), hence trivially. As before, to verify that this action

preserves the inner product let and. Then:

3) Expanding the previous example, the Lie group productacts transitively on the Special Euclidean

group (seen as a Riemannian submanifold of) as

Once again the isotropy subgroup is trivial, hence

. If then for a

given element the push forward of theaction of is given by:

Hence the action preserves the inner product since:

4) (the set of invertible matrices with real en-tries) acts transitively on (the set of pos-itive definite symmetric matrices with the inner productdescribed below) by conjugation, that is .The isotropy subgroup of the identity matrix seen as an el-ement of is the set

. So, . Lettingand and assuming the inner

product is given by , then:

5) If (the set of real matrices) withand is full rank, let the notation denote the subspacegenerated by the column vectors of . The Grassman man-ifold consists of the set of all such subspaces, i.e.

. Please notethat the elements of theGrassmanmanifold are equivalenceclasses, where the columns of and thecolumns of span the same subspace. Consider the transi-tive action of on defined as

the isotropy subgroup of is the set

hence . The actionis G-invariant since once again if and

then:

Let be the Lie algebra of and be the Lie sub-algebraof . The next step consists of finding a Lie subspace suchthat and . This step must be done byinspection but it is usually not hard to accomplish.Examples:1) Let denote the set of skew sym-metric matrices with real entries. For the coset space

, and

. By inspection (due to

the requirement that ) a logical candidate for

is the set . Since

(5)

results that is indeed a Lie subspace.

Page 8: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 641

2) When and , and(the trivial vector space). Hence, the obvious choice

is , which is obviously invariant under .3) The same happens when considering and

. Hence, for ,.

4) If and as is the case for, the corresponding Lie algebras are

( real matrices) and . A natural candi-date for the Lie subspace is the set of symmetric matrices

, and indeed if andthen hence it isinvariant.

5) For the case of the Grassmann, and(as seen previously). The

corresponding Lie algebras are and

.

Then, by inspection, the obvious choice for the Lie sub-space is

It is easily checked that this choice is invariant.All that remains to be done is to verify if the construction

verifies the property

for

Since is identified with , the dot product is the pull-backby of the dot product on . If the property is not satisfied,another construction with another Lie group acting onmight be tried, or it is possible that does not admit an NRHSstructure.Examples:1) Continuing the previous examples consider thesphere, where .

Let and

where . Then it results that

. Thus the

required result is trivially verified. If needed, the corre-sponding dot product on can be found by noting that

. Hence

2) For , the tangent vectors in both man-ifolds are canonically identified, hence ifthe inner product on is given in the usual way:

So the required property is once again satisfied

3) When and , the innerproduct on is found in the same manner as in the firstexample. Thus if , ,

The Lie bracket on the product group is given by theproduct of the Lie brackets. Hence

Then to check the required property:

(check the example for the case for details of thelast step).

4) In the case of the symmetric positive definite matriceswhere the Lie bracket results in a skewsymmetric matrix, hence the projection back to resultsin a null vector. Hence the requirement is trivially satisfied.

5) Noting that for the Grassmann manifoldthen The

needed property is once again trivially verified.Hence all five manifolds have been described as naturally re-

ductive homogeneous spaces.

B. Operations on NRHS Manifolds

This section details why the structure of these manifolds isimportant to the Riemannian optimization process. A proposi-tion in [9] states:Theorem A.2: If is a naturally reductive ho-

mogeneous space, its geodesic starting at with tangent vectorare given by for all , where

is the one parameter subgroup of identified as an ele-ment of .Hence the Riemannian exponential map follows directly

from the Lie group’s exponential map which in our examplesis the standard matrix exponential (since is eitheror ). Geodesics starting at any other point of can befound by translation of since acts transitively as an isom-etry. The Riemannian logarithm map follows from inversion.On a manifold with NRHS structure, the curvature endomor-

phism is also computable as seen for example in [26]:

(6)

Page 9: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

642 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

Examples: To finish the examples, we provide a final sum-mary of the functions needed for each of the considered mani-folds. Note that these are simplified and concise versions of theresults obtained by the above theorem and expression 6.1) The Sphere : This n-dimensional manifold is described

as the set whose tangent spaceat a point is . Let

, and is the norm of . It canbe shown that for the ambient metric :• .• where

.• .2) Special Orthogonal Group : This di-

mensional manifold represents the set of rotations of andis described as whosetangent space at a point is

, where denotes the set ofskew-symmetric matrices. The metric comes naturally from theRiemannian embedding as• , where exp denotes the matrixexponential function.

• where log denotes the matrix loga-rithm function.

• .3) Special Euclidean Group : The Euclidean

group, characterized as the product manifold ,inherits these manifolds’ properties. Hence, at a point

, the tangent spacewith dot product given by

.• , whereexp denotes the matrix exponential function.

• , where log de-notes the matrix logarithm function.

•where

the brackets denote the Lie bracket only onsince is flat.

4) Symmetric Positive Definite Matrices :This dimensional manifold is describedas the set

whose tangent space ata point is ,where denotes the set of symmetric matrices.Let , . Whenconsidering the metric thefollowing expressions hold• .• .• , where

.5) The Grassmann Manifold : The Grassmann is an

dimensional manifold of dimensional linear sub-spaces in . It is naturally described as a quotient manifoldwith the previously mentioned equivalence relation, and a point

is described by a representative whichis the set of the first columns of the elements in for

. The tangent space at a point is.

Let , .For the ambient metric (the simplifiedexpressions shown next are obtained from [2], [27] and the cur-vature endomorphism follows directly from equation 6):• , where isthe compact SVD of .

• , where is theSVD decomposition of , are the diagonal elementsof and

.• .

APPENDIX BPROOF OF THE THEOREMS

This section proves theorem IV.1 and corollary IV.2, amain contribution of this paper but delayed to an appendixfor readability. The proof is mostly a chain of known resultswhich can be found in texts such as [9], [28]. With the intentof finding the Hessian of the function , recall that (see forexample [9])

(7)

where is any local extension of . Note that from the prop-erties of a connection, its value depends only of at , butfor the expression to be formally correct the extension mustbe considered. Knowing that the gradient operator is linear andthat for any two smooth functions defined on anopen set satisfies the point-wise multiplication property

, allows for the simplification.

Defining as the unit normed radial vector field whenwritten in normal coordinates centered at , a corollary toGauss’s Lemma [25] states that . Hence theformer expression is written as . Gauss’sLemma also allows for the decomposition of any vector field

as , where is a vector fieldparallel to and is orthogonal to it. Thesestatements, along with the properties of a connection, are usedto write

Noting that for any vector field ,and since is parallel to

, there is a smooth function such that

Page 10: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 643

. Since is tangent to unit speedgeodesics emanating from , . Hence

(8)

Now let be a curve in the geodesic spherewith

and . In normal coordinates centered at , considerthe geodesic variation for some , given by

. Here the hat notation denotes a coordinaterepresentation, hence if is the normal coordinatefunction then , , and .Defining and

, the corresponding Jacobi field(see [25]) is given in coordinates by

. Note that is normal to the unit-speed geodesicand that . Also, notice

that since is a geodesic. Inorder to ease notation the coordinate representation for theseobjects will be hidden, although not forgotten. Hence, at apoint ,where the fact that is used (see [25] Lemma 6.3).Substituting in (8), again at , results in

Substituting back into (7), yields the known expression for theHessian in terms of Jacobi fields [12], [23]

(9)

All that remains to do is to find an expression for the Jacobi fieldand take its covariant derivative. This leads to a rather lengthydiscussion so it is stated here as a couple of lemmas:Lemma B.1: The solution of the ODE , ,

is given by where

Proof: Picard’s existence theorem guarantees uniquenessand direct substitution of the result in the differential equationproves the result.LemmaB.2: Let be a locally-symmetric Riemannianman-

ifold and be a geodesic ball centered at . Ifis the intrinsic distance of a point to and

is the unit speed radial geodesic runningfrom to , given a tangent vector orthogonal to

, then the normal Jacobi field alongwhich satisfies and is given by

(10)

where is the parallel transport along of thetangent space’s orthonormal basis which di-agonalizes the linear operator defined as

, i.e., . Note thatdenotes the curvature tensor of and is defined as in the

previous lemma.Proof: As stated in [25] the Jacobi field specified by its

two endpoints and exists and is unique ifis not conjugate to along . Consider the Jacobi equation

(11)

Choosing an orthonormal basis for the tangent spaceand creating the vector fields along by parallel

translation of , one can write with respect to this basis(note that the set is a basis for ) as

, where . Hence the lefthand side of (11) can be written as

(12)

where the identity is used, whichfollows from the fact that the vector fields are parallel.The goal is now to solve this ordinary differential equation.Start by defining the linear operator as

and note that due to the symme-tries of the Riemannian curvature tensor, this operator is self-ad-joint, i.e. . This guarantees thatthere is an orthonormal basis such that

for some as described next. Write as alinear combination as . Sinceany two vector spaces with the same dimension are isomor-phic there is an isomorphism, taking to

. In this vector space the operator may be written as. Hence, writing in matrix notation by defining

and since is linear in a finite dimen-sional vector field, can be described as a matrix. Hence,

......

. . ....

The fact that the operator is self-adjoint makes a symmetricmatrix and as such, it admits an eigenvalue decomposition

where is a diagonal matrix

Page 11: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

644 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013

such that are the eigenvalues of and is an orthogonal ma-trix with the normalized eigenvectors of as its columns. Hence

hence, if arethe entries of the matrix, is the orthonormalbasis such that . Using the isomorphism onceagain, this means that is the basis that di-agonalizes the operator as . Define asthe parallel transport of along . Define as well the vectorfields which, since is locallysymmetric, are parallel along [9]. It follows then, using thefact that parallel transport preserves inner products andis an orthonormal basis

Writing (12) in terms of the new basis:

Hence, the Jacobi equation decouples into scalar equationswith the corresponding two-point boundary conditions:

andand

...and

Invoking lemma B.1, the solution of the th equationis given by Hence

.This expression for can be substituted in (9), by making

. Taking the covariant derivative of evaluatedat , considering that since is parallelalong the geodesic, results in

Hence, defining by pointwise division,

Although (2) provides a way to compute the Hessian, it isnot very implementation-friendly. This section proves corollaryIV.2 which re-writes the equation in matrix form once a tangentbasis is fixed.From the symmetries of the curvature tensor follows that any

vector parallel to (for example ) belongs to the kernelof the operator . Without loss of generality assume thatis parallel to (hence , and

. Then, since , (2)can be re-written:

where the matrices are defined in the theorem statement.

REFERENCES

[1] D. Gabay, “Minimizing a differentiable function over a differentialmanifold,” J. Optim. Theory Appl., vol. 37, no. 2, pp. 177–219,1982.

[2] A. Edelman, T. A. Arias, and S. T. Smith, “The geometry of algorithmswith orthogonality constraints,” SIAM J. Matrix Anal. Appl., vol. 20,no. 2, pp. 303–353, 1998.

[3] J. H. Manton, “Optimisation algorithms exploiting unitary con-straints,” IEEE Trans. Signal Process., vol. 50, no. 3, pp. 635–650,Mar. 2002.

[4] C. Belta and V. Kumar, “An svd-based projectionmethod for interpola-tion on SE(3),” IEEE Trans. Robot. Autom., vol. 18, no. 3, pp. 334–345,Jun. 2002.

[5] J. H. Manton, “A centroid (Karcher mean) approach to the joint ap-proximate diagonalization problem: The real symmetric case,” DigitalSignal Process., vol. 5, pp. 468–478, 2005.

[6] U. Helmke, K. Hüper, P. Lee, and J. Moore, “Essential matrix estima-tion via Newton-type methods,” in Proc. 16th Int. Symp. Math. TheoryNetw. Syst. (MTNS), Leuven, Belgium, 2004.

[7] K. Krakowski , K. Hüper, and J. Manton, “On the computation ofthe Karcher mean on spheres and special orthogonal groups,” in Proc.Workshop Robot. Math. (RoboMat ’07, 2007.

[8] P. Absil, R. Mahony, and R. Sepulchre, “Riemannian geometry ofgrassmann manifolds with a view on algorithmic computation,” ActaAppl. Math, vol. 80, no. 2, pp. 199–220, Jan. 2004.

[9] B. O’Neil, Semi-Riemannian Geometry. New York, NY, USA: Aca-demic, 1983.

[10] A. Arvanitoyeorgos, An Introduction to Lie Groups and the Geometryof Homogeneous Spaces. Providence, RI, USA: Amer. Math. Soc.,1999.

[11] M. Moakher, “Means and averaging in the group of rotations,” SIAMJ. Matrix Anal. Appl., vol. 24, no. 1, pp. 1–16, 2002.

[12] J. H. Manton, “A globally convergent numerical algorithm forcomputing the centre of mass on compact Lie groups,” in Proc. 8thInt. Conf. Control, Autom., Robot., Vis., Kunming, China, Dec. 2004.

Page 12: 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL … · 2015. 7. 2. · 634 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 7, NO. 4, AUGUST 2013 Newton Algorithms for Riemannian

FERREIRA et al.: NEWTON ALGORITHMS FOR RIEMANNIAN DISTANCE RELATED PROBLEMS 645

[13] X. Pennec, P. Fillard, and N. Ayache, “A Riemannian framework fortensor computing,” Res. Rep. 5255, INRIA, Jul. 2004, Int. J. Comput.Vis..

[14] P. Fletcher, C. Lu, S. Pizer, and S. Joshi, “Principal geodesic analysisfor the study of nonlinear statistics of shape,” IEEE Trans. Med. Imag.,vol. 23, no. 8, pp. 995–1005, Aug. 2004.

[15] D. Groisser, “Newton’s method, zeroes of vector fields, and the Rie-mannian center of mass,” Adv. Appl. Math., vol. 33, pp. 95–135, 2004.

[16] R. Ferreira, J. Xavier, J. Costeira, and V. Barroso, “Newton methodfor Riemannian centroid computation in naturally reductive homoge-neous spaces,” in Proc. Int. Conf. Acoust., Speech, Signal Process.(ICASSP’06), May 2006, pp. 704–707.

[17] R. Ferreira and J. Xavier, “Hessian of the Riemannian squared distancefunction on connected locally symmetric spaces with applications,” inProc. Controlo ’06.

[18] W. M. Boothby, An Introduction to Differentiable Manifolds and Rie-mannian Geometry. New York, NY, USA: Academic, 1975.

[19] M. P. do Carmo, Riemannian Geometry. Boston, MA, USA:Birkhäuser, 1992.

[20] S. Boyd and L. Vandenberghe, Convex Optimization. New York, NY,USA: Cambridge Univ. Press, 2004.

[21] D. P. Bertsekas, Nonlinear Programming, 2nd ed. Belmont, MA,USA: Athena Scientific, 1999.

[22] K. Hüper and J. Trumpf, “Newton-like methods for numerical op-timization on manifolds,” in Proc. Asilomar Conf. Signals, Syst.,Comput., 2004, 2006, pp. 136–139.

[23] H. Karcher, “Riemannian center of mass and mollifier smoothing,”Comm. Pure Appl. Math., vol. 30, pp. 509–541, 1977.

[24] B. Afsari, “Riemannian Lp center of mass: Existence, uniqueness, andconvexity,” Proc. Amer. Math. Soc., vol. 139, pp. 655–673, 2011.

[25] J. M. Lee, Riemannian Manifolds: An Introduction to Curvature.New York, NY, USA: Springer, 1997.

[26] I. Agricola, “Connections on naturally reductive spaces, their diracoperator and homogeneous models in string theory,” Commun. Math.Phys., no. 232, pp. 535–563, 2003.

[27] D. Sepiashvili, J. M. F. Moura, and V. H. S. Ha, “Affine-permuta-tion symmetry: Invariance and shape space,” in Proc. IEEE WorkshopStatist. Signal Process., Sep.–1 Oct. 2003, pp. 307–310.

[28] J. Cheeger and D. G. Ebin, American Mathematical Society, Compar-ison Theorems in RiemannianGeometry. Providence, RI, USA:AMSChelsea Publishing, 1975.

Ricardo Ferreira graduated in Electrical andComputer Engineering in 2004 at Instituto SuperiorTécnico (IST). In 2006 he received his M.Sc. degreein studying underwater stereo reconstructions of 3Dscenes when observed through an air-water inter-face. His Ph.D. studies focused on reconstructingpaperlike surfaces from multiple camera images, andwere concluded in 2010. Both his M.Sc. and Ph.D.degrees were obtained at IST. His current researchinterests focus on general geometric problemsarising in robotics and computer vision.

João Xavier received the PhD degree in electricaland computer engineering from the Instituto SuperiorTecnico (IST), Lisbon, Portugal, in 2002. Currently,he is an assistant professor in the Department of Elec-trical and Computer Engineering, IST. He is also aresearcher at the Institute of Systems and Robotics(ISR), Lisbon, Portugal. His current research inter-ests are in the area of optimization, sensor networks,and signal processing on manifolds. He is a memberof the IEEE.

João P. Costeira received the Licenciatura de-gree, MS degree, and PhD degree in electricaland computer engineering from Instituto SuperiorTécnico (IST), Lisbon, Portugal in 1985, 1989, and1995, respectively. He was a visiting scientist atthe Robotics Institute, Carnegie Mellon Universitybetween 1992 and 1995, where he developed work inthe area of multiple motion segmentation. Currentlyhe is associate professor at IST and a researcher atInstituto de Sistemas e Robótica working in the areaof computer vision.

Victor Barroso (M’89–SM’00) was born in Lisbon,Portugal, on December 22, 1952. He received theE.E. degree in 1976 and the Ph.D. degree in electricaland computer engineering in 1990, both from Insti-tuto Superior Técnico (IST), Lisbon, Portugal. Hereceived the title of “Professor com Agregação” inElectrical and Computer Engineering in 2003 fromUniversidade Técnica de Lisboa-IST. In 1976, hejoined the faculty of the Department of Electrical andComputer Engineering, IST. He was promoted toAssistant Professor in 1990, to Associate Professor

in 1993, and to Full Professor in 2004. He has taught courses in systems andsignal theory, control systems, signal and array processing, computer networks,and communications. He was Vice-President of the IST’s Scientific Councilfrom November 2001 to March 2007. From April 2007 to March 2009, he wasthe Director of the Portuguese pole of the Information and CommunicationTechnology Institute (ICTI@Portugal), a partnership Involving several Por-tuguese Universities, research and development institutions, and companieswith the Carnegie Mellon University. He was also the Director of the Institutefor Systems and Robotics (ISR), Lisbon, Portugal, from September 2009 toDecember 2012. His research interests are in statistical signal/image and arrayprocessing and wireless communications. He has published over 100 journaland conference papers.Dr. Barroso was an Associate Editor for the IEEE TRANSACTIONS ON SIGNAL

PROCESSING and for the IEEE SIGNAL PROCESSING LETTERS. He was the rep-resentative of the IEEE Signal Processing Society at the IEEE Wireless Peri-odicals Working Group in 2000 and the General Chair of the IEEE Interna-tional Workshop on Signal Processing Advances in Wireless Communicationsin 2004. He is affiliated with IEEE Signal Processing Society.