Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations

Similar documents
Mathematics 6 12 Section 26

Montana Instructional Alignment HPS Critical Competencies Mathematics Honors Pre-Calculus

Approximating Square Roots

Alternating Projections

Lesson 29: Fourier Series and Recurrence Relations

5.5 Newton s Approximation Method

Content Standard 1: Numbers, Number Sense, and Computation

A TESSELLATION FOR ALGEBRAIC SURFACES IN CP 3

WAYNESBORO AREA SCHOOL DISTRICT CURRICULUM ACCELERATED GEOMETRY (June 2014)

Matrix algorithms: fast, stable, communication-optimizing random?!

Face two-colourable triangulations of K 13

mywbut.com Diffraction

Solving Traveling Salesman Problem Using Parallel Genetic. Algorithm and Simulated Annealing

Mathematics Appendix 1: Examples of formal written methods for addition, subtraction, multiplication and division

Prentice Hall Mathematics: Geometry 2007 Correlated to: Arizona Academic Standards for Mathematics (Grades 9-12)

Objectives and Homework List

GEOMETRIC TOOLS FOR COMPUTER GRAPHICS

Which n-venn diagrams can be drawn with convex k-gons?

Geometry Curriculum Guide Lunenburg County Public Schools June 2014

PITSCO Math Individualized Prescriptive Lessons (IPLs)

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

Game Mathematics. (12 Week Lesson Plan)

MAT 3271: Selected Solutions to the Assignment 6

Geometry Curriculum Map

Simplified Voronoi diagrams for motion planning of quadratically-solvable Gough-Stewart platforms

Journal of Engineering Research and Studies E-ISSN

Killingly Public Schools. Grades Draft Sept. 2002

Texas High School Geometry

Conjectures concerning the geometry of 2-point Centroidal Voronoi Tessellations

Graphical Approach to Solve the Transcendental Equations Salim Akhtar 1 Ms. Manisha Dawra 2

Closest Points, Moving Surfaces, and Algebraic Geometry

Michigan Edition. correlated to the. Michigan Merit Curriculum Course / Credit Requirements Geometry

Lofting 3D Shapes. Abstract

Accelerating the Hessian-free Gauss-Newton Full-waveform Inversion via Preconditioned Conjugate Gradient Method

Preliminaries: Size Measures and Shape Coordinates

Matrices. Chapter Matrix A Mathematical Definition Matrix Dimensions and Notation

The Helly Number of the Prime-coordinate Point Set

A High-Order Accurate Unstructured GMRES Solver for Poisson s Equation

Structured System Theory

The Problem of Calculating Vertex Normals for Unevenly Subdivided Smooth Surfaces

B.Stat / B.Math. Entrance Examination 2017

DEVELOPMENT OF POSITION MEASUREMENT SYSTEM FOR CONSTRUCTION PILE USING LASER RANGE FINDER

Parallel Evaluation of Hopfield Neural Networks

THE COMPUTER MODELLING OF GLUING FLAT IMAGES ALGORITHMS. Alekseí Yu. Chekunov. 1. Introduction

PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION

Foundations of Computing

High School Geometry

A Geometric Approach to the Bisection Method

Construction of planar triangulations with minimum degree 5

GEOMETRY CURRICULUM GUIDE Overview and Scope & Sequence

Isometries. 1 Identifying Isometries

coding of various parts showing different features, the possibility of rotation or of hiding covering parts of the object's surface to gain an insight

Curriculum Map: Mathematics

FOUNDATION HIGHER. F Autumn 1, Yr 9 Autumn 2, Yr 9 Spring 1, Yr 9 Spring 2, Yr 9 Summer 1, Yr 9 Summer 2, Yr 9

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1

Use throughout the course: for example, Parallel and Perpendicular Lines Proving Lines Parallel. Polygons and Parallelograms Parallelograms

6-12 Math Course Sequence Effective

Vector Addition. Qty Item Part Number 1 Force Table ME-9447B 1 Mass and Hanger Set ME Carpenter s level 1 String

Bezier Curves. An Introduction. Detlef Reimers

GCSE Higher Revision List

Mathematics Masters Examination

Honors Precalculus: Solving equations and inequalities graphically and algebraically. Page 1

correlated to the Michigan High School Content Expectations Geometry

Rectangular Matrix Multiplication Revisited

Local Linear Approximation for Kernel Methods: The Railway Kernel

Optimal Region for Binary Search Tree, Rotation and Polytope

Richard Lander School Maths Key Stage 3

MATHEMATICS Grade 7 Advanced Standard: Number, Number Sense and Operations

Suggested problems - solutions

MATHEMATICS Curriculum Grades 10 to 12

Department: Course: Chapter 1

Reals 1. Floating-point numbers and their properties. Pitfalls of numeric computation. Horner's method. Bisection. Newton's method.

Archbold Area Schools Math Curriculum Map

Revision Problems for Examination 2 in Algebra 1

THE COMPUTER MODELLING OF GLUING FLAT IMAGES ALGORITHMS. Alekseí Yu. Chekunov. 1. Introduction

Visualization of the Triangle Algorithm in R2

Year 9: Long term plan

Suggested Foundation Topics for Paper 2

AN EXTENSION OF A THEOREM OF BARLOTTI TO MULTIPLE DIMENSIONS

Mathematics. Jaehyun Park. CS 97SI Stanford University. June 29, 2015

International Journal of Emerging Technologies in Computational and Applied Sciences (IJETCAS)

Reachability on a region bounded by two attached squares

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Solving Systems of Spline Equations: A Linear Programming-based Approach

Geometric Algebra. 8. Conformal Geometric Algebra. Dr Chris Doran ARM Research

Efficient MR Image Reconstruction for Compressed MR Imaging

SECTION I: NUMBER AND NUMERATION

Introduction to Geometry

Robust Pole Placement using Linear Quadratic Regulator Weight Selection Algorithm

AQA GCSE Maths - Higher Self-Assessment Checklist

Social Network Analysis

On Minimum Weight Pseudo-Triangulations

Ellipse fitting using orthogonal hyperbolae and Stirling s oval

Convex Optimization M2

Curriki Geometry Glossary

SELECTIVE ALGEBRAIC MULTIGRID IN FOAM-EXTEND

Graph Crossing Number and Isomorphism SPUR Final Paper, Summer 2012

Deformable Registration of Cortical Structures via Hybrid Volumetric and Surface Warping

Gardener s spline curve

inter.noise 2000 The 29th International Congress and Exhibition on Noise Control Engineering August 2000, Nice, FRANCE

Transcription:

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations Ben Jeuris and Raf Vandebril KU Leuven, Dept. of Computer Science, 3001 Leuven(Heverlee), Belgium {ben.jeuris,raf.vandebril}@cs.kuleuven.be Abstract. The geometric mean of a series of positive numbers a 1,...,a n is defined as the nth root of its product: n a 1 a n. Generalizing this concept to positive definite matrices is not straightforward due to the noncommutativity. Based on a list of desired properties the ALM-list initially some recursive algorithms were proposed. Unfortunately, these algorithms were typically quite time consuming. A novel optimization based approach leads to the concept of the Karcher mean and is nowadays favored as the generalization towards matrices of the geometric mean. It is not so well-known that one can design for two scalars an algorithm quadratically convergent to the geometric mean, solely relying on arithmetic and harmonic means. The procedure is straightforward, compute the arithmetic and harmonic mean, and iterate this procedure with these newly computed means. Apparently, this procedure also works for two matrices. In these notes, we will propose a whole series of fast possible generalizations, and conduct numerical experiments with these algorithms. Unfortunately most of them do not satisfy all imposed ALM-constraints, but on the other hand, they converge rapidly and approximate the Karcher mean. As such they can be used to initiate manifold optimization procedures for computing the Karcher mean. 1 Introduction The idea of the algorithms proposed in this article is not to satisfy all ALMproperties [1] in a full theoretical sense. We do, however, aim at providing numerical means, up to a certain accuracy. More precisely, suppose that for a practical application only 4 accurate digits are required, and one of the algorithms proposed in this article is capable of computing a matrix, equal to the actual Karcher mean up to four accurate digits. Then this mean can replace the Karcher mean in the application, increasing the computational efficiency. Arithmetic-Harmonic Mean for Two Elements.1 The Scalar Case Let a 1,a be positive numbers and let us define the harmonic, arithmetic, and geometric means as follows: F. Nielsen and F. Barbaresco (Eds.): GSI 013, LNCS 8085, pp. 785 793, 013. c Springer-Verlag Berlin Heidelberg 013

786 B. Jeuris and R. Vandebril ( a 1 1 + a 1 ) 1 h(a 1,a )=, ( ) a1 + a a(a 1,a )=, g(a 1,a )= a 1 a. Then one can prove [, 3] that Algorithm 1 converges quadratically to the geometric mean. Algorithm 1. Let a 1,a be two positive numbers while a 1 a b 1 = a(a 1,a ); b = h(a 1,a ); a 1 = b 1 ; a = b ;. The Matrix Setting Let A 1,A be positive definite (positive for short) matrices and let us define the harmonic, arithmetic, and geometric means as follows: ( A 1 1 + A 1 ) 1 H (A 1,A )=, ( ) A1 + A A(A 1,A )=, ( ) 1/ G(A 1,A )=A 1/ 1 A 1/ 1 A A 1/ 1/ ( 1 A 1 = A 1 A 1 1 A 1/ ). In [4] it was proved that the algorithm proposed in Section.1, replacing the positive scalars by matrices, converges to the matrix geometric mean. 3 Beyond Two Variables In fact, there are quite some possibilities in generalizing the deduction of the geometric mean based on only using the arithmetic and harmonic mean to more than two variables. In this section, some straightforward, as well as more advanced iterative algorithms will be proposed. 3.1 Circular Mean Consider a sequence of matrices, whose geometric mean we wish to compute. A simple, but effective and convergent procedure, as discussed in [5], is the following.

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations 787 0.304 0.30 0.3 Karcher Circular Results 0.98 0.96 0.94 0.9 0.9 0.88 0.6 0.6 0.64 0.66 0.68 0.7 0.7 0.74 Fig. 1. Simplified representation of the iterations of the Circular mean (Section 3.1) and the Karcher mean. The results of both methods are clearly different. Algorithm. Let A 1,...,A n be n positive matrices while not converged For all i set B i = G(A i,a (i mod n)+1 ); For all i set A i = B i. This procedure was first proposed in [6], and a conjecture was made that this mean was possibly converging to the Karcher mean. However, this was afterwards proven to be false, as shown in Fig. 1. Though this procedure neatly converges to an approximate mean, its convergence is rather slow. This is explained in [7], where it is proven that an iteration like this one on the corners of a random polygon converges to an elliptic setting. Obviously, once an ellips is reached, size reduction of the ellips, and convergence, becomes rather slow. Let us graphically illustrate by an example that this behavior also occurs when considering Algorithm. The convergence towards this elliptic behavior is displayed in Fig. using the consecutive eigenvalues in the iterations of the Circular mean of ten matrices. Figure (b) clearly shows this slow behavior with iterations 00 and 01. The spiraling convergence of the Circular mean in Fig. 1 is another indication of this phenomenon.

788 B. Jeuris and R. Vandebril Second eigenvalue 1. 1.1 1 Iteration Iteration 0 Second eigenvalue 1.0 1.0 1.0 Iteration 00 Iteration 01 Iteration 50 0.9 0.11 0.1 0.13 0.14 First eigenvalue (a) Iteration and 0. 0.14 0.14 0.14 0.1 First eigenvalue (b) Iteration 00, 01, and 50. Fig.. The evolution of the eigenvalues of ten matrices over the iterations of the Circular mean (Section 3.1) A simple, though effective remedy, is to randomly permute the matrices. Algorithm 3. Let A 1,...,A n be n positive matrices while not converged For all i set B i = G(A i,a (i mod n)+1 ); For all i set A p(i) = B i,withp a random permutation of [1,...,n]. Of course, this algorithm will never give the exact same result because of the random permutations. However, as shown in Fig. 3, the results can be found in the vicinity of the fixed-order result (the result obtained from Algorithm, without randomizing). Another interesting observation is that all the (randomized) results appear in a plane (Fig. 3(b)), as mentioned for fixed order in [7]. 3. Crude Midpoint Guess Instead of continuing the iteration process with n matrices, one can reduce them right from the start to two. Algorithm 4. Let A 1,...,A n be n positive matrices B 1 = A(A 1,...,A n ); B = H (A 1,...,A n ); The solution is G(B 1,B ). The result of this somewhat simplistic algorithm will be a very rough approximation to the Karcher mean. However, it is very cheap to compute and can be seen to preserve some of the desired properties of geometric means, such as inversion invariance (the mean of the inverses is the inverse of the original mean).

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations 789 0.81 0.8 0.8 0.78 0.3 0.9 0.6 0.6 0.6 0.7 (a) The results of the random algorithm are close to the result of the fixed order algorithm. 0.79 0.3 0.3 0.9 0.7 0.7 0.6 0.6 (b) The results (both randomized and fixed order) lie within the plane mentioned in [7]. Fig. 3. Results of the fixed order Circular mean (Algorithm ) and 100 runs with the randomized Circular mean (Algorithm 3) for the same ten matrices.thetwo figures display the same results, but for a different angle of view.) 3.3 A Harmonic and Arithmetic Circular Iteration Since the original Algorithm 1 continuously combines the arithmetic and harmonic means, it would be interesting to generalize this towards more than two matrices. The next algorithm gives a possibility for such a generalization. We start by duplicating the set of matrices, one focusing on the harmonic and one on the arithmetic mean. These two sets are then iteratively combined in arithmetic and harmonic operations. Algorithm 5. Let A 1,...,A n be n positive matrices For all i set B i = A i and C i = A i ; while not converged For all i set B i = H (B i,c (i mod n)+1 ); For all i set C i = A(B i,c (i mod n)+1 ); For all i set C i = C i, B i = B i. When experimenting with this algorithm, we notice that the vertices of the two sets B i and C i converge fast towards each other. Afterwards, the nearly equal sets t to exhibit the elliptic behavior again, resulting in a slow total convergence. As before, we try to resolve this using a random permutation of the order of the matrices (see Algorithm 3). These results display the same behavior we noticed in Section 3.1, meaning that the results of the randomized algorithm all appear in the same plane and in the vicinity of the fixed order result. 3.4 Symmetric Function Means In [8], two different means are presented based on an electrical networks approach, which are referred to as the T-mean and P-mean. For both means, when

790 B. Jeuris and R. Vandebril there are n matrices, equally many positive functions T i,n or P i,n, i =1,...,n, are constructed and used to create a new n-tuple of matrices in each iteration. What makes these functions interesting in our context is that the functions T 1,n and P 1,n are the same and are equal to the arithmetic mean. The functions T n,n and P n,n, on the other hand, are both equal to the harmonic mean. Furthermore, each of the functions is constructed using operations reminiscent of the arithmetic and harmonic mean (serial and parallel sums). Below we show the algorithm for the T-mean. The algorithm for the P-mean is completely similar and for the definition of the functions T i,n and P i,n we refer to [8]. Algorithm 6. Let A 1,...,A n be n positive matrices while not converged For all i set B i = T i,n (A 1,...,A n ); For all i set A i = B i. 4 Numerical Experiments In this section we will present numerical experiments to give more numerical insight in the behavior of all these algorithms. Both the speed of the algorithms and the proximity of their results to the Karcher mean will be discussed. Afterwards, we test whether the proximity of the various algorithms to the Karcher mean makes them a good initial guess to start the manifold optimization. As a comparison, we also consider some standard initializers, namely the arithmetic, harmonic and cheap mean [9]. 4.1 Distance to the Karcher Mean and Computational Speed First, we compare the difference in performance between those algorithms which operate with a fixed order and their counterparts where the graph, which determines the sequence of the matrices, is randomized in each iteration. This concerns the fixed order version and randomized version of the algorithms in Sections 3.1 and 3.3. In terms of proximity to the Karcher mean, all four algorithms give nearly identical results. In Fig. 4(a), we display these results for the two versions of the algorithm in Section 3.3. For reasonable condition of the matrices, this proximity decreases only slightly as the condition of the matrices deteriorates. In terms of computational time, however, there is a clear difference. As can be seen in Fig. 4(b), the algorithm in Section 3.3 is faster when using a randomized graph than with a fixed order. A very similar acceleration could be seen for the Circular mean (Section 3.1). When comparing the algorithms of the two Sections, it is clear that the algorithm from Section 3.3 outperforms the Circular mean, since it has comparable proximity to the Karcher mean and requires less computational time.

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations 791 Relative distance to the Karcher mean 10 0 10 1 10 T crude HA fix HA rand 10 1 10 3 10 5 10 7 10 9 Order of magnitude of the condition number (a) Relative distance of some of the suggested means to the Karcher mean where the condition number of the matrices is varied. The figure displays only an average of such distances to avoid clutter. In the leg, T indicates the T-mean, crude the mean in Section 3., and HA the mean of Section 3.3, both for fixed order and a randomized graph. Computational time (s) 10 3 10 1 10 1 10 3 P Circ fix HA fix HA rand crude 4 6 8 10 Number of matrices (b) Computational time of some of the suggested means for a varying number of matrices. The figure displays only an average of such timings to avoid clutter. In the leg, P indicates the P- mean, Circ fix the mean in Section 3.1 with fixed order, HA the mean in Section 3.3, and crude the one in Section 3.. Fig. 4. Numerical properties of the various means As for the T- and P-mean, we observe a proximity to the Karcher mean which is slightly worse than for the previous methods. However, the functions T i,n and P i,n are recursively defined, which causes the required computational time for these means to increase rapidly as the number of matrices starts to grow significantly large. Finally, the crude guess from Section 3. behaves as its name suggests. It has a low computational time, but it also has a bad proximity to the Karcher mean. In Fig. 5, we display the location of the various means with respect to the original matrices for a simple example of three positive matrices. We added the geodesics, the lines of shortest distance in the Riemannian space, between these original matrices. It is interesting to show these since the non-euclidean geometry of the space of positive matrices is the cause of the computational challenge of the Karcher mean. 4. Initial Guesses for the Karcher Mean Algorithm After analyzing the various proposed means in terms of performance, we now test their appropriateness as an initial guess to the Conjugate Gradient (CG) algorithm for computing the Karcher mean. In Table 1, the number of iterations the CG algorithm requires until convergence are shown for good conditioned and bad conditioned matrices. These results are rounded averages of a number of repetitions of the test. In general, the required number of iterations is lower as the condition of the matrices becomes worse. However, this is an observation that is related to the dynamics of the Karcher mean, and not as much to the initial guesses, so we will not discuss this in these notes.

79 B. Jeuris and R. Vandebril Table 1. Number of iterations required for the Conjugate Gradient algorithm for the Karcher mean to converge, using the specified mean as an initial guess (averaged over some repetitions). HA refers to the mean in Section 3.3. condition number 1e1 1e8 Circular, fixed 15 14 Circular, random 15 15 Crude guess 18 17 HA, fixed 14 11 HA, random 16 13 T-mean 16 15 P-mean 18 17 Arithmetic mean 19 16 Harmonic mean 1 17 CHEAP mean 15 16 1 0.8 0.6 0.4 1 0.5 00.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 Fig. 5. Representation in 3D space of three positive matrices. The curved triangle represents the geodesics, the lines of shortest distance, between the three matrices, which are the vertices. As can be seen, the various means appear very close to the Karcher mean in this simple example of well-conditioned matrices.

Geometric Mean Algorithms Based on Harmonic and Arithmetic Iterations 793 After inspecting the table, we can see that in general both the CHEAP mean and the two fixed order means (Circular and HA) result in a low number of iterations. On the other hand, the arithmetic and harmonic mean, and the crude guess appear to give the worst results, and are therefore least suited as an initial guess to the Karcher mean. If we look at the bad conditioning separately, we also notice that the mean of Section 3.3 (HA) performs especially well, both with fixed order and randomized graph. 5 Conclusion There is no fixed standard on how to generalize the arithmetic-harmonic algorithm for the geometric mean towards more than two matrices. We have examined some existing approaches and newly developed algorithms, judging their performance based on computational speed and proximity to the Karcher mean. A remarkable observation is that when working with algorithms that dep on a particular order of the matrices, randomizing this order in every iteration provides a significant speed-up while maintaining the proximity to the Karcher mean. Condering proximity to the Karcher mean, computational speed, and appropriateness as an initial guess to the Karcher mean optimization problem, the algorithm in Section 3.3 appears to provide optimal results. Acknowledgements. The authors would like to thank Dario A. Bini for the interesting discussions during the Structured Matrix Days in Limoges, 01; and Bruno Iannazzo and Pierre-Antoine Absil for their useful comments. References 1. Ando, T., Li, C., Mathias, R.: Geometric means. Linear Algebra and its Applications 385, 305 334 (004). Borwein, J., Borwein, P.: Pi and the AGM: a study in the analytic number theory and computational complexity. Wiley-Interscience (1987) 3. Foster, D., Phillips, G.: The arithmetic-harmonic mean. Mathematics of Computation 4, 183 191 (1984) 4. Atteia, M., Raissouli, M.: Self dual operators on convex functionals; geometric mean and square root of convex functionals. Journal of Convex Analysis 8(1), 3 40 (001) 5. Palfia, M.: A multivariable extension of two-variable matrix means. SIAM Journal on Matrix Analysis and Applications 3(), 385 393 (011) 6. Paige, C.: A useful form of unitary matrix obtained from any sequence of unit - norm n-vectors. SIAM Journal on Matrix Analysis and Applications 31(), 565 583 (009) 7. Elmachtoub, A., Van Loan, C.: From random polygon to ellipse: an eigenanalysis. SIAM Review 5(1), 151 170 (010) 8. Anderson, W., Morley, T., Trapp, G., et al.: Symmetric function means of positive operators. Linear Algebra and its Applications 60, 19 143 (1984) 9. Bini, D., Iannazzo, B.: A note on computing matrix geometric means. Advances in Computational Mathematics 35, 175 19 (011)