Approximate Joint Diagonalization and Geometric Mean of Symmetric Positive Definite Matrices

by   Marco Congedo, et al.

We explore the connection between two problems that have arisen independently in the signal processing and related fields: the estimation of the geometric mean of a set of symmetric positive definite (SPD) matrices and their approximate joint diagonalization (AJD). Today there is a considerable interest in estimating the geometric mean of a SPD matrix set in the manifold of SPD matrices endowed with the Fisher information metric. The resulting mean has several important invariance properties and has proven very useful in diverse engineering applications such as biomedical and image data processing. While for two SPD matrices the mean has an algebraic closed form solution, for a set of more than two SPD matrices it can only be estimated by iterative algorithms. However, none of the existing iterative algorithms feature at the same time fast convergence, low computational complexity per iteration and guarantee of convergence. For this reason, recently other definitions of geometric mean based on symmetric divergence measures, such as the Bhattacharyya divergence, have been considered. The resulting means, although possibly useful in practice, do not satisfy all desirable invariance properties. In this paper we consider geometric means of co-variance matrices estimated on high-dimensional time-series, assuming that the data is generated according to an instantaneous mixing model, which is very common in signal processing. We show that in these circumstances we can approximate the Fisher information geometric mean by employing an efficient AJD algorithm. Our approximation is in general much closer to the Fisher information geometric mean as compared to its competitors and verifies many invariance properties. Furthermore, convergence is guaranteed, the computational complexity is low and the convergence rate is quadratic. The accuracy of this new geometric mean approximation is demonstrated by means of simulations.


page 5

page 11

page 13

page 14

page 15

page 17

page 18

page 20


Information geometry and asymptotic geodesics on the space of normal distributions

The family N of n-variate normal distributions is parameterized by the c...

Fast (1+ε)-approximation of the Löwner extremal matrices of high-dimensional symmetric matrices

Matrix data sets are common nowadays like in biomedical imaging where th...

Krylov Iterative Methods for the Geometric Mean of Two Matrices Times a Vector

In this work, we are presenting an efficient way to compute the geometri...

Differential geometry with extreme eigenvalues in the positive semidefinite cone

Differential geometric approaches to the analysis and processing of data...

Averaging symmetric positive-definite matrices on the space of eigen-decompositions

We study extensions of Fréchet means for random objects in the space Sym...

Joint Approximate Diagonalization under Orthogonality Constraints

Joint diagonalization of a set of positive (semi)-definite matrices has ...

Towards a median signal detector through the total Bregman divergence and its robust analysis

A novel family of geometric signal detectors are proposed through median...

Please sign up or login with your details

Forgot password? Click here to reset