Robust Differentiable SVD

04/08/2021
by   Wei Wang, et al.
0

Eigendecomposition of symmetric matrices is at the heart of many computer vision algorithms. However, the derivatives of the eigenvectors tend to be numerically unstable, whether using the SVD to compute them analytically or using the Power Iteration (PI) method to approximate them. This instability arises in the presence of eigenvalues that are close to each other. This makes integrating eigendecomposition into deep networks difficult and often results in poor convergence, particularly when dealing with large matrices. While this can be mitigated by partitioning the data into small arbitrary groups, doing so has no theoretical basis and makes it impossible to exploit the full power of eigendecomposition. In previous work, we mitigated this using SVD during the forward pass and PI to compute the gradients during the backward pass. However, the iterative deflation procedure required to compute multiple eigenvectors using PI tends to accumulate errors and yield inaccurate gradients. Here, we show that the Taylor expansion of the SVD gradient is theoretically equivalent to the gradient obtained using PI without relying in practice on an iterative process and thus yields more accurate gradients. We demonstrate the benefits of this increased accuracy for image classification and style transfer.

READ FULL TEXT

page 3

page 13

page 15

research
06/21/2019

Backpropagation-Friendly Eigendecomposition

Eigendecomposition (ED) is widely used in deep networks. However, the ba...
research
05/06/2021

Why Approximate Matrix Square Root Outperforms Accurate SVD in Global Covariance Pooling?

Global covariance pooling (GCP) aims at exploiting the second-order stat...
research
01/29/2022

Fast Differentiable Matrix Square Root and Inverse Square Root

Computing the matrix square root and its inverse in a differentiable man...
research
01/21/2022

Fast Differentiable Matrix Square Root

Computing the matrix square root or its inverse in a differentiable mann...
research
06/25/2020

An Analysis of SVD for Deep Rotation Estimation

Symmetric orthogonalization via SVD, and closely related procedures, are...
research
11/24/2021

Softmax Gradient Tampering: Decoupling the Backward Pass for Improved Fitting

We introduce Softmax Gradient Tampering, a technique for modifying the g...
research
07/09/2022

Batch-efficient EigenDecomposition for Small and Medium Matrices

EigenDecomposition (ED) is at the heart of many computer vision algorith...

Please sign up or login with your details

Forgot password? Click here to reset