Simple parameter-free self-attention approximation

07/22/2023
by   Yuwen Zhai, et al.
0

The hybrid model of self-attention and convolution is one of the methods to lighten ViT. The quadratic computational complexity of self-attention with respect to token length limits the efficiency of ViT on edge devices. We propose a self-attention approximation without training parameters, called SPSA, which captures global spatial features with linear complexity. To verify the effectiveness of SPSA combined with convolution, we conduct extensive experiments on image classification and object detection tasks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset