Temporally Resolution Decrement: Utilizing the Shape Consistency for Higher Computational Efficiency

12/02/2021
by   Tianshu Xie, et al.
0

Image resolution that has close relations with accuracy and computational cost plays a pivotal role in network training. In this paper, we observe that the reduced image retains relatively complete shape semantics but loses extensive texture information. Inspired by the consistency of the shape semantics as well as the fragility of the texture information, we propose a novel training strategy named Temporally Resolution Decrement. Wherein, we randomly reduce the training images to a smaller resolution in the time domain. During the alternate training with the reduced images and the original images, the unstable texture information in the images results in a weaker correlation between the texture-related patterns and the correct label, naturally enforcing the model to rely more on shape properties that are robust and conform to the human decision rule. Surprisingly, our approach greatly improves the computational efficiency of convolutional neural networks. On ImageNet classification, using only 33 training image to 112×112 within 90 from 76.32 the training image to 112 x 112 within 50 78.18

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset