Semantic-Aware Generation for Self-Supervised Visual Representation Learning

11/25/2021
by   Yunjie Tian, et al.
1

In this paper, we propose a self-supervised visual representation learning approach which involves both generative and discriminative proxies, where we focus on the former part by requiring the target network to recover the original image based on the mid-level features. Different from prior work that mostly focuses on pixel-level similarity between the original and generated images, we advocate for Semantic-aware Generation (SaGe) to facilitate richer semantics rather than details to be preserved in the generated image. The core idea of implementing SaGe is to use an evaluator, a deep network that is pre-trained without labels, for extracting semantic-aware features. SaGe complements the target network with view-specific features and thus alleviates the semantic degradation brought by intensive data augmentations. We execute SaGe on ImageNet-1K and evaluate the pre-trained models on five downstream tasks including nearest neighbor test, linear classification, and fine-scaled image recognition, demonstrating its ability to learn stronger visual representations.

READ FULL TEXT

page 4

page 8

page 12

research
11/17/2020

Can Semantic Labels Assist Self-Supervised Visual Representation Learning?

Recently, contrastive learning has largely advanced the progress of unsu...
research
12/30/2022

Improving Visual Representation Learning through Perceptual Understanding

We present an extension to masked autoencoders (MAE) which improves on t...
research
02/27/2020

Learning Representations by Predicting Bags of Visual Words

Self-supervised representation learning targets to learn convnet-based i...
research
03/29/2022

Self-Supervised Image Representation Learning with Geometric Set Consistency

We propose a method for self-supervised image representation learning un...
research
04/20/2023

Contrastive Tuning: A Little Help to Make Masked Autoencoders Forget

Masked Image Modeling (MIM) methods, like Masked Autoencoders (MAE), eff...
research
12/06/2021

General Facial Representation Learning in a Visual-Linguistic Manner

How to learn a universal facial representation that boosts all face anal...
research
03/22/2016

Learning Representations for Automatic Colorization

We develop a fully automatic image colorization system. Our approach lev...

Please sign up or login with your details

Forgot password? Click here to reset