ZiCo: Zero-shot NAS via Inverse Coefficient of Variation on Gradients

01/26/2023
by   Guihong Li, et al.
0

Neural Architecture Search (NAS) is widely used to automatically design the neural network with the best performance among a large number of candidate architectures. To reduce the search time, zero-shot NAS aims at designing training-free proxies that can predict the test performance of a given architecture. However, as shown recently, none of the zero-shot proxies proposed to date can actually work consistently better than a naive proxy, namely, the number of network parameters (#Params). To improve this state of affairs, as the main theoretical contribution, we first reveal how some specific gradient properties across different samples impact the convergence rate and generalization capacity of neural networks. Based on this theoretical analysis, we propose a new zero-shot proxy, ZiCo, the first proxy that works consistently better than #Params. We demonstrate that ZiCo works better than State-Of-The-Art (SOTA) proxies on several popular NAS-Benchmarks (NASBench101, NATSBench-SSS/TSS, TransNASBench-101) for multiple applications (e.g., image classification/reconstruction and pixel-level prediction). Finally, we demonstrate that the optimal architectures found via ZiCo are as competitive as the ones found by one-shot and multi-shot NAS methods, but with much less search time. For example, ZiCo-based NAS can find optimal architectures with 78.1 and 1000M FLOPs on ImageNet within 0.4 GPU days.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/05/2023

Zero-Shot Neural Architecture Search: Challenges, Solutions, and Opportunities

Recently, zero-shot (or training-free) Neural Architecture Search (NAS) ...
research
06/08/2023

Generalizable Lightweight Proxy for Robust NAS against Diverse Perturbations

Recent neural architecture search (NAS) frameworks have been successful ...
research
05/25/2021

The Nonlinearity Coefficient - A Practical Guide to Neural Architecture Design

In essence, a neural network is an arbitrary differentiable, parametrize...
research
05/21/2020

Powering One-shot Topological NAS with Stabilized Share-parameter Proxy

One-shot NAS method has attracted much interest from the research commun...
research
09/15/2022

EZNAS: Evolving Zero Cost Proxies For Neural Architecture Scoring

Neural Architecture Search (NAS) has significantly improved productivity...
research
12/11/2019

A Variational-Sequential Graph Autoencoder for Neural Architecture Performance Prediction

In computer vision research, the process of automating architecture engi...
research
12/22/2020

Multi-shot NAS for Discovering Adversarially Robust Convolutional Neural Architectures at Targeted Capacities

Convolutional neural networks (CNNs) are vulnerable to adversarial examp...

Please sign up or login with your details

Forgot password? Click here to reset