Neural Network Compression for Noisy Storage Devices

02/15/2021
by   Berivan Isik, et al.
13

Compression and efficient storage of neural network (NN) parameters is critical for applications that run on resource-constrained devices. Although NN model compression has made significant progress, there has been considerably less investigation in the actual physical storage of NN parameters. Conventionally, model compression and physical storage are decoupled, as digital storage media with error correcting codes (ECCs) provide robust error-free storage. This decoupled approach is inefficient, as it forces the storage to treat each bit of the compressed model equally, and to dedicate the same amount of resources to each bit. We propose a radically different approach that: (i) employs analog memories to maximize the capacity of each memory cell, and (ii) jointly optimizes model compression and physical storage to maximize memory utility. We investigate the challenges of analog storage by studying model storage on phase change memory (PCM) arrays and develop a variety of robust coding strategies for NN model storage. We demonstrate the efficacy of our approach on MNIST, CIFAR-10 and ImageNet datasets for both existing and novel compression methods. Compared to conventional error-free digital storage, our method has the potential to reduce the memory size by one order of magnitude, without significantly compromising the stored model's accuracy.

READ FULL TEXT
research
11/20/2020

Noisy Neural Network Compression for Analog Storage Devices

Efficient compression and storage of neural network (NN) parameters is c...
research
06/24/2020

Design And Develop Network Storage Virtualization By Using GNS3

Virtualization is an emerging and optimistic prospect in the IT industry...
research
05/10/2023

Compressing neural network by tensor network with exponentially fewer variational parameters

Neural network (NN) designed for challenging machine learning tasks is i...
research
05/22/2021

Denoising Noisy Neural Networks: A Bayesian Approach with Compensation

Noisy neural networks (NoisyNNs) refer to the inference and training of ...
research
09/20/2023

Containing Analog Data Deluge at Edge through Frequency-Domain Compression in Collaborative Compute-in-Memory Networks

Edge computing is a promising solution for handling high-dimensional, mu...
research
11/28/2019

Neural Network-Inspired Analog-to-Digital Conversion to Achieve Super-Resolution with Low-Precision RRAM Devices

Recent works propose neural network- (NN-) inspired analog-to-digital co...
research
07/12/2023

DeepMapping: The Case for Learned Data Mapping for Compression and Efficient Query Processing

Storing tabular data in a way that balances storage and query efficienci...

Please sign up or login with your details

Forgot password? Click here to reset