Relaxed Quantization for Discretized Neural Networks

Open Access
Authors
Publication date 21-02-2019
Book title ICLR 2019
Book subtitle International Conference on Learning Representations : New Orleans, Louisiana, United States, May 6-May 9, 2019
Event 7th International Conference on Learning Representations
Number of pages 15
Publisher OpenReview
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Neural network quantization has become an important research area due to its great impact on deployment of large models on resource constrained devices. In order to train networks that can be effectively discretized without loss of performance, we introduce a differentiable quantization procedure. Differentiability can be achieved by transforming continuous distributions over the weights and activations of the network to categorical distributions over the quantization grid. These are subsequently relaxed to continuous surrogates that can allow for efficient gradient-based optimization. We further show that stochastic rounding can be seen as a special case of the proposed approach and that under this formulation the quantization grid itself can also be optimized with gradient descent. We experimentally validate the performance of our method on MNIST, CIFAR 10 and Imagenet classification.
Document type Conference contribution
Note Poster presentations.
Language English
Published at https://openreview.net/forum?id=HkxjYoCqKX
Other links https://openreview.net/group?id=ICLR.cc/2019/Conference
Downloads
Permalink to this page
Back