Timezone: »
Neural network quantization methods often involve simulating the quantization process during training, making the trained model highly dependent on the target bit-width and precise way quantization is performed. Robust quantization offers an alternative approach with improved tolerance to different classes of data-types and quantization policies. It opens up new exciting applications where the quantization process is not static and can vary to meet different circumstances and implementations. To address this issue, we propose a method that provides intrinsic robustness to the model against a broad range of quantization processes. Our method is motivated by theoretical arguments and enables us to store a single generic model capable of operating at various bit-widths and quantization policies. We validate our method's effectiveness on different ImageNet Models. A reference implementation accompanies the paper.
Author Information
moran shkolnik (Technion- Israel Institute of Technology)
Brian Chmiel (Intel)
Ron Banner (Intel - Artificial Intelligence Products Group (AIPG))
Gil Shomron (Technion - Israel Institute of Technology)
Yury Nahshan (Intel - Artificial Intelligence Products Group (AIPG))
Alex Bronstein (Technion)
Uri Weiser (Technion - Israel Institute of Technology)
More from the Same Authors
-
2021 Poster: Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks »
Itay Hubara · Brian Chmiel · Moshe Island · Ron Banner · Joseph Naor · Daniel Soudry -
2021 Poster: Post-Training Sparsity-Aware Quantization »
Gil Shomron · Freddy Gabbay · Samer Kurzum · Uri Weiser -
2019 Poster: Post training 4-bit quantization of convolutional networks for rapid-deployment »
Ron Banner · Yury Nahshan · Daniel Soudry -
2018 Poster: Norm matters: efficient and accurate normalization schemes in deep networks »
Elad Hoffer · Ron Banner · Itay Golan · Daniel Soudry -
2018 Spotlight: Norm matters: efficient and accurate normalization schemes in deep networks »
Elad Hoffer · Ron Banner · Itay Golan · Daniel Soudry -
2018 Poster: Delta-encoder: an effective sample synthesis method for few-shot object recognition »
Eli Schwartz · Leonid Karlinsky · Joseph Shtok · Sivan Harary · Mattias Marder · Abhishek Kumar · Rogerio Feris · Raja Giryes · Alex Bronstein -
2018 Spotlight: Delta-encoder: an effective sample synthesis method for few-shot object recognition »
Eli Schwartz · Leonid Karlinsky · Joseph Shtok · Sivan Harary · Mattias Marder · Abhishek Kumar · Rogerio Feris · Raja Giryes · Alex Bronstein -
2018 Poster: Scalable methods for 8-bit training of neural networks »
Ron Banner · Itay Hubara · Elad Hoffer · Daniel Soudry