Poster
|
Thu 9:00
|
Towards Efficient Post-training Quantization of Pre-trained Language Models
Haoli Bai · Lu Hou · Lifeng Shang · Xin Jiang · Irwin King · Michael R Lyu
|
|
Poster
|
|
Leveraging Inter-Layer Dependency for Post -Training Quantization
changbao wang · DanDan Zheng · Yuanliu Liu · Liang Li
|
|
Workshop
|
Fri 3:50
|
Post-Training Neural Network Compression With Variational Bayesian Quantization
Zipei Tan · Robert Bamler
|
|
Workshop
|
|
Post-Training Neural Network Compression With Variational Bayesian Quantization
Zipei Tan · Robert Bamler
|
|
Poster
|
Thu 9:00
|
Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning
Elias Frantar · Dan Alistarh
|
|
Workshop
|
Fri 8:30
|
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Song Han
|
|
Poster
|
Thu 14:00
|
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
Zhewei Yao · Reza Yazdani Aminabadi · Minjia Zhang · Xiaoxia Wu · Conglong Li · Yuxiong He
|
|
Poster
|
Thu 9:00
|
A Fast Post-Training Pruning Framework for Transformers
Woosuk Kwon · Sehoon Kim · Michael Mahoney · Joseph Hassoun · Kurt Keutzer · Amir Gholami
|
|
Poster
|
Tue 14:00
|
FP8 Quantization: The Power of the Exponent
Andrey Kuzmin · Mart van Baalen · Yuwei Ren · Markus Nagel · Jorn Peters · Tijmen Blankevoort
|
|