Skip to content
/ gptaq Public
forked from IST-DASLab/gptq

Code for paper "GPTAQ - activation quantization with cross layer equalisation for GPTQ quantization framework for neural networks".

License

Notifications You must be signed in to change notification settings

Mikkicon/gptaq

 
 

Repository files navigation

GPTAQ

Neural Network Quantization Framework based on GPTQ With addition of:

  • Activations quantization (RTN + weight reoptimization + Token-wise)
  • Hessian Eigenvalues in sensitivity params
  • Cross-layer equalization


Algorithm

GPTAQ Algorithm

Experiments

Experiments

About

Code for paper "GPTAQ - activation quantization with cross layer equalisation for GPTQ quantization framework for neural networks".

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 60.0%
  • Jupyter Notebook 37.5%
  • Cuda 2.2%
  • C++ 0.3%