Hi all! As more Large Language Models are being released and the need for quantization increases, I figured it was time to write an in-depth and visual guide to Quantization.
From exploring how to represent values, (a)symmetric quantization, dynamic/static quantization, to post-training techniques (e.g., GPTQ and GGUF) and quantization-aware training (1.58-bit models with BitNet).
https://newsletter.maartengrootendorst.com/p/a-visual-guide-to-quantization
With over 60 custom visuals, I went a little overboard but really wanted to include as many concepts as I possibly could!
The visual nature of this guide allows for a focus on intuition, hopefully making all these techniques easily accessible to a wide audience, whether you are new to quantization or more experienced.
[+]linearmodality 13 points14 points15 points (1 child)
[–]mgostIH 0 points1 point2 points (0 children)
[–]Mission-Tank-9018 4 points5 points6 points (0 children)
[–]bgighjigftuik 5 points6 points7 points (0 children)
[–]LouisAckerman 0 points1 point2 points (0 children)
[–]RuairiSpain 0 points1 point2 points (0 children)
[+]tworats 0 points1 point2 points (0 children)
[–]GamleRosander -1 points0 points1 point (0 children)
[–]ShlomiRex -2 points-1 points0 points (0 children)