A plain-English explanation of Quantization (Quantization) — what it means, why it matters, and how it is used in AI.
Also known as: Model quantization, weight quantization, INT8, INT4
Quantization is essential for running large models on limited hardware — enabling local deployment on laptops and consumer GPUs.
The best way to remember Quantization is to practice unscrambling it. AI Terminology Scrambler uses spaced repetition to help you learn and retain AI vocabulary in just a few minutes a day.
Practice Quantization now →