when a trained model must be smaller or faster at inference time
Post-Training Quantization AI Prompt
This prompt applies post-training quantization and, if needed, quantization-aware training to reduce model size and improve inference speed. It includes validation steps so compression gains can be weighed against accuracy loss.
Apply post-training quantization (PTQ) to reduce model size and inference latency.
1. INT8 static quantization (PyTorch):
- Prepare model: torch.quantization.prepare with a QConfig
- Calibrate on a representative dataset (100–1000 samples): run forward passes to collect activation statistics
- Convert: torch.quantization.convert to replace float ops with int8 ops
- Save and measure: model size before vs after, inference latency before vs after
2. INT8 dynamic quantization:
- torch.quantization.quantize_dynamic for models where activation ranges vary greatly
- Suitable for: LSTMs, linear layers in NLP models
- No calibration step needed
3. Quantization-aware training (QAT) if accuracy drops > 1%:
- Insert fake quantization nodes during training
- Fine-tune for {{qat_epochs}} epochs at a lower learning rate
- Convert to fully quantized model after training
4. Accuracy validation:
- Evaluate quantized model on the full validation set
- Acceptable accuracy drop: < 1% for most production use cases
- If accuracy drops significantly: try QAT, or quantize only the later layers
5. ONNX + ONNX Runtime INT8:
- Export to ONNX, then apply ONNXRuntime quantization
- ort.quantization.quantize_dynamic or quantize_static
- Often faster than PyTorch native quantization on CPU
Return: PTQ implementation, QAT setup, accuracy comparison table, and latency/size improvement metrics.When to use this prompt
when evaluating static or dynamic INT8 quantization in PyTorch or ONNX Runtime
when calibration and accuracy validation are required
when QAT is a fallback if PTQ reduces accuracy too much
What the AI should return
PTQ implementation, optional QAT setup, and a comparison of accuracy, model size, and inference latency before and after quantization.
How to use this prompt
Open your data context
Load your dataset, notebook, or working environment so the AI can operate on the actual project context.
Copy the prompt text
Use the copy button above and paste the prompt into the AI assistant or prompt input area.
Review the output critically
Check whether the result matches your data, assumptions, and desired format before moving on.
Chain into the next prompt
Once you have the first result, continue deeper with related prompts in Model Compression.
Frequently asked questions
What does the Post-Training Quantization prompt do?+
It gives you a structured model compression starting point for ml engineer work and helps you move faster without starting from a blank page.
Who is this prompt for?+
It is designed for ml engineer workflows and marked as beginner, so it works well as a guided starting point for that level of experience.
What type of prompt is this?+
Post-Training Quantization is a single prompt. You can copy it as-is, adapt it, or use it as one step inside a larger workflow.
Can I use this outside MLJAR Studio?+
Yes. The prompt text works in other AI tools too, but MLJAR Studio is the best fit when you want local execution, visible Python code, and reusable notebooks.
What should I open next?+
Natural next steps from here are Compression Pipeline Chain, Knowledge Distillation, ONNX Export and Validation.