Int4WeightOnlyQATQuantizer¶
- class torchao.quantization.qat.Int4WeightOnlyQATQuantizer(groupsize: int = 256, inner_k_tiles: Optional[int] = 8, precision: dtype = torch.bfloat16, scales_precision: dtype = torch.bfloat16)[source]¶
Quantizer for performing QAT on a model, where linear layers have int4 fake quantized grouped per channel weights.