Shortcuts

FromIntXQuantizationAwareTrainingConfig

class torchao.quantization.qat.FromIntXQuantizationAwareTrainingConfig[source]

Object that knows how to convert a model with fake quantized modules, such as FakeQuantizedLinear() and FakeQuantizedEmbedding(), back to model with the original, corresponding modules without fake quantization. This should be used with quantize_().

Example usage:

from torchao.quantization import quantize_
quantize_(
    model_with_fake_quantized_linears,
    FromIntXQuantizationAwareTrainingConfig(),
)

Docs

Access comprehensive developer documentation for PyTorch

View Docs

Tutorials

Get in-depth tutorials for beginners and advanced developers

View Tutorials

Resources

Find development resources and get your questions answered

View Resources