Closed
Description
🚀 tl;dr
Attached is a proposal for graph mode quantization in pytorch (model_quantizer) that provides end to end post training quantization support for both mobile and server backends. Model quantization supports fp32 and int8 precisions as a starting point and will expand to support other precision types based on customer needs. Details can be found in the attached pdf doc:
Metadata
Metadata
Assignees
Labels
A request for a proper, new feature.A request for a proper, new feature.Quantization support in PyTorchQuantization support in PyTorchThis issue has been looked at a team member, and triaged and prioritized into an appropriate moduleThis issue has been looked at a team member, and triaged and prioritized into an appropriate module