Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Model Quantization for PyTorch (Proposal) #18318

Copy link
Copy link
Closed
@jspisak

Description

@jspisak
Issue body actions

🚀 tl;dr

Attached is a proposal for graph mode quantization in pytorch (model_quantizer) that provides end to end post training quantization support for both mobile and server backends. Model quantization supports fp32 and int8 precisions as a starting point and will expand to support other precision types based on customer needs. Details can be found in the attached pdf doc:

Model Quantization for Pytorch.pdf

cc @soumith, @gchanan, @raghuramank100

Metadata

Metadata

Assignees

Labels

featureA request for a proper, new feature.A request for a proper, new feature.high priorityoncall: quantizationQuantization support in PyTorchQuantization support in PyTorchtriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate moduleThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions

    Morty Proxy This is a proxified and sanitized view of the page, visit original site.