Make quantize_pt2 return an ExportedProgram instead of a GraphModule #10644
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Summary:
This will help differentiating the fp32 models from the quantized models, and prevent people from using the wrong APIs.
For fp32 cases, we have a
torch.nn.Module
, which we trace and then lower. For quantized cases, we trace, quantize, and lower.After this diff,
export_to_<edge, executorch>
will ONLY handle non-quantized cases, and importantly, the sequence ofquantize_pt2
and thenexport_to_<edge, executorch>
will not work anymore. Those cases should use the (existing)lower_ep_to_<edge, executorch>
instead.Differential Revision: D73722640