Skip to content

Make quantize_pt2 return an ExportedProgram instead of a GraphModule #10644

New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Merged
merged 1 commit into from
May 3, 2025

Conversation

mcremon-meta
Copy link
Contributor

Summary:
This will help differentiating the fp32 models from the quantized models, and prevent people from using the wrong APIs.
For fp32 cases, we have a torch.nn.Module, which we trace and then lower. For quantized cases, we trace, quantize, and lower.

After this diff, export_to_<edge, executorch> will ONLY handle non-quantized cases, and importantly, the sequence of quantize_pt2 and then export_to_<edge, executorch> will not work anymore. Those cases should use the (existing) lower_ep_to_<edge, executorch> instead.

Differential Revision: D73722640

@mcremon-meta mcremon-meta requested a review from tarun292 as a code owner May 2, 2025 00:51
Copy link

pytorch-bot bot commented May 2, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/10644

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

✅ No Failures

As of commit f3ed763 with merge base 4cc6532 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label May 2, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D73722640

…10644)

Summary:

This will help differentiating the fp32 models from the quantized models, and prevent people from using the wrong APIs.
For fp32 cases, we have a `torch.nn.Module`, which we trace and then lower. For quantized cases, we trace, quantize, and lower.

After this diff, `export_to_<edge, executorch>` will ONLY handle non-quantized cases, and importantly, the sequence of `quantize_pt2` and then `export_to_<edge, executorch>` will not work anymore. Those cases should use the (existing) `lower_ep_to_<edge, executorch>` instead.

Note that in subsequent diffs, both `quantize_pt2` and `lower_ep_to<edge, executorch` should be inner calls, not exposed to users

Reviewed By: Vysarat, zonglinpeng

Differential Revision: D73722640
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D73722640

@facebook-github-bot facebook-github-bot merged commit 94f7b10 into main May 3, 2025
85 of 88 checks passed
@facebook-github-bot facebook-github-bot deleted the export-D73722640 branch May 3, 2025 00:11
jhelsby pushed a commit to jhelsby/executorch that referenced this pull request May 9, 2025
Differential Revision: D73722640

Pull Request resolved: pytorch#10644
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported topic: not user facing
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants