Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

QAT Support for new Framework with QuantizationModifier Testing #1763

Merged
merged 15 commits into from
Oct 16, 2023

Conversation

Satrat
Copy link
Contributor

@Satrat Satrat commented Oct 12, 2023

This PR implements QAT in the new modifier framework, adds a lifecycle testing harness for testing modifiers, and fills out unit tests for the QuantizationModifier in both one-shot and training runs. It also fixes a small bug in recipe arg collection.

Example

To run an end to end test script with quantization and pruning:

cd integrations/torchvision/modifiers_refactor_example
python e2e_test.py

Testing

Wrote unit tests for QuantizationModifier and QuantizationModifierPytorch

@Satrat Satrat merged commit f889bb8 into main Oct 16, 2023
11 checks passed
@Satrat Satrat deleted the quantization_modifier branch October 16, 2023 21:04
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants