-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model Compression] fix compression doc and update a notebook example #3955
Conversation
@@ -31,7 +31,7 @@ The specification of configuration can be found `here <./Tutorial.rst#specify-th | |||
Step2. Choose a pruner and compress the model |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I suggest that we can put the notebook into documentation to increase visibility.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
good suggestion, add this into the doc.
} | ||
], | ||
"source": [ | ||
"from nni.algorithms.compression.pytorch.quantization import DoReFaQuantizer\n", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Recommend using QAT_Quantizer
instead of DoReFaQuantizer
since quantization speedup only support QAT_Quantizer
right now.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
got it, change to QAT_Quantizer
No description provided.