Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[GPTQ Modifier UX] Update tests to use GPTQModifier for obcq style quantization #2294

Merged
merged 3 commits into from
May 20, 2024

Conversation

rahul-tuli
Copy link
Member

This PR updates test recipes and readme to use new GPTQ modifier for quantization

Base automatically changed from create-gptq-modifier to quant-modifier-ux May 20, 2024 18:56
@rahul-tuli rahul-tuli merged commit c695567 into quant-modifier-ux May 20, 2024
@rahul-tuli rahul-tuli deleted the update-tests branch May 20, 2024 18:56
bfineran pushed a commit that referenced this pull request May 22, 2024
* Split WandaPruningModifier and SparseGPTModifier
Make sparsegpt not inherit from wanda modifier
Decouple SparseGPTModifierPyTorch from WandaPruningModifier
Fix docstrings

* Split SparseGPT and GPTQ modifiers (#2272)

* Update OBCQ

* Extract GPTQ Modifier

* [GPTQ Modifier UX] Update tests to use GPTQModifier for obcq style quantization (#2294)

* Update OBCQ

* Extract GPTQ Modifier

* Update test recipes

* GPTQ UX config groups support (#2273)

* Update OBCQ

* Extract GPTQ Modifier

* Update test recipes

* Add config_groups support to GPTQModifier

* mask_structure preservation test (#2284)

* test

* Preserve weight sparsity if greater than threshold

* Add argument to preserve sparsity mask in SPARSEGPT

* fix case when mask is none

* Add test to check mask_structure
- initial mask structure should be preserved
b/w consecutive runs; added test to check this

* Update tensor_follows_mask_structure to check for atleast n zeros

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>

* PR comments

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>

* Fix default case

* Update test to use new vLLMQuantizationModifier

* Style

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants