Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

mask_structure preservation test #2284

Merged
merged 6 commits into from
May 17, 2024

Conversation

rahul-tuli
Copy link
Member

This pull request introduces an integration check to ensure the preservation of mask structure across consecutive runs. The process includes:

  • Initial pruning of the model using a mask structure: "N:M".
  • Further pruning and quantization of the saved model.
  • Verification that the original "N:M" sparsity structure is maintained, with additional sparsity applied on top.

Base automatically changed from preserve-sparsity-sparsegpt to gptq-ux-config-groups May 17, 2024 16:15
@rahul-tuli rahul-tuli merged commit 440661b into gptq-ux-config-groups May 17, 2024
@rahul-tuli rahul-tuli deleted the preserve-mask-structure-test branch May 17, 2024 16:16
rahul-tuli added a commit that referenced this pull request May 20, 2024
* test

* Preserve weight sparsity if greater than threshold

* Add argument to preserve sparsity mask in SPARSEGPT

* fix case when mask is none

* Add test to check mask_structure
- initial mask structure should be preserved
b/w consecutive runs; added test to check this

* Update tensor_follows_mask_structure to check for atleast n zeros

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>
rahul-tuli added a commit that referenced this pull request May 20, 2024
* Update OBCQ

* Extract GPTQ Modifier

* Update test recipes

* Add config_groups support to GPTQModifier

* mask_structure preservation test (#2284)

* test

* Preserve weight sparsity if greater than threshold

* Add argument to preserve sparsity mask in SPARSEGPT

* fix case when mask is none

* Add test to check mask_structure
- initial mask structure should be preserved
b/w consecutive runs; added test to check this

* Update tensor_follows_mask_structure to check for atleast n zeros

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>

* PR comments

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>
bfineran pushed a commit that referenced this pull request May 22, 2024
* Split WandaPruningModifier and SparseGPTModifier
Make sparsegpt not inherit from wanda modifier
Decouple SparseGPTModifierPyTorch from WandaPruningModifier
Fix docstrings

* Split SparseGPT and GPTQ modifiers (#2272)

* Update OBCQ

* Extract GPTQ Modifier

* [GPTQ Modifier UX] Update tests to use GPTQModifier for obcq style quantization (#2294)

* Update OBCQ

* Extract GPTQ Modifier

* Update test recipes

* GPTQ UX config groups support (#2273)

* Update OBCQ

* Extract GPTQ Modifier

* Update test recipes

* Add config_groups support to GPTQModifier

* mask_structure preservation test (#2284)

* test

* Preserve weight sparsity if greater than threshold

* Add argument to preserve sparsity mask in SPARSEGPT

* fix case when mask is none

* Add test to check mask_structure
- initial mask structure should be preserved
b/w consecutive runs; added test to check this

* Update tensor_follows_mask_structure to check for atleast n zeros

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>

* PR comments

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>

* Fix default case

* Update test to use new vLLMQuantizationModifier

* Style

---------

Co-authored-by: Sara Adkins <sara@neuralmagic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants