You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi team, I am opening this issue to request support for the Google Gemma 2 models.
Recently, Google released two models: google/gemma-2-27b and google/gemma-2-9b. For an initial trial, we attempted to use the existing Gemma path for these new models, but it didn't work as expected. Specifically, when I tried to quantize google/gemma-2-9b, the model just produce non-sense outputs.
Could someone please investigate and add support to gemma2?
Thank you very much!!!
The text was updated successfully, but these errors were encountered:
I made an initial attempt that did not work. main...gemma2. Unfortunately, I do not have enough time at the moment to do further research on how to support the new architecture.
The biggest change I see for quantizing the model is that it now has a pre-feedforward and post-feedforward layernorm. So there is some challenge in trying to correctly quantize with AWQ. Maybe @TechxGenus or someone else can help contribute
Hi team, I am opening this issue to request support for the Google Gemma 2 models.
Recently, Google released two models: google/gemma-2-27b and google/gemma-2-9b. For an initial trial, we attempted to use the existing Gemma path for these new models, but it didn't work as expected. Specifically, when I tried to quantize google/gemma-2-9b, the model just produce non-sense outputs.
Could someone please investigate and add support to gemma2?
Thank you very much!!!
The text was updated successfully, but these errors were encountered: