Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

don't resize embeddings to multiples of 32x by default #313

Merged
merged 1 commit into from
Jul 22, 2023

Conversation

winglian
Copy link
Collaborator

resizing the embeddings to larger than the actual number of tokens causes problems with llama.cpp downstream. let's not enable this by default for safety.

@NanoCode012
Copy link
Collaborator

YES! I had to modify downstream code for this 😂

@NanoCode012
Copy link
Collaborator

While on this topic, I was wondering whether to consider using the "smart resize" code floating around in fast chat and other repos.

@winglian winglian merged commit 3ffb018 into main Jul 22, 2023
6 checks passed
@winglian winglian deleted the tokenizer-llama2-embeddings branch July 22, 2023 08:10
mkeoliya pushed a commit to mkeoliya/axolotl that referenced this pull request Dec 15, 2023
…/tokenizer-llama2-embeddings

don't resize embeddings to multiples of 32x by default
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants