Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

separate between token-lengths and sub-token lengths #2990

Merged

Conversation

helpmefindaname
Copy link
Collaborator

This PR should speed up the transformer word embeddings by a tiny bit (and also saves a little gpu usage), as the output embeddings are now of the token-length instead of being expanded to the sub-token-length.

I haven't tested how big the speed impact actually is, this would be a further todo.

Besides this, this PR fixes the Onnx Export ( #2964 #2930 ) by enforcing the order of inputs to be right

@alanakbik
Copy link
Collaborator

Thanks @helpmefindaname - I'll merge this now and fix the test in the next PR.

@alanakbik alanakbik merged commit 4788db1 into flairNLP:master Nov 20, 2022
@helpmefindaname helpmefindaname deleted the faster_transformer_word_embeddings branch November 28, 2022 10:45
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants