Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Expose stop_sequence at command line #196

Open
lewtun opened this issue May 27, 2024 · 1 comment
Open

Expose stop_sequence at command line #196

lewtun opened this issue May 27, 2024 · 1 comment
Assignees

Comments

@lewtun
Copy link
Member

lewtun commented May 27, 2024

Models like llama-3 use a chat template where the expected stop sequence is <|eot_id|> instead of the common EOS token used in other models. This means that generative benchmarks like ifeval continue generating past the EOS token and give incorrect results.

One way to handle this would be to either include this special token as a default in the generative benchmarks or alternatively expose --stop_sequence as an argument in the main script that users can control.

@clefourrier clefourrier self-assigned this Jul 17, 2024
@clefourrier
Copy link
Member

clefourrier commented Jul 17, 2024

Hi @lewtun , just to clarify, is <|eot_id|> a different stop token for the chat template specifically?
And is the model using <|end_of_text|> for the rest?

Other question, do you think fixing #16 would be enough to solve this too?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants