We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
perhaps not for batch inference
The text was updated successfully, but these errors were encountered:
For the llama runs we do not use mattext tokenizers though.
Sorry, something went wrong.
Ah I see now. There was this issue of Llama tokenizer not including pad token. So we set tokenizer.pad_token = tokenizer.eos_token ref.
tokenizer.pad_token = tokenizer.eos_token
We also tried adding a token, this then resized the vocab and creates a set of problems
This is not an issue for the serial interface that is in the code at the moment. For batched inference this might be important in the future
No branches or pull requests
perhaps not for batch inference
The text was updated successfully, but these errors were encountered: