Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

meta/meta-llama-3-70b ignores max_tokens #305

Open
johny-b opened this issue May 24, 2024 · 0 comments
Open

meta/meta-llama-3-70b ignores max_tokens #305

johny-b opened this issue May 24, 2024 · 0 comments

Comments

@johny-b
Copy link

johny-b commented May 24, 2024

I'm pretty sure I'm sending max_tokens and:

  • I get much more tokens
  • I also don't see this max_tokens when looking at my prediction in the browser

When I use exactly the same code for e.g. meta/llama-2-70b this does not happen, i.e. I really get the requested number of tokens.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant