We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
CreateChatCompletionRequest.max_tokens is a Option<u16> as of 0.23.1.
CreateChatCompletionRequest.max_tokens
Option<u16>
The newer models such as gpt-4o has a context window of 128,000 tokens. This context window limit is the sum of input and output tokens.
I believe the max_tokens field should be Option<u32> to allow numbers as high as 128,000.
max_tokens
Option<u32>
The text was updated successfully, but these errors were encountered:
Thank you for reporting the bug, a PR is most welcome!
Sorry, something went wrong.
Successfully merging a pull request may close this issue.
CreateChatCompletionRequest.max_tokens
is aOption<u16>
as of 0.23.1.The newer models such as gpt-4o has a context window of 128,000 tokens. This context window limit is the sum of input and output tokens.
I believe the
max_tokens
field should beOption<u32>
to allow numbers as high as 128,000.The text was updated successfully, but these errors were encountered: