-
-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[Bug]: KeyError: 'type'. when inferencing Llama 3.2 3B Instruct #8855
Comments
It looks like vLLM wasn't installed properly. How did you install it? |
Nvm, it's a bug on our end (fixed by #8894). Is your |
@WoosukKwon I think Colab uses TPU? In that case you may have a better idea |
Weird, I was sure I had responded here and closed the issue. Anyway:
Also, I was able to get running with this:
then:
|
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
Error:
Reproduction (T4 on Colab):
There seem to be quite a few different errors with different models, like
google/gemma-2-2b-it
which gives:or
microsoft/Phi-3-mini-128k-instruct
gives:Before submitting a new issue...
The text was updated successfully, but these errors were encountered: