Skip to content

Commit 360ddbd

Browse files
authored
[Misc] Update Pixtral example (#8431)
1 parent a480939 commit 360ddbd

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

examples/offline_inference_pixtral.py

+3-2
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
# - Server:
1212
#
1313
# ```bash
14-
# vllm serve mistralai/Pixtral-12B-2409 --tokenizer_mode mistral --limit_mm_per_prompt 'image=4' --max_num_batched_tokens 16384
14+
# vllm serve mistralai/Pixtral-12B-2409 --tokenizer-mode mistral --limit-mm-per-prompt 'image=4' --max-model-len 16384
1515
# ```
1616
#
1717
# - Client:
@@ -45,6 +45,7 @@ def run_simple_demo():
4545
model_name = "mistralai/Pixtral-12B-2409"
4646
sampling_params = SamplingParams(max_tokens=8192)
4747

48+
# Lower max_num_seqs or max_model_len on low-VRAM GPUs.
4849
llm = LLM(model=model_name, tokenizer_mode="mistral")
4950

5051
prompt = "Describe this image in one sentence."
@@ -83,7 +84,7 @@ def run_advanced_demo():
8384
model=model_name,
8485
tokenizer_mode="mistral",
8586
limit_mm_per_prompt={"image": max_img_per_msg},
86-
max_num_batched_tokens=max_img_per_msg * max_tokens_per_img,
87+
max_model_len=max_img_per_msg * max_tokens_per_img,
8788
)
8889

8990
prompt = "Describe the following image."

0 commit comments

Comments
 (0)