Skip to content

How to infer multiple prompts(bs>1) at the same time? #1623

New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Closed
vicwer opened this issue May 28, 2023 · 2 comments
Closed

How to infer multiple prompts(bs>1) at the same time? #1623

vicwer opened this issue May 28, 2023 · 2 comments

Comments

@vicwer
Copy link

vicwer commented May 28, 2023

No description provided.

@KerfuffleV2
Copy link
Collaborator

Possibly with the server example: https://github.com/ggerganov/llama.cpp/tree/master/examples/server

You would need a script or something to manage the queries and collect the result.

@vicwer vicwer closed this as completed Jun 5, 2023
@liuxiaohao-xn
Copy link

Possibly with the server example: https://github.com/ggerganov/llama.cpp/tree/master/examples/server

You would need a script or something to manage the queries and collect the result.

i have the same question.

I have multiple prompts, I want to feed them all at once to the model to generate the outputs,can you tell me how to achieve it ?

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants