Skip to content

Add LlamaCppInferenceEngine #567

New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

vs4vijay
Copy link

Related to #167

@AlexCheema
Copy link
Contributor

Hi Vijay, this would be a great addition - we need better support for CPU in exo!

Did you forget to commit some files?

@lexasub
Copy link

lexasub commented Jan 23, 2025

@vs4vijay , any updates?

@pcfreak30
Copy link

👀 @vs4vijay definitely interested in this PR, though based on prev comments some files are missing?

@vs4vijay
Copy link
Author

Hey @AlexCheema , @lexasub , @pcfreak30 - I got carried away with other work. I plan to pick this up this week.

@shellfyred
Copy link

@vs4vijay sorry to bug :), but I have an interest in this as well. Would love to look at using this to extend support to Intel iGPU/dGPU

https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/llama_cpp_quickstart.md

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants