Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

llama : fix loading models with shared tok_embd and output #5651

Merged
merged 1 commit into from
Feb 21, 2024

Conversation

slaren
Copy link
Member

@slaren slaren commented Feb 21, 2024

No description provided.

Copy link
Collaborator

@cebtenzzre cebtenzzre left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Seems to work on CPU, CUDA, and Kompute.

@slaren slaren merged commit 973053d into master Feb 21, 2024
46 of 62 checks passed
@slaren slaren deleted the sl/fix-extra-tensor branch February 21, 2024 23:42
cebtenzzre pushed a commit to nomic-ai/llama.cpp that referenced this pull request Feb 22, 2024
jordankanter pushed a commit to jordankanter/llama.cpp that referenced this pull request Mar 13, 2024
hodlen pushed a commit to hodlen/llama.cpp that referenced this pull request Apr 1, 2024
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants