Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Model File Formats: .pth, .bin vs. GGUF #20

Open
sebvannistel opened this issue Sep 27, 2023 · 0 comments
Open

Model File Formats: .pth, .bin vs. GGUF #20

sebvannistel opened this issue Sep 27, 2023 · 0 comments

Comments

@sebvannistel
Copy link

Hello,

I've been exploring the OmniQuant repository and am impressed with the quantization techniques provided for Large Language Models (LLMs). I noticed that the pre-trained models are available in .pth and .bin file formats from huggingface

I was wondering why these models are not available in the GGUF format, which is considered more efficient for handling large models. Is there a specific reason for this choice of file formats? Am I missing something here?

I am sure there is a reason for that I am probably just missing something.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant