Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama.cpp / GGUF support #7

Open
sammcj opened this issue Sep 4, 2024 · 7 comments
Open

llama.cpp / GGUF support #7

sammcj opened this issue Sep 4, 2024 · 7 comments

Comments

@sammcj
Copy link

sammcj commented Sep 4, 2024

It would be great to see OLMoE/OlmoeForCausalLM Llama.cpp/GGUF support.

Really neat project!

@AmitKKhanchandani
Copy link

+1, need to try this with ollama :)

@Bobetele
Copy link

Bobetele commented Sep 9, 2024

yes

@Muennighoff
Copy link
Collaborator

won't have bandwidth to do this, but if anyone is interested, that'd be amazing!

@MrDowntempo
Copy link

Yeah, this is hard to work with if it isn't in GGUF format to run locally, or available from Ollama directly. I'm looking into how to serve from Safetensors but not a lot of servers support that.

@Muennighoff
Copy link
Collaborator

also cc @2015aroras

@2015aroras
Copy link

See ggerganov/llama.cpp#9462

@Meshwa428
Copy link

See ggerganov/llama.cpp#9462

It still isn't merged 😞

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants