Skip to content

I saw that llama.cpp has an OpenAI compatible server now #118

Closed Answered by giladgd
Madd0g asked this question in Q&A
Discussion options

You must be logged in to vote

The server example in llama.cpp is only an example and is not meant for production use, nor is it exposed as part of the llama.cpp stable interface.
It also doesn't implement significant parts of OpenAI API.

An inference server is planned as part of the roadmap, but it'll take some time.
The next version (#105) will include many improvements required to implement a sufficient API for most use cases people are using OpenAI API for.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by giladgd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants