LLM inference server implementation based on llama.cpp.
-
Updated
Sep 21, 2024 - C++
LLM inference server implementation based on llama.cpp.
A Discordjs bot that uses OpenAI compatible endpoints for chatting.
A FastAPI-powered REST API offering a comprehensive suite of natural language processing services using machine learning models with PyTorch and Transformers, packaged in a Docker container to run efficiently.
Add a description, image, and links to the openai-compatible-api topic page so that developers can more easily learn about it.
To associate your repository with the openai-compatible-api topic, visit your repo's landing page and select "manage topics."