TaskingAI is a BaaS (Backend as a Service) platform for LLM-based Agent Development and Deployment. It unified the integration of hundreds of LLM models, and provides an intuitive user interface for managing your LLM application's functional modules, including tools, RAG systems, assistants, conversation history, and more.
- All-In-One LLM Platform: Access hundreds of AI models with unified APIs.
- Abundant enhancement: Enhance LLM agent performance with hundreds of customizable built-in tools and advanced Retrieval-Augmented Generation (RAG) system
- BaaS-Inspired Workflow: Separate AI logic (server-side) from product development (client-side), offering a clear pathway from console-based prototyping to scalable solutions using RESTful APIs and client SDKs.
- One-Click to Production: Deploy your AI agents with a single click to production stage, and scale them with ease. Let TaskingAI handle the rest.
- Asynchronous Efficiency: Harness Python FastAPI's asynchronous features for high-performance, concurrent computation, enhancing the responsiveness and scalability of the applications.
- Intuitive UI Console: Simplifies project management and allows in-console workflow testing.
Models: TaskingAI connects with hundreds of LLMs from various providers, including OpenAI, Anthropic, and more. We also allow users to integrate local host models through Ollama, LM Studio and Local AI.
Plugins: TaskingAI supports a wide range of built-in plugins to empower your AI agents, including Google search, website reader, stock market retrieval, and more. Users can also create custom tools to meet their specific needs.
LangChain is a tool framework for LLM application development, but it faces practical limitations:
- Statelessness: Relies on client-side or external services for data management.
- Scalability Challenges: Statelessness impacts consistent data handling across sessions.
- External Dependencies: Depends on outside resources like model SDKs and vector storage.
OpenAI's Assistant API excels in delivering GPTs-like functionalities but comes with its own constraints:
- Tied Functionalities: Integrations like tools and retrievals are tied to each assistant, not suitable for multi-tenant applications.
- Proprietary Limitations: Restricted to OpenAI models, unsuitable for diverse needs.
- Customization Limits: Users cannot customize agent configuration such as memory and retrieval system.
- Supports both stateful and stateless usages: Whether to keep track of and manage the message histories and agent conversation sessions, or just make stateless chat completion requests, TaskingAI has them both covered.
- Decoupled modular management: Decoupled the management of tools, RAGs systems, language models from the agent. And allows free combination of these modules to build a powerful AI agent.
- Multi-tenant support: TaskingAI supports fast deployment after development, and can be used in multi-tenant scenarios. No need to worry about the cloud services, just focus on the AI agent development.
- Unified API: TaskingAI provides unified APIs for all the modules, including tools, RAGs systems, language models, and more. Super easy to manage and change the AI agent's configurations.
- Interactive Application Demos
- AI Agents for Enterprise Productivity
- Multi-Tenant AI-Native Applications for Business
Please give us a FREE STAR 🌟 if you find it helpful 😇
A simple way to initiate self-hosted TaskingAI community edition is through Docker.
- Docker and Docker Compose installed on your machine.
- Git installed for cloning the repository.
- Python environment (above Python 3.8) for running the client SDK.
First, clone the TaskingAI (community edition) repository from GitHub.
git clone https://github.com/taskingai/taskingai.git
cd taskingai
Inside the cloned repository, go to the docker directory.
cd docker
-
Copy
.env.example
to.env
:cp .env.example .env
-
Edit the
.env
file: Open the.env
file in your favorite text editor and update the necessary configurations. Ensure all required environment variables are set correctly. -
Start Docker Compose: Run the following command to start all services:
docker-compose -p taskingai --env-file .env up -d
Once the service is up, access the TaskingAI console through your browser with the URL http://localhost:8080. The default username and password are admin
and TaskingAI321
.
If you have already installed TaskingAI with a previous version and want to upgrade to the latest version, first update the repository.
git pull origin master
Then stop the current docker service, upgrade to the latest version by pulling the latest image, and finally restart the service.
cd docker
docker-compose -p taskingai down
docker-compose -p taskingai pull
docker-compose -p taskingai --env-file .env up -d
Don't worry about data loss; your data will be automatically migrated to the latest version schema if needed.
Click the image above to view the TaskingAI Console Demo Video.
Once the console is up, you can programmatically interact with the TaskingAI server using the TaskingAI client SDK.
Ensure you have Python 3.8 or above installed, and set up a virtual environment (optional but recommended). Install the TaskingAI Python client SDK using pip.
pip install taskingai
Here is a client code example:
import taskingai
taskingai.init(api_key='YOUR_API_KEY', host='http://localhost:8080')
# Create a new assistant
assistant = taskingai.assistant.create_assistant(
model_id="YOUR_MODEL_ID",
memory="naive",
)
# Create a new chat
chat = taskingai.assistant.create_chat(
assistant_id=assistant.assistant_id,
)
# Send a user message
taskingai.assistant.create_message(
assistant_id=assistant.assistant_id,
chat_id=chat.chat_id,
text="Hello!",
)
# generate assistant response
assistant_message = taskingai.assistant.generate_message(
assistant_id=assistant.assistant_id,
chat_id=chat.chat_id,
)
print(assistant_message)
Note that the YOUR_API_KEY
and YOUR_MODEL_ID
should be replaced with the actual API key and chat completion model ID you created in the console.
You can learn more in the documentation.
Please see our contribution guidelines for how to contribute to the project.
Also, we’re excited to announce that TaskingAI now has an official Discord community! 🎊
• 💬 Engage in discussions about TaskingAI, share ideas, and provide feedback.
• 📚 Get support, tips, and best practices from other users and our team.
• 🚀 Stay updated on the latest news, updates, and feature releases.
• 🤝 Network with like-minded individuals who are passionate about AI and task automation.
TaskingAI is released under a specific TaskingAI Open Source License. By contributing to this project, you agree to abide by its terms.
For support, please refer to our documentation or contact us at support@tasking.ai.