Skip to content

A starter RAG application for developers to learn and start building from.

Notifications You must be signed in to change notification settings

mongodb-partners/MongoDB-RAG-Render

Repository files navigation

RAG-based chatbot using LangChain, MongoDB Atlas, and Render

This starter template implements a Retrieval-Augmented Generation (RAG) chatbot using LangChain, MongoDB Atlas, and Render. RAG combines AI language generation with knowledge retrieval for more informative responses. LangChain simplifies building the chatbot logic, while MongoDB Atlas' vector database capability provides a powerful platform for storing and searching the knowledge base that fuels the chatbot's responses. Render makes it easy to build, deploy, and scale the chatbot web service.

Setup

Follow the steps below to set up a RAG chatbot powered by data from PDF files you provide.

Prerequisites

Before you begin, make sure you have the following ready:

  • MongoDB Atlas URI: Set up your account if you don't already have one (Create Account). Then create an Atlas cluster.

  • OpenAI API Key: Set up an OpenAI account. Then retrieve your API keys here.

  • Render account: Set up a Render account.

  • A PDF of your choice. This PDF represents your knowledge base. (Here's an example PDF if you need one.)

Step 1: Configure Render Web Service

  • Fork mongodb-partners/MongoDB-RAG-Render on GitHub.

  • Create a new Web Service on Render. Choose "Build and deploy from a Git repository" and select your forked GitHub repo.

  • Use the following values during Web Service creation:

    Runtime/Language    Node
    Build Command       npm install; npm run build
    Start Command       npm run start
    
  • Populate the values of the Environment Variables as follows:

    OPENAI_API_KEY = <YOUR_OPENAI_KEY>           # API Key copied from the OpenAI portal
    MONGODB_URI = <YOUR_MONGODB_URI>             # Connection URI to MongoDB Instance
    

Step 2: Deploy Render Web Service

  • Once you have inputted the above values, create the service.
  • Wait for the service to be deployed and start serving traffic.
  • Click the URL of your new service to open your new chatbot website: image

Step 3: Give Render Web Service permission to access Atlas

You must allow your new web service to talk to your MongoDB instance.

Step 4: Upload PDF files to your chatbot

  • On your chatbot website, select the Train tab and upload a PDF document of your choice.

  • If everything is deployed correctly, your document should start uploading to your cluster under the chatter > training_data collection.

  • Your data should appear like this in the collection:

    image

Step 5: Create Vector Index on Atlas

For the RAG Question Answering (QnA) to work, you need to create a Vector Search Index on Atlas so your vector data can be fetched and served to LLMs.

Let’s head over to our MongoDB Atlas user interface to create our Vector Search Index.

  • First, click on "Atlas Search” in the sidebar of the Atlas dashboard. Select the cluster you're using for this guide. Then click “Create Search Index.”

  • You’ll be taken to this page (shown below). Here, select “JSON Editor” in the Atlas Vector Search section. Click "Next". image

  • Input the values shown below and create the vector index.

      {
        "fields": [
          {
            "type": "vector",
            "path": "text_embedding",
            "numDimensions": 1536,
            "similarity": "cosine"
          }
        ]
      }
    

    image

  • You should start seeing a vector index getting created. You should get an email once index creation is completed. image

Step 6: Ask questions

Finally, head back to your chatbot website. Select the "QnA" tab to start asking questions based on your trained data.

image

Reference Architechture

image

This architecture depicts a Retrieval-Augmented Generation (RAG) chatbot system built with LangChain, OpenAI, and MongoDB Atlas Vector Search. Let's break down its key players:

  • PDF File: This serves as the knowledge base, containing the information the chatbot draws from to answer questions. The RAG system extracts and processes this data to fuel the chatbot's responses.
  • Text Chunks: These are meticulously crafted segments extracted from the PDF. By dividing the document into smaller, targeted pieces, the system can efficiently search and retrieve the most relevant information for specific user queries.
  • LangChain: This acts as the central control unit, coordinating the flow of information between the chatbot and the other components. It preprocesses user queries, selects the most appropriate text chunks based on relevance, and feeds them to OpenAI for response generation.
  • Query Prompt: This signifies the user's question or input that the chatbot needs to respond to.
  • Actor: This component acts as the trigger, initiating the retrieval and generation process based on the user query. It instructs LangChain and OpenAI to work together to retrieve relevant information and formulate a response.
  • OpenAI Embeddings: OpenAI, a powerful large language model (LLM), takes centre stage in response generation. By processing the retrieved text chunks (potentially converted into numerical representations or embeddings), OpenAI crafts a response that aligns with the user's query and leverages the retrieved knowledge.
  • MongoDB Atlas Vector Store: This specialized database is optimized for storing and searching vector embeddings. It efficiently retrieves the most relevant text chunks from the knowledge base based on the query prompt's embedding. These retrieved knowledge nuggets are then fed to OpenAI to inform its response generation.

This RAG-based architecture seamlessly integrates retrieval and generation. It retrieves the most relevant knowledge from the database and utilizes OpenAI's language processing capabilities to deliver informative and insightful answers to user queries.

Implementation

The below components are used to build up the bot, which can retrieve the required information from the vector store, feed it to the chain and stream responses to the client.

LLM Model

    const model = new ChatOpenAI({
        temperature: 0.8,
        streaming: true,
        callbacks: [handlers],
    });

Vector Store

    const retriever = vectorStore().asRetriever({ 
        "searchType": "mmr", 
        "searchKwargs": { "fetchK": 10, "lambda": 0.25 } 
    })

Chain

   const conversationChain = ConversationalRetrievalQAChain.fromLLM(model, retriever, {
        memory: new BufferMemory({
          memoryKey: "chat_history",
        }),
      })
    conversationChain.invoke({
        "question": question
    })

About

A starter RAG application for developers to learn and start building from.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published