Star us on GitHub!
Danswer

Guide: Connect Llama3.1 to your local data and documents

8/22/2024

llama chat with docs

Many companies cannot send their internal data, documents, and conversations with an external third party such as OpenAI. In order to benefit from AI without sharing their data, these companies can turn to self-hosting open source models. In this post we’ll share how you can connect a state-of-the-art LLM to your own knowledge sources/documents and run everything locally without connecting to the internet.

Llama 3.1 405B is the first opely available model that rivals the top AI models when it comes to state-of-the-art capabilities in general knowledge, steerability, math, tool use, and multilingual translation. At the time of writing, the 405B model is ranked fifth on LMSYS Chatbot Arena Leaderboard. The upgraded versions of 8B and 70B models are multilingual and have a significantly longer context length of 128K, state-of-the-art tool use, and overall stronger reasoning capabilities.

With the increased context length, we can start making better use of these open source models by adding our private knowledge sources. To set up the Llama model locally and connect it to your company knowledge base follow these steps:

Step 1: Install and Configure Ollama

  1. Install Ollama: Ollama allows you to host LLMs locally and provides a REST API for the model. Visit Ollama’s Github for installation instructions.
  2. Run Llama Model: Start the Llama model with the following command:
ollama run llama3.1

  1. Verify API: Ensure the API is running with a curl request:

curl http://localhost:11434/api/generate -d '{ "model": "llama3.1", "prompt":"Why is the sky blue?" }'

Step 2: Configure Danswer to Use Ollama

  1. Run Danswer Locally: Follow this quickstart guide to setup and run Danswer on localhost.
  2. Add A Custom LLM: Navigate to the LLM tab in the Admin panel and add a Custom LLM. Enter any display name, and use “ollama” for the Provider Name.
An image without an alt, whoops

  1. Configure Llama3.1: For the API Base use http://localhost:11434 and use llama3.1 for the Model Name. Additionally add llama3.1 as your Default Model. After clicking save/update the model should be live in Danswer.
An image without an alt, whoops

Step 3: Integrate Your Knowledge Sources

  1. Setup Connectors: Danswer allows you to connect over 35 different knowledge sources such as Salesforce, Dropbox, and Microsoft Teams. Navigate to the Add Connector page and follow the steps for each connecter you’d like to integrate.
An image without an alt, whoops

2. Chat with Your Docs: With everything set up you can now query your company data using Llama3.1 and keep all of your data local!

An image without an alt, whoops

If you need any help setting up something similar feel free to join our Slack.