How To Talk To Your Giant Docs (2024 Guide)

In this guide, you will be able to setup a UI where you can upload and talk to HUNDREDS of files (CSV’s, Spreadsheets, PDFs, Audio Files, and more!).


This guide is for Windows 11 specifically (Also works with Windows 10). However, the general idea will work with any OS.


This guide is also to setup with your NVIDIA GPU


Download LM Studio: https://lmstudio.ai/
Download Anything LLM: https://useanything.com/download

LM Studio will act as a server for your local model. You can download and choose any model within LM Studio.

Anything LLM will be your UI interface and your Cloud Vector Solution for ingesting your documents.


  1. Setup and Install LM Studio
  • Go ahead and double click the installer (LM_Studio-0.x.x.x-Setup.exe) and install LM Studio. It will automatically open up.
  • Once LM Studio is open, click the Folder icon. Then go ahead and install Mistral_7B. Just for this example. You can choose another model to your liking later on.

  • Click on the ‘local server’ icon: <–>

  • Select the model you just downloaded

  • Make sure you set the GPU Offload Layers to maximum

  • Reload the server after every time you change settings

The model will then load into your RAM (And possibly VRAM)

Your server is now setup. We will move onto setting up the Ui and document ingest with Anything LLM

2. Install and Setup AnythingLLM

  • Go ahead and double click on the installer (AnythingLLMDesktop.exe)

  • It will open up, go ahead and click “Get Started”

  • On this page, scroll down and click “LM Studio”

  • Going back to LM Studio we need to get the URL and Token Context Window size

If you noticed, that this specific model, supports up to 32,768 tokens, that doesn’t mean you should max it out. We recommend you use 24,000 tokens max. otherwise it starts spewing weird answers.

  • Going back to Anything LLM we will type in the LM Studio server details (URL and Token Context Window, Chat Model Selection is auto loaded with what you selected in LM Studio server model selector)

Make sure to match the Token Context Window with the same settings you have in LM Studio.

If you noticed, that this specific model, supports up to 32,768 tokens, that doesn’t mean you should max it out. We recommend you use 24,000 tokens max. otherwise it starts spewing weird answers.

EXAMPLE:

The larger the context window, the more your RAM/VRAM is used.

I recommend using 2048 as a token size to make sure you have everything setup correctly. Then try a bigger token size later (2048->4096->8192, etc.). It won’t work if you run out of VRAM.

  • Use the recommended Embedding Preference

  • Use the LanceDP options

  • This is how we have it setup

  • Enter your email and click whatever applies and click next

  • Enter your workspace name



You have now setup everything you need to start talking to your documents!

  • You can highlight over the workspace and click the upload button to store your data into the Vector database and then talk to your documents.

Once you do that, you can create separate threads like you can in chatGPT and start talking to documents.

Here is my setup with the 4x 3060 12GB GPU’s

I am using openbuddy-mixtral-7bx8-v16.3-32k-GGUF

I also set the token limit to 24,000.

Make sure you talk to the bot and see if it’s replying properly before you upload your docs.

Try other models and see what runs right. There’s a lot of fine tuning with:
Context Window Size (Token Limits). If it’s too much, the bot will start talking weird. So try small first, then go bigger and bigger, if that’s what you want.


To get your desired result, you will also need to mess with these settings:


Comments

3 responses to “How To Talk To Your Giant Docs (2024 Guide)”

  1. chris Avatar
    chris

    I’m trying to use this guide to chat with spoken word audio files (interviews) but every time I try to upload a file to the workspace I get an error:

    “Failed to parse content from track.mp3.”

    Even when I get a text file to upload to the workspace, the AI does not seem to recognize it’s presence:

    I apologize again, but as a conversational AI, I don’t have the ability to see or access any file you’ve uploaded, including test.txt. Our conversation is limited to text-based input and output, so we can only discuss topics and share information through text. If you’d like to describe the contents of the file, I’d be happy to help you analyze or discuss it!

    1. Try messing with the settings in AnythingLLM. Go to the chat settings and change it to “Query” instead of “Chat” and see if that works better.

      1. So… normally AI models can read the file in workspace?
        Mine is very similar to Chris, in my case I was trying to upload pdf for ai model to read which it basically replies “As a conversational AI, I don’t have the ability to access any file you’ve uploaded”, and I tried bderkhan’s idea which did not work out.

        Maybe it is because I am using AMD Radeon? It is a RX 7600 XT.

        Anyway, if anybody have any ideas please drop it with me, I will try it eventually 🙂

Leave a Reply

Your email address will not be published. Required fields are marked *