Local LLMs with Ollama and AnythingLLM

Introduction

Large Language Models (LLMs) have revolutionized the way we interact with artificial intelligence, offering capabilities ranging from text generation and translation to complex reasoning. Traditionally, accessing these powerful models required connecting to cloud-based services. However, advancements in software like Ollama now allow users to run these sophisticated AI models directly on their local computers, providing a private, offline, and secure alternative. This local operation eliminates the need for constant internet access and protects sensitive data by ensuring it never leaves the user's machine. Furthermore, unlike many online AI services that impose usage limits, running LLMs locally with Ollama often comes without such restrictions.

To effectively interact with these locally hosted LLMs, applications like AnythingLLM provide a user-friendly interface. AnythingLLM is designed to leverage the power of LLMs in conjunction with local documents and knowledge bases, essentially creating a private, personalized AI assistant. It enables users to turn any document, resource, or piece of content into context that an LLM can reference during conversations, making it particularly useful for tasks like summarizing research papers, answering questions based on specific documents, or generating content informed by local data.

This guide aims to provide a comprehensive, step-by-step process for Windows 10 users to download, install, and configure both Ollama and AnythingLLM to work together seamlessly. By following these instructions, individuals with varying levels of technical expertise can set up their own local AI environment, unlocking the potential of LLMs for a multitude of applications while maintaining privacy and control over their data. This guide will cover everything from the initial software installation to downloading and utilizing various LLMs, troubleshooting common issues, and exploring the basic functionalities of this powerful local AI setup.

Prerequisites

Before embarking on the installation and configuration process, it is essential to ensure your Windows 10 system meets certain prerequisites. These foundational elements will ensure a smooth and successful setup of both Ollama and AnythingLLM.

Ensuring these prerequisites are met will lay the groundwork for a successful journey into running local LLMs.

Step 1: Installing Ollama on Windows 10

Ollama provides a simplified interface for downloading, managing, and running open-source LLMs on your local machine. Unlike cloud-based solutions, Ollama ensures your data never leaves your computer, offering enhanced privacy and security while eliminating subscription costs.

Finding the Official Ollama Website and Download Link

Go to https://ollama.com. Look for the "Download" button and specifically the Windows download link. The main download page will usually have clear options for macOS, Linux, and Windows.

Downloading the Ollama Installer

Click the Windows download link to download the installer (typically `OllamaSetup.exe`). Ensure you are downloading from the official `ollama.com` domain.

Running the Installer

Navigate to your Downloads folder, right-click `OllamaSetup.exe`, and select "Run as administrator". Follow the on-screen instructions. Ollama runs as a background process, indicated by an icon in the system tray (bottom right corner of your desktop).

Verifying the Installation

Open the command prompt (type "cmd" in the Windows search bar). Run these commands:

ollama --version

This should display the Ollama version number.

And:

curl http://localhost:11434

If Ollama is running, the response will be "Ollama is running".

Step 2: Installing AnythingLLM on Windows 10

AnythingLLM provides the user interface to interact with the LLMs managed by Ollama. You can create complex AI workflows with various ways to configure the models. It is available as a desktop application or a Docker container. This guide focuses on the desktop version, which is easier for most users.

Finding the Official AnythingLLM Website and Download Link

Go to https://anythingllm.com. Look for the "Download for desktop" button.

Downloading the AnythingLLM Desktop Installer

Click "Download for desktop" to get the installer (e.g., `AnythingLLMDesktop.exe`). Ensure you are downloading from the official `anythingllm.com` domain.

Running the Installer

Double-click the downloaded `AnythingLLMDesktop.exe` file. You may encounter a warning from Windows Defender because the application is unsigned. If so, click "More details" and then "Run anyway". Follow the on-screen instructions.

The installer might also install extra dependencies for GPU or NPU usage. If not, you might see a warning in the AnythingLLM UI about reduced performance.

Launching AnythingLLM

After installation, find the AnythingLLM icon on your desktop or in the Start Menu and launch it. You might be prompted to select an LLM model to download on the first launch. You can skip this since we'll connect to your existing Ollama installation.

Step 3: Configuring AnythingLLM to Connect to Local Ollama

To use LLMs managed by Ollama within AnythingLLM, you need to configure the connection.

Navigating to LLM Preferences

In AnythingLLM, find the settings or preferences section (look for "Settings", "Preferences", or an icon). Then locate the "LLM Preference" or "LLM Configuration" section.

Selecting Ollama as the LLM Provider

In the LLM preferences, select "Ollama" from the list of available providers.

Entering the Ollama Base URL

Enter the base URL of your Ollama server: http://127.0.0.1:11434 (This is the default. If you changed the Ollama port, use that instead).

Saving the Configuration

Make sure to save the changes in the AnythingLLM settings.

Verifying the Connection

AnythingLLM might automatically attempt to connect. If successful, the manual input field might be hidden. If it fails, double-check the URL and that the Ollama server is running (check the system tray icon or use `curl http://localhost:11434`).

AnythingLLM uses Ollama for both the LLM and potentially for embedding models. Make sure to configure Ollama and the base URL in both relevant sections (LLM provider and embedding provider) for full functionality.

Step 4: Downloading Large Language Models with Ollama

Ollama uses a command-line interface (CLI) to manage models.

Opening the Command Prompt

Open the command prompt (type "cmd" in the Windows search bar).

Using the ollama run Command

The main command is: ollama run

Replace `` with the name of the model (e.g., `llama3.2`, `mistral`, `codellama`).

Example:

ollama run llama3.2

Ollama will download the model (if it's the first time) and then run it. You can then interact with it directly in the command prompt. Model names are case-sensitive.

Other examples:
ollama run mistral
ollama run codellama
ollama run gemma3.2

Using the ollama pull Command (Optional)

To download a model *without* running it immediately, use:

ollama pull

Example:

ollama pull deepseek-r1

Listing Downloaded Models

To see a list of downloaded models:

ollama list

Exploring the Ollama Model Library

Visit https://ollama.com/library to explore available models. Models come in different sizes (e.g., 1B, 3B, 7B, 13B, 70B). Larger models are generally more capable but require more resources.

Model Size and RAM Considerations

Be mindful of your system's RAM. Running a model that exceeds your RAM can cause slow performance or crashes. General guidelines:

  • 1B - 3B models: at least 8 GB of RAM.
  • 3B - 13B models: 8 - 16 GB of RAM or more.
  • 13B+ models: 16 GB of RAM or more.
  • 30B+ models: 32 GB or more, and often benefit from a GPU.

Start with smaller models if unsure.

Step 5: Using Downloaded LLMs in AnythingLLM

Now you can instruct AnythingLLM to use the downloaded models.

Selecting the Ollama Provider in AnythingLLM

Ensure "Ollama" is selected as the LLM provider in AnythingLLM's settings (as in Step 3).

Choosing the Desired Model

In the LLM configuration, choose a "Chat Model". This might be a dropdown or a text field. Select or enter the exact name of the model you downloaded (e.g., `llama3.2`, `mistral`).

You'll also need to configure the "Embedder" or "Embedding Model" if you want to use Ollama for document processing. Select "Ollama" and choose a suitable embedding model (some models are specifically named with "embedding").

Interacting with the LLM through AnythingLLM

Navigate to the main chat interface. You can now type prompts and receive responses powered by the LLM running locally via Ollama.

If you've uploaded documents, the LLM can use them as context (Retrieval-Augmented Generation - RAG). AnythingLLM finds relevant snippets using the embedding model and passes them to the LLM.

Explore different chat modes like "Conversation" (retains history) and "Query" (simple Q&A).

Step 6: Troubleshooting Common Issues

Ollama Installation Issues

AnythingLLM Installation Issues

Connection Issues Between AnythingLLM and Ollama

Issues with Downloading or Running Models

AnythingLLM Interface Issues

Step 7: Managing and Updating Ollama and AnythingLLM

Updating Ollama

Ollama automatically checks for updates. A notification will appear on the system tray icon. Click the icon and select "Restart to update". You can also manually update by downloading the latest installer from ollama.com.

Updating AnythingLLM

Go to the AnythingLLM download page (anythingllm.com/desktop) and download the latest installer. Re-run the installer; this will overwrite the existing application while preserving your data.

If using Docker, stop the container, pull the latest image (`docker pull mintplexlabs/anythingllm`), and restart the container.

Managing Models in Ollama

Managing Workspaces and Documents in AnythingLLM

Use the AnythingLLM interface to create, manage, and delete workspaces and documents. Refer to the AnythingLLM documentation for details.

Step 8: Exploring Basic Functionalities of AnythingLLM with Ollama

Creating Workspaces

Create workspaces to organize documents and chats. Look for a button or menu option to create a new workspace.

Uploading and Managing Documents

Upload documents (PDF, TXT, DOCX, etc.) to your workspace. AnythingLLM will process them. Manage documents within the workspace (view, select, delete).

Chatting with Documents

In the chat interface, type questions. AnythingLLM uses the LLM (via Ollama) and document content to generate responses. Experiment with different question types.

Selecting Different LLMs

If you have multiple LLMs, switch between them in AnythingLLM's settings.

Basic Prompt Engineering

Be clear and specific in your prompts to guide the LLM.

Exploring Agent Flows (Optional)

AnythingLLM offers agent flows for automated workflows. Explore the "Agent Skills" page for more.

Step 9: System Requirements for Ollama and AnythingLLM on Windows 10

These requirements cover operating system, processor, RAM, disk space, and optionally a GPU.

Minimum System Requirements

Ollama:

AnythingLLM:

Recommended System Requirements

Ollama:

AnythingLLM:

Table of System Requirements

Requirement Ollama (Minimum) Ollama (Recommended) AnythingLLM (Minimum) AnythingLLM (Recommended)
Operating System Windows 10+ (64-bit) Windows 10+ (64-bit) Windows 10+ (64-bit) Windows 10+ (64-bit)
Processor 2 cores 4 cores or better 2 cores -
RAM 8 GB 16 GB or more 2 GB 4 GB
Disk Space 1 GB + model space 5 GB + model space 5 GB 1 GB for install
GPU (Optional) NVIDIA (Compute 5.0+) or AMD NVIDIA (Compute 6.0+) or AMD - -

RAM requirements for Ollama depend *heavily* on the LLM size.

LLM Parameter Size to RAM Requirements

RAM needed is closely related to the LLM's parameter size. Quantization (reducing precision) can reduce memory usage.

Parameter Size Approximate Minimum RAM (16-bit) Notes
1B-3B 4-8 GB Smaller models.
7B 14-16 GB Common general-purpose size.
13B 26-32 GB Better performance, more RAM.
30B-70B 60-140+ GB High-performance, requires substantial RAM and often a GPU.

These are rough estimates. Actual usage varies.

More precise estimation formula:

M = (P × (Q/8)) × 1.2

Where:

Example: 70B model, 4-bit quantization: (70 × (4/8)) × 1.2 = 42 GB RAM.

Start with smaller models and gradually increase if your system allows.

Conclusion

This guide has provided a comprehensive process for running LLMs locally on Windows 10 using Ollama and AnythingLLM. You should now have both applications installed, configured, and be ready to download and use LLMs.

Local LLMs offer significant advantages: privacy, data security, offline access, and no usage restrictions.

Explore the Ollama model library and experiment. AnythingLLM provides an intuitive interface, especially for working with documents.

Consult the official documentation for Ollama (https://ollama.com) and AnythingLLM (https://anythingllm.com) for advanced features and support.

Back to Top