DEV Community

Cover image for With OllaMan, Even Beginners Can Run LLMs
baboon
baboon

Posted on

With OllaMan, Even Beginners Can Run LLMs

A beginner-friendly guide to running AI models on your own computer. Get from zero to chatting with a......

You've probably heard of ChatGPT, Claude, and Gemini. They're powerful, but they all run in the cloud — meaning your conversations travel through someone else's servers.

What if you could run AI models entirely on your own machine? Local LLMs make this possible: complete privacy, no internet required, and zero API costs.

The catch? Setting up local models usually involves command lines, environment variables, and technical know-how that scares off most people.

That's where OllaMan comes in.

Let's clear up these two terms first:

Ollama: The Engine

Ollama is the open-source project that actually runs AI models on your computer. It supports all the popular open-source models:

  • Llama 3 — Meta's flagship open model
  • Mistral — The lightweight European alternative
  • DeepSeek — Exceptional reasoning capabilities
  • Gemma — Google's efficient open model

Ollama is fantastic, but it only offers a command-line interface. Great for developers, intimidating for everyone else.

OllaMan: The Dashboard

Think of Ollama as the kitchen, and OllaMan as the restaurant's beautiful front-of-house.

OllaMan is a desktop app that wraps Ollama in a modern graphical interface. With it, you can:

  • 👀 Browse all your installed models at a glance
  • 🖱️ Download new models with a single click
  • 💬 Chat with models like you would with ChatGPT
  • 🎨 Enjoy polished dark and light themes

Ollama runs the models. OllaMan makes it delightful.

Step 1: Install Ollama

First, get Ollama running on your machine:

  1. Visit ollama.ai
  2. Download the installer for your OS (macOS / Windows / Linux)
  3. Run the installer — it's a standard"Next, Next, Finish" setup

Once installed, Ollama runs silently in the background.

💡 Note: Don't expect a window to pop up — Ollama runs as a background service. That's normal.

Step 2: Install OllaMan

Next, grab OllaMan:

  1. Head to ollaman.com
  2. Download the app for your platform
  3. Install and launch OllaMan

OllaMan automatically detects your local Ollama service. If everything's working, you'll land on the dashboard.

Download a Model

A fresh Ollama installation has no models yet. Let's fix that:

  1. Click "Discover" in the left sidebar
  2. Browse the model library — you'll see dozens of options
  3. Pick something like Llama 3 or Mistral
  4. Click into the model details page
  5. Choose a size (we recommend 7B or 8B for beginners — lower hardware requirements)
  6. Hit the "Pull" button to start downloading

While downloading, you can:

  • Watch real-time progress on the Downloads page
  • See download speed and completion percentage
  • Queue up multiple models simultaneously

Download time: Depends on your internet speed and model size. A 4GB model takes roughly 5 minutes on a 100Mbps connection.

Start Chatting

Once downloaded, getting to your first conversation is straightforward:

  1. Click "Chat" in the sidebar
  2. Select your newly downloaded model from the top bar
  3. Type a message and hit Enter

That's it. You're now chatting with a local AI.

Create an Agent for Repeated Tasks

After using OllaMan for a while, you might notice you're typing the same instructions repeatedly:"Act as a coding assistant"or"Always respond in a friendly tone."

Agents solve this. An Agent is a pre-configured AI persona with:

  • A system prompt (the AI's role)
  • A default model
  • Custom generation parameters

To create one:

  1. Go to the Chat page
  2. Click the current Agent card in the left sidebar
  3. Click the "+" button
  4. Set a name, icon, and system prompt
  5. Save

Here are some Agent ideas:

Agent Name Use Case System Prompt Snippet
Code Buddy Programming help "You're a patient coding mentor who explains concepts clearly..."
Writing Coach Content creation "You're a creative writing assistant who helps brainstorm and polish text..."
Study Helper Learning "You're a friendly tutor who breaks down complex topics into simple terms..."

Once created, switching Agents instantly changes your AI's personality and defaults.

Tip 1: Attach Files Instead of Pasting

Need the AI to analyze code or a document? Skip the copy-paste.

Click the 📎 attachment button in the input area and select files directly. OllaMan supports:

  • Code files: .py, .js, .ts, .java, and more
  • Documents: .txt, .md, .json
  • Images (with vision models): .png, .jpg

Tip 2: Enable Thinking Mode

Some models (like DeepSeek R1 or QwQ) support "thinking mode" — they'll show their reasoning process before giving an answer.

If your model supports this, you'll see a "Think" toggle near the input. When enabled:

  • Responses split into "thinking" and "answer" sections
  • The thinking section is collapsible
  • Great for complex reasoning tasks

Tip 3: Tune Generation Parameters

The settings panel on the right side of each chat lets you adjust:

Parameter What It Does Recommendations
Temperature Controls creativity Code/factual: 0.1-0.3
Creative writing: 0.8-1.2
Top P Sampling range Usually keep at 0.9
Top K Candidate token count Usually keep at 40

Changes apply only to the current session — your Agent's defaults stay untouched.

Tip 4: Connect Multiple Servers

Got a beefy desktop at home and a thin laptop on the go? Great setup:

  1. Run Ollama on your powerful machine
  2. Connect to it remotely from OllaMan on any device

Just add the remote server address in Settings → Servers.

Q: What specs do I need?

Quick reference:

Model Size Recommended Setup
1B-3B 8GB RAM — entry level
7B-8B 16GB RAM — sweet spot
13B 32GB RAM or 8GB VRAM
70B+ Dedicated GPU required

💡 If unsure, start with a 7B model. It's the best balance of performance and quality.

Q: Where are models stored?

Ollama keeps models in:

  • macOS: ~/.ollama/models
  • Windows: C:\Users\<username>\.ollama\models
  • Linux: ~/.ollama/models

Q: Does it work offline?

Absolutely — that's the whole point!

  • Needs internet: Downloading models, browsing the model library
  • Works offline: Chatting with downloaded models

Once a model is on your machine, conversations happen entirely locally.

Running AI locally isn't just for power users anymore.

With Ollama + OllaMan:

  • No coding skills required
  • Your data never leaves your machine
  • Works without an internet connection

If you've been curious about local LLMs but intimidated by the terminal, now's the time.

5 minutes to install. An AI assistant that's truly yours.


Top comments (0)