Ollama Chat is a conversational AI chat client that uses Ollama to interact with local large language models (LLMs) entirely offline. Ideal for AI enthusiasts, developers, or anyone wanting private, offline LLM chats.
- Chat with local large language models (LLMs) entirely offline
- Prompt Commands to include files, images, and URL content
- Conversation Templates for repeating prompts with variable substitutions
- Browse, download, monitor, and select local models directly in the app
- Multiple concurrent chats
- Show/hide thinking of reasoning models
- Regenerate the most recent conversation response
- Delete the most recent conversation exchange
- View responses as Markdown or text
- Save conversations as Markdown text
- Start a conversation or template from the command line
- Platform independent - tested on macOS, Windows, and Linux
To get up and running with Ollama Chat follow these steps:
-
Install Ollama
-
Install Ollama Chat
macOS and Linux
python3 -m venv $HOME/venv --upgrade-deps . $HOME/venv/bin/activate pip install ollama-chat
Windows
python3 -m venv %USERPROFILE%\venv --upgrade-deps %USERPROFILE%\venv\Scripts\activate pip install ollama-chat
To start Ollama Chat, open a terminal prompt and follow the steps for your OS. When you start Ollama Chat, a web browser is launched and opens the Ollama Chat application.
By default, a configuration file, "ollama-chat.json", is created in the user's home directory.
. $HOME/venv/bin/activate
ollama-chat
%USERPROFILE%\venv\Scripts\activate
ollama-chat
To start a conversation from the command line, use the -m
argument:
ollama-chat -m "Why is the sky blue?"
To start a named template from the command line, use the -t
and -v
arguments:
ollama-chat -t askAristotle -v question "Why is the sky blue?"
To add a desktop launcher, follow the steps for your OS.
In Finder, locate the "ollama-chat" executable and drag-and-drop it into the lower portion of the Dock.
In File Explorer, locate the "ollama-chat" executable, right-click it, and select "Pin to Start".
-
Copy the following Ollama Chat desktop file contents:
[Desktop Entry] Name=Ollama Chat Exec=sh -c "$HOME/venv/bin/ollama-chat" Type=Application Icon=dialog-information Terminal=true Categories=Utility;
-
Create the Ollama Chat desktop file and paste the contents:
nano $HOME/.local/share/applications/ollama-chat.desktop
-
Update the "Exec" path, if necessary, and save
Conversation Templates allow you to repeat a sequence of prompts. Templates can include variable
substitutions in the title text and the prompt text (e.g., {{var}}
).
There are two ways to create a template:
- Click "Add Template" from the home page
- Click "Template" on a conversation page
To run a template, click on its title on the home page. If the template has any variables, the user is prompted for their values prior to running the template. When a template runs, a new conversation is created and each prompt is entered in sequence.
To edit a template, from the home page, click "Select" on the template you want to edit, and then click "Edit". On the template editor page you can update the template's title, set its name, add or remove variables, and add or remove prompts.
Ollama Chat supports special prompt commands that allow you to include files, images, and URL content in your prompt, among other things. The following prompt commands are available:
-
/file
- include a file/file README.md
-
/image
- include an image/image image.jpeg
-
/dir
- include files from a directory/dir src/ollama_chat py
-
/url
- include a URL resource/url https://craigahobbs.github.io/ollama-chat/README.md
-
/do
- execute a conversation template by name/do city-report -v CityState "Seattle, WA"
-
/?
- list available prompt commands/?
To get prompt command help use the -h
option:
/file -h
This package is developed using python-build. It was started using python-template as follows:
template-specialize python-template/template/ ollama-chat/ -k package ollama-chat -k name 'Craig A. Hobbs' -k email 'craigahobbs@gmail.com' -k github 'craigahobbs' -k noapi 1