Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

oslook/ollama-webui

Repository files navigation

Ollama Web UI

A modern web interface for Ollama, featuring a clean design and essential chat functionalities.

Ollama Web UI Screenshot

Features

Chat Interface

  • πŸš€ Real-time streaming responses
  • πŸ’¬ Multi-conversation management
  • πŸ”„ Conversation history
  • πŸ“ Markdown and code syntax highlighting support
  • πŸŒ“ Clean and modern UI design

Model Management

  • πŸ“‹ List available models
  • πŸ”„ Auto-select first available model
  • 🎯 Easy model switching

Settings & Configuration

  • βš™οΈ Configurable Ollama server URL
  • πŸ’Ύ Settings persistence
  • πŸ“€ Export chat history to JSON
  • πŸ“₯ Import chat history from JSON

Getting Started

Prerequisites

  • Ollama installed and running on your machine
  • Node.js 18+ installed (for development)
  • Docker (optional, for containerized deployment)

Installation

Method 1: Docker (Recommended)

  1. Using Docker Compose (includes Ollama service):
wget https://raw.githubusercontent.com/oslook/ollama-webui/main/docker-compose.yml
docker compose up -d
  1. Using Docker directly:
docker run -d -p 3000:3000 ghcr.io/oslook/ollama-webui:latest

Method 2: Development Setup

  1. Clone the repository:
git clone https://github.com/oslook/ollama-webui.git
cd ollama-webui
  1. Install dependencies:
npm install
  1. Start the development server:
npm run dev
  1. Open http://localhost:3000 in your browser

Usage

  1. Ensure your Ollama server is running (default: http://127.0.0.1:11434)
  2. Select a model from the dropdown menu
  3. Start chatting!

Configuration

  • Click the settings icon (βš™οΈ) to:
    • Configure Ollama server URL
    • Export chat history
    • Import chat history

Docker Support

Available Tags

  • latest: Latest stable release
  • main: Latest development build
  • vX.Y.Z: Specific version releases

Environment Variables

Variable Description Default
NODE_ENV Node environment production
PORT Port to run the server on 3000

Building Locally

# Build the image
docker build -t ollama-webui .
# Run the container
docker run -d -p 3000:3000 ollama-webui

Using Docker Compose

The included docker-compose.yml provides two services:

  1. ollama-webui: The web interface
  2. ollama: The Ollama server (optional)

To use only the web interface:

docker compose up -d ollama-webui

To run both services:

docker compose up -d

Built With

Contributing

Feel free to submit issues and enhancement requests!

License

This project is licensed under the MIT License - see the LICENSE file for details.


Note: This project was generated by an AI agent (Cursor) and has been human-verified for functionality and best practices. The implementation combines modern web development patterns with practical user experience considerations.

About

A modern web interface for [Ollama](https://ollama.ai/), with DeepSeek in next version.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

AltStyle γ«γ‚ˆγ£γ¦ε€‰ζ›γ•γ‚ŒγŸγƒšγƒΌγ‚Έ (->γ‚ͺγƒͺγ‚ΈγƒŠγƒ«) /