Skip to content

2.3.19 Satellite: ol1

av edited this page Sep 16, 2024 · 1 revision

Handle: ol1 URL: http://localhost:34111

ol1 screenshot

"But we have o1 at home!"

Based on the original work from:

Lightweight Streamlit-based service with Ollama as a backend, that implements the o1-like reasoning chains.

Starting

# [Optional] pre-build the image
harbor build ol1

# Start the service
harbor up ol1

# Open ol1 in the browser
harbor open ol1

Configuration

ol1 can be configured in the same way as the other Harbor services:

  • via Harbor CLI (aliases, config, or profiles)
  • via the .env file

Following configuration options are available:

# Get/set desired Ollama model for ol1
harbor ol1 model

# Manage dictionary of "options" to be
# sent to the Ollama
harbor ol1 args

# List all currently set args
harbor ol1 args
harbor ol1 args ls

# Get the value of a specific arg
harbor ol1 args get <arg_name>
harbor ol1 args get temperature

# Set the temperature
harbor ol1 args set <arg_name> <value>
harbor ol1 args set temperature 0.5

Collections of settings can be saved in Harbor profiles for future reuse.

# Perform the configuration
harbor ol1 model codestral
harbor ol1 args set temperature 0.8
harbor ol1 args set num_ctx 1024

# Save profile
harbor profile save codestral-ol1

# Later, restore the profile
harbor profile use codestral-ol1
Clone this wiki locally