ollamon is a terminal monitor for Ollama nodes.
It provides a TUI focused on operational visibility:
- installed models
- running models
- CPU, memory, disk, and GPU activity
- Ollama access-log telemetry
- lightweight operational insights
Add your screenshot file and update the path below if needed:
- Live overview for host and Ollama runtime state
- Running and installed model panels
- Access-log telemetry from Ollama
server.log - Filterable model views
- Mini trend charts for CPU, memory, disk, and GPU
- GPU metrics on macOS via
agputop
One-line install:
curl -fsSL https://raw.githubusercontent.com/hbasria/ollamon/main/scripts/install.sh | shInstall only, without launching:
curl -fsSL https://raw.githubusercontent.com/hbasria/ollamon/main/scripts/install.sh | sh -s -- --no-runInstall a specific version:
curl -fsSL https://raw.githubusercontent.com/hbasria/ollamon/main/scripts/install.sh | sh -s -- --version v0.1.0Supported targets:
darwin/amd64darwin/arm64linux/amd64linux/arm64
make build
./bin/ollamonFor development:
make runEnvironment variables:
OLLAMA_HOSTOLLAMON_INTERVAL_MSOLLAMON_REQUEST_TIMEOUT_MSOLLAMON_DISK_PATHOLLAMON_LOG_PATHOLLAMON_COMPACT
Example:
OLLAMA_HOST=http://127.0.0.1:11434 \
OLLAMON_INTERVAL_MS=2000 \
OLLAMON_REQUEST_TIMEOUT_MS=5000 \
make runBy default, ollamon reads Ollama access logs from:
~/.ollama/logs/server.logYou can override this with:
OLLAMON_LOG_PATH=/path/to/server.logOn macOS, GPU telemetry is collected from agputop using its JSON output.
Example command:
agputop --jsonIf agputop is not available, ollamon falls back to basic GPU detection.
- Access-log-derived telemetry can show API latency, request counts, and endpoint activity.
- Token throughput is only shown when the underlying telemetry source exposes it.
