This document covers all configuration options available in caro.
caro stores configuration in platform-specific locations:
| Platform | Config Path |
|---|
| macOS | ~/Library/Application Support/caro/config.toml |
| Linux | ~/.config/caro/config.toml |
| Windows | %APPDATA%\caro\config.toml |
Configuration uses TOML format:
# caro configuration file
# Model to use for inference
name = "qwen2.5-coder-1.5b-instruct"
# Ollama backend settings
host = "http://localhost:11434"
model = "qwen2.5-coder:latest"
url = "http://localhost:8000"
| Option | Type | Default | Description |
|---|
backend | string | "mlx" | Default inference backend |
color | bool | true | Enable colored terminal output |
safety_warnings | bool | true | Show safety level warnings |
confirm_execution | bool | true | Require confirmation before execution |
| Option | Type | Default | Description |
|---|
name | string | "qwen2.5-coder-1.5b-instruct" | Model name |
format | string | "gguf" | Model file format |
quantization | string | "q4_k_m" | Quantization level |
cache_dir | string | (auto) | Custom model cache directory |
| Option | Type | Default | Description |
|---|
enabled | bool | true | Enable MLX backend |
threads | int | 4 | Number of CPU threads |
gpu | bool | true | Use GPU acceleration |
| Option | Type | Default | Description |
|---|
enabled | bool | false | Enable Ollama backend |
host | string | "http://localhost:11434" | Ollama server URL |
model | string | "qwen2.5-coder:latest" | Ollama model name |
timeout | int | 30 | Request timeout in seconds |
| Option | Type | Default | Description |
|---|
enabled | bool | false | Enable vLLM backend |
url | string | "http://localhost:8000" | vLLM server URL |
timeout | int | 30 | Request timeout in seconds |
Configuration can also be set via environment variables:
# Override default backend
export CARO_BACKEND=ollama
export CARO_CACHE_DIR=~/custom/cache
export OLLAMA_HOST=http://localhost:11434
Command-line flags override configuration file settings:
caro --backend ollama "list files"
caro --no-color "list files"
caro --verbose "list files"
Model cache location:
| Platform | Cache Path |
|---|
| macOS | ~/Library/Caches/caro/models/ |
| Linux | ~/.cache/caro/models/ |
| Windows | %LOCALAPPDATA%\caro\cache\ |
# Show cache location and size
# Download specific model
caro cache download qwen2.5-coder-1.5b-instruct
confirm_execution = false
url = "http://inference-server:8000"
backend = "ollama" # Primary
host = "http://localhost:11434"
model = "qwen2.5-coder:latest"
url = "http://backup-server:8000"