config#
Read and write LLM and behavior settings.
Synopsis#
oxo-call config set <KEY> <VALUE>
oxo-call config get <KEY>
oxo-call config show
oxo-call config verify
oxo-call config path
Subcommands#
config set#
Persist a configuration value:
oxo-call config set llm.provider openai
oxo-call config set llm.api_token sk-...
oxo-call config set llm.temperature 0.2
config get#
Show the effective value (with environment variable overrides applied):
config show#
Display all stored and effective values side-by-side:
config verify#
Make a real API call to confirm the LLM configuration works:
config path#
Print the path to config.toml:
Configuration Reference#
See the Configuration tutorial for complete details on all keys, defaults, and environment variables.