Configuration¶
pyproject.toml¶
[tool.checkllm]
judge_backend = "auto" # auto, openai, anthropic, gemini, azure, ollama, litellm
judge_model = "gpt-4o"
default_threshold = 0.8
runs_per_test = 1
engine = "auto" # async, thread, process, hybrid, auto
max_concurrency = 10
budget = 10.00 # Max USD per run (optional)
cache_enabled = true
cache_ttl_seconds = 604800 # 7 days
log_level = "WARNING"
Profiles¶
[tool.checkllm.profiles.dev]
judge_model = "gpt-4o-mini"
budget = 1.0
log_level = "DEBUG"
[tool.checkllm.profiles.ci]
cache_enabled = false
budget = 10.0
engine = "async"
[tool.checkllm.profiles.prod]
judge_model = "gpt-4o"
default_threshold = 0.9
max_concurrency = 20
Activate: CHECKLLM_PROFILE=ci checkllm run tests/
Environment Variables¶
All settings support env var overrides:
| Variable | Config Key |
|---|---|
CHECKLLM_JUDGE_BACKEND |
judge_backend |
CHECKLLM_JUDGE_MODEL |
judge_model |
CHECKLLM_DEFAULT_THRESHOLD |
default_threshold |
CHECKLLM_ENGINE |
engine |
CHECKLLM_BUDGET |
budget |
CHECKLLM_CACHE_ENABLED |
cache_enabled |
CHECKLLM_MAX_CONCURRENCY |
max_concurrency |
CHECKLLM_LOG_LEVEL |
log_level |
CHECKLLM_PROFILE |
Active profile |