Loading...
Loading...
Generate a test suite of natural-language → SQL pairs that becomes the quality benchmark for a nao agent, then run it via `nao test`. Use when the user wants to start measuring agent reliability, extend an existing test suite, or add tests for new metrics. Tests are the only honest answer to "is the context working?". Do not use for writing rules (write-context-rules) or diagnosing failures (audit-context).
npx skill4agent add getnao/nao create-context-testsnao testRULES.md## Key Metrics Reference| Bad | Good |
|---|---|
"What was the churn rate from | "How's churn looking this quarter?" |
"Compute MRR as SUM( | "What's our MRR?" |
| Bad | Good |
|---|---|
| |
| |
| |
<metric>_float_0_1<metric>_percentage_0_100<metric>_<currency>_<unit><thing>_count<thing>_date_yyyy_mm_ddtemplates/test.yamltests/templates/test.yamlnao testcdnao_config.yamlnao chat &nao_config.yamlAI_APICallError: Not Foundhttps://api.anthropic.com/messages/v1/unset ANTHROPIC_BASE_URL ANTHROPIC_API_KEYsetup-contextnao test -m <model_id> -t 10 # -t = parallelismtests/outputs/write-context-rulesaudit-context<placeholder>FROMpromptRULES.mdtemplates/test.yaml