Most agent failures aren't model failures.
They're context failures.
Three production codebases. Same skill applied. Here's what changed.
Most agent failures aren't model failures.
Five things the research keeps confirming:
We applied all of this to nv:context.
Distilled from 200+ sources and three production repos.
Every line competes with the actual task for attention.
Document what agents can't discover by reading code.
Snippets with full flags beat paragraphs of explanation.
150-200 instructions max. Beyond that, attention degrades.
Load context layer by need. Not everything, not at once.
Critical rules need 100% compliance. Use hooks, not hope.
Say "MUST do X", not "don't do Y". Negation is fragile.
60% safe. 70% precision drops. 85% hallucinations begin.
One command. Configs for every AI coding tool you already use.
Each file has a job. No duplication.
Every repo gets graded on six dimensions. Here’s a sample diagnostic.
Primary sources only: academic studies, lab papers, and frontline practitioners.
Sources include
Or install all four nv: skills: