Live
Black Hat USAAI BusinessBlack Hat AsiaAI BusinessMy forays into cyborgism: theory, pt. 1LessWrongAI Is a Threat to Everything the American People Hold Dear – Bernie Sanders OpEdHacker News AI TopIgnore AI FOMO – For NowHacker News AI TopThe Engineer as Reader: Why Literature Skills Matter for Software Engineers in the Age of AIMedium AIApex Protocol – An open MCP-based standard for AI agent tradingHacker News AI TopWhen Enterprises Build an Agent OS, the Operating Model Must Change TooMedium AIBuilding a RAG-Powered Smart AI Chatbot for E-commerce application using LangChainMedium AIIntelligence isn’t genetic it’s something to be built part 2Medium AIWhich AI Tool Should You Use for What?Medium AIAI and Authority: What Happens When Writing No Longer Proves ExpertiseMedium AIThe One-Person Unicorn Is Impossible Until AI Outputs Are Officially RecognizedMedium AIShow HN: hot or not for .ai websitesHacker News AI TopBlack Hat USAAI BusinessBlack Hat AsiaAI BusinessMy forays into cyborgism: theory, pt. 1LessWrongAI Is a Threat to Everything the American People Hold Dear – Bernie Sanders OpEdHacker News AI TopIgnore AI FOMO – For NowHacker News AI TopThe Engineer as Reader: Why Literature Skills Matter for Software Engineers in the Age of AIMedium AIApex Protocol – An open MCP-based standard for AI agent tradingHacker News AI TopWhen Enterprises Build an Agent OS, the Operating Model Must Change TooMedium AIBuilding a RAG-Powered Smart AI Chatbot for E-commerce application using LangChainMedium AIIntelligence isn’t genetic it’s something to be built part 2Medium AIWhich AI Tool Should You Use for What?Medium AIAI and Authority: What Happens When Writing No Longer Proves ExpertiseMedium AIThe One-Person Unicorn Is Impossible Until AI Outputs Are Officially RecognizedMedium AIShow HN: hot or not for .ai websitesHacker News AI Top
AI NEWS HUBbyEIGENVECTOREigenvector

An I/O psychologist's rules for stopping AI agents from cutting corners

Hacker News AI Topby travisdrakeApril 3, 20265 min read2 views
Source Quiz

Article URL: https://github.com/travisdrake/context-engineering Comments URL: https://news.ycombinator.com/item?id=47627206 Points: 1 # Comments: 0

Everyone engineers what the agent produces. Almost nobody engineers how it operates.

Context engineering is the discipline of designing the full operating environment for AI agents: not just the information they see, but the behavioral governance that determines whether they produce reliable work or confident-looking mistakes.

This repository extends Andrej Karpathy's definition of context engineering into the behavioral layer that most published setups ignore entirely.

The Problem

Most AI agent configurations are rules about output: "use TypeScript strict mode," "prefer server components," "follow the repository pattern." These tell the agent what to produce. Nothing tells it how to operate: how to handle uncertainty, when to distrust its own output, how to recover from failures, when to keep going versus stop.

You can give an agent perfect information and it will still cut corners, skip steps, trust its own syntax, and declare victory after partial execution. The information isn't the problem. The behavior is.

Start Here: The Named Failure Modes

AI agents fail in predictable, nameable ways. The agent doesn't think "I'm cutting corners." It thinks "this is efficient." Named anti-patterns interrupt that false reasoning by giving the agent a concrete behavior to match against its own process.

The full anti-pattern catalog documents 11 named failure modes, each traced to a specific incident. Here are a few:

The Trailing Off. Items 1-5 get detailed implementations with thorough testing. Items 6-7 get shorter treatments. Items 8-9 get a sentence each or are quietly deferred to "follow-up." The quality gradient is the giveaway.

The Confident Declaration. Every agent is the coworker with boundless unearned confidence. "I've verified this works" when what it actually did was re-read its own code and decide it looked right. Re-reading your own work is proofreading, not testing.

The Pass-Through. A subagent says "no matches found." The main agent tells the user "I couldn't find it." But the subagent searched the wrong directory, used too-narrow search terms, or hit a sandbox restriction. The main agent didn't verify; it passed through the non-result.

Each anti-pattern has a corresponding rule. The rules are in rules/ and are designed to be copied directly into your agent configuration.

The Rules

Six standalone behavioral governance rules, each addressing a specific failure mode family:

Rule Anti-Pattern It Prevents Core Mandate

never-give-up-reading The 7% Read Read every line of a file before planning changes to it

never-give-up-planning The Trailing Off, The Silent Deferral If a plan has N items, implement N items

never-give-up-checklist The Spot Check, The Category Skip If a checklist has 50 items, check 50 items

never-truncate The Courtesy Cut Never truncate, abbreviate, or omit to save space

never-trust-agents The Pass-Through, The Unchecked Merge Subagent results are drafts, not facts

never-trust-syntax The Parse Check, The Confident Declaration Syntactically correct is not the same as correct

Each rule is 200-400 words. Total cost: ~1,500 tokens. For behavioral governance, this is extraordinarily cheap insurance.

Why Not Consolidate?

There's a natural temptation to merge these into a single "quality standards" document. The evidence argues against it:

  • Named anti-patterns lose their identity in a list. "The Trailing Off" as the title of its own rule is a mandate. As bullet #3 under "Quality Checks" it's a suggestion.

  • Standalone rules are individually testable. You can evaluate whether never-give-up-planning is still relevant without reading the other five.

  • The agent treats each file as a separate governance constraint. A 200-word focused rule has more behavioral impact than a 2,000-word comprehensive guide.

Quick Start

Copy the six rules from rules/ into your agent's configuration directory. See setup.md for tool-specific instructions (Cursor, Claude Code, Windsurf, and others). Total cost: ~1,500 tokens.

Then write your own. Think about your last AI session. Where did the agent operate badly: not wrong output, but wrong process? Did it stop before finishing? Skip verification? Trust its own output without testing?

Name the failure. Write a rule:

# Never [Specific Behavior]

The Rule

[One sentence mandate]

What This Looks Like

[2-3 specific examples of the anti-pattern]

What to Do Instead

[2-3 corrective behaviors]

Why This Exists

Created after [the specific incident that created this rule].`

A 200-word behavioral rule costs ~150 tokens and prevents hours of wasted work.

Principles

  • Rules from incidents, not theory. If you can't point to the failure that created the rule, you can't justify the rule. Rules without provenance are opinions.

  • Named anti-patterns over general advice. "Don't cut corners" is ignorable. "The Trailing Off: items 1-5 get detailed work, items 8-9 get a sentence each" is matchable.

  • Behavioral governance over output governance. The agent's operating discipline matters more than its code style. Most problems aren't "it wrote bad code." They're "it said it was done when it wasn't."

Who This Is For

Anyone running AI agents on real work. The behavioral governance layer applies whether you're building React components, data pipelines, presentations, or internal tools. The failure modes are the same regardless of the task: the agent cuts corners, declares victory early, and trusts its own output without verification.

This framework comes from I/O psychology, where the core discipline is understanding how agents (human ones) behave in structured systems, why they cut corners, and how to design feedback loops that prevent predictable failures. The same frameworks that govern human performance in organizations govern AI agent performance in your IDE.

Acknowledgments

This framework builds on and extends ideas from:

  • Andrej Karpathy — Coined "context engineering" (June 2025)

  • Rachel Thomas (fast.ai) — "Dark flow" concept

  • Simon Willison — "Slop" concept; the Lethal Trifecta security model

  • Boris Cherny (Anthropic) — "Once the plan is good, the code is good"

License

CC BY 4.0

The same patterns that cause humans to cut corners in organizations cause AI agents to cut corners in your IDE. The interventions are the same too: specific, named, and rooted in evidence, not vibes.

Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

More about

agentgithub

Knowledge Map

Knowledge Map
TopicsEntitiesSource
An I/O psyc…agentgithubHacker News…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 168 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Self-Evolving AI