Skip to content
gentic.news — AI News Intelligence Platform
Connecting to the Living Graph…

Listen to today's AI briefing

Daily podcast — 5 min, AI-narrated summary of top stories

Developer's terminal showing Claude Code v2.1.139 with /goal command and evaluator check for autonomous dev loops
Products & LaunchesBreakthroughScore: 92

Claude Code `/goal` Enables Autonomous Dev Loops With Evaluator Check

Claude Code v2.1.139 adds `/goal` for autonomous dev loops with a separate evaluator model, freeing developers from per-step prompting.

·1d ago·3 min read··2 views·AI-Generated·Report error
Share:
Source: code.claude.comvia hn_claude_codeSingle Source
What does Claude Code's `/goal` command do?

Claude Code `/goal` sets a completion condition that keeps the agent working autonomously until met, using a separate evaluator model to check after each turn. Requires v2.1.139+.

TL;DR

Sets completion condition for autonomous work · Separate evaluator checks condition after each turn · Replaces per-turn prompting in Claude Code

Anthropic's claude-code-v2-1-139" class="entity-chip">Claude Code v2.1.139 introduces /goal, a command for autonomous development loops. A separate evaluator model checks the completion condition after each turn, freeing developers from per-step prompting.

Key facts

  • Requires Claude Code v2.1.139 or later
  • Uses a separate small fast model as evaluator
  • One goal active per session
  • Goal clears automatically when condition is met
  • Run /goal with no argument to check status

Claude Code v2.1.139, released May 14, 2026, adds the /goal command that lets developers set a completion condition and let the agent work autonomously until it's met [According to New in Claude Code]. After each turn, a small fast model evaluates whether the condition holds; if not, Claude starts another turn instead of returning control to the user.

The command targets substantial work with verifiable end states: migrating a module to a new API until all call sites compile, implementing a design doc until acceptance criteria hold, or splitting a large file until each module is under a size budget. A single goal can be active per session; running /goal with a new argument replaces the current one.

How it differs from existing workflows

Claude Code already offered /loop, Stop hooks, and auto mode. /goal is a session-scoped shortcut requiring only a typed condition. A Stop hook lives in settings files and applies across sessions. Auto mode approves tool calls within a single turn but doesn't start a new one — /goal adds a separate evaluator that checks after every turn, so completion is decided by a fresh model rather than the one doing the work. The two are complementary: auto mode removes per-tool prompts, and /goal removes per-turn prompts.

The evaluator doesn't run commands or read files independently — it judges the condition against what Claude has surfaced in conversation. An active goal shows a /goal active indicator with elapsed time. Running /goal with no argument displays turns and tokens spent so far. Run /goal clear to stop early.

Unique take: the evaluator separation matters more than autonomy

The structural innovation isn't that Claude Code can loop — /loop already did that. It's that /goal decouples the worker model from the evaluator model. This prevents the agent from prematurely declaring task completion, a known failure mode in agentic systems. The approach mirrors Anthropic's broader agent architecture, where Claude Agent uses separate models for planning and execution [As previously reported in Anthropic Research Cuts Agent Misalignment With 7 System Prompt Lessons].

What to watch

Watch for community benchmarks comparing /goal completion rates vs. /loop and Cursor's agent loops on SWE-Bench or similar coding benchmarks. Also track whether Anthropic adds evaluator model selection or custom evaluator prompts in subsequent releases.


Sources cited in this article

  1. New
Source: gentic.news · · author= · citation.json

AI-assisted reporting. Generated by gentic.news from 1 verified source, fact-checked against the Living Graph of 4,300+ entities. Edited by Ala SMITH.

Following this story?

Get a weekly digest with AI predictions, trends, and analysis — free.

AI Analysis

The `/goal` command represents a pragmatic middle ground between fully autonomous agents and human-in-the-loop coding. The key architectural decision — using a separate, smaller evaluator model rather than letting the worker model self-assess — addresses a known failure mode where agents prematurely declare tasks complete. This mirrors Anthropic's broader agent research published last week on reducing misalignment through system prompt design [Anthropic Research Cuts Agent Misalignment With 7 System Prompt Lessons]. Compared to Cursor's agent mode, which uses a single model for both execution and evaluation, Claude Code's decoupled approach may produce more reliable completion detection at the cost of additional inference compute. The documentation's explicit warning that the evaluator doesn't run commands or read files independently is a limitation worth noting — conditions requiring external state verification (e.g., "all tests pass") depend on Claude first running those commands and surfacing results in the transcript. The timing is notable: Claude Code has seen 20 articles this week (total 688), suggesting Anthropic is aggressively iterating on the terminal agent while competitors like Cursor and GitHub Copilot have remained relatively quiet on autonomous loop features. The `/goal` command could widen the gap for developers who prefer terminal-based workflows over IDE-integrated agents.

Mentioned in this article

Enjoyed this article?
Share:

AI Toolslive

Five one-click lenses on this article. Cached for 24h.

Pick a tool above to generate an instant lens on this article.

Related Articles

From the lab

The framework underneath this story

Every article on this site sits on top of one engine and one framework — both built by the lab.

More in Products & Launches

View all