返回顶部
s

sprint-contract

Multi-agent development workflow with Sprint Contracts and independent QA evaluation. Use when building features, fixing complex bugs, or any task that involves spawning sub-agents to do work. Implements the Planner-Generator-Evaluator pattern (inspired by Anthropic's GAN-style harness design) to ensure quality through explicit completion criteria and independent testing. Triggers on development tasks, feature builds, bug fixes, code reviews, or when spawning coding agents.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.0.0
安全检测
已通过
91
下载量
0
收藏
概述
安装方式
版本历史

sprint-contract

# Sprint Contract — Multi-Agent Quality System Based on [Anthropic's harness design for long-running apps](https://www.anthropic.com/engineering/harness-design-long-running-apps): separate the agent doing the work from the agent judging it. ## Core Principle **Never let the builder evaluate their own work on complex tasks.** LLMs reliably praise their own output — even when it's mediocre. An independent evaluator, tuned to be skeptical, catches what self-evaluation misses. ## Architecture ``` Planner (you/human) → Generator (sub-agent) → Evaluator (independent sub-agent) ↑ | └──────────── feedback loop ←──────────────────┘ ``` ## Workflow ### 1. Write a BRIEF.md with Sprint Contract Every task gets a BRIEF.md. The Sprint Contract section is **mandatory** — it lists specific, testable completion criteria. ```markdown # Task Brief ## Background [Why this task exists] ## Objective [What to build/fix] ## Sprint Contract (Completion Criteria) - [ ] Criterion 1 (specific, verifiable) - [ ] Criterion 2 - [ ] ... ⚠️ Write criteria specific to THIS task. No generic checklists. ## Related Files [File paths relevant to the task] ## Constraints [Tech stack, prior decisions, known pitfalls] ## Handoff Requirements Write a HANDOFF.md when done, containing: - What was done (file change list) - Design decisions made (and why) - What's left / known issues - Everything needed for reporting to the human ``` ### 2. Spawn Generator (Builder) The generator receives the BRIEF.md and builds against the Sprint Contract. Key rules for the generator prompt: - Work against the Sprint Contract criteria - Self-check each criterion before handing off - Write HANDOFF.md when done - Write files first, read references second (output > research) ### 3. Spawn Evaluator (Independent QA) After the generator finishes, spawn a **separate** agent as evaluator. The evaluator prompt must include: **The Sprint Contract** — copied from BRIEF.md, to verify each criterion. **4 Evaluation Dimensions** (select what's relevant): | Dimension | What to check | |-----------|--------------| | **Functional completeness** | Every Sprint Contract criterion passes | | **User experience** | Flow is intuitive, no dead ends | | **Visual quality** | Layout, spacing, colors are professional | | **Code/content quality** | No errors, clean logic, no regressions | **The critical prompt line:** > "Your job is to find problems, not to praise. If everything looks fine, you probably didn't test carefully enough. Report issues honestly — better a false alarm than a missed bug." ### 4. Decision Gate Based on evaluator feedback: - **All criteria pass** → Ship it - **Criteria fail** → Feed evaluator report back to generator for fixes - **Architecture issues** → Escalate to human ## When to Use Each Mode | Task complexity | Generator | Evaluator | Example | |----------------|-----------|-----------|---------| | Simple (< 30 min) | Sub-agent | Self-evaluate, mark "⚠️ untested" | Fix a typo, update config | | Medium (30 min - 2 hr) | Sub-agent | Independent sub-agent | New feature, bug fix | | Complex (2+ hr) | Claude Code / ACP | Independent sub-agent + human review | Architecture change, new project | ## Sprint Contract Examples See [references/contract-examples.md](references/contract-examples.md) for project-specific contract templates. ## Key Insights from Anthropic's Research 1. **File-based communication** — Agents talk through files (BRIEF.md, HANDOFF.md), not conversation 2. **Evaluator calibration** — Default LLMs are too lenient; explicitly prompt for skepticism 3. **Sprint scoping** — One feature at a time; don't bundle unrelated work 4. **Opus 4.6 + 1M context** — Context anxiety is gone; sprint decomposition is less critical, but evaluator still adds value at task boundaries 5. **Evaluation criteria shape output** — The wording of your criteria directly steers what the generator produces

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 sprint-contract-1776021560 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 sprint-contract-1776021560 技能

通过命令行安装

skillhub install sprint-contract-1776021560

下载 Zip 包

⬇ 下载 sprint-contract v1.0.0

文件大小: 3.79 KB | 发布时间: 2026-4-13 12:08

v1.0.0 最新 2026-4-13 12:08
Initial release introducing Sprint Contract, a multi-agent workflow for high-quality development.

- Implements Planner-Generator-Evaluator architecture inspired by Anthropic's harness design.
- Enforces use of explicit Sprint Contracts with specific, testable completion criteria for every task.
- Independent evaluator agent performs QA against Sprint Contract, ensuring reliable quality checks.
- Defines clear workflow, file handoff requirements, and evaluation dimensions (functional, UX, visuals, code).
- Includes guidance for when to use each mode (simple, medium, complex tasks).

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部