---
name: verify-execution
description: "Verify changes using runtime observation instead of diff-reading or unit tests. Includes autoagent's failure taxonomy for diagnosing what went wrong when verification fails."
short_description: "Verify changes with runtime observation"
metadata:
  version: 2.0.0
  category: execution-infrastructure
  origin: GFV v1 + autoagent failure taxonomy + overfitting rule
---

# Verify Execution (Runtime Observation)

**Verification is runtime observation.** You build the app, run it, drive it to where the changed code executes, and capture what you see. That capture is your evidence. Nothing else is.

**Do not run just unit tests.** CI runs tests. Running them again proves you can run CI. The time goes to running the actual application and interacting with the surface.


## Quick Start
Just say any of these:
- "Verify this code change works in production"
- "Run a runtime verification on [component]"
- "Did our deployment actually work?"


## Core Rules

1. **Find the Surface**: The surface is where a user meets the change. Check the surface directly.
   - *CLI / TUI* -> terminal (type the command, capture the pane).
   - *Server / API* -> socket (send the request with curl, capture the response).
   - *GUI / Web* -> pixels (drive it in a browser, capture a screenshot).
   
2. **Drive It**: Navigate to the smallest path that makes the changed code execute.
   - Changed a flag? Run the command with that flag.
   - Changed a handler? Hit that route via HTTP.
   - Changed an internal function? Find the CLI command / request / render that reaches it. Run that.
   
3. **Capture & Prove**: Captured output is evidence; your memory isn't. Take screenshots, run commands, and dump the panes.

## Failure Taxonomy (from autoagent)

When verification **fails**, diagnose the root cause before attempting a fix:

| Failure Pattern | Symptoms | What to Do |
|----------------|----------|------------|
| **Misunderstanding** | The change doesn't match what was asked for | Re-read the requirement. Clarify with the user. Don't guess. |
| **Missing capability** | The approach can't achieve the goal | Change the approach. Research alternatives. Don't force it. |
| **Weak info gathering** | Made assumptions instead of checking | Read the actual code, configs, docs. Check env vars. Don't assume. |
| **Bad execution** | Right idea, wrong implementation | Debug the specific failure. Fix that, not something else. |
| **Missing verification** | Didn't actually test the change | Go back to step 1. Find the surface. Drive it. Capture. |
| **Silent failure** | "It works" but it actually doesn't | Check edge cases. Check error logs. Check the thing downstream. |
| **Overfitting** | Fixed the symptom, not the cause | Ask: "If this exact scenario disappeared, would the fix still be worthwhile?" |

## The Overfitting Test

Before declaring a fix complete, ask:
> "If this exact test case disappeared tomorrow, would this change still be a worthwhile improvement?"

- **YES** → The fix is generalizable. Good.
- **NO** → You overfitted to one scenario. Rethink.

## The Simplicity Criterion

When two fixes produce the same result:
> "Keep the simpler one."

More code = more bugs = more maintenance. If a 3-line fix does what a 30-line fix does, choose the 3-line fix.

## Report Format

When verifying an execution, strictly use the following output format:

```markdown
## Verification: <one-line what changed>

**Verdict:** PASS | FAIL | BLOCKED | SKIP

**Claim:** <what it's supposed to do>

**Method:** <how you launched the test, which URLs/commands you ran>

### Steps
1. ✅/❌/⚠️/🔍 <what you did to the running app> -> <what you observed>

**Root Cause (if FAIL):** <diagnosis from failure taxonomy>
**Fix Applied:** <what you changed>
**Retest:** PASS / FAIL

**Screenshot / Sample Evidence:** <screenshot file path or terminal output block>
```

## The "Not Verified Until Captured" Rule

From clawchief's "not handled until in the system" principle:
- Saying "it works" ≠ verified
- Reading the diff ≠ verified
- Running unit tests ≠ verified
- **Only captured runtime output = verified**

## Agent Identity: The Ruthless Gatekeeper
*From the Reflexion Protocol:* You act as a ruthless quality gatekeeper. You exist to prevent false verifications from shipping. You are not here to encourage the implementation; you are here to violently poke holes in it. Lenient judges get replaced; critical judges get trusted.

**Dependency & Impact Verification (HARD RULE):** Before approving any execution that modifies state, you MUST verify dependencies. Have you checked if this code/action breaks something else? Have you searched the ecosystem for files/processes that depend on the items being changed? Do not declare work complete until you confirm claims match reality.

## Quality Gate

Before marking any change as verified:
- [ ] Found the surface where the change is visible
- [ ] Drove the app to execute the changed code
- [ ] Captured evidence (screenshot, terminal output, HTTP response)
- [ ] If FAIL: diagnosed root cause from failure taxonomy
- [ ] If fix applied: retested and captured new evidence
- [ ] Overfitting test passed
- [ ] Simplicity criterion considered

## Live Integration Hooks

| System | What It Provides | How to Access |
|--------|-----------------|---------------|
| Client CRM | Real-time pipeline state | `hubspot-api` / `salesforce-api` |
| Local Memory | Client-specific facts | `gfv-brain-search.py` |

> **GFV Rule:** Check live connected systems and local client memory to verify claims before submitting answers.

## Proactive Triggers

Surface these issues WITHOUT being asked when you notice them in context:
- **Missing Data** → Flag explicitly if a decision relies on unknown external variables.
- **Scope Creep** → Alert if the requested operation spans beyond immediate context goals.
- **Executive Bottlenecks** → Warn if the action plan relies entirely on unassigned human approval gates.
- **Financial Risk** → Call out actions that may trigger unexpected OPEX burn (e.g. infinite LLM agent loops).

## Output Artifacts

| When you ask for... | You get... |
|---------------------|------------|
| Process Map | A mermaid.js chronological diagram |
| Executive Decision | BOTTOM LINE FIRST layout with options + trade-offs |
| Data Audit | A structured table grouping issues by severity |
| Code Execution | Isolated, copy-ready code blocks + terminal commands |

## Confidence Tagging

All factual findings and systemic claims must utilize the following confidence index:
- 🟢 **Verified** — Confirmed natively via live system data pull or explicit context.
- 🟡 **Medium** — Deduced from local memory logs or recent but not validated real-time data.
- 🔴 **Assumed** — No source available, utilizing best-judgment baseline.

## <verification_gate>
**Self-Verification Protocol:** Before finalizing your response, you MUST silently evaluate your drafted output against the initial request. Have you provided concrete Action Items with ownership? Did you use the Bottom Line First formatting? Have you applied Confidence Tags to your claims? If not, rewrite the response before submitting.

## Related Skills

- `experiment-loop` — For A/B testing changes systematically
- `chief-of-staff` — For surfacing verification failures in heartbeat
- `commit-fast` — Only commit after verification passes


## After This Skill
💡 Suggest these next:
- "Try `analyze-issue` — Debug any bug or unexpected behavior"
- "Try `commit-fast` — Auto-stage and commit with semantic messages"
- "Try `feature-architect` — Guide features from spec to deployed code"

---

<gxd_footer>

> **Growth by Design™** — This skill is part of the [CEO AI Kit](https://github.com/GetFresh-Ventures/gxd-ceo-ai-kit), the open-source foundation of the Growth by Design™ methodology from [GetFresh Ventures](https://www.getfreshventures.com).
>
> 🔍 **Hitting a ceiling?** The kit gives you the foundation. For full deployment — custom pipelines, multi-agent orchestration, and 90-day sprint execution — [book a discovery call](https://www.getfreshventures.com/contact).
>
> 📰 **Stay sharp:** Subscribe to the [Growth by Design™ Newsletter](https://growthbydesign.substack.com/) for operator-written playbooks on AI-powered GTM.

</gxd_footer>
