Reality
Question whether the proxy still points toward the truth — or has it drifted?
Reality is the instrument panel. Strategy sets the intention — where to position, what to bet on. Reality reads whether you got there. KPIs are not goals. They are instruments: bearing and position against strategic intentions.
Performance is second in the Tight Five — after Principles, before Platform. The sequence matters. Without principles you don't know what to measure. Without a picture of good, every feedback loop is aimless.
Dig Deeper
📄️ Commissioning
Does the output match the spec — or does it just look like it does?
📄️ Content
Is the content pipeline producing work that lands?
📄️ Crypto
Understanding value creation is crucial for assessing the potential and sustainability of projects.
📄️ Economics
KPIs for Economics and Business Cycle
🗃️ Financials
7 items
🗃️ Human Resources
2 items
📄️ Product Design
How do you know your design is getting better?
📄️ Retail Industry
Measure performance of retail stores.
🗃️ SaaS Valuation
1 item
🗃️ Software
3 items
Instruments
Two loops, one system. The inner loop forms intention — what you believe good looks like. The outer loop tests it against reality. The instruments tell you where you are, not where to point.
| Loop | Question | Without Instruments |
|---|---|---|
| Inner (intention) | What should this produce? | Building blind |
| Outer (validation) | Did it match the spec? | Feedback without signal |
| Closed (compounding) | What's better this time? | Activity without progress |
| Bearing (direction) | Where am I pointed? | Position without intention is drift |
Obstacles shift bearing temporarily — the navigator corrects by checking the northstar, not by changing it. Instruments tell you where you are. First principles tell you where to point. A metric is always a proxy: question regularly whether it still indicates the truth it was chosen to reflect.
The VVFL only compounds when the gauge reads true. Validated — against reality, not assumption. Virtuous — serving beyond self. Feedback — output changes input. Loop — each cycle starts higher.
Commissioning is the discipline: the builder never validates their own work. Syntactically correct is not functionally correct.
The Gauge
The scoreboard is an output. You can't move it directly. You win micro-moments — collisions where standards get followed, thresholds get checked, decisions get made right.
| Level | What It Is | Example |
|---|---|---|
| Micro-moment | The collision you can win | Developer checks contrast before committing |
| Leading indicator | Signal it's compounding | Token resolution rate trending to 100% |
| Scoreboard | The number that moves | Lighthouse score hits 90+ |
| Perspective | What you chose to track | Reveals what you actually value |
What you track reveals your perspective. The ability to perceive which micro-moments matter — that's the competitive advantage.
Production Instruments
What reads reality in the live system. Each MCP or tool feeds a specific VVFL station.
| Instrument | What It Reads | VVFL Station | Threshold |
|---|---|---|---|
| Commissioning API | Capability L-levels (L0-L4) | Systems | 85 capabilities tracked |
| Playwright + intent specs | Feature verification | Value | All critical paths green |
| Sentry (planned) | Runtime errors, stack traces | Reflect | Error rate under 5% |
| PostHog (planned) | Feature usage, activation | Distribute | Active users per feature |
| Agent receipts | Skill invocations, gate pass rate | Evolve | Receipt count per skill |
Commissioning Velocity
How fast capabilities move through L0 → L4. The gauge for VVFL health.
| Metric | Current | Target |
|---|---|---|
| Total capabilities | 85 | — |
| At L4 (commissioned) | 0 (0%) | First capability by end of Phase 2 |
| At L3 (tested) | 9 (12%) | 15+ |
| Average days L0 → L2 | Not tracked | Under 14 days |
| Average days L2 → L4 | Not tracked | Under 7 days |
The bottleneck is L3 → L4. L3 requires intent specs (engineering). L4 requires independent walkthrough (human commissioner). The commissioning work chart defines who does what at each level.
Benchmark Standards
If you don't know widely accepted thresholds for "good," you have no valid context to judge reality. Show me the incentive and I'll show you the outcome.
| Minimum | Why |
|---|---|
| Industry baseline | Where does everyone start? |
| Internal baseline | Where do you start? |
| Target threshold | What does good look like? |
| Measurement method | How do you read the gauge? |
| Review cadence | How fast does the loop close? |
Questions
What is the difference between a metric that tells you where you are and a principle that tells you where to point?
- Which of your current KPIs has drifted furthest from the truth it was chosen to indicate?
- When an obstacle forces a bearing change, what is the mechanism that returns you to the northstar — and does it exist?
- If your instruments showed green while your strategy said something was wrong, which do you trust?
Context
- Commissioning — Verify the output matches the spec, not just the code
- Strategy — What the instruments are measuring against
- Pricing — The market's instrument: price signals whether reality matches strategy
- VVFL — The feedback loop that compounds when the gauge reads true
- Control System — PID mechanics: how controllers decide how hard to correct
- Consciousness — Inner space is where better intentions form before outer expression
- Incentive Engineering — Show me the incentive, I'll show you the outcome
- Standards — Thresholds that define what "good" looks like
- Predictions — Betting on which standards win