Skip to main content
L2inner-loop

Agent Platform

When agents need identity, memory, scaffold generators, and boundary enforcement to operate safely and improve autonomously — the PUMP that powers the factory.

1,200
Priority Score
Pain × Demand × Edge × Trend × Conversion
Customer Journey

Why should I care?

Five cards that sell the dream

1Why

Agents repeat, humans correct.

What's the cost of agents that can't remember?

The friction: 67 CLI commands, 542 tests, 8 auditor dimensions. But agents create files in wrong directories, repeat the same mistakes across sessions, and can't recall what worked before.

The desire: Agents that know who they are, what they can touch, and what they learned last time. Identity + memory + boundaries = trust.

The proof: Every inner-loop PRD depends on agent quality. The factory runs on agents. If agents don't improve, the factory doesn't improve.

1 / 5

Same five positions. Different seat. The operator asks "can I trust the dashboard?" The agent asks "what am I allowed to touch?"

Feature Dev Journey

How do we build this?

Five cards that sell the process

1Job

Fix the surface first.

What already exists that just needs wiring?

11 build rows across 3 jobs. Plans UI fix (3 rows). Scaffold CLI + boundaries (5 rows). Pattern extraction + memory (3 rows). ~60% wiring, ~40% new.

1 / 5
Situation

67+ CLI commands, 8 auditor dimensions, 542+ tests. But agents create files in wrong directories, repeat mistakes across sessions, and can't recall what worked before. Plans dashboard exists but math is wrong. Scaffold generators exist as functions but have no CLI surface.

Intention

One platform where agents have identity (who am I), memory (what do I know), scaffolds (how do I create), and boundaries (what can't I touch). The PUMP that powers every other inner-loop PRD.

Obstacle

Agent capability is scattered across 4 repos, 7 skill files, and 3 database schemas. No unified surface. The boundary between 'agent can do this' and 'agent must not do this' is implicit, not declared.

Hardest Thing

Agent boundaries that are too tight prevent useful work. Too loose and agents break things. The boundary must be declared per-agent, enforced by hooks, and learnable from patterns.

Priority (5P)

5/5
Pain
4/5
Demand
4/5
Edge
5/5
Trend
3/5
Convert

Readiness (5R)

Principles4 / 5
Performance2 / 5
Platform4 / 5
Process3 / 5
Players3 / 5

What Exists

ComponentStateGap
Plans dashboard UIStubPage exists at /plans. Math wrong (5 issues). No drill-down. No project grouping.
Scaffold generator functionsWorkingFunctions exist in scaffold-generators.ts. Not wired to drmg CLI. No content-type registry.
Agent boundary hooksStubOne proof-of-concept (src-post-edit.sh). No per-agent scope declarations.
Virtue auditor (pattern tracking)Working8 dimensions track trends. No cross-run extraction. No prevention proposals.
Agent memory DB schemaWorkingagent_memory_stores table with vector column exists. No write pipeline. No recall query.
DRMG CLI (67+ commands)WorkingUnified binary works. Scaffold namespace not yet added.
Agent config (.claude/agents/)WorkingAgent definitions exist. No scope declarations per agent.

Kill Signal

Boundary hooks block >30% of legitimate agent actions after 30 days. Agent task completion rate drops below current baseline.

Questions

Where does automation end and human judgment begin for agent boundaries?

  • If agents can extract their own patterns, will they converge on the same rules humans would write?
  • Should memory be per-agent or shared across all agents in a session?
  • At what point does scaffold templating become over-engineering — when does an agent just write the file directly?
  • Can boundary violations be the training signal for better scope declarations?