Sales Dev Agent Spec
How do we think slowly so the agent gets smarter and the outreach gets cleaner each cycle?
Build Contract
| # | Feature | Function | Outcome | Job | State |
|---|---|---|---|---|---|
| 1 | Prospect Research | DocumentIntelligence + industry data → prospect profile | 5 prospects in 60s, not 3 hours | Research | Gap |
| 2 | Lead Scoring | ICP fit via sales-forecasting + explore-exploit | Know which 5 of 20 deserve a call | Research | Dormant |
| 3 | Outreach Composer | Personalized draft from research + ICP pain mapping | Messages worth reading, not templates | Compose | Gap |
| 4 | SPCL Quality Gate | Score draft before human review | Never send robot-speak | Compose | Dormant |
| 5 | Follow-up Sequencer | Multi-touch cadence via CRM tasks + adapters | Follow-ups on schedule, not memory | Deliver | Gap |
| 6 | Email Delivery | Send via Resend with tracking (config only) | Know: delivered, opened, clicked | Deliver | Built |
| 7 | LinkedIn Actions | Real API + OAuth replacing mock client | Multi-channel, one system | Deliver | Partial |
| 8 | CRM Activity Logging | Auto-log every touchpoint | Full history, no data entry | Deliver | Dormant |
| 9 | Channel Performance | Send/open/reply by channel, segment, type | Learn what converts | Learn | Gap |
| 10 | Explore-Exploit | Test variants, shift to what converts | Outreach improves itself | Learn | Dormant |
| 11 | Pipeline Acceleration | Agent-sourced deals, touch-to-meeting time | Prove revenue impact | Learn | Dormant |
| 12 | Agent Profile | SOUL memory with sales playbook | Context, personality, learning | Core | Gap |
| 13 | Outreach WorkChart | Research → Score → Compose → Sequence → Measure | End-to-end orchestration | Core | Gap |
Principles
What truths constrain how this agent operates?
The Job: When the kill date is 28 days away and there are zero paying customers, help the founder find qualified prospects, score them, compose personalized outreach, and sequence follow-ups — so human time goes to discovery calls, not research and admin.
| Trigger | Current Failure | Desired Progress |
|---|---|---|
| Need 5 discovery calls this month | LinkedIn browsing, scattered email, no system | Agent surfaces, scores, drafts, sequences |
| RFP opportunity identified | Cold email from scratch each time | Personalized message from ICP data |
| Discovery call completed | Notes in notebook, follow-up forgotten | Auto-log, create tasks, update probability |
| No pipeline visibility | Gut feeling about who's warm | Scored leads with recommendations |
| Outreach unanswered | Same message to everyone | Explore-exploit adapts channels and copy |
| Agency needs qualified leads | Volume providers, unqualified contacts | Pre-qualify through conversation before handover |
Hidden objection: "AI outreach feels spammy." Agent composes — human reviews every message before send. Research is automated, relationships are not.
Why now: The product is ready. The demand generation isn't. Nine algorithms, four adapters, a CRM, an RFP tool — all built, none generating demand. The outreach IS the commissioning test.
The convergence: Eddy Whatt at Sneakers Media identified the same pattern from the buy side. Every lead gen provider delivers volume, not quality. Lead gen is a data flow problem dressed as a marketing problem:
OLD: Advertiser → Agency → Lead Gen Provider → Database → Leads (unqualified)
NEW: Advertiser → Agent → Pre-Qualified Leads (with conversation context)
The NZ angle: Small market (5M) makes lead gen quality critical. Relationship-based culture means bad leads damage agency reputation. Proving ground.
Performance
How do we know it's working?
Priority Score
| Dimension | Score | Evidence |
|---|---|---|
| Pain | 4 | 9 algorithms x 0 UI = 0 value. 28 days to kill date. 3hrs/week manual research. Sneakers confirms volume-not-quality. |
| Demand | 4 | Sneakers independently identified same problem. 5 test clients (HRV, Kitchen Studio, Asset Factory, BizCover, MAS). Internal dogfood demand. |
| Edge | 3 | Three Flows convergence. Existing platform (9 algos, CRM, adapters) = composition advantage. No proprietary data yet — edge grows with usage. |
| Trend | 5 | AI SDR is hottest B2B SaaS category. Every lead gen provider will be AI-native within 2 years. |
| Conversion | 2 | CPL model clear. No pricing tested, no pilot signed, no invoice sent. Path plausible, unproven. |
| Composite | 480 | Strong candidate. Demand and trend validated, conversion needs proof. |
Success Criteria
Functional:
| # | Criterion | Verification | Job |
|---|---|---|---|
| F1 | Prospect profile from company name in <60s | Timed test | Research |
| F2 | 10 prospects ranked by ICP fit with reasoning | Output check | Research |
| F3 | Draft scores >3.5 SPCL before human review | Automated gate | Compose |
| F4 | Follow-up creates CRM tasks with correct timing | CRM check | Deliver |
| F5 | Email delivery + open tracking end-to-end | Resend webhook | Deliver |
| F6 | Channel dashboard shows real data within 24h | Data check | Learn |
Outcome (Kill: 2026-03-24):
| # | Criterion | Target | Now |
|---|---|---|---|
| O1 | Prospects researched and scored/week | >= 10 | 0 |
| O2 | Reply rate on agent-drafted outreach | >= 15% | N/A |
| O3 | Scored contacts in CRM pipeline | >= 20 | 0 |
| O4 | Follow-up completion rate | >= 90% | N/A |
Team-level outcomes (calls booked, pipeline value, first customer) tracked in SPO.
Business Dev
| Layer | Decision | Assumption | Evidence Needed |
|---|---|---|---|
| ICP | Who first? | Construction/solar EPC, 10+ staff, active RFPs, spreadsheet pain | 5 calls where they name the pain unprompted |
| Offer | Opening message? | "Win more bids. Your answer library compounds." | Reply rate >15% |
| Channel | How to reach? | LinkedIn + personalized email | Compare reply rates by channel |
| Proof | What earns a meeting? | Industry RFP pain + live auto-fill demo | Meeting-to-pilot >30% |
| Conversion | What closes? | Free 30-day pilot on one real RFP | Pilot-to-paid >20% |
Platform
What do we control directly?
Components
This agent is one instrument in the SPO orchestra.
| Component | Role | State |
|---|---|---|
| CRM Contacts + Deals | Prospect database, deal pipeline | Live |
| CRM Activities + Tasks | Activity logging, follow-up tracking | Live |
| Sales Forecasting Algo | Lead scoring, pipeline prediction | Wire |
| Explore-Exploit Algo | Channel/message optimization | Wire |
| SPCL Scoring Algo | Outreach quality gate | Wire |
| Compound Rate Tracking | Engagement velocity | Wire |
| Email Adapter (Resend) | Delivery + tracking | Live |
| LinkedIn Adapter | Post/connect/message + analytics | Partial |
| Document Intelligence | Prospect research, doc analysis | Live |
Build ratio: ~80% composition, ~20% new code.
Commissioning
| Component | Schema | API | UI | Tests | % |
|---|---|---|---|---|---|
| Prospect Research | Pending | Pending | Pending | Pending | 0% |
| Lead Scoring | Pending | Pending | Pending | Pending | 0% |
| Outreach Composer | Pending | Pending | Pending | Pending | 0% |
| SPCL Quality Gate | Done | Pending | Pending | Pending | 15% |
| Follow-up Sequencer | Pending | Pending | Pending | Pending | 0% |
| Email Delivery | Done | Done | Done | Done | 95% |
| LinkedIn Actions | Done | Done | Done | Pending | 85% |
| CRM Activity Logging | Done | Done | Partial | Pending | 50% |
| Channel Performance | Pending | Pending | Pending | Pending | 0% |
| Explore-Exploit | Done | Pending | Pending | Pending | 15% |
| Pipeline Acceleration | Partial | Partial | Pending | Pending | 10% |
| Agent Profile | Pending | Pending | N/A | Pending | 0% |
| Outreach WorkChart | Pending | Pending | N/A | Pending | 0% |
Protocols
How does the system coordinate?
Agent/Human Split
AGENT (AI-Led + AI-Only) HUMAN (Human-Led + Human-Only)
──────────────────────── ────────────────────────────
Prospect research Discovery calls
Lead scoring + ranking Relationship building
Outreach drafting Message review + send
Follow-up sequencing Deal negotiation
Activity logging Go/No-Go decisions
Channel optimization Trust building
WorkChart
RESEARCH → SCORE → COMPOSE → SEQUENCE → MEASURE
↑ |
└──── Feedback improves next cycle ──────┘
| Score | Route | Agent | Human |
|---|---|---|---|
| Hot (>80%) | Direct | Personalized email + LinkedIn connect | Review, personalize, send |
| Warm (50-80%) | Nurture | 3-touch sequence | Review sequence, approve |
| Cold (<50%) | Content | Distribution list, no direct outreach | Monthly review for warming |
Build Sequence
| Sprint | What | Effort | Depends On | Acceptance |
|---|---|---|---|---|
| -1 | Prove channels: Resend config, LinkedIn OAuth, CRM auto-log | 3d | — | Real email delivered + opened. Real LinkedIn post. Both logged to CRM. |
| S0 | Agent profile + prospect research | 3.5d | Agent Platform ETL | "Acme Construction" → profile with pain points in <60s |
| S1 | Score + compose + quality gate | 5d | S0 | 10 ranked, draft >3.5 SPCL, references prospect pain |
| S2 | Sequencer + WorkChart orchestration | 5d | Sprint -1, S1 | Company → end-to-end with human review step |
| S3 | Channel performance dashboard | 1d | Webhooks live | Real send/open/reply within 24h |
| Park | Explore-exploit + pipeline acceleration | — | 50+ messages | Volume data needed |
Horizon Model
HORIZON 1: PROVE (now → 90 days) ← THIS PRD
Sales Dev Agent — one instrument
│ feeds
▼
HORIZON 2: ORCHESTRATE (90d → 6 months)
Sales Process Optimisation — the orchestra
│ generates data for
▼
HORIZON 3: PROTOCOL (6m → 18 months)
Trust Commerce — the protocol
Risks
| Risk | Mitigation |
|---|---|
| LinkedIn app approval delayed | Register day 1 of Sprint -1. Manual-publish fallback. |
| AI outreach feels spammy | Human reviews every message. SPCL gate enforces floor. |
| Construction ICP unresponsive | Test solar EPC in parallel. Explore-exploit switches. |
| Agent slower than manual | Start manual now. Agent replaces when ready. |
| Calls don't convert to pilots | Product/pitch problem, not agent. Separate kill signal. |
Kill signal: 50 messages, 0 replies, 30 days — ICP or message is wrong. Diagnose with explore-exploit data before iterating.
Players
Who creates harmony?
Job 1: Find Prospects Worth Talking To
| Element | Detail |
|---|---|
| Struggling moment | 3 hours to find 5 prospects, then equal time on high-fit and low-fit |
| Workaround | LinkedIn search, industry events, chase whoever responds |
| Progress | Agent surfaces 10 scored/week ranked by ICP fit with recommended action |
| Hidden objection | "AI-found prospects won't be as good as my network referrals" |
| Switch trigger | Pipeline empty AND 4 of 5 calls with wrong buyer persona |
Job 2: Say Something Worth Reading
| Element | Detail |
|---|---|
| Struggling moment | Blank compose window, writing same ineffective message |
| Workaround | Copy template, change name, hope |
| Progress | Agent drafts from research + pain points, human reviews and sends |
| Hidden objection | "AI messages sound like AI messages" |
| Switch trigger | 50 emails, 0 replies |
Job 3: Follow Up Without Forgetting
| Element | Detail |
|---|---|
| Struggling moment | Meant to follow up Wednesday, it's Friday, window closing |
| Workaround | Mental note, calendar reminder, hope |
| Progress | Agent schedules: Day 3 email, Day 7 LinkedIn, Day 14 value-add |
| Hidden objection | "Automated follow-up feels like spam" |
| Switch trigger | Warm prospect goes cold because nobody followed up for 2 weeks |
Job 4: Learn What Converts
| Element | Detail |
|---|---|
| Struggling moment | Month of outreach, no idea what worked |
| Workaround | Remember which emails "felt good" |
| Progress | Dashboard: sent, opened, replied, meetings — by channel, segment, type |
| Hidden objection | "Not enough volume for data to be meaningful" |
| Switch trigger | Same approach keeps failing, no data to diagnose |
Job 5: Pre-Qualify for Media Clients
| Element | Detail |
|---|---|
| Struggling moment | Every lead gen campaign: volume not quality. Client sales team complains. Agency absorbs blame. |
| Workaround | Pad volume, rotate providers, absorb complaints |
| Progress | Agent qualifies via conversation before handover — leads arrive with context |
| Hidden objection | "Every AI tool promises better leads — heard it before" |
| Switch trigger | Client threatens to pull media budget because leads waste their time |
ICP: NZ Media Agency
Archetype. Sneakers Media exemplifies the segment.
| Attribute | Specification |
|---|---|
| Role | Agency owner/MD or media director |
| Context | NZ media agency, 5-50 staff, $500K-$20M annual media spend |
| Geography | New Zealand, primarily Auckland and Wellington |
| Shared Pain | Lead gen = volume, not quality. Client teams complain. Agency reputation hit. |
Psycho-logic:
| They Say | They Mean |
|---|---|
| "Lead quality is always poor" | "We get blamed after handover" |
| "We need to test it first" | "Show proof, not promises" |
| "Our client's sales team complains" | "When they complain, our contract is at risk" |
The real problem: Not "higher quality leads" — leads their client's sales team doesn't complain about. The complaint loop costs more than the CPL.
Revenue model: Pilot CPL ($20-100/lead) → Platform competing with lead gen providers → Partnership (media + creative + AI/data).
Two Deployment Contexts
Same WorkChart, different jobs. One instrument, multiple ventures. BOaaS proven.
| Context | ICP | Job | Revenue Path |
|---|---|---|---|
| Internal (dogfood) | Construction/solar EPC | Discovery calls for Stackmates | Product validation + first customer |
| External (Sneakers) | NZ homeowners via HRV | Pre-qualified leads for agency clients | CPL from existing media budgets |
Context
- Sales CRM & RFP — The convert/retain half
- Sales Process Optimisation — The orchestra (Horizon 2)
- Trust Commerce — The protocol (Horizon 3)
- The Three Flows — Advertising, data, AI convergence
- Advertising Industry — The convergence thesis in context
- Sales Work Chart — Human/AI split
- Lead Qualification — Pre-qualification method
Questions
If lead gen is a data flow problem dressed as a marketing problem, what breaks when the data flow is better than the marketing?
- When does the agent's qualifying conversation become the product — not a feature of the product?
- What data from Horizon 1 (this agent) must flow to Horizon 2 (SPO) for the orchestra to tune itself?
- If the NZ proving ground works, does it prove the model or prove the geography?