The Coding Canary
What happens when elite programmers stop looking at the code they generate—and become more productive than ever?
"I haven't seen a line of code in nine months."
This isn't a confession of laziness. It's not the admission of a hobbyist or someone who stepped away from the keyboard. I've written more code this year than in my entire career—thousands of lines of production software, complex integrations, entire applications from scratch. And I haven't read most of it.
The statement sounds impossible, perhaps even irresponsible. How can you code without seeing code? How can you ship software you haven't reviewed line by line? Yet this is precisely what's happening at the cutting edge of software development. And the pattern emerging there will reshape every knowledge-work interface in the next five years.
The Elite Programmers Are Doing It Too
I'm not alone in this shift. Some of the most accomplished programmers in the world are publicly describing the same experience—and they're processing it in real time, with all the nuance and contradiction that implies.
"Vibe coding is maybe the most democratizing idea of the whole year. It's this threshold we crossed where people could build really impressive programs just by explaining what they wanted in plain English."
— Andrej KarpathyAndrej Karpathy—former director of AI at Tesla, OpenAI co-founder, and one of the most respected minds in machine learning—coined the term "vibe coding" in February 2025. He described it as fully giving in to the vibes, embracing exponentials, and forgetting that the underlying code even exists.
But Karpathy's relationship with this new paradigm is more nuanced than the viral term suggests. By year's end, when reflecting on his 8,000-line nanochat project, he admitted something striking:
"It's basically entirely hand-written (with tab autocomplete). I tried to use claude/codex agents a few times but they just didn't work well enough at all and net unhelpful." — Andrej Karpathy, on his nanochat project1
This isn't a contradiction—it's a crucial distinction. Vibe coding works brilliantly for exploratory work, rapid prototyping, and certain kinds of application development. For production foundations—the load-bearing walls of a system—the pattern is different. The lesson isn't that AI coding tools don't work; it's that knowing when to supervise versus when to hand-craft is the new core skill.
The Numbers from Elite Practice
Theo Browne—500,000+ subscribers, elite developer, founder of multiple successful projects—documented what this looks like in practice:
lines of production code generated without opening an IDE
Source: Medium, "Claude Code is Redefining Software Engineering"2
of personal code now AI-generated
Source: LinkedIn3
Browne's workflow isn't "write less code." It's "write differently." More time in taste, judgement, and product thinking. Less time in keystrokes. The phrase he uses: "Stop micromanaging, describe outcomes."
From Code Producer to Creative Director
What's happening isn't just a productivity boost—it's a redefinition of the role itself.
"The role of 'software developer' just changed. Developers describe their role less as 'code producer' and more as 'creative director'. The shift: Strategy (define what's needed) → Delegation (let AI draft) → Verification (ensure it works)." — LinkedIn, 20254
This three-part workflow—Strategy → Delegation → Verification—is the new shape of knowledge work. It's not unique to coding. It's a pattern that will propagate across every domain where humans produce artifacts from intent.
The Abstraction Layer
What made this possible? AI tools created a new abstraction layer between intent and artifact.
Claude Code
Shows summaries and test results, not line-by-line code. You see what changed and whether it works.
Cursor
Presents diffs, not files. You review deltas from current state to desired state.
GitHub Copilot
Suggests completions in context. You accept, reject, or iterate—but you don't type from scratch.
The interface shifted from "here's all the code" to "here's what changed and whether it works." This is the fundamental move: from artifact display to outcome verification.
The New Workflow Pattern
The Workflow Transformation
❌ Old Workflow
✓ New Workflow
The code exists. But you don't need to look at it. You're supervising an execution layer, not hand-crafting artifacts. The tests pass. The behavior is correct. The diff is clean. Why would you read every line?
Why "Not Looking at Code" Works
There's a counterintuitive insight hidden in the senior developer experience: AI's first pass is often genius-level. Fresh generation from a clear prompt produces surprisingly coherent, well-structured code.
But editing existing code? That's where AI struggles. It tacks on bits. It misses holistic design. It adds a date field by squishing it into the corner rather than redesigning the whole page.
delete_code();
regenerate_from_spec();
} // Don't patch. Rebuild.
This inverts the traditional relationship between code and design. In the old world, code was the asset—you accumulated it carefully, protected it, maintained it. In the new world, design is the asset. Code is ephemeral. Specifications and acceptance tests are durable. Code can be regenerated whenever the spec changes.
This is why delete-and-regenerate beats patch-and-accumulate. It's why senior developers trust the AI's first pass but get suspicious when they're asking it to "just add this one thing."
The Canary in the Coal Mine
Coding got here first for three reasons:
Economic Focus
AI companies invested heavily in coding tools early. Good use case, revenue from API calls, visible ROI.
Early Adopters
Developers are the natural first users of AI tools. They understand the technology and have high tolerance for experimentation.
Clear Success Criteria
Code either works or it doesn't. Tests pass or fail. This makes verification straightforward.
But the pattern isn't unique to coding. It will propagate to every data-heavy, artifact-producing domain. Anywhere humans currently navigate raw information to produce an output, the same shift will occur.
The Question This Raises
If senior programmers no longer need to see code to code...
- → Do CRM users need to see 80 fields to sell?
- → Do support agents need to see 3 years of ticket history to resolve an issue?
- → Do analysts need to navigate 50 reports to make a decision?
- → Does anyone need to see all the data to do their job?
The coding world stopped looking at the artifact.
It started looking at the outcome.
What happens when every other knowledge-work interface does the same?
Key Takeaways
- 1 Elite programmers have stopped reading the code they generate—and they're more productive. The code exists; they just don't need to see it.
- 2 The shift is from "code producer" to "creative director"—Strategy → Delegation → Verification. The role has been redefined.
- 3 Code is now ephemeral; design and intent are the durable assets. Delete and regenerate beats patch and accumulate.
- 4 Coding is the early warning signal for every other knowledge-work interface. The same pattern will propagate.
CRMs Are Databases With Cosplay
The numbers reveal a brutal truth: your sales team spends most of their time navigating, not selling.
of time spent actually selling
navigating and administering
Source: Revenue Grid, "How to Customize Salesforce for Maximum Efficiency"6
Sales reps spend only 34% of their time actually selling. The rest—nearly two-thirds of every working day—is consumed by administrative tasks and navigating poorly optimized systems. For a team of 10 reps, that translates to over 25,000 hours annually spent on activities that don't close deals.
This isn't a people problem. It's not a training problem. It's not even a data problem. It's an interface problem. The tools we've given salespeople were designed for a different era—and they're drowning in them.
The Time Sink Quantified
hours per year
wasted on manual data entry per rep
Source: LinkedIn/xdge7
Let's put that 66% into concrete terms. The average rep wastes 286 hours per year on manual data entry alone—not selling, not building relationships, not closing deals. Just typing data into fields.
That's over seven weeks of full-time work. Every year. For each rep. Just entering data.
And even after all that typing, the data is incomplete. The system designed to capture your customer relationships is leaking context at every interaction.
The Context Loss Problem
Here's a number that should alarm every revenue leader:
words spoken
per sales call
words captured
in the CRM
Less than 1% of the conversation makes it into your database.
Source: Gong, "Why AI Systems of Action Replace Traditional CRM Workflows"8
An average sales call generates about 6,000 words of conversation. The prospect mentions their procurement timeline. The budget holder raises a concern. The champion reveals internal politics. The decision-maker hints at a competing priority.
What makes it into the CRM? "Good call. Follow up next week."
Your database is filled with surface-level summaries that miss what actually matters. The nuance, the objections, the relationship signals—all lost. And when the rep changes roles, or the account gets transferred, that context evaporates entirely.
"CRM should be a profit center, not a cost center. If Salesforce only records activity, it is an expense. If it enforces revenue flow, it becomes an asset." — Ameet Patra, LinkedIn9
What CRMs Were Actually Designed For
Traditional CRMs were built on a simple premise: show the user all the data, let them figure out what to do.
The CRM presents itself as a "system of record"—a faithful mirror of reality. Every interaction logged. Every field populated. Every relationship documented. The interface is a database with a pretty face.
The Traditional CRM Navigation Model
Eight steps. And the user is responsible for synthesis at every transition.
Why This Made Sense (Once)
Before AI, humans were the only processors of context. You needed to see the data to make sense of it. The assumption was straightforward: more fields = more information = better decisions.
The CRM was a filing cabinet you could search. A shared memory for the organization. A place to store everything so nothing got lost.
And for its time, this was progress. Better than spreadsheets. Better than post-it notes. Better than each rep keeping their own mental database of customers.
Why This Fails Now
But humans aren't good at synthesizing 80 fields and 3 years of history. We're terrible at it, actually. We miss things. We get tired. We see what we expect to see. We forget what we read two screens ago.
Power Users
Develop workarounds. Ignore most fields. Build mental shortcuts. Create their own systems alongside the official one.
New Users
Overwhelmed from day one. "Where do I even look?" Training takes months. Mistakes are inevitable.
The interface forces cognitive labor that could be automated. Every click, every scroll, every "where was that field again?" is friction between intention and action.
System of Record vs System of Action
The distinction that matters:
System of Record
"Here's everything that happened"
- • Mirror of reality
- • Comprehensive history
- • All fields visible
- • User synthesizes
An expense if it only records.
System of Action
"Here's what you should do next"
- • Decision engine
- • Relevant context surfaced
- • Proposals prepared
- • System synthesizes
An asset if it enforces revenue flow.
Most CRMs are systems of record pretending to be systems of action. They have "next best action" features bolted on the side—an AI suggestion widget floating in the corner of the same old record view. It's an afterthought, not an architecture.
The Interface Is the Problem
It's not the data that's wrong. Companies have invested millions in data quality, in integrations, in enrichment services. The data is better than ever.
It's how the data is presented. Showing everything means the user has to filter everything. The cognitive load sits on the human, not the system.
Every click is friction.
Every scroll is friction.
Every "where was that field?" is friction.
Every tab-switch is friction.
Every manual synthesis is friction.
Friction accumulates. It's the hidden tax on every interaction. And the bill comes due in that 66% of time spent not selling.
The Parallel to Coding
Here's where Chapter 1 connects to the present. Just as code was a proxy for intent—lines and functions standing in for desired behavior—fields are a proxy for understanding. They're the low-level representation of something higher: customer relationships, deal momentum, buying signals.
The Parallel
In Coding
Developers stopped looking at code and started looking at outcomes.
In CRM
Users need to stop looking at fields and start looking at decisions.
The artifact (record) isn't the point. The action is.
CRMs are databases wearing a costume.
It's time to redesign for what users actually need:
not "show me everything"
but "tell me what to do next."
Key Takeaways
- 1 Sales reps spend 66% of their time NOT selling—navigating and administering. The interface is the bottleneck.
- 2 6,000 words of conversation → 30-60 words in CRM. Massive context loss at every interaction.
- 3 CRMs were built as systems of record; users need systems of action. The distinction is architectural, not cosmetic.
- 4 The interface is the problem, not the data. Showing everything forces humans to filter everything.
From Record Navigation to Decision Navigation
The shift isn't about better UI. It's about a different question entirely.
Record Navigation
"What do you want to see?"
Decision Navigation
"What do you want to decide?"
The shift happening in coding isn't coding-specific. It's a change in the fundamental unit of interaction—and it applies to every knowledge-work interface.
This isn't just cleaner UI. It's not better color schemes or more intuitive layouts. It's a different design primitive altogether. And understanding the distinction is the key to designing interfaces for the AI era.
The Framework
| Dimension | Record Navigation | Decision Navigation |
|---|---|---|
| Primary Question | "What do you want to see?" | "What do you want to decide?" |
| Primary Object | The record | The proposal |
| User's Job | Navigate to find context | Evaluate the recommendation |
| AI's Job | Search and retrieve | Synthesize and recommend |
| Detail Model | Default (show everything) | On-demand (show when needed) |
| Cognitive Load | On the human | On the system |
| Success Metric | "Did you find it?" | "Did you decide well?" |
This framework is the core lens for everything that follows. Subsequent chapters apply this distinction to specific patterns and implementations.
What Record Navigation Looks Like
In record navigation, the user is a spelunker—searching for relevant information in a sea of data.
The Spelunking Workflow
Every decision requires assembly. Pull from here. Pull from there. Synthesize mentally. Hope you didn't miss anything.
The system is a warehouse. The user is the picker.
What Decision Navigation Looks Like
In decision navigation, the AI did the spelunking. The user is a governor—reviewing the synthesis, not creating it.
The Governance Workflow
Every decision is pre-assembled. "Here's what to do. Here's why. Here's the evidence."
The system is a recommender. The user is the governor.
The Design Primitive Shift
Old Primitive
The Record
A container of fields. All data visible. User navigates.
New Primitive
The Proposal
A recommendation with evidence. Detail on-demand. User governs.
This changes everything: layout, workflow, what's visible by default, what's hidden. It's not about removing features—every field is still accessible if you need it. It's about changing the default. Most of the time, you don't need it.
Why This Isn't Just "Better UX"
Cleaner UI
- • Same model, fewer distractions
- • Record view, but prettier
- • User still navigates
- • Cognitive load unchanged
Polishing the same paradigm.
Decision Navigation
- • Different model, different interaction
- • Proposal view with drill-down
- • User governs, not navigates
- • Cognitive load on system
Replacing navigation with supervision.
The user's role changes from finder to judge. That's not a cosmetic difference. It's a fundamental redefinition of what the interface is for.
The Direction: Generative UI
Jakob Nielsen—the godfather of usability research—made a striking prediction for 2026:
"The concept of a static interface where every user sees the same menu options, buttons, and layout, as determined by a UX designer in advance, is becoming obsolete. Software interfaces are no longer hard-coded; they are drawn in real-time based on the user's intent, context, and history." — Jakob Nielsen, "18 Predictions for 2026"10
Nielsen calls this "Generative UI"—interfaces that assemble themselves based on what the user is trying to accomplish. Not one-size-fits-all screens, but bespoke micro-interfaces that dissolve when the task is complete.
Decision navigation is the first step toward this future. Instead of hard-coded record views, you get contextual proposal cards. The interface responds to intent, not just clicks.
The Pattern: Hide the Artifact, Surface the Summary
The Same Pattern, Two Domains
In Coding
Hide: The code
Surface: The diff summary
Drill-down: Full file when suspicious
In CRM
Hide: The fields
Surface: The proposal card
Drill-down: Full record when needed
When the AI is right, you move fast. When it's wrong, you can quickly see where.
What This Means for Design
If you're designing interfaces, the shift requires a new starting point:
Stop designing screens around records.
The record is the artifact, not the goal.
Start designing screens around decisions.
What action does the user take from here?
Ask: What do they need to see to act confidently?
Not everything—just what supports the decision.
Everything else is drill-down.
Available, but not blocking the flow.
The Cognitive Load Transfer
The fundamental shift is where cognitive load lives:
Record Navigation
User does synthesis
(High cognitive load)
Decision Navigation
System does synthesis
(Low load on user)
This isn't about making users dumber. It's about making systems smarter. Humans are better at judgment than navigation. Let them do what they're good at.
Record navigation asks:
"What do you want to see?"
Decision navigation asks:
"What do you want to decide?"
The second question is the right question.
Key Takeaways
- 1 The shift is in the unit of interaction: from record to proposal. The primary object changes; everything follows.
- 2 Record navigation puts cognitive load on humans; decision navigation puts it on systems.
- 3 This isn't just cleaner UI—it's a different design primitive. Cleaner record views miss the point.
- 4 Design around decisions, not records. Everything else is drill-down.
Two-Pane Cognition
The architecture for human-AI pair interfaces: what if human and AI each had their own workspace—and they talked to each other?
"What if the human and AI each had their own pane—
and they talked to each other?"
We've established what we're shifting to: decision navigation instead of record navigation. Now we need how. What's the architecture for human-AI pair interfaces? How do you actually design a screen where human and AI work together?
The answer isn't a chatbot bolted onto your existing interface. It's not an AI assistant widget floating in the corner. It's a fundamentally different screen architecture12—one that gives both human and AI their own space to work, with a clear contract between them.
I call it Two-Pane Cognition.
Where This Comes From
The pattern emerges from observation: how do senior engineers actually work with AI coding tools?
They don't read all the code. As we saw in Chapter 1, they've moved past that. Instead, they:
Read the proposal (the diff summary, the suggested change)
Check the evidence (tests pass, behavior correct)
Make a judgment (approve, reject, or modify)
The tool does preparation. The human does decision.19 That's the pattern we're extracting—and it applies far beyond coding.
The Two-Pane Model
Two-Pane Cognition divides the interface into three distinct areas, each with a clear purpose:
Human Pane
What only humans can decide
Outcomes
What are we trying to achieve?
Priorities
What matters most right now?
Approvals
Yes / no / modify on proposed actions
Exceptions
Edge cases and judgment calls
High-signal summaries
The "so what?" distillation
AI Pane
What AI prepares
Drafts
Pre-written artifacts ready to use
Candidate actions
"Here's what I think you should do"
Retrieved evidence
The specific data points used
Uncertainty signals
Confidence levels, what's missing
"Here's what I tried"
The reasoning trace
Shared Surface
The contract between human and AIConstraints
What AI is NOT allowed to do
Permissions
What tools/actions AI can invoke
Acceptance Criteria
How we know "good enough"
Feedback Handles
Where to correct when wrong
How This Looks in CRM
Let's make this concrete. Compare a traditional CRM screen to a Two-Pane CRM screen:
❌ Traditional CRM Screen
Activity Timeline (200 entries)
User navigates → synthesizes → decides. Cognitive load on human.
✓ Two-Pane CRM Screen
Human Pane: Decision Queue
"These accounts need attention today"
AI Pane: Proposal for Acme Corp
Action: Call John, ask about Q2 budget
Why: 17 days since touch • Champion role change
Draft: Call agenda ready →
Evidence: 3 items ▼
"Here's what I propose, here's why"
AI prepares → human governs. Cognitive load on system.
The Interaction Flow
Here's how the two panes interact:
Human selects from decision queue
The queue is prioritized by AI based on urgency, opportunity, and strategy alignment.
AI pane shows the proposal
Proposed action + draft artifact + reasoning + evidence. Everything needed to decide.
Human reviews and responds
Approve / modify / reject / ask for more. The human governs; the AI awaits instruction.
If approved: action executes
Logged with full trace. The system knows what happened, why, and who approved it.
If rejected: feedback captured
AI learns at the right layer. The playbook improves, not just this one interaction.
Why "Bolted-On Assistants" Fail
Most CRM AI today is a chatbox in the corner of the same old record screen. The assistant is an add-on, not a co-pilot.16 You still navigate the record; the AI just answers questions about it.
"The assistant bolted onto the old UI feels lame because it's a chatbox glued onto a cockpit—but the cockpit wasn't designed for a co-pilot that can actually fly."
The fundamental problem: the interaction model hasn't changed. You're still spelunking through records. The AI is just a slightly faster way to search.
Two-Pane Cognition changes the model itself. The AI doesn't just answer questions—it prepares proposals. The human doesn't navigate—they govern.17
The Right Division of Labor
| AI Does Well | Human Does Well |
|---|---|
| Retrieval — finding relevant information in vast datasets14 | Judgment — knowing when rules don't apply |
| Synthesis — combining data points into summaries | Exceptions — handling edge cases with nuance |
| Drafting — generating first-pass artifacts | Priorities — deciding what matters most |
| Pattern-matching — recognizing similar situations | Relationships — understanding human dynamics15 |
| Consistency — applying rules uniformly | Strategy — setting direction and goals |
The principle: don't ask AI to do what humans do well, and don't ask humans to do what AI does well.13 Navigating 200 activities to find the relevant one? AI's job. Deciding whether this account needs a different approach because the champion just left? Human's job.
Example: A Sales Rep's Morning
❌ Old Workflow
- 1.Open CRM
- 2.Check dashboard
- 3.Pick an account
- 4.Open record
- 5.Read history
- 6.Read emails
- 7.Check contacts
- 8.Decide what to do
- 9.Draft email
- 10.Send
- 11.Log activity
~20 minutes per account
✓ Two-Pane Workflow
- 1.Open CRM → see prioritized decision queue (AI prepared overnight)
- 2.Select top account → see proposal card
- 3.See evidence cards: "17 days since touch; champion changed roles"
- 4.See draft: call agenda already written
- 5.Approve → call happens → AI logs with trace
~3 minutes per account
Designing for Two-Pane
When designing a Two-Pane interface, work through these questions in order:
1. What decision does the user make here?
This is your human pane. The decision queue. The judgment point.
2. What does the user need to see to decide confidently?
This is your proposal structure. Action, reasoning, evidence.
3. What can AI prepare in advance?
This is your AI pane. Drafts, synthesis, retrieved context.
4. What's the minimum human input to approve/modify/reject?
This is your interaction design. One click? Edit field? Selection?
The human is still in control.
But the human isn't navigating.
The human is supervising.
Key Takeaways
- 1 Two-Pane Cognition: Human pane (decisions) + AI pane (proposals) + Shared surface (contract).
- 2 Right division of labor: AI does retrieval, synthesis, drafting. Human does judgment, exceptions, priorities.
- 3 Bolted-on assistants fail because they don't change the interaction model. Chatbox ≠ co-pilot.
- 4 Design screens around decisions, not records. Start with "what judgment does the user make here?"
Proposal Cards: The New Primitive
Every card has four parts. Learn them, and you can redesign any interface.
We've established the architecture: Two-Pane Cognition, with human and AI each owning their own space. We've distinguished record navigation from decision navigation. But what, exactly, does the AI pane show?
In record navigation, the primitive is obvious: the record. A container of fields. All data visible. User navigates. This is the atom of traditional enterprise software—every screen is built from records, related records, lists of records.
Decision navigation needs its own atom. Its own fundamental unit. Something as universal as the record, but designed for human-AI pairs instead of humans alone.
That unit is the proposal card.
"A record is a container of data. A proposal card is a container of decision."
The Anatomy of a Proposal Card
Every proposal card has exactly four components. Not three. Not five. Four. Each serves a distinct function, and removing any of them breaks the system.
The Four Components
ACTION
What the AI suggests doing. The headline. What you're being asked to approve.
"Call John tomorrow at 2pm"
"Send follow-up email"
"Update stage to Negotiation"
"Schedule demo with technical team"
DRAFT
The prepared artifact. Already written. Ready to send or use.
Personalized email copy
Call agenda with talking points
Proposal outline with key sections
Meeting invite with context
WHY
The reasoning. Compact, scannable, decision-relevant. Shows the logic without verbose explanation.
"17 days since last touch"
"Procurement mentioned Q1 budget"
"Champion changed roles"
"Competitor mentioned in last call"
EVIDENCE
The specific data points used. Collapsible, on-demand. The receipts for verification.
Link to email where budget mentioned
Snippet from call transcript
LinkedIn update showing role change
CRM field: "Last Activity: 17 days ago"
Why Four Components?
This isn't arbitrary. Each component serves a function that, if missing, causes the system to fail:
Without this, there's nothing to approve. The card is pointless.
Without this, the human does the work. The AI prepared nothing. What was the point?
Without this, AI is a black box. Trust fails. "Why should I do this?" goes unanswered.
Without this, you can't verify. Governance fails. Skeptics are right to reject.
Adding more components creates cognitive overload.22 Removing any creates a gap that undermines the whole pattern. Four is the magic number—not by accident, but by function.
A Proposal Card in the Wild
Let's see what this looks like in practice. Here's a proposal card as it might appear in a sales rep's morning queue:
1. Congratulate on new VP role
2. Re-confirm Q2 budget availability
3. Address SSO integration questions from Dec call
Notice what's present and what's absent. The rep sees everything needed to decide in seconds: who, what, why. The full agenda expands on click. The evidence is available but collapsed. The full account record—all 80 fields—is nowhere to be seen.
That's not hiding information. That's progressive disclosure20: showing what you need at each level, with more detail available on demand.
The Progressive Disclosure Principle
This is perhaps the most important insight in proposal card design: detail is progressive, not default.
The Three Layers of Disclosure
Default View
~5 seconds to scanAction + Why — The headline and the reasoning. "Call John because 17 days and budget timing." Decision-ready at a glance.
One Click
~30 seconds to reviewDraft + Evidence summary — The prepared artifact and the receipts. "Here's the full agenda. Here's where I got the budget info."
Another Click
As long as neededFull record + complete history — All 80 fields. Every activity. The full timeline. Everything, if you want it.
Most of the time, you never need to go past Layer 1. The Action and Why are enough. You approve and move on. Layer 2 is for when something feels off and you want to verify. Layer 3 is for the rare edge case where you need the complete picture.
"Show the decision-ready summary. Drill down when suspicious. Full record is one click away but not blocking the flow."
This is how trust scales. You don't verify everything by default—that's exhausting and slow. You verify when you need to, and the system makes verification easy.23 Trust but verify, with verify on demand.
The Coding Tool Parallel
This pattern will feel familiar if you've used modern coding tools. They follow the exact same structure:
In Coding Tools
In Proposal Cards
The pattern is identical because the problem is identical. Developers were drowning in code the way CRM users are drowning in fields. The solution in both domains: summarize by default, drill down on demand.
Most of the time, the diff summary is enough. You see "+42 lines, tests passing, no conflicts" and you merge. You don't read every character. The same economy applies to proposal cards: most of the time, "Call John because 17 days" is enough. You don't need all 80 fields.
The User's Decision Tree
When a user reviews a proposal card, they're running through a mental checklist. Understanding this checklist helps you design better cards:
The Approval Flowchart
The key insight: most cards pass steps 1-3 quickly. Evidence drill-down is the exception, not the rule. When the AI is well-calibrated, the majority of cards get approved at Layer 1 in seconds.
This is the productivity gain: not doing everything faster, but not doing most things at all. The AI did the work. The human governs.
When Cards Get Rejected
Rejection isn't failure—it's feedback. And how you capture rejection feedback determines whether the system learns.
The critical principle: feed back at the right layer. If the playbook is wrong, fix the playbook—don't just reject this one card and leave the playbook broken. That's how systems learn.
This mirrors the coding insight from Chapter 1: fix the spec, not the code. When AI generates wrong code, you don't patch the code—you clarify the specification and regenerate. When AI generates wrong proposals, you don't just skip this one—you update the playbook so all future proposals improve.
The Confidence Signal
Some teams add a fifth element to their proposal cards: a confidence indicator. This isn't required, but it can be useful for prioritization.
The confidence signal can be explicit (a visible badge on each card) or implicit (sort the queue so high-confidence items are at the top). Either way, it helps users allocate their attention: spend more time on uncertain recommendations, less on confident ones.
A well-calibrated system should have most cards at high confidence. If everything is medium or low, the AI isn't providing enough value—it's just creating more work. The goal is a queue where most items sail through with a single approval click.
The Delete-and-Regenerate Test
Here's a quality test for your proposal cards: Can the AI regenerate this card from scratch?
This test matters because it ensures your system is stateless at the right layer. The proposal can be regenerated anytime. The durable assets are the playbook (your decision rules) and the data (your CRM + external signals). The card itself is ephemeral—a computed view, not a stored artifact.
This is the same insight from coding: code is ephemeral, specs are durable. Proposals are ephemeral, playbooks are durable.
Designing Proposal Cards for Your Domain
Ready to design your own? Start with three questions:
What action does the user typically take?
This becomes your ACTION component. What are they actually doing when they complete a task in this workflow?
What do they need to see to feel confident?
This becomes your WHY + EVIDENCE components. What context makes them say "yes, this is the right thing to do"?
What artifact do they produce?
This becomes your DRAFT component. What output results from the action—an email, a document, an update?
Assemble those answers into a card structure. Test with real users. Iterate. The four-component pattern is universal, but the specifics—what counts as evidence, how detailed the draft should be, what "why" means in your context—are domain-specific.
Common Mistakes to Avoid
When teams first implement proposal cards, they often stumble on these:
❌ Showing too much by default
Cognitive overload. The whole point is progressive disclosure. If Layer 1 has 20 data points, you've recreated record navigation with extra steps.
❌ Hiding the evidence entirely
Opacity kills trust.21 Users need to be able to verify. Even if evidence is collapsed by default, it must be one click away. Black-box recommendations get rejected.
❌ No drill-down path
Users feel trapped. "Wait, I can't see the full account record?" Progressive disclosure requires that Layer 3 exists. You can hide it, but not remove it.
❌ No feedback capture on rejection
The system doesn't learn. Every rejection should ask "why?" and feed back to the right layer—data, playbook, template, or context.
Get these right, and your proposal cards will feel natural—like they were always how the system should have worked. Get them wrong, and you've just created a new kind of frustration layered on top of the old one.
"A record is a container of data. A proposal card is a container of decision. The card says: 'Here's what to do, here's why, and here's the proof.' That's the new primitive."
Key Takeaways
- 1 The proposal card has four parts: Action, Draft, Why, Evidence. Each serves a function; remove any and something breaks.
- 2 Progressive disclosure is the key: summary by default, evidence on demand, full record available. Most decisions happen at Layer 1.
- 3 Rejection is feedback, not failure. Capture why, and feed back to the right layer—data, playbook, template, or context.
- 4 Design cards by asking: What action? What confidence? What artifact? Assemble, test, iterate.
The Evidence Layer
Trust doesn't come from showing everything. Trust comes from showing the receipts.
The Trust Objection
By now, you might be feeling a familiar objection rising. It goes something like this:
"Users want to see the data. They don't trust AI. If you hide all the fields behind proposal cards, they'll reject everything because they can't verify anything."
This objection is valid. Users should be skeptical of AI recommendations. Black-box suggestions get rejected—and rightly so. Nobody wants to approve actions they can't understand or verify.
But the solution isn't "show everything." The solution is show the receipts.
Trust comes from inspectable evidence, not from raw data volume. And there's a world of difference between the two.
Why "Show Everything" Doesn't Build Trust
The intuition seems reasonable: more data means more informed decisions. If users can see everything, they can verify everything. Complete access equals complete confidence.
But this intuition is wrong. Here's what actually happens when you show everything:22
What "Show Everything" Produces
- → Cognitive overload: 80 fields and 200 activities = paralysis, not clarity
- → Scanning and skipping: Users skim when overwhelmed; they miss things
- → Illusion of completeness: "I saw it all" without actual understanding
- → Decision fatigue: Each field is a micro-decision about relevance
What Actually Builds Trust
- → Inspectable reasoning: "Here's why I suggested this"
- → Verifiable evidence: "Here's what I looked at"
- → Admitted uncertainty: "Here's what I don't know"
- → Consistent behavior: "I follow the same logic every time"
Trust requires understanding. Understanding requires synthesis. When you dump raw data on users, you're asking them to do the synthesis work—the exact work that AI should be doing for them.
The Evidence Layer solves this. It's not about hiding data. It's about presenting data in a way that enables efficient verification.18
Anatomy of an Evidence Card
In Chapter 5, we established that every proposal card has an EVIDENCE component. Now let's examine what that component actually contains.
An evidence card isn't a dump of all related data. It's a curated snippet that answers a specific question: "What did the AI look at to reach this conclusion?"
Evidence Card Structure
The Specific Data Point
Not the whole record—just the relevant snippet. The actual quote, the actual number, the actual date.
Where It Came From
Source attribution: email, call transcript, CRM field, LinkedIn, news article. Users need to know the provenance.
Why It Matters
The connection to the decision. "This email mentions budget timing, which supports the urgency of outreach."
Link to Full Source
One click to the original. If you want the full email, the complete transcript, the entire record—it's there.
Evidence Cards in Practice
Let's see what this looks like for a real proposal. Remember the "Call John" proposal from Chapter 5? Here are its evidence cards:
Notice what's present: specific quotes, specific dates, specific sources, specific relevance. Notice what's absent: the other 200 emails, the 50 other calls, the 80 CRM fields that weren't relevant to this particular decision.
The evidence layer shows the receipts—the specific inputs that led to this specific recommendation. It doesn't show the whole filing cabinet.
The Drill-Down Principle
Evidence works in layers. Most users stop early. Some dig deeper. The system should accommodate both without forcing either.
The Verification Funnel
This matches how trust actually works. You don't verify everything by default—that's exhausting and slow. You verify when something feels off. The system makes verification easy without making it mandatory.
What Makes Evidence Trustworthy
Not all evidence is equally convincing. Here's what separates trustworthy evidence from noise:
Specificity
"The email from Dec 8" not "recent communications." Vague references undermine trust.
✗ "Budget was discussed recently"
Recency
Clear timestamps on everything. Old data with no date feels unreliable.
✗ "at some point"
Relevance
Why this matters for THIS decision. Connection to the recommendation should be explicit.
✗ [no explanation]
Source Clarity
Where did this come from? Email, call, CRM, LinkedIn, news? Users need provenance.
✗ "Based on data"
Scope Signal
"Based on 4 signals" tells the user how much was considered. Missing scope feels incomplete.
✗ [no count]
Verifiability
One click to the original. If users can't verify, they can't trust.
✗ [no link]
Confidence Signals
Not all evidence is equal, and not all proposals deserve the same level of scrutiny. Some recommendations are high-confidence—clear signals, consistent pattern, straightforward decision. Others are low-confidence—sparse data, ambiguous signals, judgment call required.
A well-designed system surfaces this distinction so users can calibrate their review depth:
Confidence-Based Review Depth
High Confidence
Strong signals, consistent pattern, clear recommendation
Action: Approve quickly, minimal review
Medium Confidence
Mixed signals, some uncertainty, judgment helpful
Action: Review the Why, check key evidence
Low Confidence
Sparse data, ambiguous signals, high-risk decision
Action: Verify all evidence, consider alternatives
This can be shown explicitly (a badge on each card) or implicitly (sort the queue so high-confidence items appear first). Either way, users know where to focus their limited attention.
A well-calibrated system should have most proposals at high confidence. If everything is medium or low, the AI isn't providing enough value—it's just creating more work with different packaging.
The John West Principle
There's a deeper layer of trust that goes beyond "here's what I looked at." It's: "Here's what I considered but didn't recommend."
When a proposal card says "I recommend calling John, and here's why," that's helpful. But when it also says "I considered emailing first, but rejected it because response rates are 40% lower for this contact type"—now you know the AI actually evaluated options. The reasoning is comparative, not arbitrary.
This lets users second-guess intelligently. They can see the trade-offs. They can say "actually, I think email is right here because..." and that disagreement is informed, not blind.
We'll explore the John West principle in depth in Chapter 8, where overnight batch processing makes it economically feasible to generate and evaluate multiple candidates for every proposal. For now, know that the evidence layer can include not just "what I used" but "what I considered and rejected."
When Evidence Reveals Problems
Sometimes, when you drill into evidence, you discover something wrong. The AI used bad data. It missed something important. The reasoning doesn't hold up.
This is good. Better to catch it at review than after execution.
Transparency enables correction. When users can see the evidence, they can identify when it's wrong. When they can flag what's wrong, the system can learn.
The Correction Feedback Loop
User reviews evidence card
User spots problem: "This email quote is out of context"
User flags the issue with specifics
System captures feedback at the right layer (data, playbook, or model)
Future proposals improve
Opacity hides problems. When users can't see the reasoning, they can't catch errors. The errors persist. Trust erodes slowly as recommendations fail for reasons nobody can diagnose.21
Privacy and Security in Evidence
A brief but important note: evidence shouldn't expose more than necessary.
The principle is progressive rehydration. The AI reasons on abstracted data—role titles, relationship patterns, timing signals. The evidence layer shows specifics only when users drill down, and only to authorized users.
relationship_status: "Champion"
last_touch: "17 days"
budget_signal: "Q1 timing mentioned"
company: "Acme Corporation"
email_snippet: "...finalizing Q1 budget..."
This is architecture, not policy. The AI doesn't need to see "Jane Smith at 14 Smith Street" to reason about timing and relationship signals. It sees abstracted tokens. Real names and details appear only in the evidence layer, only for authorized viewers.
This principle—called SiloOS in our broader framework—ensures that privacy and security are built into the system, not bolted on as an afterthought.
The Evidence Layer as Governance
Here's a benefit that compliance teams love: every proposal card with evidence is an automatic audit trail.
When someone asks "Why did the system recommend X?"—the answer is built in. Here's the evidence, here's the reasoning, here's who approved it and when. No need to reconstruct after the fact. No forensic archaeology through logs. The explanation exists as a first-class artifact.
Governance Built-In
What Compliance Gets
- • Evidence of what data was used
- • Record of reasoning applied
- • Timestamp of recommendation
- • Record of human approval/rejection
- • Feedback captured on rejected proposals
What This Replaces
- • Manual audit log reconstruction
- • "Why did we do that?" archaeology
- • Separate documentation requirements
- • Post-hoc explanation generation
- • Defensive documentation theater
Governance becomes a feature, not a burden. The same evidence that builds user trust also satisfies auditor questions. The transparency that enables correction also enables compliance.23
"The evidence layer isn't about proving AI is right. It's about making it easy to check. When verification is efficient, trust follows."
Key Takeaways
- 1 Trust comes from inspectable evidence, not from showing everything. More data creates overload, not confidence.
- 2 Evidence cards contain: specific data point + source + relevance + link to full source. Curated, not comprehensive.
- 3 Drill-down principle: summary by default, evidence on click, full record available. Most users stop early.
- 4 Confidence signals help users calibrate review depth. High-confidence proposals sail through; low-confidence gets scrutiny.
- 5 Transparency enables correction: when users can see evidence, they can catch errors. The evidence layer is governance built-in.
Supervise, Don't Navigate
Old CRM day: 8 steps. New CRM day: 4 steps. The workflow itself has changed.
We've covered the architecture: Two-Pane Cognition, with human and AI each owning their workspace. We've established the primitive: the Proposal Card, with its four components of Action, Draft, Why, and Evidence. We've explored how the Evidence Layer builds trust through inspectable reasoning.
Now the question becomes practical: What does the actual workflow look like?
The answer isn't just "faster." It's structurally different. Fewer steps. Different cognitive load. A fundamentally different relationship between human and system.
The Old Workflow
Here's what a typical CRM day looks like under record navigation—the model most systems still use:
The 8-Step Workflow (Record Navigation)
Notice where the time goes. Steps 3, 4, and 5—the navigation steps—consume the majority of effort. The actual decision (step 6) is a sliver. Most of the work is finding information, not using it.
This is the 34% selling time from Chapter 2 made concrete. Sixty percent of the workflow is finding. Ten percent is deciding. The proportion is inverted from what it should be.
The New Workflow
Now here's what the same work looks like under decision navigation:
The 4-Step Workflow (Decision Navigation)
Four steps instead of eight. But the difference isn't just fewer clicks—it's where the cognitive load sits.
The proportion has flipped. Seventy percent of the workflow is now deciding—reviewing, judging, refining. Navigation has collapsed to 10%. The human spends most of their time on what humans are good at: judgment.
The Visual Comparison
Old Workflow (Record Navigation)
New Workflow (Decision Navigation)
The Cognitive Load Shift
This isn't just about time efficiency. It's about what kind of work the human is doing.22
Record Navigation Demands
- → Finding: Where is the information I need?
- → Synthesizing: What does all this data mean?
- → Remembering: What did I see three screens ago?
- → Filtering: Which of these 80 fields matters?
AI does this better than humans.
Decision Navigation Demands
- → Judging: Is this the right action for this situation?
- → Refining: How can I improve this draft?
- → Exception handling: This is a special case because...
- → Strategy: Are we pursuing the right goals?
Humans do this better than AI.
This is playing to strengths, not replacing humans. AI is good at retrieval, synthesis, and pattern-matching. Humans are good at judgment, exceptions, and strategy.14 The new workflow assigns each to their strength.
The Bike Metaphor
"Look mum, no hands!" isn't a phrase about recklessness. It's about trust in the system.
"Riding a bike without hands isn't reckless—it's trusting the system enough to operate at a higher level. Your hands aren't on the handlebars because you don't need them there. The bike is stable. You're supervising, not gripping."
When a child first learns to ride, every input requires conscious attention. Steer left. Steer right. Don't wobble. Watch the road. It's exhausting.
An experienced cyclist doesn't think about these things. The lower-level operations are delegated to muscle memory and mechanical stability. The rider's attention is free for higher-level concerns: route planning, traffic awareness, enjoying the scenery.
Riding with no hands is the visible proof that this delegation has succeeded. You're not gripping the handlebars because the system handles steering autonomously. Your attention has moved up a level.
That's what's happening with AI pair interfaces. You're not navigating through 80 fields because the system handles context synthesis. Your attention has moved up a level—to judgment, to exceptions, to strategy. Same human, different level of operation.
Human Over the Loop vs Human In the Loop
There's a spectrum of human involvement in AI-assisted workflows. Understanding where you sit on this spectrum—and where you should sit—is crucial.
The Autonomy Spectrum
Human reviews every proposal before execution
Human reviews high-stakes / low-confidence proposals
Agent executes routine; human audits batches
Human sets strategy; agent executes within guardrails
The terminology matters here:
Human in the Loop (HITL)
The human is embedded in the execution flow. Every action pauses for human approval. The human is assessing each step.
Human Over the Loop (HOTL)
The human is above the execution flow. The agent operates autonomously within boundaries. The human audits and governs.17
"The human role is fundamentally shifting from assessment to oversight... At the highest level of autonomy, analysts transition from reviewing individual alerts to managing a team of agents, auditing their output weekly or monthly."24 — Detection at Scale
Most CRM workflows should live at Level 2-3. Not life-or-death, but reputation matters. High-stakes deals get full review; routine follow-ups get batch approval or automated execution with audit.
Where CRM Fits on the Spectrum
Not all CRM actions are equal. A well-designed system knows the difference and routes appropriately:
High-Stakes Actions
Level 1-2Enterprise deals, sensitive accounts, contract renewals, executive outreach
→ Human reviews every proposal. No autonomous execution.
Medium-Stakes Actions
Level 2-3Standard follow-ups, qualification calls, stage updates
→ Human reviews low-confidence proposals; high-confidence executed or batch-approved.
Routine Actions
Level 3-4Data hygiene, activity logging, meeting summaries, internal notes
→ Agent executes autonomously. Human audits samples periodically.
The system should route proposals to the right level automatically, based on deal size, account sensitivity, confidence score, and organizational policy. A rep shouldn't have to manually decide what needs review—the system knows.
The Exception Flow
Not every proposal gets approved. Rejections and modifications aren't failures—they're signals. And how you handle them determines whether the system learns.
The Exception Handling Loop
User rejects or modifies a proposal
System captures: What was wrong? (data / logic / context / goal)
Feedback routed to the right layer (update playbook, fix data, adjust model)
Future proposals improve based on accumulated feedback
Strategy Adjustments
The human doesn't just approve individual cards. The human also steers the overall approach. This is the "human pane" from Chapter 4 in action—setting priorities and constraints that apply across all proposals.
"Focus more on enterprise accounts this week"
→ Adjusts queue prioritization
"De-prioritize cold outreach, focus on warm leads"
→ Changes proposal generation rules
"This account is sensitive—I want to review everything"
→ Overrides autonomy level for specific account
"Pause all outreach to Acme Corp until I say otherwise"
→ Constraint that blocks proposals
This is supervision at the strategy level, not the action level. The human isn't approving every email; the human is setting the rules for which emails get proposed.
What "Supervise" Actually Means
If you're not navigating, what are you doing? Here's what supervisory work looks like:
Review the Queue
Is prioritization making sense? Are the right things at the top?
Spot-Check Evidence
Are the reasons good? Does the logic hold up?
Handle Edge Cases
Where AI judgment isn't enough. Exceptions that need human nuance.
Adjust Strategy
Are we pursuing the right goals? Does the playbook need updating?
Provide Feedback
When something is wrong, capture why. Help the system learn.
Add Human Touch
Personalization, relationship nuance, judgment that AI can't replicate.
The Best People Become Better Editors
Under record navigation, the best reps were fast clickers with good memory. They knew where everything was. They could navigate five systems in parallel. They remembered what was said three calls ago.
Under decision navigation, the best reps have good judgment. They spot when a proposal is wrong. They know when to override. They can edit a draft to make it perfect. They see patterns in exceptions.
This is exactly what happened with senior engineers (Chapter 1). They stopped being fast typists. They became creative directors. Strategy → Delegation → Verification.
The skill shifts from "fast execution" to "good decisions." Less typing, more taste. Less navigation, more judgment.
"The human remains in control. But the loop itself has changed—less time triaging, more time governing. That's what supervision looks like."
Key Takeaways
- 1 Old workflow: 8 steps, 60% spent navigating. New workflow: 4 steps, 70% spent deciding.
- 2 Cognitive load shifts from finding/synthesizing to judging/refining. AI does what AI is good at; humans do what humans are good at.
- 3 Human over the loop: audit and govern, not review every action. Different stakes warrant different autonomy levels.
- 4 Exceptions are feedback; the system learns from rejections. Never to zero, but improving over time.
- 5 Best performers become good editors with good judgment. Less typing, more taste.
The Overnight Advantage
Real-time copilots are cute. Overnight agentic precompute is where you start printing leverage.
Everything we've discussed so far has an implicit assumption: the user clicks, and the system responds. Real-time. Interactive. The proposal card appears because the user asked for it.
But real-time has a constraint. A brutal one.
When a user clicks and expects a response, you have maybe 300 milliseconds before they feel friction.25 That's not much time. Certainly not enough for deep context retrieval, for multi-step reasoning, for generating and evaluating multiple candidates, for adversarial review.
What if we removed that constraint?
The Batch Opportunity
Here's the insight that changes everything: most CRM work isn't urgent.
The sales rep doesn't need the answer in 300 milliseconds. They need it in the morning. They need a queue of proposals ready when they sit down at their desk. The work can happen overnight.
The Constraint Flip
Real-Time Constraint
- → 300ms response time
- → Limited retrieval depth
- → Single-pass generation
- → No adversarial review
- → One answer, take it or leave it
Overnight Opportunity
- → 8 hours of compute time
- → Deep context from everywhere
- → Generate → Critique → Refine
- → Full adversarial gauntlet
- → Multiple options with trade-offs
Overnight batch processing isn't about making the experience slower. It's about making the preparation deeper. Batch processing delivers 40-60% cost savings versus real-time processing for AI workloads with acceptable latency tolerance26—but the real advantage isn't cost. When you move from fast to slow, you don't get less AI—you get AI that can finally think.
The Ralph Wiggum Protocol
This pattern emerged from observing how senior engineers get the best results from AI coding tools. Elite developers now run 8-10 hour overnight AI sessions that produce working code, passing tests, and PR-ready results by morning.30 They don't have one long conversation that accumulates context. They run loops. Fresh start, focused task, clean exit, repeat.
We call it the Ralph Wiggum Protocol—after The Simpsons character who sits on the bus and cheerfully says "I'm in danger!" The AI starts fresh each time, blissfully unaware of previous confusion, and that's exactly what makes it work.
The Ralph Wiggum Loop
Why Fresh Beats Accumulated
There's a counterintuitive insight hidden in this pattern: AI's first pass is genius-level. Editing existing work is where it falls apart. Research shows that performance improves dramatically when models are given time to think—a small model with thinking time can outperform a model 14× larger with instant response.28
"AI's first pass at code is genius-level. But when it's editing existing code, it misses bits, tacks on patches, loses the holistic view. So the workflow becomes: generate fresh → validate → regenerate if wrong."
When you give AI a clear task and a clean context, it produces surprisingly coherent, well-structured output. But ask it to "just add one thing" to existing work, and it gets lost. It patches rather than redesigns. It tacks on rather than integrates.
This is why the Ralph Wiggum Protocol works:
Fresh context avoids accumulated confusion
Each loop starts clean. No leftover state from three tasks ago. No confusion about what's current.
The startup protocol rehydrates precisely
Instead of remembering everything, read exactly what's needed: current spec, current state, recent actions.
Delete and regenerate beats patch and accumulate
When something's wrong, don't fix it—regenerate from scratch with better constraints.
This inverts the traditional relationship between output and specification. Code (or proposals) are ephemeral—they can be regenerated anytime. Specifications and playbooks are durable. Design is the asset; the artifact is disposable.
Applying Ralph Wiggum to CRM
The same loop structure works for CRM, just with different domain specifics:
Fresh Context
Clean slate for each account. No carryover from the previous one.
Read Current State
Account record, recent activity, contact changes, external signals (LinkedIn, news, funding).
Pick Best Action
"What should we do with this account right now?" Based on playbook + signals.
Generate Proposal
Action + Draft + Why + Evidence. The full proposal card, pre-assembled.
Exit and Move On
Store the proposal. Clear context. Pick up the next account.
Loop Through Portfolio
Repeat for every account. Entire portfolio processed overnight.
By morning, the rep has a stack of pre-thought proposals waiting. Not "what should I do today?" but "here's what's ready—approve or adjust."
What Overnight Enables
When you have 8 hours instead of 300 milliseconds, entirely different capabilities become possible:
1. Deep Context Retrieval
Real-time can do:
- • Last email
- • CRM fields
- • Basic contact lookup
Overnight can do:
- • Full email thread history
- • Web signals: funding, hiring, press, leadership changes
- • Pattern matching across similar accounts
- • Relationship mapping across contacts
2. Multiple Candidate Actions
Real-time gives you:
- • One suggestion
- • Take it or leave it
Overnight gives you:
- • Three options with trade-offs
- • Ranked by confidence, risk, expected outcome
- • Let the human choose, not just approve/reject
3. Adversarial Review
Multi-model approaches—using several AI models to check each other's work—have been shown to reduce hallucinations and improve reliability.27
Real-time:
- • Single-pass generation
- • No second opinion
Overnight:
- • Generator produces candidates
- • Critic finds holes
- • Bias auditor checks patterns
- • Policy gate verifies compliance
4. Deeper Reasoning
Real-time:
- • Surface-level pattern matching
- • Simple rules-based suggestions
Overnight:
- • Journey inference: "Where is this account really?"
- • Pattern matching: "Accounts like this usually need X"
- • Scenario planning: "If we do X, likely response is Y"
The Fast/Slow Split
This creates a natural division of labor between daytime and overnight:
The Fast Lane and Slow Lane
Fast Lane (Daytime)
- → Reviewing proposals
- → Tweaking drafts
- → Approving actions
- → Executing communications
- → Handling exceptions
Human time is expensive; use it for judgment.
Slow Lane (Overnight)
- → Heavy thinking
- → Deep retrieval
- → Web enrichment
- → Adversarial review
- → Multi-candidate generation
Compute time is cheap; use it for preparation.
Moving from fast to slow enables more AI, not less. You're not making the experience slower—you're making the preparation deeper. Daytime becomes the decision surface; overnight becomes the planning and verification pipeline.
The John West Strategy
Here's where overnight processing unlocks something that fundamentally changes trust and governance: keeping track of rejected alternatives.
Remember the John West principle from Chapter 6: "It's the fish John West rejects that makes John West the best." When you have time to generate multiple candidates, you can show not just what you recommended, but what you considered and rejected.
Risks Detected
- • Champion just changed roles — may have different priorities
- • Competitor mentioned in last call — risk of losing deal
What Would Change This Decision
- • If John is traveling → switch to email
- • If budget delayed → wait and monitor
This transforms AI from "magic eight-ball" into traceable decision process:
Without John West
"The model suggested X"
- • Single thread
- • No alternatives visible
- • No accountability
- • Black box
With John West
"Here's what was considered"
- • Top 3 candidates visible
- • Why #1 won the ranking
- • Why #2 and #3 rejected
- • Governance built-in
"Instead of 'the model suggested X' (single-thread, no accountability), you have: top 3 candidates, why #1 won, why #2 and #3 were rejected, what risks were detected, what evidence was used. That's governance built into the interaction."
The Adversarial Gauntlet
With overnight processing, you can run proposals through multiple stages before they ever reach the human queue:
The Five-Stage Gauntlet
Each stage improves quality. The proposal that reaches the human has already survived critique, passed compliance checks, and been ranked against alternatives. Real-time can't afford this; overnight can.
The Morning Experience
Here's what the sales rep sees when they sit down at 8am:
Each proposal card is:
- ✓ Pre-researched: Deep context from CRM, email, web signals
- ✓ Pre-drafted: Email or call agenda ready to edit
- ✓ Pre-validated: Passed adversarial review and compliance checks
- ✓ Pre-evidenced: Receipts attached, sources linked
The thinking already happened. The deciding is all that's left.
Batch Enables Personalization at Scale
There's a business model insight hidden in overnight processing: mass customization becomes economically feasible. Mass customization that feels personal, meaningful, and seamless is already happening across industries—e-commerce platforms tailoring in real-time, manufacturers allowing custom configurations, healthcare delivering individualized recommendations.29
With real-time generation, personalization is expensive. Every customized proposal costs human attention or immediate AI compute. So companies default to generic templates with mail-merge fields.
With overnight batch processing, the equation flips. Compute time is cheap. You can afford to spend significant AI time on every account, every night. The result is genuinely bespoke reasoning—not "Dear {FirstName}" but "Based on your transition from VP Sales to CRO, and your mention of SSO integration concerns in the November call..."
When to Use Batch vs Real-Time
Not everything should be batch. Here's the decision matrix:
| Use Case | Real-Time | Overnight |
|---|---|---|
| User clicks account, needs immediate response | ✓ | — |
| Daily action queue for all accounts | — | ✓ |
| Simple field lookup | ✓ | — |
| Deep journey analysis | — | ✓ |
| Single quick proposal | ✓ | — |
| Multiple options with trade-offs | — | ✓ |
| Quick draft for immediate need | ✓ | — |
| Adversarially reviewed draft | — | ✓ |
The default pattern: batch prepares, real-time supplements. Overnight does the heavy thinking. Daytime handles the exceptions and immediate needs.
"Real-time AI is constrained by what you can do in 300ms. Overnight AI is constrained only by what you want to invest. When you move from fast to slow, you don't get less AI—you get AI that can finally think."
Key Takeaways
- 1 Ralph Wiggum Protocol: clean start → read state → pick best task → work → exit → loop. Fresh context beats accumulated confusion.
- 2 Overnight enables: deep retrieval, multiple candidates, adversarial review, deeper reasoning. Things you can't do in 300ms.
- 3 John West strategy: keep top 3 candidates + rejection reasons = governance built into the interaction.
- 4 Fast/slow split: overnight prepares (planning pipeline), daytime decides (decision surface).
- 5 Morning experience: pre-thought proposals waiting, decisions ready to make. The thinking already happened.
Variant: Sales Development
SDRs spend 80% of their time researching. What if research was already done when they arrived?
We've established the core architecture: Two-Pane Cognition, Proposal Cards, the Evidence Layer, supervisory workflow, and overnight batch processing. Now let's see how these patterns apply to a specific domain.
Sales Development Representatives (SDRs) do the front-line work of prospecting and lead qualification. They find potential customers, research them, reach out, qualify interest, and hand off to account executives. It's the top of the sales funnel.
And it's drowning in research work.31
The SDR Time Problem
Here's what an SDR's day actually looks like:
The Traditional SDR Workflow
Notice where the time goes. Steps 2, 3, and 4—the research and crafting steps—consume 70-80% of effort.32 The actual outreach and conversation (steps 5-6) are a sliver.
This is exactly backwards. The SDR's value is in conversation—qualifying, building rapport, understanding need. But they spend most of their time doing work that AI can do: research and first-draft writing.34
The Framework Applied
Let's apply the architecture from Part II to sales development:
Two-Pane Cognition for SDR
Human Pane
- • Prioritized prospect queue
- • Response handling
- • Qualification judgment
- • Relationship signals
- • Handoff decisions
AI Pane
- • Research synthesis
- • Trigger event detection
- • Personalized drafts
- • Evidence links
- • Confidence scores
Shared Surface
- • ICP criteria
- • Messaging guidelines
- • Compliance rules
- • Brand voice
- • Territory boundaries
The Proposal Card for Prospecting
Here's what a proposal card looks like for an SDR's morning queue:
"Sarah — saw your comment on [Sarah Guo's post] about the challenges of scaling ops without adding headcount.
We've been helping similar-stage companies automate their operational workflows without the typical integration nightmare. Just helped [similar company] cut their ops load by 40%.
Worth a 15-min chat? Timing might be good given the Series B."
This proposal card has everything the SDR needs to decide:
- → ACTION: What to do (connect on LinkedIn) and how (reference the specific post)
- → DRAFT: Message already written, personalized, ready to send or tweak
- → WHY: Trigger events that make this the right moment (funding, activity, fit)
- → EVIDENCE: The receipts—links to the post, funding announcement, company profile
What the Evidence Looks Like
When the SDR clicks "View" on the evidence section, they see the specific signals that drove this recommendation:
Overnight Batch for SDR
Applying the Ralph Wiggum loop (Chapter 8) to sales development:
Load prospect list
Territory assignment, marketing handoffs, enrichment queue
For each prospect: Fresh context
Clean slate, load ICP criteria and messaging playbook
Research and enrich
LinkedIn activity, company news, funding, hiring signals, trigger events
Generate proposal
Personalized angle + draft message + evidence links + confidence score
Quality check
Does this pass threshold? Strong signal, valid angle, fresh trigger?
Add to queue or skip
High-quality → morning queue. Low-quality → hold for later refresh.
Run this on 200 prospects overnight. The SDR reviews the best 30-50 in the morning—the ones with strong signals and ready drafts.
The Morning Experience
Compare the old SDR morning to the new one:
Old SDR Morning
- 1. Open CRM, look at list
- 2. Pick someone, open LinkedIn
- 3. Scroll their profile, recent posts
- 4. Check company news, funding
- 5. Find an angle, write message
- 6. Send, log, repeat
New SDR Morning
- 1. Open queue → 35 pre-researched prospects
- 2. Review top prospect → see proposal card
- 3. Check why + evidence (if needed)
- 4. Edit draft if needed
- 5. Approve → send
- 6. Next prospect
What the SDR Actually Does
The role shifts from researcher to editor/approver. Here's what that looks like:
What the SDR Does
- ✓ Reviews draft: "Is this angle good? Does it land?"
- ✓ Checks evidence: "Is this trigger real? Timing right?"
- ✓ Edits: Adjusts tone, adds personal touch, fixes errors
- ✓ Approves: Sends or queues for sending
- ✓ Rejects: Skips bad fits, flags data issues
- ✓ Handles responses: The actual conversations
What AI Handles
- → Researching each prospect's background
- → Finding trigger events and signals
- → Identifying the best angle
- → Crafting first-draft messages
- → Tracking who's been contacted
- → Ranking by likelihood to respond
The Feedback Loop
Like all good systems, this one learns from use:
When the SDR Rejects a Proposal
Capture why:
- • Bad fit: Not actually ICP
- • Wrong timing: Trigger is stale
- • Weak angle: Personalization doesn't land
- • Bad data: Info is wrong or outdated
Feed back to improve:
- • Refine ICP matching criteria
- • Adjust trigger freshness thresholds
- • Update angle generation prompts
- • Flag data source issues
When Prospects Respond
Track what worked:
- • Which angles got replies
- • Which triggers were most predictive
- • Which message styles resonated
- • Which channels performed best
Feed success patterns back:
- • Weight trigger types by response rate
- • Favor successful messaging patterns
- • Prioritize high-response segments
- • Refine confidence scoring
Exception rate should drop over time as the system learns what works for this SDR, this territory, this product.
The Metrics That Change
Here's what moves when you apply decision navigation to sales development:
| Metric | Before | After |
|---|---|---|
| Prospects researched/day | 15-20 | 50-100 (AI pre-research) |
| Personalized messages/day | 15-20 | 50-80 |
| Time per prospect | 20-30 min | 5-10 min |
| Response rate | Baseline | Higher (better personalization)38 |
| SDR territory coverage | 1 territory | 2-3x territory |
"The SDR's job doesn't change—qualify and hand off. What changes is how they spend their time. Less research, more judgment. Less crafting, more deciding. That's decision navigation for sales development."
Key Takeaways
- 1 SDRs spend 80% on research; overnight batch flips this ratio. Research happens while they sleep.
- 2 Proposal cards for prospecting: Action + Draft + Why (trigger events) + Evidence (links to signals).
- 3 Morning queue: 30-50 pre-researched, pre-drafted prospects ranked by confidence and trigger strength.
- 4 SDR role shifts from researcher to editor/approver. Same conversations, less prep.
- 5 Feedback loop: Rejections and responses improve the system over time.
Variant: Customer Support
Support agents spend 60% of their time gathering context. What if context was already gathered when the ticket arrived?
Customer support agents handle tickets, resolve issues, and maintain relationships. They're the front line of customer experience—the human voice when something goes wrong.
Their job requires empathy, judgment, and problem-solving. But most of their time isn't spent on any of those things. It's spent gathering context.39
Sound familiar?
The Context Gathering Problem
Here's the traditional support workflow:
Traditional Ticket Handling
Steps 3, 4, and 5—the context gathering steps—consume 50-60% of handle time.40 The agent is playing detective before they can play helper.
The ratio is backwards. Agents are hired for their judgment and empathy—not their ability to navigate five different systems.41
The Framework Applied
Applying Two-Pane Cognition to customer support:
Two-Pane Cognition for Support
Human Pane
- • Ticket queue prioritization
- • Resolution decisions
- • Tone and empathy
- • Edge case judgment
- • Escalation handling
AI Pane
- • Context synthesis
- • Similar ticket matching
- • Draft responses
- • Policy lookups
- • Sentiment detection
Shared Surface
- • Response policies
- • Escalation rules
- • Tone guidelines
- • Approval thresholds
- • SLA requirements
The Proposal Card for Support
Here's what a proposal card looks like for a support ticket:
Hi Sarah,
I'm so sorry about the billing confusion—I completely understand the frustration. I've looked into this, and you're right: there was an error in how our system applied the promotion code.
I've issued a $50 credit to your account. You'll see it reflected within 24 hours. I've also flagged your account to make sure this doesn't happen again.
Is there anything else I can help with?
Context Synthesis vs Context Gathering
The fundamental shift is from gathering to reviewing:
Old: Context Gathering
- 1. Agent opens CRM record
- 2. Agent opens billing system
- 3. Agent searches previous tickets
- 4. Agent checks product docs
- 5. Agent mentally synthesizes
- 6. Agent decides what to do
New: Context Synthesis
- 1. AI pre-gathers from all systems
- 2. AI synthesizes full picture
- 3. AI recommends resolution
- 4. Agent reviews proposal
- 5. Agent approves or adjusts
The Evidence Layer for Support
When the agent clicks to view evidence, they see everything the AI considered:
Handling Escalations
Not every ticket gets a "resolve now" proposal. Some need escalation—and escalation is a proposal type, not a failure mode.
Issue: Customer disputes early termination fee of $547. Claims he was told the contract was month-to-month.
Customer value: 2-year customer, $8,200 LTV, Enterprise plan
Churn risk: High — explicit mention of switching to competitor
The escalation proposal includes a handoff summary so Tier 2 doesn't start from scratch. Context travels with the ticket.
The Queue Transformation
Decision navigation changes what the queue looks like:43
Old Support Queue
- • Oldest tickets first (or random priority)
- • Agent opens ticket, sees only the message
- • Each ticket is a mini research project
- • No sense of complexity until you're in it
New Support Queue
- • Pre-triaged by urgency, value, complexity
- • Each ticket shows proposal + synthesis + draft
- • Confidence score visible before opening
- • Escalations flagged, handoffs ready
Governance in Support
Different tickets warrant different levels of human involvement:44
Low-Stakes + High-Confidence
Auto-send with auditPassword resets, basic how-to questions, acknowledgment responses
→ Send automatically. Human audits samples weekly.
Medium-Stakes or Medium-Confidence
Human reviews before sendRefund requests, technical troubleshooting, feature questions
→ AI drafts. Human reviews and approves. Standard flow.
High-Stakes or Low-Confidence
Human must approveLarge disputes, churn threats, legal mentions, VIP accounts
→ AI synthesizes context. Human makes all decisions.
The audit trail is automatic:
- → Every response: What was proposed, what was sent, who approved
- → Customer feedback: Did this resolve the issue? Satisfaction rating?
- → Quality metrics: Resolution rate, CSAT, escalation rate by agent and issue type
The Metrics That Change
These are representative improvements based on industry implementations:45
| Metric | Before | After |
|---|---|---|
| Average handle time | 8-12 min | 3-5 min |
| Context gathering time | 4-6 min | 30 sec (review synthesis) |
| Tickets per agent per day | 30-50 | 60-100 |
| First-contact resolution | Baseline | Higher (better context) |
| Customer satisfaction | Baseline | Higher (faster, personalized) |
"Support agents are hired for their judgment and empathy—not their ability to navigate five systems. Decision navigation lets them focus on what they're actually good at."
Key Takeaways
- 1 Support agents spend 60% gathering context; AI synthesis flips this ratio.
- 2 Proposal cards for support: Action + Draft + Why + Evidence (including customer context and similar tickets).
- 3 Pre-triage and pre-synthesis enable faster, more consistent responses with better personalization.
- 4 Escalation is a proposal type, not a failure mode. Handoff summaries travel with the ticket.
- 5 Agent role shifts from researcher to reviewer/personalizer. Judgment and empathy, not system navigation.
Designing Your Own Proposal Cards
Pick one workflow. Design the card. Hide the fields.
Chapters 4-8 gave you the architecture. Chapters 9-10 showed domain-specific examples. Now: how do you design proposal cards for your workflows?
This is the practical chapter. A step-by-step guide for turning any navigation-heavy workflow into a decision navigation interface.
The Three Questions
Every proposal card design starts with three questions. Answer them clearly, and the card almost designs itself.
The Three Questions
What action does the user typically take?
This becomes your ACTION component.
- • Be specific: "Call John about Q2 budget" not "follow up"
- • The action should be approvable/rejectable
- • One action per card (not a list of options)
What do they need to see to feel confident?
This becomes your WHY and EVIDENCE components.
- • Not everything—just what moves them from uncertain to confident
- • What would they check manually if AI wasn't there?
- • Usually 3-5 key signals, with source links
What artifact do they produce?
This becomes your DRAFT component.
- • Email, report, agenda, proposal, update—something tangible
- • What would the human otherwise create from scratch?
- • Some actions have no artifact (data updates); that's fine
Worksheet: The Three Questions
Step-by-Step Design Process
Here's the full process for designing a proposal card from scratch:
Step 1: Identify the Workflow
Pick ONE workflow to start. Not the whole system—one specific, repeatable process.
- • High-volume (done many times per day)
- • Routine (similar pattern each time)
- • Currently time-consuming (lots of navigation)
Example: "Daily outreach to warm leads"
Step 2: Map the Current State
Document exactly what the user does today, step by step.
- • What systems do they open?
- • What screens do they navigate?
- • What data do they look at?
- • What do they produce?
- • How long does it take?
- • Tab-switching patterns
- • Copy-paste workflows
- • "Wait, where was that?" moments
- • Manual synthesis steps
- • Repeated research
Step 3: Identify the Decision Point
Find the actual decision being made. Not "fill out this form" but the judgment call.
- • Not "research this account" → "decide whether to contact this account"
- • Not "look at the ticket" → "decide how to resolve this ticket"
- • Not "check the pipeline" → "decide which deal needs attention"
Step 4: Design the Card Components
Using your answers to the three questions, specify each component:
What's the specific, approvable action? One verb + one target + optional context.
What artifact would the user create? Ready to send/use with minor edits.
What 3-5 reasons would convince them? Short phrases, not paragraphs.
What receipts would they check? Links to source data.
Step 5: Define the Drill-Down
Specify what appears at each layer of progressive disclosure:
Step 6: Test with Users
Show them the card design and ask the key question:
"Would you feel confident approving this?"
If no: "What's missing?"
Their answer tells you what to add to Why/Evidence. Iterate until they say yes.
Common Card Patterns
Most proposal cards fall into a few patterns. Here are templates you can adapt:
Pattern 1: Outreach Card
ACTION: Contact [person] via [channel]
DRAFT: Message/email content
WHY: Trigger events, timing signals, relationship status
EVIDENCE: Source of signals, previous interactions
Examples: Sales outreach, re-engagement, meeting follow-up
Pattern 2: Update Card
ACTION: Change [field] to [value]
DRAFT: N/A (or explanation note)
WHY: Data discrepancy detected, event occurred
EVIDENCE: Source of new information
Examples: Stage changes, contact updates, data hygiene
Pattern 3: Escalation Card
ACTION: Escalate [item] to [person/team]
DRAFT: Summary for receiving party
WHY: Threshold exceeded, policy triggered, complexity
EVIDENCE: What triggered the escalation
Examples: Support escalation, deal approval, exception handling
Pattern 4: Review Card
ACTION: Approve/reject [item]
DRAFT: N/A (item is the artifact)
WHY: Quality signals, risk assessment, completeness
EVIDENCE: What was evaluated
Examples: Content review, proposal approval, quality check
Progressive Disclosure Design
Getting the layers right is crucial. Here's a detailed breakdown:
What Each Layer Contains
- • Action headline
- • 2-3 line "Why" summary
- • Approve / Modify / Reject buttons
User can decide on 80% of cards without leaving this view.
- • Full draft artifact
- • Evidence cards (3-5 items)
- • Confidence indicator (optional)
User can verify AI reasoning, edit draft, check receipts.
- • Full record/history
- • All fields
- • Complete audit trail
User can access anything, for edge cases and deep investigation.
The Delete-and-Regenerate Test
Here's a quality check for your proposal card design:
Can the AI regenerate this exact card from:
Card is well-designed. It's derivable, consistent, auditable.
Card depends on something fragile. Find what's missing and add it to state or playbook.
Common Mistakes to Avoid
Here are the pitfalls that catch most first-time card designers:
Mistake 1: Showing too much by default
Fix: Ruthlessly cut to 3-5 most relevant points. Everything else is Layer 2.
Mistake 2: Hiding the evidence
Fix: Make evidence one click away, always. Never hidden or inaccessible.
Mistake 3: No drill-down path
Fix: Every summary should link to fuller view. Layer 3 must exist.
Mistake 4: Generic actions
Fix: Be as specific as the context allows. Verb + target + reason.
Mistake 5: No feedback capture
Fix: Every rejection should capture why. Feed back to improve the system.
Governance Checklist
Before deploying a proposal card design, verify these governance requirements:
Pre-Deployment Checklist
Your First Card: An Exercise
Let's design a proposal card together. Here's the workflow:
Workflow: Daily review of accounts that haven't been touched in 14+ days
Example Answer
- • 17 days since last touch (relationship cooling)
- • Open opportunity at proposal stage ($45k)
- • Decision timeline was "end of Q1" — we're approaching that
- • Last email: Jan 7 (link)
- • Opportunity record (link)
- • Notes from Dec 15 call mentioning Q1 timeline (link)
"The framework is simple: Action, Draft, Why, Evidence. Progressive disclosure: summary first, evidence on click, full record available. Test with users, iterate, deploy. That's how you turn any workflow into decision navigation."
Key Takeaways
- 1 Three questions: What action? What confidence needs? What artifact? These give you the four components.
- 2 Design each component deliberately; test progressive disclosure. 80% should decide at Layer 1.
- 3 Delete-and-regenerate test: Can AI recreate the card from state + playbook + evidence?
- 4 Common mistakes: Too much default, hidden evidence, no drill-down, generic actions, no feedback capture.
- 5 Start with one workflow, get it working end-to-end, then expand.
What Changes If You Get This Right
For individuals. For teams. For the industry.
We've covered the shift: from record navigation to decision navigation. The architecture: Two-Pane Cognition and proposal cards. The amplifier: overnight batch processing. The applications: sales development, customer support, and a framework for designing your own.
Now: what actually changes if you implement this?
The impact cascades from individuals to teams to the entire industry. This isn't incremental improvement—it's a different operating model.
For Individuals: 10x Faster Time-to-Decision
The shift changes what work feels like, moment to moment.
The Cognitive Shift
Old Mental Model
Cognitive load on: finding, synthesizing, creating
New Mental Model
Cognitive load on: evaluating, judging
Here's what this feels like in practice:
Morning starts with pre-thought proposals, not a blank slate. The system has already done overnight preparation.
Each decision is 2 minutes instead of 20. You review, judge, act—not research, synthesize, create.
Mental energy goes to hard problems, not routine research. You save your judgment for where it matters.
End of day: more decisions made, less exhaustion. The cognitive lightness compounds.
"The shift isn't just faster—it's cognitively lighter. When the system does the finding, you can focus on the judging."
One decision faster is nice. Every decision faster is transformative. You process more, with higher quality, with less fatigue. The compound effect builds—research shows AI adoption leaders see performance improvements 3.8x higher than the bottom half of adopters.49
For Teams: Standardized Format, Faster Handoffs
Decision navigation creates consistency across the team—not by forcing compliance, but by making the right pattern the easy pattern.
Consistency
Everyone uses the same proposal card format: Action + Draft + Why + Evidence. Every time.
No more "how does Sarah organize her notes?"
Trivial Handoffs
Account handoff = here are the pending cards. Shift handoff = here's the queue. Escalation = here's the card that triggered it.
No more "let me bring you up to speed" meetings.
Measurable Quality
Approval rates by rep, by card type, by segment. Rejection reasons. Response rates. Time to decision.
What's working? Where are bottlenecks?
For the Industry: System of Record → System of Action
This is the paradigm shift happening across enterprise software:
What CRMs were built as
- • Faithful mirror of what happened
- • All data visible and navigable
- • User synthesizes and decides
- • The record is the interface
What the future demands
- • Engine that recommends what to do
- • Proposals visible, records on demand
- • System synthesizes, user decides
- • The proposal is the interface
The record doesn't disappear—it becomes the foundation. But users interact with proposals, not fields.
"Enterprise software is shifting from 'system of record' to 'system of action.' The record becomes infrastructure; the proposal becomes the interface."
This applies to every enterprise tool. The vendors that make this shift will win; the rest will be navigated around or replaced. As Gong documented, systems of action don't just store data—they turn 6,000 words of conversation into actionable recommendations rather than 30-60 words of notes.8
The Competitive Dynamic
→ Early adopters of decision navigation gain compounding advantage—BCG research shows the gap between leaders and laggards is widening rapidly51
→ Every day you're spelunking, competitors are supervising
→ The gap widens with scale: 10% efficiency at 100 reps = 10 rep-equivalents
The Skills That Matter Now
Decision navigation changes which skills are valuable:
Becomes Less Valuable
- → Fast clicking and navigation
- → Memorizing where things are
- → Manually synthesizing from multiple sources
- → Creating artifacts from scratch
- → Speed-of-execution tasks
Becomes More Valuable
- → Judgment: Is this proposal right?
- → Editing: How do I make this better?
- → Exception handling: When is the system wrong?
- → Strategy: What should we optimize for?
- → Feedback: How do I correct the system?
The senior role redefines—just like it did in coding. Less "doer," more "director." Less execution, more judgment. Less navigation, more supervision. As Bain puts it: "You can't automate your way to transformation. You have to rethink the work itself."52
The best people operate at a higher level. They're not freed from work; they're freed for higher-value work.
The Governance Advantage
Compliance teams should love decision navigation:
What Governance Gets
- ✓ Every proposal logged with evidence
- ✓ Every decision traceable
- ✓ Rejection reasons captured
- ✓ Audit trail built into workflow
- ✓ "Why did this happen?" has an answer
- ✓ Policy enforcement in generation
Governance becomes a feature, not a burden. Not forms and approvals slowing everything down. Built into the interaction pattern. Compliance by architecture, not by process. This matters: McKinsey research shows 40% of organizations identify explainability as a key AI risk, but only 17% actively work to mitigate it.21
The Risks If You Don't
What happens if you stay with record navigation while the world moves on?
Competitive erosion
Competitors using AI-native interfaces will outpace you. Same market, different operating speed.
Talent flight
Your best people will leave for environments where they're not spelunking. Why navigate when you could supervise?
Training costs stay high
Onboarding stays slow when there's no standard format. "Shadow Sarah" doesn't scale.
Data remains underutilized
You have the data. It's an asset. But if it's hard to access, it's an asset that doesn't compound.
It won't be one catastrophic failure. It'll be gradual: slightly slower, slightly less efficient, slightly less attractive. Death by a thousand clicks. Research from Gartner predicts 15% of day-to-day work decisions will be made autonomously by 2028, up from near 0% in 202453—those who haven't redesigned their interfaces by then will be working against the current.
The Path Forward
Here's how to start:
- • Choose: high-volume, routine, currently time-consuming
- • Design the proposal card (Chapter 11)
- • Build the overnight batch (Chapter 8)
- • Deploy, measure, iterate
- • Apply the pattern to related processes
- • Reuse the architecture you built
- • Train the team on decision navigation
- • New interfaces start with proposal cards
- • Record views become drill-down, not default
- • Batch processing becomes standard infrastructure
The Coding World Learned This by Accident
Remember Chapter 1? Senior developers stopped looking at code. They didn't plan it. AI tools made it possible, and they discovered it worked.
The workflow changed without anyone designing it. Engineers stumbled into decision navigation because the tools abstracted the artifact layer. They saw diffs instead of files. They supervised instead of spelunking.
You don't have to stumble into it. You can design for it from the start.
"The coding world stopped looking at the artifact. It started looking at the outcome. They didn't plan it—AI tools made it possible, and engineers discovered it worked. You don't have to stumble into it. You can design for it from the start."
Look Mum, No Hands
Remember the bike metaphor from Chapter 7?
That moment when a child rides without hands—it's not reckless. It's confident. They trust the system enough to let go. Their hands aren't on the handlebars because they don't need them there.
Your users can have that moment. Design interfaces that let them ride without hands. Not passive—supervising. Not reckless—confident. Operating at a higher level because the system handles the lower level.
Design for proposals.
Hide the fields.
Let your users ride without hands.
"The coding world learned this by accident. You can learn it on purpose."
Key Takeaways
- 1 For individuals: 10x faster decisions. Cognitive load shifts from finding to judging.
- 2 For teams: Standardized format, trivial handoffs, measurable quality, faster onboarding.
- 3 For the industry: System of record → system of action. The proposal becomes the interface.
- 4 Skills shift: Less navigation, more judgment and editing. The senior role redefines.
- 5 Governance becomes a feature, built into the interaction pattern. Compliance by architecture.
- 6 Start now: Pick one workflow. Design the card. Build the batch. You can learn this on purpose.
References & Sources
The evidence and external sources that informed this guide.
This ebook synthesizes original frameworks with external research and industry data. Below are the primary sources referenced throughout, organized by category. Where inline citations appear (e.g., 1), they link to these numbered references.
Industry Analysis
[1] "Even the Inventor of Vibe Coding Says Vibe Coding Can't Cut It"
Gizmodo. Andrej Karpathy's reflection on vibe coding limitations—his nanochat project was "basically entirely hand-written" because AI agents "just didn't work well enough."
https://gizmodo.com/even-the-inventor-of-vibe-coding-says-vibe-coding-cant-cut-it-2000672821
[2] "Claude Code is Redefining Software Engineering"
Medium. Documents the workflow shift where developers generated 11,900 lines of production code without opening an IDE.
https://medium.com/@ajepym/claude-code-is-redefining-software-engineering-faster-than-were-ready-for-da1a110b5567
[24] "AI Security Operations 2025 Patterns"
Detection at Scale. Documents the shift from human-in-the-loop assessment to human-over-the-loop governance in security operations—analysts transition from reviewing individual alerts to managing teams of agents.
https://www.detectionatscale.com/p/ai-security-operations-2025-patterns
[26] "Real-Time vs Batch Processing Architecture"
Zen van Riel. Analysis showing 40-60% cost reduction for batch AI processing versus real-time, with architectural guidance on when to use each approach.
https://zenvanriel.nl/ai-engineer-blog/should-i-use-real-time-or-batch-processing-for-ai-complete-guide/
[30] "How the Creator of Claude Code Actually Uses It"
Dev.to. Documents Boris Cherny's workflow for running 10+ hour overnight AI coding sessions that produce working code, passing tests, and PR-ready results by morning.
https://dev.to/a_shokn/how-the-creator-of-claude-code-actually-uses-it-32df
LinkedIn Commentary
[3] "Switching to Claude Code Boosts Dev Team Productivity"
LinkedIn. Documents the shift to 90% AI-generated code and the role redefinition from "code producer" to "creative director."
https://www.linkedin.com/posts/md-mukthar-8239b2278_i-quietly-switched-my-remote-dev-team-and-activity-7413165586949525504-A7aE
[4] "Switching to Claude Code Boosts Dev Team Productivity"
LinkedIn. Source for the "creative director" role shift quote: "Strategy → Delegation → Verification."
https://www.linkedin.com/posts/md-mukthar-8239b2278_i-quietly-switched-my-remote-dev-team-and-activity-7413165586949525504-A7aE
[7] "Automate Salesforce with xdge for Real-Time CRM Accuracy"
LinkedIn/xdge. Source for the 286 hours per year wasted on manual data entry per sales rep.
https://www.linkedin.com/posts/xdge-ai_at-xdge-our-sales-operations-run-on-autopilot-activity-7413418154560761859-S-XN
[9] Ameet Patra on CRM as Profit Center
LinkedIn. "CRM should be a profit center, not a cost center. If Salesforce only records activity, it is an expense. If it enforces revenue flow, it becomes an asset."
https://www.linkedin.com/posts/ameet-patra_a-different-perspective-on-servicenow-vs-activity-7411369617350225920-O6DF
[29] "Personalization at Scale: How AI Enabling Mass"
LinkedIn/Nicolas Babin. Analysis of how AI enables mass customization across industries—e-commerce, manufacturing, and healthcare delivering individualized experiences at scale.
https://www.linkedin.com/pulse/personalization-scale-how-ai-enabling-mass-nicolas-babin-tsuge
Primary Research
[5] "Claude Code Token Limits: A Guide for Engineering Leaders"
Faros.ai. RCT finding: experienced developers were 19% slower with AI tools but perceived themselves as 20% faster—a 40 percentage point perception gap.
https://www.faros.ai/blog/claude-code-token-limits
[6] "How to Customize Salesforce for Maximum Efficiency"
Revenue Grid. Source for the 34% selling time statistic—sales reps spend only 34% of their time actually selling.
https://revenuegrid.com/blog/enhance-your-experience-how-to-customize-salesforce-for-maximum-efficiency/
[8] "Why AI Systems of Action Replace Traditional CRM Workflows"
Gong. Documents the context loss problem: 6,000 words of conversation reduced to 30-60 words in CRM notes.
https://www.gong.io/blog/crm-system-of-record-vs-ai-system-of-action
[10] "18 Predictions for 2026"
Jakob Nielsen. Prediction that static interfaces will become obsolete, replaced by generative UI that assembles based on user intent and context.
https://jakobnielsenphd.substack.com/p/2026-predictions
[11] "Top 100 Developer Productivity Statistics with AI Coding Tools (2026)"
Index.dev. Source for the 10-30% average productivity gains from AI tools, with some studies showing up to 81% improvement among active users.
https://www.index.dev/blog/developer-productivity-statistics-with-ai-tools
[25] "Latency as UX: Why 200ms Matters for Perceived Intelligence"
Alex Smale. Research showing 200-300ms as the threshold for perceived responsiveness in interactive systems—the constraint that limits real-time AI capabilities.
https://www.alexsmale.com/latency-as-ux-why-200ms-matters-for-perceived-intelligence/
[28] "Test-Time Compute Analysis"
Hugging Face. Research demonstrating that small model + thinking time can outperform a model 14× larger with instant response—the foundation for why overnight batch processing unlocks AI capability.
https://huggingface.co/blog/Kseniase/testtimecompute
AI Design Patterns & UX
[12] "Guide to AI Integration Patterns for Enterprise-Scale Systems"
Transcenda. Documents the side-panel copilot pattern: AI lives in a collapsible panel providing space for history, reasoning, and deeper synthesis while leaving the main canvas intact.
https://www.transcenda.com/insights/guide-to-ai-integration-patterns-for-enterprise-scale-systems
[16] "AI Design Patterns Enterprise Dashboards"
AufaitUX. Modern dashboards should function as cognitive collaborators that summarize context and predict outcomes rather than just display data.
https://www.aufaitux.com/blog/ai-design-patterns-enterprise-dashboards/
[18] "Designing AI User Interfaces That Foster Trust and Transparency"
UXMatters. Interfaces can present context-sensitive explanations that show the reasoning behind AI outputs without overwhelming users with complex algorithmic details.
https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php
Academic Research on Human-AI Collaboration
[13] "Beyond Human-in-the-Loop"
ScienceDirect. AI offers unparalleled capabilities in processing large-scale data; humans contribute strategic thinking, contextual judgment, and ethical reasoning. Human-over-the-loop shifts humans to a supervisory role.
https://www.sciencedirect.com/science/article/pii/S2666188825007166
[14] "Human- versus Artificial Intelligence"
PMC/NIH. People are better suited than AI systems for a broader spectrum of cognitive and social tasks under unforeseen circumstances. AI excels at data processing while humans handle social-psychosocial interaction.
https://pmc.ncbi.nlm.nih.gov/articles/PMC8108480/
[15] "AI won't make the call: Why human judgment still drives innovation"
Harvard Business School. Human experience and judgment remain critical to making decisions because AI can't reliably distinguish good ideas from mediocre ones or guide long-term business strategies.
https://www.hbs.edu/bigs/artificial-intelligence-human-jugment-drives-innovation
[17] "Beyond Human-in-the-Loop" (HOTL pattern)
ScienceDirect. Human-over-the-loop (HOTL) shifts humans to a supervisory role, allowing AI to handle routine tasks while reserving human input for complex decisions.
https://www.sciencedirect.com/science/article/pii/S2666188825007166
[19] "Brain Cache: A Generative AI-powered Cognitive Exoskeleton"
MIT GenAI & HCI Conference. AI achieves cognitive augmentation through three mechanisms: externalizing biological memory, structuring knowledge, and activating insights.
https://generativeaiandhci.github.io/papers/2025/genaichi2025_51.pdf
[20] "Designing AI User Interfaces That Foster Trust and Transparency"
UXMatters. Context-sensitive explanations show reasoning behind AI outputs without overwhelming users with complex algorithmic details—the foundation of progressive disclosure in AI interfaces.
https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php
[21] "Building AI Trust: The Key Role of Explainability"
McKinsey. Research shows 40% of organizations identify explainability as a key AI risk, but only 17% actively work to mitigate it—evidence that opacity undermines trust.
https://www.mckinsey.com/capabilities/quantumblack/our-insights/building-ai-trust-the-key-role-of-explainability
[22] "Cognitive Load Theory"
John Sweller / ScienceDirect. The foundational research showing that human working memory has limited capacity—our brains can only handle so much at once before performance drops.
https://www.sciencedirect.com/topics/psychology/cognitive-load-theory
[23] "AI UX Design: Creating Transparency and Trust in AI Products"
Wildnet Edge. Documents how 2025 UI toolkits integrate explainability widgets that update dynamically, keeping users informed without demanding deep technical knowledge.
https://www.wildnetedge.com/blogs/ai-ux-design-creating-transparency-trust-in-ai-products
[27] "Mitigating AI Hallucinations: The Power of Multi-Model Approaches"
AI Sutra. Research showing that multi-model approaches—using several AI models together to check each other's work—reduce hallucinations and improve system trustworthiness.
https://aisutra.com/mitigating-ai-hallucinations-the-power-of-multi-model-approaches-2393a2ee109b
Sales Development & SDR Productivity
[31] "AI Agents for Outbound Sales: Complete Guide (2025)"
Smartlead.ai. Analysis showing SDRs spend 70-80% of time on repetitive research tasks, with only 20-30% on relationship-building and strategic thinking. Manual SDRs achieve 50-100 touches/day vs 200-300 for AI-assisted.
https://www.smartlead.ai/blog/ai-agents-for-outbound-sales
[32] "140+ Sales Statistics | 2026 Update"
Spotio. Sales reps spend only 30% of their time actively selling, with 70% consumed by administrative tasks, data entry, and research. Based on Salesforce State of Sales Report data.
https://spotio.com/blog/sales-statistics/
[33] "How to Enable Your SDRs with Data Enrichment: The Playbook"
Databar.ai. SDRs spend up to 40% of their workweek searching for contacts, with 10-15 minutes per prospect on manual research before personalization.
https://databar.ai/blog/article/how-to-enable-your-sdrs-with-data-enrichment-the-playbook
[34] "AI SDRs and Real-Time Lead Data Explained"
Salesforge.ai. Human SDRs spend only 35% of their time actively selling, with the rest consumed by research, data entry, and routine outreach. Real-time lead data improves lead scoring accuracy by up to 40%.
https://www.salesforge.ai/blog/ai-sdrs-real-time-lead-data-explained
[35] "The Best Sales Insights of 2025"
Gong. Analysis of 7.1 million opportunities showing sellers using AI generate 77% more revenue than those not using AI. AI enables personalized outreach at scale that previously required extensive manual research.
https://www.gong.io/blog/the-best-sales-insights-of-2025
[36] "AI Agents for Outbound Sales: Complete Guide"
Smartlead.ai. Comparison of manual vs AI-assisted SDR performance: manual SDRs achieve 50-100 touches/day at 5-10 minutes research per prospect; AI-assisted SDRs achieve 200-300 touches with research automated to seconds.
https://www.smartlead.ai/blog/ai-agents-for-outbound-sales
[37] "AI Sales Tools 2026: Transform Your Sales Performance"
MarketsandMarkets. Companies using AI report productivity jumped 30%, revenue grew 13-15%, and sales cycles became 68% shorter. 83% of sales teams with AI saw revenue growth vs 66% without.
https://www.marketsandmarkets.com/AI-sales/ai-sales-tools-whats-changing
[38] "What is AI Cold Calling? Tools, Strategies & Best Practices"
Outreach. 54% of sales teams now use AI for personalized outreach. Personalization that previously required 15-20 minutes of research per prospect now happens automatically.
https://www.outreach.io/resources/blog/ai-cold-calling
Customer Support & AI
[39] "Call Center Voice AI: How AI is Transforming Support Operations"
Assembled. Studies show agent after-call work dropped by 35% with AI; 81% of agents at companies without AI were overwhelmed by information during calls, compared to 53% at companies with AI deployed.
https://www.assembled.com/page/call-center-voice-ai
[40] "McKinsey: Seizing the Agentic AI Advantage"
McKinsey. In layered AI approaches, AI handles specific steps autonomously—classifies tickets, identifies root causes, resolves simple issues—delivering 20-40% savings in time and 30-50% backlog reduction for internal support operations.
https://www.mckinsey.com/capabilities/quantumblack/our-insights/seizing-the-agentic-ai-advantage
[41] "33 Crucial Customer Service Statistics (2026)"
Shopify/Katana. 49% of customers prefer talking to a live human over an AI chatbot for support; more than half say they'll switch to a competitor after just one bad experience.
https://www.shopify.com/blog/customer-service-statistics
[42] "10 Proven Ways AI Reduces Average Handle Time (2025)"
Dialzara. AI automation cuts documentation time by 50%; Observe.AI achieved 50% reduction in after-call work time; Zendesk's AI assistant improved first-call resolution by 18%.
https://dialzara.com/blog/10-ways-ai-reduces-average-handle-time-aht-in-2024
[43] "Boost SaaS Onboarding with AI-Powered Guidance"
Navigable AI (LinkedIn). Smart ticket classification cuts support time by 25-40%; ticket resolution time down 40% with AI-powered triage and routing.
https://www.linkedin.com/posts/navigable-ai_5-onboarding-questions-your-ai-assistant-activity-7411262970443481088-ECdO
[44] "AI Helpdesk vs Traditional IT Support: The Complete 2025 Comparison"
Dialzara. AI tools boost customer satisfaction by up to 66%; per-ticket costs drop from $17-22 to $3 with AI automation; Zendesk AI agents can automate up to 80% of customer interactions.
https://dialzara.com/blog/ai-help-desk-tools-vs-traditional-support-systems
[45] "How AI Changes CRM: What Businesses Need to Know"
Rapid Online. HubSpot's Customer Agent resolves over 50% of support tickets automatically with 40% less time closing tickets; Reddit cut support resolution times by 84% with Salesforce Agentforce.
https://www.rapidionline.com/blog/ai-crm
UI/UX Design Principles
[46] "Cognitive Load Theory"
John Sweller / ScienceDirect. Foundational research demonstrating that human working memory has limited capacity—excessive information causes performance degradation. Informs the "3-5 key signals" design principle.
https://www.sciencedirect.com/topics/psychology/cognitive-load-theory
[47] "Building AI Trust: The Key Role of Explainability"
McKinsey. Research showing 40% of organizations identify explainability as a key AI risk—hidden evidence undermines user trust, leading to override behavior.
https://www.mckinsey.com/capabilities/quantumblack/our-insights/building-ai-trust-the-key-role-of-explainability
[48] "Designing AI User Interfaces That Foster Trust and Transparency"
UXMatters. Interfaces presenting context-sensitive explanations that show reasoning behind AI outputs—without overwhelming users—form the foundation of progressive disclosure in AI interfaces.
https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php
Strategic Impact & Competitive Dynamics
[49] "The State of AI in 2025"
McKinsey. AI adoption leaders see performance improvements 3.8x higher than the bottom half of adopters, demonstrating the compounding advantage of early AI interface adoption.
https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai
[50] "Beyond Code Generation: How AI is Changing Tech Teams' Dynamics"
Stack Overflow. Traditional onboarding takes 3-6 months to full productivity; AI-enabled onboarding with standardized formats achieves productivity in 1-2 weeks.
https://stackoverflow.blog/2025/10/06/beyond-code-generation-how-ai-is-changing-tech-teams-dynamics/
[51] "The Widening AI Value Gap: Build for the Future 2025"
BCG. The gap between AI leaders and laggards is widening rapidly—early adopters gain compounding advantages that late movers struggle to close.
https://web-assets.bcg.com/e4/14/440d165648a29f02376854919b4d/indias-triple-ai-imperative.pdf
[52] "Unsticking Your AI Transformation"
Bain & Company. "You can't automate your way to transformation. You have to rethink the work itself." Process redesign—not technology—creates most of the transformation value.
https://www.bain.com/insights/unsticking-your-ai-transformation/
[53] "Agentic AI Predictions 2028"
Gartner. 15% of day-to-day work decisions will be made autonomously by 2028, up from near 0% in 2024—a fundamental shift in how knowledge work interfaces operate.
https://www.gartner.com/en/articles/intelligent-agent-in-ai
Note on Sources
This guide draws on a combination of primary research, industry reports, and practitioner observations from LinkedIn and other platforms. Where possible, we cite specific statistics with their original sources. Framework concepts (Two-Pane Cognition, Proposal Cards, Ralph Wiggum Protocol, etc.) are original to LeverageAI's methodology library unless otherwise attributed.
All sources were accessed in January 2026. URLs are provided for verification; content may have been updated since access.