Designing for Human-AI Pairs

Look Mum No Hands

Using CRM and Not Looking at Fields — The Shift from Record Navigation to Decision Navigation

Why proposal cards are the new UI primitive for AI-augmented knowledge work

What You'll Learn

LeverageAI

January 2026

01
Part I: The Shift

The Coding Canary

What happens when elite programmers stop looking at the code they generate—and become more productive than ever?

"I haven't seen a line of code in nine months."

This isn't a confession of laziness. It's not the admission of a hobbyist or someone who stepped away from the keyboard. I've written more code this year than in my entire career—thousands of lines of production software, complex integrations, entire applications from scratch. And I haven't read most of it.

The statement sounds impossible, perhaps even irresponsible. How can you code without seeing code? How can you ship software you haven't reviewed line by line? Yet this is precisely what's happening at the cutting edge of software development. And the pattern emerging there will reshape every knowledge-work interface in the next five years.

The Elite Programmers Are Doing It Too

I'm not alone in this shift. Some of the most accomplished programmers in the world are publicly describing the same experience—and they're processing it in real time, with all the nuance and contradiction that implies.

On Vibe Coding

"Vibe coding is maybe the most democratizing idea of the whole year. It's this threshold we crossed where people could build really impressive programs just by explaining what they wanted in plain English."

— Andrej Karpathy

Andrej Karpathy—former director of AI at Tesla, OpenAI co-founder, and one of the most respected minds in machine learning—coined the term "vibe coding" in February 2025. He described it as fully giving in to the vibes, embracing exponentials, and forgetting that the underlying code even exists.

But Karpathy's relationship with this new paradigm is more nuanced than the viral term suggests. By year's end, when reflecting on his 8,000-line nanochat project, he admitted something striking:

"It's basically entirely hand-written (with tab autocomplete). I tried to use claude/codex agents a few times but they just didn't work well enough at all and net unhelpful." — Andrej Karpathy, on his nanochat project1

This isn't a contradiction—it's a crucial distinction. Vibe coding works brilliantly for exploratory work, rapid prototyping, and certain kinds of application development. For production foundations—the load-bearing walls of a system—the pattern is different. The lesson isn't that AI coding tools don't work; it's that knowing when to supervise versus when to hand-craft is the new core skill.

The Numbers from Elite Practice

Theo Browne—500,000+ subscribers, elite developer, founder of multiple successful projects—documented what this looks like in practice:

11,900

lines of production code generated without opening an IDE

Source: Medium, "Claude Code is Redefining Software Engineering"2

90%

of personal code now AI-generated

Source: LinkedIn3

Browne's workflow isn't "write less code." It's "write differently." More time in taste, judgement, and product thinking. Less time in keystrokes. The phrase he uses: "Stop micromanaging, describe outcomes."

From Code Producer to Creative Director

What's happening isn't just a productivity boost—it's a redefinition of the role itself.

"The role of 'software developer' just changed. Developers describe their role less as 'code producer' and more as 'creative director'. The shift: Strategy (define what's needed) → Delegation (let AI draft) → Verification (ensure it works)." — LinkedIn, 20254

This three-part workflow—Strategy → Delegation → Verification—is the new shape of knowledge work. It's not unique to coding. It's a pattern that will propagate across every domain where humans produce artifacts from intent.

The Abstraction Layer

What made this possible? AI tools created a new abstraction layer between intent and artifact.

Claude Code

Shows summaries and test results, not line-by-line code. You see what changed and whether it works.

Cursor

Presents diffs, not files. You review deltas from current state to desired state.

GitHub Copilot

Suggests completions in context. You accept, reject, or iterate—but you don't type from scratch.

The interface shifted from "here's all the code" to "here's what changed and whether it works." This is the fundamental move: from artifact display to outcome verification.

The New Workflow Pattern

The Workflow Transformation

❌ Old Workflow

Write code Debug Test Ship

✓ New Workflow

Describe intent Generate Validate outcome Approve/reject

The code exists. But you don't need to look at it. You're supervising an execution layer, not hand-crafting artifacts. The tests pass. The behavior is correct. The diff is clean. Why would you read every line?

Why "Not Looking at Code" Works

There's a counterintuitive insight hidden in the senior developer experience: AI's first pass is often genius-level. Fresh generation from a clear prompt produces surprisingly coherent, well-structured code.

But editing existing code? That's where AI struggles. It tacks on bits. It misses holistic design. It adds a date field by squishing it into the corner rather than redesigning the whole page.

The Regeneration Principle
if (design_changed_materially) {
delete_code();
regenerate_from_spec();
} // Don't patch. Rebuild.

This inverts the traditional relationship between code and design. In the old world, code was the asset—you accumulated it carefully, protected it, maintained it. In the new world, design is the asset. Code is ephemeral. Specifications and acceptance tests are durable. Code can be regenerated whenever the spec changes.

This is why delete-and-regenerate beats patch-and-accumulate. It's why senior developers trust the AI's first pass but get suspicious when they're asking it to "just add this one thing."

The Canary in the Coal Mine

Coding got here first for three reasons:

🎯
Economic Focus

AI companies invested heavily in coding tools early. Good use case, revenue from API calls, visible ROI.

🚀
Early Adopters

Developers are the natural first users of AI tools. They understand the technology and have high tolerance for experimentation.

Clear Success Criteria

Code either works or it doesn't. Tests pass or fail. This makes verification straightforward.

But the pattern isn't unique to coding. It will propagate to every data-heavy, artifact-producing domain. Anywhere humans currently navigate raw information to produce an output, the same shift will occur.

The Question This Raises

If senior programmers no longer need to see code to code...

  • Do CRM users need to see 80 fields to sell?
  • Do support agents need to see 3 years of ticket history to resolve an issue?
  • Do analysts need to navigate 50 reports to make a decision?
  • Does anyone need to see all the data to do their job?

The coding world stopped looking at the artifact.
It started looking at the outcome.

What happens when every other knowledge-work interface does the same?

Key Takeaways

  • 1 Elite programmers have stopped reading the code they generate—and they're more productive. The code exists; they just don't need to see it.
  • 2 The shift is from "code producer" to "creative director"—Strategy → Delegation → Verification. The role has been redefined.
  • 3 Code is now ephemeral; design and intent are the durable assets. Delete and regenerate beats patch and accumulate.
  • 4 Coding is the early warning signal for every other knowledge-work interface. The same pattern will propagate.
02
Part I: The Shift

CRMs Are Databases With Cosplay

The numbers reveal a brutal truth: your sales team spends most of their time navigating, not selling.

34%

of time spent actually selling

66%

navigating and administering

Source: Revenue Grid, "How to Customize Salesforce for Maximum Efficiency"6

Sales reps spend only 34% of their time actually selling. The rest—nearly two-thirds of every working day—is consumed by administrative tasks and navigating poorly optimized systems. For a team of 10 reps, that translates to over 25,000 hours annually spent on activities that don't close deals.

This isn't a people problem. It's not a training problem. It's not even a data problem. It's an interface problem. The tools we've given salespeople were designed for a different era—and they're drowning in them.

The Time Sink Quantified

286

hours per year

wasted on manual data entry per rep

Source: LinkedIn/xdge7

Let's put that 66% into concrete terms. The average rep wastes 286 hours per year on manual data entry alone—not selling, not building relationships, not closing deals. Just typing data into fields.

That's over seven weeks of full-time work. Every year. For each rep. Just entering data.

And even after all that typing, the data is incomplete. The system designed to capture your customer relationships is leaking context at every interaction.

The Context Loss Problem

Here's a number that should alarm every revenue leader:

6,000

words spoken

per sales call

30-60

words captured

in the CRM

Less than 1% of the conversation makes it into your database.

Source: Gong, "Why AI Systems of Action Replace Traditional CRM Workflows"8

An average sales call generates about 6,000 words of conversation. The prospect mentions their procurement timeline. The budget holder raises a concern. The champion reveals internal politics. The decision-maker hints at a competing priority.

What makes it into the CRM? "Good call. Follow up next week."

Your database is filled with surface-level summaries that miss what actually matters. The nuance, the objections, the relationship signals—all lost. And when the rep changes roles, or the account gets transferred, that context evaporates entirely.

"CRM should be a profit center, not a cost center. If Salesforce only records activity, it is an expense. If it enforces revenue flow, it becomes an asset." — Ameet Patra, LinkedIn9

What CRMs Were Actually Designed For

Traditional CRMs were built on a simple premise: show the user all the data, let them figure out what to do.

The CRM presents itself as a "system of record"—a faithful mirror of reality. Every interaction logged. Every field populated. Every relationship documented. The interface is a database with a pretty face.

The Traditional CRM Navigation Model

Find account Open record Read 80 fields Scroll activity history Check contacts Decide Act Log

Eight steps. And the user is responsible for synthesis at every transition.

Why This Made Sense (Once)

Before AI, humans were the only processors of context. You needed to see the data to make sense of it. The assumption was straightforward: more fields = more information = better decisions.

The CRM was a filing cabinet you could search. A shared memory for the organization. A place to store everything so nothing got lost.

And for its time, this was progress. Better than spreadsheets. Better than post-it notes. Better than each rep keeping their own mental database of customers.

Why This Fails Now

But humans aren't good at synthesizing 80 fields and 3 years of history. We're terrible at it, actually. We miss things. We get tired. We see what we expect to see. We forget what we read two screens ago.

Power Users

Develop workarounds. Ignore most fields. Build mental shortcuts. Create their own systems alongside the official one.

New Users

Overwhelmed from day one. "Where do I even look?" Training takes months. Mistakes are inevitable.

The interface forces cognitive labor that could be automated. Every click, every scroll, every "where was that field again?" is friction between intention and action.

System of Record vs System of Action

The distinction that matters:

System of Record

"Here's everything that happened"

  • • Mirror of reality
  • • Comprehensive history
  • • All fields visible
  • • User synthesizes

An expense if it only records.

System of Action

"Here's what you should do next"

  • • Decision engine
  • • Relevant context surfaced
  • • Proposals prepared
  • • System synthesizes

An asset if it enforces revenue flow.

Most CRMs are systems of record pretending to be systems of action. They have "next best action" features bolted on the side—an AI suggestion widget floating in the corner of the same old record view. It's an afterthought, not an architecture.

The Interface Is the Problem

It's not the data that's wrong. Companies have invested millions in data quality, in integrations, in enrichment services. The data is better than ever.

It's how the data is presented. Showing everything means the user has to filter everything. The cognitive load sits on the human, not the system.

1

Every click is friction.

2

Every scroll is friction.

3

Every "where was that field?" is friction.

4

Every tab-switch is friction.

5

Every manual synthesis is friction.

Friction accumulates. It's the hidden tax on every interaction. And the bill comes due in that 66% of time spent not selling.

The Parallel to Coding

Here's where Chapter 1 connects to the present. Just as code was a proxy for intent—lines and functions standing in for desired behavior—fields are a proxy for understanding. They're the low-level representation of something higher: customer relationships, deal momentum, buying signals.

The Parallel

In Coding

Developers stopped looking at code and started looking at outcomes.

In CRM

Users need to stop looking at fields and start looking at decisions.

The artifact (record) isn't the point. The action is.

CRMs are databases wearing a costume.

It's time to redesign for what users actually need:
not "show me everything"
but "tell me what to do next."

Key Takeaways

  • 1 Sales reps spend 66% of their time NOT selling—navigating and administering. The interface is the bottleneck.
  • 2 6,000 words of conversation → 30-60 words in CRM. Massive context loss at every interaction.
  • 3 CRMs were built as systems of record; users need systems of action. The distinction is architectural, not cosmetic.
  • 4 The interface is the problem, not the data. Showing everything forces humans to filter everything.
03
Part I: The Shift

From Record Navigation to Decision Navigation

The shift isn't about better UI. It's about a different question entirely.

Record Navigation

"What do you want to see?"

Decision Navigation

"What do you want to decide?"

The shift happening in coding isn't coding-specific. It's a change in the fundamental unit of interaction—and it applies to every knowledge-work interface.

This isn't just cleaner UI. It's not better color schemes or more intuitive layouts. It's a different design primitive altogether. And understanding the distinction is the key to designing interfaces for the AI era.

The Framework

Dimension Record Navigation Decision Navigation
Primary Question "What do you want to see?" "What do you want to decide?"
Primary Object The record The proposal
User's Job Navigate to find context Evaluate the recommendation
AI's Job Search and retrieve Synthesize and recommend
Detail Model Default (show everything) On-demand (show when needed)
Cognitive Load On the human On the system
Success Metric "Did you find it?" "Did you decide well?"

This framework is the core lens for everything that follows. Subsequent chapters apply this distinction to specific patterns and implementations.

What Record Navigation Looks Like

In record navigation, the user is a spelunker—searching for relevant information in a sea of data.

The Spelunking Workflow

Open account Scan fields Open activities Read emails Check contacts Decide

Every decision requires assembly. Pull from here. Pull from there. Synthesize mentally. Hope you didn't miss anything.

The system is a warehouse. The user is the picker.

What Decision Navigation Looks Like

In decision navigation, the AI did the spelunking. The user is a governor—reviewing the synthesis, not creating it.

The Governance Workflow

Open queue See proposal card Review evidence Approve / Modify / Reject

Every decision is pre-assembled. "Here's what to do. Here's why. Here's the evidence."

The system is a recommender. The user is the governor.

The Design Primitive Shift

Old Primitive

The Record

A container of fields. All data visible. User navigates.

New Primitive

The Proposal

A recommendation with evidence. Detail on-demand. User governs.

This changes everything: layout, workflow, what's visible by default, what's hidden. It's not about removing features—every field is still accessible if you need it. It's about changing the default. Most of the time, you don't need it.

Why This Isn't Just "Better UX"

Cleaner UI
  • • Same model, fewer distractions
  • • Record view, but prettier
  • • User still navigates
  • • Cognitive load unchanged

Polishing the same paradigm.

Decision Navigation
  • • Different model, different interaction
  • • Proposal view with drill-down
  • • User governs, not navigates
  • • Cognitive load on system

Replacing navigation with supervision.

The user's role changes from finder to judge. That's not a cosmetic difference. It's a fundamental redefinition of what the interface is for.

The Direction: Generative UI

Jakob Nielsen—the godfather of usability research—made a striking prediction for 2026:

"The concept of a static interface where every user sees the same menu options, buttons, and layout, as determined by a UX designer in advance, is becoming obsolete. Software interfaces are no longer hard-coded; they are drawn in real-time based on the user's intent, context, and history." — Jakob Nielsen, "18 Predictions for 2026"10

Nielsen calls this "Generative UI"—interfaces that assemble themselves based on what the user is trying to accomplish. Not one-size-fits-all screens, but bespoke micro-interfaces that dissolve when the task is complete.

Decision navigation is the first step toward this future. Instead of hard-coded record views, you get contextual proposal cards. The interface responds to intent, not just clicks.

The Pattern: Hide the Artifact, Surface the Summary

The Same Pattern, Two Domains

In Coding

Hide: The code

Surface: The diff summary

Drill-down: Full file when suspicious

In CRM

Hide: The fields

Surface: The proposal card

Drill-down: Full record when needed

When the AI is right, you move fast. When it's wrong, you can quickly see where.

What This Means for Design

If you're designing interfaces, the shift requires a new starting point:

1

Stop designing screens around records.

The record is the artifact, not the goal.

2

Start designing screens around decisions.

What action does the user take from here?

3

Ask: What do they need to see to act confidently?

Not everything—just what supports the decision.

4

Everything else is drill-down.

Available, but not blocking the flow.

The Cognitive Load Transfer

The fundamental shift is where cognitive load lives:

Record Navigation
🧠

User does synthesis

(High cognitive load)

Decision Navigation
🤖

System does synthesis

(Low load on user)

This isn't about making users dumber. It's about making systems smarter. Humans are better at judgment than navigation. Let them do what they're good at.

Record navigation asks:
"What do you want to see?"

Decision navigation asks:
"What do you want to decide?"

The second question is the right question.

Key Takeaways

  • 1 The shift is in the unit of interaction: from record to proposal. The primary object changes; everything follows.
  • 2 Record navigation puts cognitive load on humans; decision navigation puts it on systems.
  • 3 This isn't just cleaner UI—it's a different design primitive. Cleaner record views miss the point.
  • 4 Design around decisions, not records. Everything else is drill-down.
04
Part II: The Architecture

Two-Pane Cognition

The architecture for human-AI pair interfaces: what if human and AI each had their own workspace—and they talked to each other?

"What if the human and AI each had their own pane—
and they talked to each other?"

We've established what we're shifting to: decision navigation instead of record navigation. Now we need how. What's the architecture for human-AI pair interfaces? How do you actually design a screen where human and AI work together?

The answer isn't a chatbot bolted onto your existing interface. It's not an AI assistant widget floating in the corner. It's a fundamentally different screen architecture12—one that gives both human and AI their own space to work, with a clear contract between them.

I call it Two-Pane Cognition.

Where This Comes From

The pattern emerges from observation: how do senior engineers actually work with AI coding tools?

They don't read all the code. As we saw in Chapter 1, they've moved past that. Instead, they:

1

Read the proposal (the diff summary, the suggested change)

2

Check the evidence (tests pass, behavior correct)

3

Make a judgment (approve, reject, or modify)

The tool does preparation. The human does decision.19 That's the pattern we're extracting—and it applies far beyond coding.

The Two-Pane Model

Two-Pane Cognition divides the interface into three distinct areas, each with a clear purpose:

🧠

Human Pane

What only humans can decide

Outcomes

What are we trying to achieve?

Priorities

What matters most right now?

Approvals

Yes / no / modify on proposed actions

Exceptions

Edge cases and judgment calls

High-signal summaries

The "so what?" distillation

🤖

AI Pane

What AI prepares

Drafts

Pre-written artifacts ready to use

Candidate actions

"Here's what I think you should do"

Retrieved evidence

The specific data points used

Uncertainty signals

Confidence levels, what's missing

"Here's what I tried"

The reasoning trace

🤝

Shared Surface

The contract between human and AI

Constraints

What AI is NOT allowed to do

Permissions

What tools/actions AI can invoke

Acceptance Criteria

How we know "good enough"

Feedback Handles

Where to correct when wrong

How This Looks in CRM

Let's make this concrete. Compare a traditional CRM screen to a Two-Pane CRM screen:

❌ Traditional CRM Screen
Account Name
Industry
Revenue
Employees
Website
Phone
... 74 more fields ...

Activity Timeline (200 entries)

Email sent Jan 15...
Call logged Jan 12...
Meeting Dec 28...
↓ scroll for more ↓

User navigates → synthesizes → decides. Cognitive load on human.

✓ Two-Pane CRM Screen

Human Pane: Decision Queue

Acme Corp — needs attention
TechStart Inc — follow up
GlobalTech — proposal review

"These accounts need attention today"

AI Pane: Proposal for Acme Corp

Action: Call John, ask about Q2 budget

Why: 17 days since touch • Champion role change

Draft: Call agenda ready →

Evidence: 3 items ▼

"Here's what I propose, here's why"

AI prepares → human governs. Cognitive load on system.

The Interaction Flow

Here's how the two panes interact:

1

Human selects from decision queue

The queue is prioritized by AI based on urgency, opportunity, and strategy alignment.

2

AI pane shows the proposal

Proposed action + draft artifact + reasoning + evidence. Everything needed to decide.

3

Human reviews and responds

Approve / modify / reject / ask for more. The human governs; the AI awaits instruction.

4

If approved: action executes

Logged with full trace. The system knows what happened, why, and who approved it.

5

If rejected: feedback captured

AI learns at the right layer. The playbook improves, not just this one interaction.

Why "Bolted-On Assistants" Fail

Most CRM AI today is a chatbox in the corner of the same old record screen. The assistant is an add-on, not a co-pilot.16 You still navigate the record; the AI just answers questions about it.

"The assistant bolted onto the old UI feels lame because it's a chatbox glued onto a cockpit—but the cockpit wasn't designed for a co-pilot that can actually fly."

The fundamental problem: the interaction model hasn't changed. You're still spelunking through records. The AI is just a slightly faster way to search.

Two-Pane Cognition changes the model itself. The AI doesn't just answer questions—it prepares proposals. The human doesn't navigate—they govern.17

The Right Division of Labor

AI Does Well Human Does Well
Retrieval — finding relevant information in vast datasets14 Judgment — knowing when rules don't apply
Synthesis — combining data points into summaries Exceptions — handling edge cases with nuance
Drafting — generating first-pass artifacts Priorities — deciding what matters most
Pattern-matching — recognizing similar situations Relationships — understanding human dynamics15
Consistency — applying rules uniformly Strategy — setting direction and goals

The principle: don't ask AI to do what humans do well, and don't ask humans to do what AI does well.13 Navigating 200 activities to find the relevant one? AI's job. Deciding whether this account needs a different approach because the champion just left? Human's job.

Example: A Sales Rep's Morning

Old Workflow

  1. 1.Open CRM
  2. 2.Check dashboard
  3. 3.Pick an account
  4. 4.Open record
  5. 5.Read history
  6. 6.Read emails
  7. 7.Check contacts
  8. 8.Decide what to do
  9. 9.Draft email
  10. 10.Send
  11. 11.Log activity

~20 minutes per account

Two-Pane Workflow

  1. 1.Open CRM → see prioritized decision queue (AI prepared overnight)
  2. 2.Select top account → see proposal card
  3. 3.See evidence cards: "17 days since touch; champion changed roles"
  4. 4.See draft: call agenda already written
  5. 5.Approve → call happens → AI logs with trace

~3 minutes per account

Designing for Two-Pane

When designing a Two-Pane interface, work through these questions in order:

1. What decision does the user make here?

This is your human pane. The decision queue. The judgment point.

2. What does the user need to see to decide confidently?

This is your proposal structure. Action, reasoning, evidence.

3. What can AI prepare in advance?

This is your AI pane. Drafts, synthesis, retrieved context.

4. What's the minimum human input to approve/modify/reject?

This is your interaction design. One click? Edit field? Selection?

The human is still in control.
But the human isn't navigating.

The human is supervising.

Key Takeaways

  • 1 Two-Pane Cognition: Human pane (decisions) + AI pane (proposals) + Shared surface (contract).
  • 2 Right division of labor: AI does retrieval, synthesis, drafting. Human does judgment, exceptions, priorities.
  • 3 Bolted-on assistants fail because they don't change the interaction model. Chatbox ≠ co-pilot.
  • 4 Design screens around decisions, not records. Start with "what judgment does the user make here?"
05
Part II: The Architecture

Proposal Cards: The New Primitive

Every card has four parts. Learn them, and you can redesign any interface.

We've established the architecture: Two-Pane Cognition, with human and AI each owning their own space. We've distinguished record navigation from decision navigation. But what, exactly, does the AI pane show?

In record navigation, the primitive is obvious: the record. A container of fields. All data visible. User navigates. This is the atom of traditional enterprise software—every screen is built from records, related records, lists of records.

Decision navigation needs its own atom. Its own fundamental unit. Something as universal as the record, but designed for human-AI pairs instead of humans alone.

That unit is the proposal card.

"A record is a container of data. A proposal card is a container of decision."

The Anatomy of a Proposal Card

Every proposal card has exactly four components. Not three. Not five. Four. Each serves a distinct function, and removing any of them breaks the system.

The Four Components

1

ACTION

What the AI suggests doing. The headline. What you're being asked to approve.

"Call John tomorrow at 2pm"

"Send follow-up email"

"Update stage to Negotiation"

"Schedule demo with technical team"

2

DRAFT

The prepared artifact. Already written. Ready to send or use.

Personalized email copy

Call agenda with talking points

Proposal outline with key sections

Meeting invite with context

3

WHY

The reasoning. Compact, scannable, decision-relevant. Shows the logic without verbose explanation.

"17 days since last touch"

"Procurement mentioned Q1 budget"

"Champion changed roles"

"Competitor mentioned in last call"

4

EVIDENCE

The specific data points used. Collapsible, on-demand. The receipts for verification.

Link to email where budget mentioned

Snippet from call transcript

LinkedIn update showing role change

CRM field: "Last Activity: 17 days ago"

Why Four Components?

This isn't arbitrary. Each component serves a function that, if missing, causes the system to fail:

ACTION

Without this, there's nothing to approve. The card is pointless.

DRAFT

Without this, the human does the work. The AI prepared nothing. What was the point?

WHY

Without this, AI is a black box. Trust fails. "Why should I do this?" goes unanswered.

EVIDENCE

Without this, you can't verify. Governance fails. Skeptics are right to reject.

Adding more components creates cognitive overload.22 Removing any creates a gap that undermines the whole pattern. Four is the magic number—not by accident, but by function.

A Proposal Card in the Wild

Let's see what this looks like in practice. Here's a proposal card as it might appear in a sales rep's morning queue:

Acme Corporation
Call John Morrison
High priority
Action
Call tomorrow 2pm — discuss Q2 budget timing
Why now
17 days since touch Champion role change Budget timing: Q2
Draft agenda

1. Congratulate on new VP role

2. Re-confirm Q2 budget availability

3. Address SSO integration questions from Dec call

Evidence
3 items supporting this recommendation

Notice what's present and what's absent. The rep sees everything needed to decide in seconds: who, what, why. The full agenda expands on click. The evidence is available but collapsed. The full account record—all 80 fields—is nowhere to be seen.

That's not hiding information. That's progressive disclosure20: showing what you need at each level, with more detail available on demand.

The Progressive Disclosure Principle

This is perhaps the most important insight in proposal card design: detail is progressive, not default.

The Three Layers of Disclosure

L1
Default View
~5 seconds to scan

Action + Why — The headline and the reasoning. "Call John because 17 days and budget timing." Decision-ready at a glance.

L2
One Click
~30 seconds to review

Draft + Evidence summary — The prepared artifact and the receipts. "Here's the full agenda. Here's where I got the budget info."

L3
Another Click
As long as needed

Full record + complete history — All 80 fields. Every activity. The full timeline. Everything, if you want it.

Most of the time, you never need to go past Layer 1. The Action and Why are enough. You approve and move on. Layer 2 is for when something feels off and you want to verify. Layer 3 is for the rare edge case where you need the complete picture.

"Show the decision-ready summary. Drill down when suspicious. Full record is one click away but not blocking the flow."

This is how trust scales. You don't verify everything by default—that's exhausting and slow. You verify when you need to, and the system makes verification easy.23 Trust but verify, with verify on demand.

The Coding Tool Parallel

This pattern will feel familiar if you've used modern coding tools. They follow the exact same structure:

In Coding Tools

Layer 1
Diff summary: "+42 lines, -18 lines"
Layer 2
The actual changes (green/red)
Layer 3
Full file with all context

In Proposal Cards

Layer 1
Action + Why summary
Layer 2
Draft artifact + Evidence
Layer 3
Full record with all fields

The pattern is identical because the problem is identical. Developers were drowning in code the way CRM users are drowning in fields. The solution in both domains: summarize by default, drill down on demand.

Most of the time, the diff summary is enough. You see "+42 lines, tests passing, no conflicts" and you merge. You don't read every character. The same economy applies to proposal cards: most of the time, "Call John because 17 days" is enough. You don't need all 80 fields.

The User's Decision Tree

When a user reviews a proposal card, they're running through a mental checklist. Understanding this checklist helps you design better cards:

The Approval Flowchart
Step 1: Is the ACTION right?
If no → Reject or modify
↓ yes
Step 2: Is the DRAFT good enough?
If no → Edit
↓ yes
Step 3: Does the WHY make sense?
If no → Check evidence
↓ yes
Step 4: Does the EVIDENCE check out?
If no → Reject or investigate
↓ yes
APPROVE

The key insight: most cards pass steps 1-3 quickly. Evidence drill-down is the exception, not the rule. When the AI is well-calibrated, the majority of cards get approved at Layer 1 in seconds.

This is the productivity gain: not doing everything faster, but not doing most things at all. The AI did the work. The human governs.

When Cards Get Rejected

Rejection isn't failure—it's feedback. And how you capture rejection feedback determines whether the system learns.

The critical principle: feed back at the right layer. If the playbook is wrong, fix the playbook—don't just reject this one card and leave the playbook broken. That's how systems learn.

This mirrors the coding insight from Chapter 1: fix the spec, not the code. When AI generates wrong code, you don't patch the code—you clarify the specification and regenerate. When AI generates wrong proposals, you don't just skip this one—you update the playbook so all future proposals improve.

The Confidence Signal

Some teams add a fifth element to their proposal cards: a confidence indicator. This isn't required, but it can be useful for prioritization.

🟢
High Confidence
Approve quickly. AI is certain.
🟡
Medium Confidence
Review the Why. Something's uncertain.
🔴
Low Confidence
Verify evidence. High-risk recommendation.

The confidence signal can be explicit (a visible badge on each card) or implicit (sort the queue so high-confidence items are at the top). Either way, it helps users allocate their attention: spend more time on uncertain recommendations, less on confident ones.

A well-calibrated system should have most cards at high confidence. If everything is medium or low, the AI isn't providing enough value—it's just creating more work. The goal is a queue where most items sail through with a single approval click.

The Delete-and-Regenerate Test

Here's a quality test for your proposal cards: Can the AI regenerate this card from scratch?

This test matters because it ensures your system is stateless at the right layer. The proposal can be regenerated anytime. The durable assets are the playbook (your decision rules) and the data (your CRM + external signals). The card itself is ephemeral—a computed view, not a stored artifact.

This is the same insight from coding: code is ephemeral, specs are durable. Proposals are ephemeral, playbooks are durable.

Designing Proposal Cards for Your Domain

Ready to design your own? Start with three questions:

1
What action does the user typically take?

This becomes your ACTION component. What are they actually doing when they complete a task in this workflow?

2
What do they need to see to feel confident?

This becomes your WHY + EVIDENCE components. What context makes them say "yes, this is the right thing to do"?

3
What artifact do they produce?

This becomes your DRAFT component. What output results from the action—an email, a document, an update?

Assemble those answers into a card structure. Test with real users. Iterate. The four-component pattern is universal, but the specifics—what counts as evidence, how detailed the draft should be, what "why" means in your context—are domain-specific.

Common Mistakes to Avoid

When teams first implement proposal cards, they often stumble on these:

❌ Showing too much by default

Cognitive overload. The whole point is progressive disclosure. If Layer 1 has 20 data points, you've recreated record navigation with extra steps.

❌ Hiding the evidence entirely

Opacity kills trust.21 Users need to be able to verify. Even if evidence is collapsed by default, it must be one click away. Black-box recommendations get rejected.

❌ No drill-down path

Users feel trapped. "Wait, I can't see the full account record?" Progressive disclosure requires that Layer 3 exists. You can hide it, but not remove it.

❌ No feedback capture on rejection

The system doesn't learn. Every rejection should ask "why?" and feed back to the right layer—data, playbook, template, or context.

Get these right, and your proposal cards will feel natural—like they were always how the system should have worked. Get them wrong, and you've just created a new kind of frustration layered on top of the old one.

"A record is a container of data. A proposal card is a container of decision. The card says: 'Here's what to do, here's why, and here's the proof.' That's the new primitive."

Key Takeaways

  • 1 The proposal card has four parts: Action, Draft, Why, Evidence. Each serves a function; remove any and something breaks.
  • 2 Progressive disclosure is the key: summary by default, evidence on demand, full record available. Most decisions happen at Layer 1.
  • 3 Rejection is feedback, not failure. Capture why, and feed back to the right layer—data, playbook, template, or context.
  • 4 Design cards by asking: What action? What confidence? What artifact? Assemble, test, iterate.
06
Part II: The Architecture

The Evidence Layer

Trust doesn't come from showing everything. Trust comes from showing the receipts.

The Trust Objection

By now, you might be feeling a familiar objection rising. It goes something like this:

"Users want to see the data. They don't trust AI. If you hide all the fields behind proposal cards, they'll reject everything because they can't verify anything."

This objection is valid. Users should be skeptical of AI recommendations. Black-box suggestions get rejected—and rightly so. Nobody wants to approve actions they can't understand or verify.

But the solution isn't "show everything." The solution is show the receipts.

Trust comes from inspectable evidence, not from raw data volume. And there's a world of difference between the two.

Why "Show Everything" Doesn't Build Trust

The intuition seems reasonable: more data means more informed decisions. If users can see everything, they can verify everything. Complete access equals complete confidence.

But this intuition is wrong. Here's what actually happens when you show everything:22

What "Show Everything" Produces

  • Cognitive overload: 80 fields and 200 activities = paralysis, not clarity
  • Scanning and skipping: Users skim when overwhelmed; they miss things
  • Illusion of completeness: "I saw it all" without actual understanding
  • Decision fatigue: Each field is a micro-decision about relevance

What Actually Builds Trust

  • Inspectable reasoning: "Here's why I suggested this"
  • Verifiable evidence: "Here's what I looked at"
  • Admitted uncertainty: "Here's what I don't know"
  • Consistent behavior: "I follow the same logic every time"

Trust requires understanding. Understanding requires synthesis. When you dump raw data on users, you're asking them to do the synthesis work—the exact work that AI should be doing for them.

The Evidence Layer solves this. It's not about hiding data. It's about presenting data in a way that enables efficient verification.18

Anatomy of an Evidence Card

In Chapter 5, we established that every proposal card has an EVIDENCE component. Now let's examine what that component actually contains.

An evidence card isn't a dump of all related data. It's a curated snippet that answers a specific question: "What did the AI look at to reach this conclusion?"

Evidence Card Structure

1
The Specific Data Point

Not the whole record—just the relevant snippet. The actual quote, the actual number, the actual date.

2
Where It Came From

Source attribution: email, call transcript, CRM field, LinkedIn, news article. Users need to know the provenance.

3
Why It Matters

The connection to the decision. "This email mentions budget timing, which supports the urgency of outreach."

4
Link to Full Source

One click to the original. If you want the full email, the complete transcript, the entire record—it's there.

Evidence Cards in Practice

Let's see what this looks like for a real proposal. Remember the "Call John" proposal from Chapter 5? Here are its evidence cards:

📅 Last Touch Date CRM Field
17 days ago
→ Supports: Relationship cooling. Typical re-engagement threshold is 14 days.
📧 Email Snippet Dec 8, 2024
"...we're finalizing Q1 budget next week. Would be good to have pricing locked in by then..."
→ Supports: Timing urgency. Budget finalization is imminent.
💼 LinkedIn Update Jan 3, 2025
John Morrison promoted to VP of Engineering
→ Supports: Champion role change. Re-establish relationship at new level; may have expanded authority.
📞 Call Transcript Nov 22, 2024
"...we're also looking at [Competitor]. They've been pretty aggressive on pricing..."
→ Supports: Competitive pressure. Urgency signal; don't let deal stall.

Notice what's present: specific quotes, specific dates, specific sources, specific relevance. Notice what's absent: the other 200 emails, the 50 other calls, the 80 CRM fields that weren't relevant to this particular decision.

The evidence layer shows the receipts—the specific inputs that led to this specific recommendation. It doesn't show the whole filing cabinet.

The Drill-Down Principle

Evidence works in layers. Most users stop early. Some dig deeper. The system should accommodate both without forcing either.

The Verification Funnel
Layer 1: Summary (WHY)
80% of users stop here
"17 days since touch • Budget timing • Champion role change"
↓ click
Layer 2: Evidence Cards
15% of users stop here
Specific snippets, sources, and relevance for each signal
↓ click
Layer 3: Full Source
5% of users need this
Complete email thread, full transcript, entire record
The verification funnel: each layer provides more detail, but most decisions don't require going deep.

This matches how trust actually works. You don't verify everything by default—that's exhausting and slow. You verify when something feels off. The system makes verification easy without making it mandatory.

What Makes Evidence Trustworthy

Not all evidence is equally convincing. Here's what separates trustworthy evidence from noise:

🎯
Specificity

"The email from Dec 8" not "recent communications." Vague references undermine trust.

✓ "John mentioned Q1 budget"
✗ "Budget was discussed recently"
🕐
Recency

Clear timestamps on everything. Old data with no date feels unreliable.

✓ "Dec 8, 2024"
✗ "at some point"
🔗
Relevance

Why this matters for THIS decision. Connection to the recommendation should be explicit.

✓ "→ Supports: timing urgency"
✗ [no explanation]
📍
Source Clarity

Where did this come from? Email, call, CRM, LinkedIn, news? Users need provenance.

✓ "From: email"
✗ "Based on data"
📊
Scope Signal

"Based on 4 signals" tells the user how much was considered. Missing scope feels incomplete.

✓ "4 evidence items"
✗ [no count]
🔍
Verifiability

One click to the original. If users can't verify, they can't trust.

✓ "View full email →"
✗ [no link]

Confidence Signals

Not all evidence is equal, and not all proposals deserve the same level of scrutiny. Some recommendations are high-confidence—clear signals, consistent pattern, straightforward decision. Others are low-confidence—sparse data, ambiguous signals, judgment call required.

A well-designed system surfaces this distinction so users can calibrate their review depth:

Confidence-Based Review Depth

High Confidence

Strong signals, consistent pattern, clear recommendation

Action: Approve quickly, minimal review

?
Medium Confidence

Mixed signals, some uncertainty, judgment helpful

Action: Review the Why, check key evidence

!
Low Confidence

Sparse data, ambiguous signals, high-risk decision

Action: Verify all evidence, consider alternatives

This can be shown explicitly (a badge on each card) or implicitly (sort the queue so high-confidence items appear first). Either way, users know where to focus their limited attention.

A well-calibrated system should have most proposals at high confidence. If everything is medium or low, the AI isn't providing enough value—it's just creating more work with different packaging.

The John West Principle

There's a deeper layer of trust that goes beyond "here's what I looked at." It's: "Here's what I considered but didn't recommend."

When a proposal card says "I recommend calling John, and here's why," that's helpful. But when it also says "I considered emailing first, but rejected it because response rates are 40% lower for this contact type"—now you know the AI actually evaluated options. The reasoning is comparative, not arbitrary.

This lets users second-guess intelligently. They can see the trade-offs. They can say "actually, I think email is right here because..." and that disagreement is informed, not blind.

We'll explore the John West principle in depth in Chapter 8, where overnight batch processing makes it economically feasible to generate and evaluate multiple candidates for every proposal. For now, know that the evidence layer can include not just "what I used" but "what I considered and rejected."

When Evidence Reveals Problems

Sometimes, when you drill into evidence, you discover something wrong. The AI used bad data. It missed something important. The reasoning doesn't hold up.

This is good. Better to catch it at review than after execution.

Transparency enables correction. When users can see the evidence, they can identify when it's wrong. When they can flag what's wrong, the system can learn.

The Correction Feedback Loop

1

User reviews evidence card

2

User spots problem: "This email quote is out of context"

3

User flags the issue with specifics

4

System captures feedback at the right layer (data, playbook, or model)

5

Future proposals improve

Opacity hides problems. When users can't see the reasoning, they can't catch errors. The errors persist. Trust erodes slowly as recommendations fail for reasons nobody can diagnose.21

Privacy and Security in Evidence

A brief but important note: evidence shouldn't expose more than necessary.

The principle is progressive rehydration. The AI reasons on abstracted data—role titles, relationship patterns, timing signals. The evidence layer shows specifics only when users drill down, and only to authorized users.

# How the AI sees it (abstracted)
contact_role: "Decision-maker"
relationship_status: "Champion"
last_touch: "17 days"
budget_signal: "Q1 timing mentioned"
# How evidence rehydrates (for authorized user)
contact_name: "John Morrison"
company: "Acme Corporation"
email_snippet: "...finalizing Q1 budget..."

This is architecture, not policy. The AI doesn't need to see "Jane Smith at 14 Smith Street" to reason about timing and relationship signals. It sees abstracted tokens. Real names and details appear only in the evidence layer, only for authorized viewers.

This principle—called SiloOS in our broader framework—ensures that privacy and security are built into the system, not bolted on as an afterthought.

The Evidence Layer as Governance

Here's a benefit that compliance teams love: every proposal card with evidence is an automatic audit trail.

When someone asks "Why did the system recommend X?"—the answer is built in. Here's the evidence, here's the reasoning, here's who approved it and when. No need to reconstruct after the fact. No forensic archaeology through logs. The explanation exists as a first-class artifact.

Governance Built-In

What Compliance Gets
  • • Evidence of what data was used
  • • Record of reasoning applied
  • • Timestamp of recommendation
  • • Record of human approval/rejection
  • • Feedback captured on rejected proposals
What This Replaces
  • • Manual audit log reconstruction
  • • "Why did we do that?" archaeology
  • • Separate documentation requirements
  • • Post-hoc explanation generation
  • • Defensive documentation theater

Governance becomes a feature, not a burden. The same evidence that builds user trust also satisfies auditor questions. The transparency that enables correction also enables compliance.23

"The evidence layer isn't about proving AI is right. It's about making it easy to check. When verification is efficient, trust follows."

Key Takeaways

  • 1 Trust comes from inspectable evidence, not from showing everything. More data creates overload, not confidence.
  • 2 Evidence cards contain: specific data point + source + relevance + link to full source. Curated, not comprehensive.
  • 3 Drill-down principle: summary by default, evidence on click, full record available. Most users stop early.
  • 4 Confidence signals help users calibrate review depth. High-confidence proposals sail through; low-confidence gets scrutiny.
  • 5 Transparency enables correction: when users can see evidence, they can catch errors. The evidence layer is governance built-in.
07
Part II: The Architecture

Supervise, Don't Navigate

Old CRM day: 8 steps. New CRM day: 4 steps. The workflow itself has changed.

We've covered the architecture: Two-Pane Cognition, with human and AI each owning their workspace. We've established the primitive: the Proposal Card, with its four components of Action, Draft, Why, and Evidence. We've explored how the Evidence Layer builds trust through inspectable reasoning.

Now the question becomes practical: What does the actual workflow look like?

The answer isn't just "faster." It's structurally different. Fewer steps. Different cognitive load. A fundamentally different relationship between human and system.

The Old Workflow

Here's what a typical CRM day looks like under record navigation—the model most systems still use:

The 8-Step Workflow (Record Navigation)

1
Open CRM
Log in, wait for dashboard to load
~30 sec
2
Check dashboard, pick an account
Scan list, decide where to start, click through
~2 min
3
Open the account record
Wait for 80 fields to load
navigation
4
Read activity history
Scroll through timeline, scan emails, read notes
navigation
5
Check contacts, opportunities, related records
Tab-switch to contacts, review opportunities, cross-reference
navigation
6
Decide what to do
Mental synthesis: "Based on all that, I should..."
decision
7
Execute action
Write the email, make the call, update the fields
execution
8
Log the activity
Record what you did so the next person knows
admin

Notice where the time goes. Steps 3, 4, and 5—the navigation steps—consume the majority of effort. The actual decision (step 6) is a sliver. Most of the work is finding information, not using it.

60%
Navigation & Synthesis
Steps 3-5
10%
Actual Decision
Step 6
30%
Execution & Admin
Steps 7-8

This is the 34% selling time from Chapter 2 made concrete. Sixty percent of the workflow is finding. Ten percent is deciding. The proportion is inverted from what it should be.

The New Workflow

Now here's what the same work looks like under decision navigation:

The 4-Step Workflow (Decision Navigation)

1
Open CRM → See prioritized decision queue
AI has ranked your accounts by urgency, opportunity, and strategy alignment. The queue is already built.
2
Select top item → See proposal card
Action + Draft + Why + Evidence, all pre-assembled. No navigation needed.
3
Review and decide: Approve / Modify / Reject
This is where judgment lives. Edit the draft if needed. Check evidence if suspicious. Make the call.
4
Handle exceptions and strategy adjustments
Edge cases, priority overrides, "this account is special." The human-over-the-loop governance layer.

Four steps instead of eight. But the difference isn't just fewer clicks—it's where the cognitive load sits.

10%
Queue Selection
Steps 1-2
70%
Judgment & Refinement
Step 3
20%
Exception Handling
Step 4

The proportion has flipped. Seventy percent of the workflow is now deciding—reviewing, judging, refining. Navigation has collapsed to 10%. The human spends most of their time on what humans are good at: judgment.

The Visual Comparison

Old Workflow (Record Navigation)
Navigate & Synthesize (60%)
Decide
Execute & Admin (30%)
Open → Navigate → Read → Synthesize
Decide
Draft → Execute → Log
New Workflow (Decision Navigation)
Queue
Judge & Refine (70%)
Exceptions (20%)
Open → Queue
Review → Approve/Modify/Reject
Handle Edge Cases
The workflow transformation: from navigation-heavy to judgment-heavy.

The Cognitive Load Shift

This isn't just about time efficiency. It's about what kind of work the human is doing.22

Record Navigation Demands

  • Finding: Where is the information I need?
  • Synthesizing: What does all this data mean?
  • Remembering: What did I see three screens ago?
  • Filtering: Which of these 80 fields matters?

AI does this better than humans.

Decision Navigation Demands

  • Judging: Is this the right action for this situation?
  • Refining: How can I improve this draft?
  • Exception handling: This is a special case because...
  • Strategy: Are we pursuing the right goals?

Humans do this better than AI.

This is playing to strengths, not replacing humans. AI is good at retrieval, synthesis, and pattern-matching. Humans are good at judgment, exceptions, and strategy.14 The new workflow assigns each to their strength.

The Bike Metaphor

"Look mum, no hands!" isn't a phrase about recklessness. It's about trust in the system.

"Riding a bike without hands isn't reckless—it's trusting the system enough to operate at a higher level. Your hands aren't on the handlebars because you don't need them there. The bike is stable. You're supervising, not gripping."

When a child first learns to ride, every input requires conscious attention. Steer left. Steer right. Don't wobble. Watch the road. It's exhausting.

An experienced cyclist doesn't think about these things. The lower-level operations are delegated to muscle memory and mechanical stability. The rider's attention is free for higher-level concerns: route planning, traffic awareness, enjoying the scenery.

Riding with no hands is the visible proof that this delegation has succeeded. You're not gripping the handlebars because the system handles steering autonomously. Your attention has moved up a level.

That's what's happening with AI pair interfaces. You're not navigating through 80 fields because the system handles context synthesis. Your attention has moved up a level—to judgment, to exceptions, to strategy. Same human, different level of operation.

Human Over the Loop vs Human In the Loop

There's a spectrum of human involvement in AI-assisted workflows. Understanding where you sit on this spectrum—and where you should sit—is crucial.

The Autonomy Spectrum

Level 1
Full Review

Human reviews every proposal before execution

Level 2
Selective Review

Human reviews high-stakes / low-confidence proposals

Level 3
Batch Audit

Agent executes routine; human audits batches

Level 4
Strategy Only

Human sets strategy; agent executes within guardrails

The terminology matters here:

Human in the Loop (HITL)

The human is embedded in the execution flow. Every action pauses for human approval. The human is assessing each step.

Human Over the Loop (HOTL)

The human is above the execution flow. The agent operates autonomously within boundaries. The human audits and governs.17

"The human role is fundamentally shifting from assessment to oversight... At the highest level of autonomy, analysts transition from reviewing individual alerts to managing a team of agents, auditing their output weekly or monthly."24 — Detection at Scale

Most CRM workflows should live at Level 2-3. Not life-or-death, but reputation matters. High-stakes deals get full review; routine follow-ups get batch approval or automated execution with audit.

Where CRM Fits on the Spectrum

Not all CRM actions are equal. A well-designed system knows the difference and routes appropriately:

High-Stakes Actions
Level 1-2

Enterprise deals, sensitive accounts, contract renewals, executive outreach

→ Human reviews every proposal. No autonomous execution.

Medium-Stakes Actions
Level 2-3

Standard follow-ups, qualification calls, stage updates

→ Human reviews low-confidence proposals; high-confidence executed or batch-approved.

Routine Actions
Level 3-4

Data hygiene, activity logging, meeting summaries, internal notes

→ Agent executes autonomously. Human audits samples periodically.

The system should route proposals to the right level automatically, based on deal size, account sensitivity, confidence score, and organizational policy. A rep shouldn't have to manually decide what needs review—the system knows.

The Exception Flow

Not every proposal gets approved. Rejections and modifications aren't failures—they're signals. And how you handle them determines whether the system learns.

The Exception Handling Loop

1

User rejects or modifies a proposal

2

System captures: What was wrong? (data / logic / context / goal)

3

Feedback routed to the right layer (update playbook, fix data, adjust model)

4

Future proposals improve based on accumulated feedback

Strategy Adjustments

The human doesn't just approve individual cards. The human also steers the overall approach. This is the "human pane" from Chapter 4 in action—setting priorities and constraints that apply across all proposals.

"Focus more on enterprise accounts this week"

→ Adjusts queue prioritization

"De-prioritize cold outreach, focus on warm leads"

→ Changes proposal generation rules

"This account is sensitive—I want to review everything"

→ Overrides autonomy level for specific account

"Pause all outreach to Acme Corp until I say otherwise"

→ Constraint that blocks proposals

This is supervision at the strategy level, not the action level. The human isn't approving every email; the human is setting the rules for which emails get proposed.

What "Supervise" Actually Means

If you're not navigating, what are you doing? Here's what supervisory work looks like:

📋
Review the Queue

Is prioritization making sense? Are the right things at the top?

🔍
Spot-Check Evidence

Are the reasons good? Does the logic hold up?

🎯
Handle Edge Cases

Where AI judgment isn't enough. Exceptions that need human nuance.

🧭
Adjust Strategy

Are we pursuing the right goals? Does the playbook need updating?

💬
Provide Feedback

When something is wrong, capture why. Help the system learn.

Add Human Touch

Personalization, relationship nuance, judgment that AI can't replicate.

The Best People Become Better Editors

Under record navigation, the best reps were fast clickers with good memory. They knew where everything was. They could navigate five systems in parallel. They remembered what was said three calls ago.

Under decision navigation, the best reps have good judgment. They spot when a proposal is wrong. They know when to override. They can edit a draft to make it perfect. They see patterns in exceptions.

This is exactly what happened with senior engineers (Chapter 1). They stopped being fast typists. They became creative directors. Strategy → Delegation → Verification.

The skill shifts from "fast execution" to "good decisions." Less typing, more taste. Less navigation, more judgment.

"The human remains in control. But the loop itself has changed—less time triaging, more time governing. That's what supervision looks like."

Key Takeaways

  • 1 Old workflow: 8 steps, 60% spent navigating. New workflow: 4 steps, 70% spent deciding.
  • 2 Cognitive load shifts from finding/synthesizing to judging/refining. AI does what AI is good at; humans do what humans are good at.
  • 3 Human over the loop: audit and govern, not review every action. Different stakes warrant different autonomy levels.
  • 4 Exceptions are feedback; the system learns from rejections. Never to zero, but improving over time.
  • 5 Best performers become good editors with good judgment. Less typing, more taste.
08
Part II: The Architecture

The Overnight Advantage

Real-time copilots are cute. Overnight agentic precompute is where you start printing leverage.

Everything we've discussed so far has an implicit assumption: the user clicks, and the system responds. Real-time. Interactive. The proposal card appears because the user asked for it.

But real-time has a constraint. A brutal one.

When a user clicks and expects a response, you have maybe 300 milliseconds before they feel friction.25 That's not much time. Certainly not enough for deep context retrieval, for multi-step reasoning, for generating and evaluating multiple candidates, for adversarial review.

What if we removed that constraint?

The Batch Opportunity

Here's the insight that changes everything: most CRM work isn't urgent.

The sales rep doesn't need the answer in 300 milliseconds. They need it in the morning. They need a queue of proposals ready when they sit down at their desk. The work can happen overnight.

The Constraint Flip

Real-Time Constraint
  • 300ms response time
  • Limited retrieval depth
  • Single-pass generation
  • No adversarial review
  • One answer, take it or leave it
Overnight Opportunity
  • 8 hours of compute time
  • Deep context from everywhere
  • Generate → Critique → Refine
  • Full adversarial gauntlet
  • Multiple options with trade-offs

Overnight batch processing isn't about making the experience slower. It's about making the preparation deeper. Batch processing delivers 40-60% cost savings versus real-time processing for AI workloads with acceptable latency tolerance26—but the real advantage isn't cost. When you move from fast to slow, you don't get less AI—you get AI that can finally think.

The Ralph Wiggum Protocol

This pattern emerged from observing how senior engineers get the best results from AI coding tools. Elite developers now run 8-10 hour overnight AI sessions that produce working code, passing tests, and PR-ready results by morning.30 They don't have one long conversation that accumulates context. They run loops. Fresh start, focused task, clean exit, repeat.

We call it the Ralph Wiggum Protocol—after The Simpsons character who sits on the bus and cheerfully says "I'm in danger!" The AI starts fresh each time, blissfully unaware of previous confusion, and that's exactly what makes it work.

The Ralph Wiggum Loop

1
Clean Start
Fresh context, no cruft
2
Read State
Specs, last actions, where you're at
3
Pick Best Task
"What's most important?"
4
Work on It
Generate, test, validate
5
Exit
End session cleanly
6
Loop Forever
Repeat until done

Why Fresh Beats Accumulated

There's a counterintuitive insight hidden in this pattern: AI's first pass is genius-level. Editing existing work is where it falls apart. Research shows that performance improves dramatically when models are given time to think—a small model with thinking time can outperform a model 14× larger with instant response.28

"AI's first pass at code is genius-level. But when it's editing existing code, it misses bits, tacks on patches, loses the holistic view. So the workflow becomes: generate fresh → validate → regenerate if wrong."

When you give AI a clear task and a clean context, it produces surprisingly coherent, well-structured output. But ask it to "just add one thing" to existing work, and it gets lost. It patches rather than redesigns. It tacks on rather than integrates.

This is why the Ralph Wiggum Protocol works:

Fresh context avoids accumulated confusion

Each loop starts clean. No leftover state from three tasks ago. No confusion about what's current.

The startup protocol rehydrates precisely

Instead of remembering everything, read exactly what's needed: current spec, current state, recent actions.

Delete and regenerate beats patch and accumulate

When something's wrong, don't fix it—regenerate from scratch with better constraints.

This inverts the traditional relationship between output and specification. Code (or proposals) are ephemeral—they can be regenerated anytime. Specifications and playbooks are durable. Design is the asset; the artifact is disposable.

Applying Ralph Wiggum to CRM

The same loop structure works for CRM, just with different domain specifics:

The CRM Overnight Loop
1
Fresh Context

Clean slate for each account. No carryover from the previous one.

2
Read Current State

Account record, recent activity, contact changes, external signals (LinkedIn, news, funding).

3
Pick Best Action

"What should we do with this account right now?" Based on playbook + signals.

4
Generate Proposal

Action + Draft + Why + Evidence. The full proposal card, pre-assembled.

5
Exit and Move On

Store the proposal. Clear context. Pick up the next account.

6
Loop Through Portfolio

Repeat for every account. Entire portfolio processed overnight.

By morning, the rep has a stack of pre-thought proposals waiting. Not "what should I do today?" but "here's what's ready—approve or adjust."

What Overnight Enables

When you have 8 hours instead of 300 milliseconds, entirely different capabilities become possible:

1. Deep Context Retrieval

Real-time can do:

  • • Last email
  • • CRM fields
  • • Basic contact lookup

Overnight can do:

  • • Full email thread history
  • • Web signals: funding, hiring, press, leadership changes
  • • Pattern matching across similar accounts
  • • Relationship mapping across contacts

2. Multiple Candidate Actions

Real-time gives you:

  • • One suggestion
  • • Take it or leave it

Overnight gives you:

  • • Three options with trade-offs
  • • Ranked by confidence, risk, expected outcome
  • • Let the human choose, not just approve/reject

3. Adversarial Review

Multi-model approaches—using several AI models to check each other's work—have been shown to reduce hallucinations and improve reliability.27

Real-time:

  • • Single-pass generation
  • • No second opinion

Overnight:

  • • Generator produces candidates
  • • Critic finds holes
  • • Bias auditor checks patterns
  • • Policy gate verifies compliance

4. Deeper Reasoning

Real-time:

  • • Surface-level pattern matching
  • • Simple rules-based suggestions

Overnight:

  • • Journey inference: "Where is this account really?"
  • • Pattern matching: "Accounts like this usually need X"
  • • Scenario planning: "If we do X, likely response is Y"

The Fast/Slow Split

This creates a natural division of labor between daytime and overnight:

The Fast Lane and Slow Lane

☀️
Fast Lane (Daytime)
  • Reviewing proposals
  • Tweaking drafts
  • Approving actions
  • Executing communications
  • Handling exceptions

Human time is expensive; use it for judgment.

🌙
Slow Lane (Overnight)
  • Heavy thinking
  • Deep retrieval
  • Web enrichment
  • Adversarial review
  • Multi-candidate generation

Compute time is cheap; use it for preparation.

Moving from fast to slow enables more AI, not less. You're not making the experience slower—you're making the preparation deeper. Daytime becomes the decision surface; overnight becomes the planning and verification pipeline.

The John West Strategy

Here's where overnight processing unlocks something that fundamentally changes trust and governance: keeping track of rejected alternatives.

Remember the John West principle from Chapter 6: "It's the fish John West rejects that makes John West the best." When you have time to generate multiple candidates, you can show not just what you recommended, but what you considered and rejected.

The Audit Trail of Thinking
#1 Recommended: Call John tomorrow
Confidence: 85% • Expected outcome: Re-establish relationship, get budget confirmation
#2 Considered: Send email first
Rejected because: Email response rates 40% lower for this contact type. Call is higher-value.
#3 Considered: Wait another week
Rejected because: Budget timeline creates urgency. Waiting increases risk of losing position.
Risks Detected
  • • Champion just changed roles — may have different priorities
  • • Competitor mentioned in last call — risk of losing deal
What Would Change This Decision
  • • If John is traveling → switch to email
  • • If budget delayed → wait and monitor

This transforms AI from "magic eight-ball" into traceable decision process:

Without John West

"The model suggested X"

  • • Single thread
  • • No alternatives visible
  • • No accountability
  • • Black box
With John West

"Here's what was considered"

  • • Top 3 candidates visible
  • • Why #1 won the ranking
  • • Why #2 and #3 rejected
  • • Governance built-in
"Instead of 'the model suggested X' (single-thread, no accountability), you have: top 3 candidates, why #1 won, why #2 and #3 were rejected, what risks were detected, what evidence was used. That's governance built into the interaction."

The Adversarial Gauntlet

With overnight processing, you can run proposals through multiple stages before they ever reach the human queue:

The Five-Stage Gauntlet

1
Generator
Produce candidate actions + drafts + evidence for each account
2
Critic
Find holes — missing context, weak reasoning, overconfidence, factual errors
3
Bias/Fairness Auditor
Check prioritization patterns, language, recommended tactics for systematic bias
4
Policy/Risk Gate
Verify compliance — privacy, competition law, contractual constraints, brand guidelines
5
Ranker
Final prioritization based on strategy, urgency, confidence for morning queue

Each stage improves quality. The proposal that reaches the human has already survived critique, passed compliance checks, and been ranked against alternatives. Real-time can't afford this; overnight can.

The Morning Experience

Here's what the sales rep sees when they sit down at 8am:

Good morning, Sarah
Your Decision Queue
12
proposals ready
Acme Corp — Call John
High confidence • Budget timing
Pre-researched
TechStart Inc — Send proposal
High confidence • Ready to close
Draft ready
GlobalTech — Re-engage champion
Medium confidence • Role change detected
Review recommended
+ 9 more proposals...
Processed overnight: 47 accounts • 12 proposals generated • 35 no-action-needed View all →

Each proposal card is:

The thinking already happened. The deciding is all that's left.

Batch Enables Personalization at Scale

There's a business model insight hidden in overnight processing: mass customization becomes economically feasible. Mass customization that feels personal, meaningful, and seamless is already happening across industries—e-commerce platforms tailoring in real-time, manufacturers allowing custom configurations, healthcare delivering individualized recommendations.29

With real-time generation, personalization is expensive. Every customized proposal costs human attention or immediate AI compute. So companies default to generic templates with mail-merge fields.

With overnight batch processing, the equation flips. Compute time is cheap. You can afford to spend significant AI time on every account, every night. The result is genuinely bespoke reasoning—not "Dear {FirstName}" but "Based on your transition from VP Sales to CRO, and your mention of SSO integration concerns in the November call..."

When to Use Batch vs Real-Time

Not everything should be batch. Here's the decision matrix:

Use Case Real-Time Overnight
User clicks account, needs immediate response
Daily action queue for all accounts
Simple field lookup
Deep journey analysis
Single quick proposal
Multiple options with trade-offs
Quick draft for immediate need
Adversarially reviewed draft

The default pattern: batch prepares, real-time supplements. Overnight does the heavy thinking. Daytime handles the exceptions and immediate needs.

"Real-time AI is constrained by what you can do in 300ms. Overnight AI is constrained only by what you want to invest. When you move from fast to slow, you don't get less AI—you get AI that can finally think."

Key Takeaways

  • 1 Ralph Wiggum Protocol: clean start → read state → pick best task → work → exit → loop. Fresh context beats accumulated confusion.
  • 2 Overnight enables: deep retrieval, multiple candidates, adversarial review, deeper reasoning. Things you can't do in 300ms.
  • 3 John West strategy: keep top 3 candidates + rejection reasons = governance built into the interaction.
  • 4 Fast/slow split: overnight prepares (planning pipeline), daytime decides (decision surface).
  • 5 Morning experience: pre-thought proposals waiting, decisions ready to make. The thinking already happened.
09
Part III: The Variants

Variant: Sales Development

SDRs spend 80% of their time researching. What if research was already done when they arrived?

We've established the core architecture: Two-Pane Cognition, Proposal Cards, the Evidence Layer, supervisory workflow, and overnight batch processing. Now let's see how these patterns apply to a specific domain.

Sales Development Representatives (SDRs) do the front-line work of prospecting and lead qualification. They find potential customers, research them, reach out, qualify interest, and hand off to account executives. It's the top of the sales funnel.

And it's drowning in research work.31

The SDR Time Problem

Here's what an SDR's day actually looks like:

The Traditional SDR Workflow

1
Get target list
Account list from marketing or territory assignment
2
Research each prospect
LinkedIn profile, company website, news, trigger events
~15 min each33
3
Find the angle
What's relevant? What's the hook? Why should they care?
~5 min each
4
Craft personalized message
Write the email or LinkedIn message with research-based personalization
~10 min each
5
Send and track
Execute outreach, log activity
6
Follow up, qualify, hand off
Continue conversation, qualify fit, pass to AE

Notice where the time goes. Steps 2, 3, and 4—the research and crafting steps—consume 70-80% of effort.32 The actual outreach and conversation (steps 5-6) are a sliver.

80%
Research & Crafting
Finding, reading, synthesizing, writing
20%
Actual Conversations
Outreach, follow-up, qualifying

This is exactly backwards. The SDR's value is in conversation—qualifying, building rapport, understanding need. But they spend most of their time doing work that AI can do: research and first-draft writing.34

The Framework Applied

Let's apply the architecture from Part II to sales development:

Two-Pane Cognition for SDR

🧠

Human Pane

  • • Prioritized prospect queue
  • • Response handling
  • • Qualification judgment
  • • Relationship signals
  • • Handoff decisions
🤖

AI Pane

  • • Research synthesis
  • • Trigger event detection
  • • Personalized drafts
  • • Evidence links
  • • Confidence scores
🤝

Shared Surface

  • • ICP criteria
  • • Messaging guidelines
  • • Compliance rules
  • • Brand voice
  • • Territory boundaries

The Proposal Card for Prospecting

Here's what a proposal card looks like for an SDR's morning queue:

Acme Corp • Series B • 85 employees
Sarah Chen, VP Operations
Champion match
Action
Connect via LinkedIn — reference scaling challenges post
Why now
Champion match Funding trigger: 2 weeks ago Active on LinkedIn ICP fit: 50-200 employees
Draft message

"Sarah — saw your comment on [Sarah Guo's post] about the challenges of scaling ops without adding headcount.

We've been helping similar-stage companies automate their operational workflows without the typical integration nightmare. Just helped [similar company] cut their ops load by 40%.

Worth a 15-min chat? Timing might be good given the Series B."

Evidence
3 items supporting this recommendation

This proposal card has everything the SDR needs to decide:

What the Evidence Looks Like

When the SDR clicks "View" on the evidence section, they see the specific signals that drove this recommendation:

💬 LinkedIn Activity 3 days ago
Commented on Sarah Guo's post about scaling operations: "This resonates. We're hitting the same wall at 80 people—more process, same headcount isn't sustainable."
→ Signal: Active engagement, expressed specific pain point matching our solution
💰 Funding Event 2 weeks ago
Acme Corp raised $25M Series B led by Benchmark
→ Signal: Budget available, growth phase, likely investing in operations infrastructure
👤 Role Match ICP criteria
VP Operations at 85-person company. Decision-maker for ops tooling.
→ Signal: Champion profile match. Right seniority, right function, right company size.

Overnight Batch for SDR

Applying the Ralph Wiggum loop (Chapter 8) to sales development:

The SDR Overnight Loop
1
Load prospect list

Territory assignment, marketing handoffs, enrichment queue

2
For each prospect: Fresh context

Clean slate, load ICP criteria and messaging playbook

3
Research and enrich

LinkedIn activity, company news, funding, hiring signals, trigger events

4
Generate proposal

Personalized angle + draft message + evidence links + confidence score

5
Quality check

Does this pass threshold? Strong signal, valid angle, fresh trigger?

6
Add to queue or skip

High-quality → morning queue. Low-quality → hold for later refresh.

Run this on 200 prospects overnight. The SDR reviews the best 30-50 in the morning—the ones with strong signals and ready drafts.

The Morning Experience

Compare the old SDR morning to the new one:

Old SDR Morning

  • 1. Open CRM, look at list
  • 2. Pick someone, open LinkedIn
  • 3. Scroll their profile, recent posts
  • 4. Check company news, funding
  • 5. Find an angle, write message
  • 6. Send, log, repeat
~30 min/prospect
15-20 prospects/day36

New SDR Morning

  • 1. Open queue → 35 pre-researched prospects
  • 2. Review top prospect → see proposal card
  • 3. Check why + evidence (if needed)
  • 4. Edit draft if needed
  • 5. Approve → send
  • 6. Next prospect
~5 min/prospect
50-80 prospects/day37

What the SDR Actually Does

The role shifts from researcher to editor/approver. Here's what that looks like:

What the SDR Does
  • Reviews draft: "Is this angle good? Does it land?"
  • Checks evidence: "Is this trigger real? Timing right?"
  • Edits: Adjusts tone, adds personal touch, fixes errors
  • Approves: Sends or queues for sending
  • Rejects: Skips bad fits, flags data issues
  • Handles responses: The actual conversations
What AI Handles
  • Researching each prospect's background
  • Finding trigger events and signals
  • Identifying the best angle
  • Crafting first-draft messages
  • Tracking who's been contacted
  • Ranking by likelihood to respond

The Feedback Loop

Like all good systems, this one learns from use:

When the SDR Rejects a Proposal

Capture why:
  • Bad fit: Not actually ICP
  • Wrong timing: Trigger is stale
  • Weak angle: Personalization doesn't land
  • Bad data: Info is wrong or outdated
Feed back to improve:
  • • Refine ICP matching criteria
  • • Adjust trigger freshness thresholds
  • • Update angle generation prompts
  • • Flag data source issues

When Prospects Respond

Track what worked:
  • • Which angles got replies
  • • Which triggers were most predictive
  • • Which message styles resonated
  • • Which channels performed best
Feed success patterns back:
  • • Weight trigger types by response rate
  • • Favor successful messaging patterns
  • • Prioritize high-response segments
  • • Refine confidence scoring

Exception rate should drop over time as the system learns what works for this SDR, this territory, this product.

The Metrics That Change

Here's what moves when you apply decision navigation to sales development:

Metric Before After
Prospects researched/day 15-20 50-100 (AI pre-research)
Personalized messages/day 15-20 50-80
Time per prospect 20-30 min 5-10 min
Response rate Baseline Higher (better personalization)38
SDR territory coverage 1 territory 2-3x territory

"The SDR's job doesn't change—qualify and hand off. What changes is how they spend their time. Less research, more judgment. Less crafting, more deciding. That's decision navigation for sales development."

Key Takeaways

  • 1 SDRs spend 80% on research; overnight batch flips this ratio. Research happens while they sleep.
  • 2 Proposal cards for prospecting: Action + Draft + Why (trigger events) + Evidence (links to signals).
  • 3 Morning queue: 30-50 pre-researched, pre-drafted prospects ranked by confidence and trigger strength.
  • 4 SDR role shifts from researcher to editor/approver. Same conversations, less prep.
  • 5 Feedback loop: Rejections and responses improve the system over time.
10
Part III: The Variants

Variant: Customer Support

Support agents spend 60% of their time gathering context. What if context was already gathered when the ticket arrived?

Customer support agents handle tickets, resolve issues, and maintain relationships. They're the front line of customer experience—the human voice when something goes wrong.

Their job requires empathy, judgment, and problem-solving. But most of their time isn't spent on any of those things. It's spent gathering context.39

Sound familiar?

The Context Gathering Problem

Here's the traditional support workflow:

Traditional Ticket Handling

1
Ticket arrives in queue
Customer submitted issue via email, chat, or form
2
Open ticket, read message
"What are they asking? What's the problem?"
3
Look up customer
CRM → account status, history, plan tier, previous tickets
~2 min
4
Check product context
Known issues? Recent changes? Documentation?
~2 min
5
Diagnose
What's actually wrong? Have we seen this before?
~2 min
6
Draft response
Explain, solve, or escalate
7
Send and close (or escalate)
Resolve or hand off

Steps 3, 4, and 5—the context gathering steps—consume 50-60% of handle time.40 The agent is playing detective before they can play helper.

60%
Context Gathering
Looking up, cross-referencing, synthesizing
40%
Actually Solving
Responding, resolving, escalating

The ratio is backwards. Agents are hired for their judgment and empathy—not their ability to navigate five different systems.41

The Framework Applied

Applying Two-Pane Cognition to customer support:

Two-Pane Cognition for Support

🧠

Human Pane

  • • Ticket queue prioritization
  • • Resolution decisions
  • • Tone and empathy
  • • Edge case judgment
  • • Escalation handling
🤖

AI Pane

  • • Context synthesis
  • • Similar ticket matching
  • • Draft responses
  • • Policy lookups
  • • Sentiment detection
🤝

Shared Surface

  • • Response policies
  • • Escalation rules
  • • Tone guidelines
  • • Approval thresholds
  • • SLA requirements

The Proposal Card for Support

Here's what a proposal card looks like for a support ticket:

Ticket #8847 • Billing • 23 min old
Sarah Miller — Billing Dispute
High-value customer
Recommended Action
Issue $50 credit + apologize for billing error
Why this resolution
3-year customer $5k LTV Similar issue: resolved with credit Negative sentiment
Draft response

Hi Sarah,

I'm so sorry about the billing confusion—I completely understand the frustration. I've looked into this, and you're right: there was an error in how our system applied the promotion code.

I've issued a $50 credit to your account. You'll see it reflected within 24 hours. I've also flagged your account to make sure this doesn't happen again.

Is there anything else I can help with?

Evidence
4 items: account history, previous ticket, billing log, policy

Context Synthesis vs Context Gathering

The fundamental shift is from gathering to reviewing:

Old: Context Gathering

  • 1. Agent opens CRM record
  • 2. Agent opens billing system
  • 3. Agent searches previous tickets
  • 4. Agent checks product docs
  • 5. Agent mentally synthesizes
  • 6. Agent decides what to do
5-10 min just to understand

New: Context Synthesis

  • 1. AI pre-gathers from all systems
  • 2. AI synthesizes full picture
  • 3. AI recommends resolution
  • 4. Agent reviews proposal
  • 5. Agent approves or adjusts
1-2 min to review and decide

The Evidence Layer for Support

When the agent clicks to view evidence, they see everything the AI considered:

👤 Customer Profile
Tenure
3 years, 2 months
Plan
Business Pro
LTV
$5,240
→ Implication: High-value, long-term customer. Worth retaining. Precedent: resolve billing issues generously.
📋 Previous Ticket #4523 8 months ago
Similar issue: Promotion code not applied correctly
Resolution: Issued $35 credit, customer satisfied
→ Implication: Same issue type. Previous resolution worked. Match or exceed.
💳 Billing Log
Invoice #9847: Promotion code SAVE20 was in cart but not applied to final charge
→ Implication: Clear billing error. Customer is correct. Straightforward resolution.
😤 Sentiment Analysis
Negative (75%)
"This is the second time this has happened. I'm really frustrated—I specifically checked that the code was applied..."
→ Implication: Customer is upset. Lead with empathy. Resolve quickly.

Handling Escalations

Not every ticket gets a "resolve now" proposal. Some need escalation—and escalation is a proposal type, not a failure mode.

Ticket #8901 • Billing Dispute • $500+
Marcus Chen — Contract Dispute
Escalation recommended
Recommended Action
Escalate to Tier 2 — requires manager approval
Why escalation
Dispute amount ($547) exceeds auto-resolve threshold ($200)
Customer mentioning cancellation: "considering switching providers"
Contract interpretation question requires legal/policy review
Handoff Summary for Tier 2

Issue: Customer disputes early termination fee of $547. Claims he was told the contract was month-to-month.

Customer value: 2-year customer, $8,200 LTV, Enterprise plan

Churn risk: High — explicit mention of switching to competitor

The escalation proposal includes a handoff summary so Tier 2 doesn't start from scratch. Context travels with the ticket.

The Queue Transformation

Decision navigation changes what the queue looks like:43

Old Support Queue
  • Oldest tickets first (or random priority)
  • Agent opens ticket, sees only the message
  • Each ticket is a mini research project
  • No sense of complexity until you're in it
New Support Queue
  • Pre-triaged by urgency, value, complexity
  • Each ticket shows proposal + synthesis + draft
  • Confidence score visible before opening
  • Escalations flagged, handoffs ready

Governance in Support

Different tickets warrant different levels of human involvement:44

Low-Stakes + High-Confidence
Auto-send with audit

Password resets, basic how-to questions, acknowledgment responses

→ Send automatically. Human audits samples weekly.

Medium-Stakes or Medium-Confidence
Human reviews before send

Refund requests, technical troubleshooting, feature questions

→ AI drafts. Human reviews and approves. Standard flow.

High-Stakes or Low-Confidence
Human must approve

Large disputes, churn threats, legal mentions, VIP accounts

→ AI synthesizes context. Human makes all decisions.

The audit trail is automatic:

The Metrics That Change

These are representative improvements based on industry implementations:45

Metric Before After
Average handle time 8-12 min 3-5 min
Context gathering time 4-6 min 30 sec (review synthesis)
Tickets per agent per day 30-50 60-100
First-contact resolution Baseline Higher (better context)
Customer satisfaction Baseline Higher (faster, personalized)

"Support agents are hired for their judgment and empathy—not their ability to navigate five systems. Decision navigation lets them focus on what they're actually good at."

Key Takeaways

  • 1 Support agents spend 60% gathering context; AI synthesis flips this ratio.
  • 2 Proposal cards for support: Action + Draft + Why + Evidence (including customer context and similar tickets).
  • 3 Pre-triage and pre-synthesis enable faster, more consistent responses with better personalization.
  • 4 Escalation is a proposal type, not a failure mode. Handoff summaries travel with the ticket.
  • 5 Agent role shifts from researcher to reviewer/personalizer. Judgment and empathy, not system navigation.
11
Part III: The Variants

Designing Your Own Proposal Cards

Pick one workflow. Design the card. Hide the fields.

Chapters 4-8 gave you the architecture. Chapters 9-10 showed domain-specific examples. Now: how do you design proposal cards for your workflows?

This is the practical chapter. A step-by-step guide for turning any navigation-heavy workflow into a decision navigation interface.

The Three Questions

Every proposal card design starts with three questions. Answer them clearly, and the card almost designs itself.

The Three Questions

1

What action does the user typically take?

This becomes your ACTION component.

  • • Be specific: "Call John about Q2 budget" not "follow up"
  • • The action should be approvable/rejectable
  • • One action per card (not a list of options)
2

What do they need to see to feel confident?

This becomes your WHY and EVIDENCE components.

  • • Not everything—just what moves them from uncertain to confident
  • • What would they check manually if AI wasn't there?
  • • Usually 3-5 key signals, with source links
3

What artifact do they produce?

This becomes your DRAFT component.

  • • Email, report, agenda, proposal, update—something tangible
  • • What would the human otherwise create from scratch?
  • • Some actions have no artifact (data updates); that's fine

Worksheet: The Three Questions

WORKFLOW:
_______________________
1. What action?
_______________________
2. What confidence needs?
_______________________
3. What artifact?
_______________________

Step-by-Step Design Process

Here's the full process for designing a proposal card from scratch:

Step 1: Identify the Workflow

Pick ONE workflow to start. Not the whole system—one specific, repeatable process.

Good first candidates:
  • • High-volume (done many times per day)
  • • Routine (similar pattern each time)
  • • Currently time-consuming (lots of navigation)

Example: "Daily outreach to warm leads"

Step 2: Map the Current State

Document exactly what the user does today, step by step.

Ask:
  • • What systems do they open?
  • • What screens do they navigate?
  • • What data do they look at?
  • • What do they produce?
  • • How long does it take?
Watch for:
  • • Tab-switching patterns
  • • Copy-paste workflows
  • • "Wait, where was that?" moments
  • • Manual synthesis steps
  • • Repeated research

Step 3: Identify the Decision Point

Find the actual decision being made. Not "fill out this form" but the judgment call.

The decision is what we're supporting:
  • • Not "research this account" → "decide whether to contact this account"
  • • Not "look at the ticket" → "decide how to resolve this ticket"
  • • Not "check the pipeline" → "decide which deal needs attention"

Step 4: Design the Card Components

Using your answers to the three questions, specify each component:

ACTION:

What's the specific, approvable action? One verb + one target + optional context.

DRAFT:

What artifact would the user create? Ready to send/use with minor edits.

WHY:

What 3-5 reasons would convince them? Short phrases, not paragraphs.

EVIDENCE:

What receipts would they check? Links to source data.

Step 5: Define the Drill-Down

Specify what appears at each layer of progressive disclosure:

L1 Summary: Action + Why (3-5 reasons). Scannable in 5-10 seconds.
L2 Expanded: Full draft + Evidence cards. 30-60 seconds to review.
L3 Full record: All fields, complete history. As long as needed.

Step 6: Test with Users

Show them the card design and ask the key question:

"Would you feel confident approving this?"

If no: "What's missing?"

Their answer tells you what to add to Why/Evidence. Iterate until they say yes.

Common Card Patterns

Most proposal cards fall into a few patterns. Here are templates you can adapt:

Pattern 1: Outreach Card

ACTION: Contact [person] via [channel]

DRAFT: Message/email content

WHY: Trigger events, timing signals, relationship status

EVIDENCE: Source of signals, previous interactions

Examples: Sales outreach, re-engagement, meeting follow-up

Pattern 2: Update Card

ACTION: Change [field] to [value]

DRAFT: N/A (or explanation note)

WHY: Data discrepancy detected, event occurred

EVIDENCE: Source of new information

Examples: Stage changes, contact updates, data hygiene

Pattern 3: Escalation Card

ACTION: Escalate [item] to [person/team]

DRAFT: Summary for receiving party

WHY: Threshold exceeded, policy triggered, complexity

EVIDENCE: What triggered the escalation

Examples: Support escalation, deal approval, exception handling

Pattern 4: Review Card

ACTION: Approve/reject [item]

DRAFT: N/A (item is the artifact)

WHY: Quality signals, risk assessment, completeness

EVIDENCE: What was evaluated

Examples: Content review, proposal approval, quality check

Progressive Disclosure Design

Getting the layers right is crucial. Here's a detailed breakdown:

What Each Layer Contains

Layer 1 Default View 5-10 seconds to scan
Contains:
  • • Action headline
  • • 2-3 line "Why" summary
  • • Approve / Modify / Reject buttons
Goal:

User can decide on 80% of cards without leaving this view.

Layer 2 One Click 30-60 seconds to review
Contains:
  • • Full draft artifact
  • • Evidence cards (3-5 items)
  • • Confidence indicator (optional)
Goal:

User can verify AI reasoning, edit draft, check receipts.

Layer 3 Second Click As long as needed
Contains:
  • • Full record/history
  • • All fields
  • • Complete audit trail
Goal:

User can access anything, for edge cases and deep investigation.

The Delete-and-Regenerate Test

Here's a quality check for your proposal card design:

Can the AI regenerate this exact card from:

📊
Current State
Data in the system now
📋
Playbook
Rules for what to recommend
🔍
Evidence
External signals and triggers
If YES:

Card is well-designed. It's derivable, consistent, auditable.

If NO:

Card depends on something fragile. Find what's missing and add it to state or playbook.

Common Mistakes to Avoid

Here are the pitfalls that catch most first-time card designers:

Mistake 1: Showing too much by default
Symptom: Layer 1 has 10+ data points
Problem: Cognitive overload, users stop reading46

Fix: Ruthlessly cut to 3-5 most relevant points. Everything else is Layer 2.

Mistake 2: Hiding the evidence
Symptom: Users can't verify AI reasoning
Problem: Trust fails, users override everything47

Fix: Make evidence one click away, always. Never hidden or inaccessible.

Mistake 3: No drill-down path
Symptom: Summary only, no way to see more
Problem: Users feel trapped, can't investigate48

Fix: Every summary should link to fuller view. Layer 3 must exist.

Mistake 4: Generic actions
Symptom: "Follow up" instead of "Call John about Q2 budget"
Problem: Action isn't specific enough to be useful

Fix: Be as specific as the context allows. Verb + target + reason.

Mistake 5: No feedback capture
Symptom: Rejections disappear into void
Problem: System can't learn from mistakes

Fix: Every rejection should capture why. Feed back to improve the system.

Governance Checklist

Before deploying a proposal card design, verify these governance requirements:

Pre-Deployment Checklist

Your First Card: An Exercise

Let's design a proposal card together. Here's the workflow:

Workflow: Daily review of accounts that haven't been touched in 14+ days

ACTION:
[What specific action?]
DRAFT:
[What artifact, if any?]
WHY:
[3 reasons this account needs attention]
EVIDENCE:
[What data points support this?]

Example Answer

ACTION:
Send re-engagement email to Acme Corp
DRAFT:
Personalized email checking in, referencing the proposal discussion from your last call and the Q1 decision timeline mentioned
WHY:
  • • 17 days since last touch (relationship cooling)
  • • Open opportunity at proposal stage ($45k)
  • • Decision timeline was "end of Q1" — we're approaching that
EVIDENCE:
  • • Last email: Jan 7 (link)
  • • Opportunity record (link)
  • • Notes from Dec 15 call mentioning Q1 timeline (link)

"The framework is simple: Action, Draft, Why, Evidence. Progressive disclosure: summary first, evidence on click, full record available. Test with users, iterate, deploy. That's how you turn any workflow into decision navigation."

Key Takeaways

  • 1 Three questions: What action? What confidence needs? What artifact? These give you the four components.
  • 2 Design each component deliberately; test progressive disclosure. 80% should decide at Layer 1.
  • 3 Delete-and-regenerate test: Can AI recreate the card from state + playbook + evidence?
  • 4 Common mistakes: Too much default, hidden evidence, no drill-down, generic actions, no feedback capture.
  • 5 Start with one workflow, get it working end-to-end, then expand.
12
Part III: The Variants

What Changes If You Get This Right

For individuals. For teams. For the industry.

We've covered the shift: from record navigation to decision navigation. The architecture: Two-Pane Cognition and proposal cards. The amplifier: overnight batch processing. The applications: sales development, customer support, and a framework for designing your own.

Now: what actually changes if you implement this?

The impact cascades from individuals to teams to the entire industry. This isn't incremental improvement—it's a different operating model.

For Individuals: 10x Faster Time-to-Decision

The shift changes what work feels like, moment to moment.

The Cognitive Shift

Old Mental Model
1 Find the information
2 Synthesize it mentally
3 Decide what to do
4 Create the artifact
5 Act

Cognitive load on: finding, synthesizing, creating

New Mental Model
1 Review the proposal
2 Decide
3 Act

Cognitive load on: evaluating, judging

Here's what this feels like in practice:

Morning starts with pre-thought proposals, not a blank slate. The system has already done overnight preparation.

Each decision is 2 minutes instead of 20. You review, judge, act—not research, synthesize, create.

Mental energy goes to hard problems, not routine research. You save your judgment for where it matters.

End of day: more decisions made, less exhaustion. The cognitive lightness compounds.

"The shift isn't just faster—it's cognitively lighter. When the system does the finding, you can focus on the judging."

One decision faster is nice. Every decision faster is transformative. You process more, with higher quality, with less fatigue. The compound effect builds—research shows AI adoption leaders see performance improvements 3.8x higher than the bottom half of adopters.49

For Teams: Standardized Format, Faster Handoffs

Decision navigation creates consistency across the team—not by forcing compliance, but by making the right pattern the easy pattern.

📋

Consistency

Everyone uses the same proposal card format: Action + Draft + Why + Evidence. Every time.

No more "how does Sarah organize her notes?"

🔄

Trivial Handoffs

Account handoff = here are the pending cards. Shift handoff = here's the queue. Escalation = here's the card that triggered it.

No more "let me bring you up to speed" meetings.

📊

Measurable Quality

Approval rates by rep, by card type, by segment. Rejection reasons. Response rates. Time to decision.

What's working? Where are bottlenecks?

For the Industry: System of Record → System of Action

This is the paradigm shift happening across enterprise software:

System of Record

What CRMs were built as

  • • Faithful mirror of what happened
  • • All data visible and navigable
  • • User synthesizes and decides
  • • The record is the interface
System of Action

What the future demands

  • • Engine that recommends what to do
  • • Proposals visible, records on demand
  • • System synthesizes, user decides
  • • The proposal is the interface

The record doesn't disappear—it becomes the foundation. But users interact with proposals, not fields.

"Enterprise software is shifting from 'system of record' to 'system of action.' The record becomes infrastructure; the proposal becomes the interface."

This applies to every enterprise tool. The vendors that make this shift will win; the rest will be navigated around or replaced. As Gong documented, systems of action don't just store data—they turn 6,000 words of conversation into actionable recommendations rather than 30-60 words of notes.8

The Competitive Dynamic

Early adopters of decision navigation gain compounding advantage—BCG research shows the gap between leaders and laggards is widening rapidly51

Every day you're spelunking, competitors are supervising

The gap widens with scale: 10% efficiency at 100 reps = 10 rep-equivalents

The Skills That Matter Now

Decision navigation changes which skills are valuable:

Becomes Less Valuable

  • Fast clicking and navigation
  • Memorizing where things are
  • Manually synthesizing from multiple sources
  • Creating artifacts from scratch
  • Speed-of-execution tasks

Becomes More Valuable

  • Judgment: Is this proposal right?
  • Editing: How do I make this better?
  • Exception handling: When is the system wrong?
  • Strategy: What should we optimize for?
  • Feedback: How do I correct the system?

The senior role redefines—just like it did in coding. Less "doer," more "director." Less execution, more judgment. Less navigation, more supervision. As Bain puts it: "You can't automate your way to transformation. You have to rethink the work itself."52

The best people operate at a higher level. They're not freed from work; they're freed for higher-value work.

The Governance Advantage

Compliance teams should love decision navigation:

What Governance Gets

  • Every proposal logged with evidence
  • Every decision traceable
  • Rejection reasons captured
  • Audit trail built into workflow
  • "Why did this happen?" has an answer
  • Policy enforcement in generation

Governance becomes a feature, not a burden. Not forms and approvals slowing everything down. Built into the interaction pattern. Compliance by architecture, not by process. This matters: McKinsey research shows 40% of organizations identify explainability as a key AI risk, but only 17% actively work to mitigate it.21

The Risks If You Don't

What happens if you stay with record navigation while the world moves on?

Competitive erosion

Competitors using AI-native interfaces will outpace you. Same market, different operating speed.

Talent flight

Your best people will leave for environments where they're not spelunking. Why navigate when you could supervise?

Training costs stay high

Onboarding stays slow when there's no standard format. "Shadow Sarah" doesn't scale.

Data remains underutilized

You have the data. It's an asset. But if it's hard to access, it's an asset that doesn't compound.

It won't be one catastrophic failure. It'll be gradual: slightly slower, slightly less efficient, slightly less attractive. Death by a thousand clicks. Research from Gartner predicts 15% of day-to-day work decisions will be made autonomously by 2028, up from near 0% in 202453—those who haven't redesigned their interfaces by then will be working against the current.

The Path Forward

Here's how to start:

Phase 1 Pick One Workflow
  • • Choose: high-volume, routine, currently time-consuming
  • • Design the proposal card (Chapter 11)
  • • Build the overnight batch (Chapter 8)
  • • Deploy, measure, iterate
Phase 2 Expand to Adjacent Workflows
  • • Apply the pattern to related processes
  • • Reuse the architecture you built
  • • Train the team on decision navigation
Phase 3 Make It the Default
  • • New interfaces start with proposal cards
  • • Record views become drill-down, not default
  • • Batch processing becomes standard infrastructure

The Coding World Learned This by Accident

Remember Chapter 1? Senior developers stopped looking at code. They didn't plan it. AI tools made it possible, and they discovered it worked.

The workflow changed without anyone designing it. Engineers stumbled into decision navigation because the tools abstracted the artifact layer. They saw diffs instead of files. They supervised instead of spelunking.

You don't have to stumble into it. You can design for it from the start.

"The coding world stopped looking at the artifact. It started looking at the outcome. They didn't plan it—AI tools made it possible, and engineers discovered it worked. You don't have to stumble into it. You can design for it from the start."

Look Mum, No Hands

Remember the bike metaphor from Chapter 7?

That moment when a child rides without hands—it's not reckless. It's confident. They trust the system enough to let go. Their hands aren't on the handlebars because they don't need them there.

Your users can have that moment. Design interfaces that let them ride without hands. Not passive—supervising. Not reckless—confident. Operating at a higher level because the system handles the lower level.

Design for proposals.

Hide the fields.

Let your users ride without hands.

"The coding world learned this by accident. You can learn it on purpose."

Key Takeaways

  • 1 For individuals: 10x faster decisions. Cognitive load shifts from finding to judging.
  • 2 For teams: Standardized format, trivial handoffs, measurable quality, faster onboarding.
  • 3 For the industry: System of record → system of action. The proposal becomes the interface.
  • 4 Skills shift: Less navigation, more judgment and editing. The senior role redefines.
  • 5 Governance becomes a feature, built into the interaction pattern. Compliance by architecture.
  • 6 Start now: Pick one workflow. Design the card. Build the batch. You can learn this on purpose.
R

References & Sources

The evidence and external sources that informed this guide.

This ebook synthesizes original frameworks with external research and industry data. Below are the primary sources referenced throughout, organized by category. Where inline citations appear (e.g., 1), they link to these numbered references.

Industry Analysis

[1] "Even the Inventor of Vibe Coding Says Vibe Coding Can't Cut It"

Gizmodo. Andrej Karpathy's reflection on vibe coding limitations—his nanochat project was "basically entirely hand-written" because AI agents "just didn't work well enough."

https://gizmodo.com/even-the-inventor-of-vibe-coding-says-vibe-coding-cant-cut-it-2000672821

[2] "Claude Code is Redefining Software Engineering"

Medium. Documents the workflow shift where developers generated 11,900 lines of production code without opening an IDE.

https://medium.com/@ajepym/claude-code-is-redefining-software-engineering-faster-than-were-ready-for-da1a110b5567

[24] "AI Security Operations 2025 Patterns"

Detection at Scale. Documents the shift from human-in-the-loop assessment to human-over-the-loop governance in security operations—analysts transition from reviewing individual alerts to managing teams of agents.

https://www.detectionatscale.com/p/ai-security-operations-2025-patterns

[26] "Real-Time vs Batch Processing Architecture"

Zen van Riel. Analysis showing 40-60% cost reduction for batch AI processing versus real-time, with architectural guidance on when to use each approach.

https://zenvanriel.nl/ai-engineer-blog/should-i-use-real-time-or-batch-processing-for-ai-complete-guide/

[30] "How the Creator of Claude Code Actually Uses It"

Dev.to. Documents Boris Cherny's workflow for running 10+ hour overnight AI coding sessions that produce working code, passing tests, and PR-ready results by morning.

https://dev.to/a_shokn/how-the-creator-of-claude-code-actually-uses-it-32df

LinkedIn Commentary

[3] "Switching to Claude Code Boosts Dev Team Productivity"

LinkedIn. Documents the shift to 90% AI-generated code and the role redefinition from "code producer" to "creative director."

https://www.linkedin.com/posts/md-mukthar-8239b2278_i-quietly-switched-my-remote-dev-team-and-activity-7413165586949525504-A7aE

[4] "Switching to Claude Code Boosts Dev Team Productivity"

LinkedIn. Source for the "creative director" role shift quote: "Strategy → Delegation → Verification."

https://www.linkedin.com/posts/md-mukthar-8239b2278_i-quietly-switched-my-remote-dev-team-and-activity-7413165586949525504-A7aE

[7] "Automate Salesforce with xdge for Real-Time CRM Accuracy"

LinkedIn/xdge. Source for the 286 hours per year wasted on manual data entry per sales rep.

https://www.linkedin.com/posts/xdge-ai_at-xdge-our-sales-operations-run-on-autopilot-activity-7413418154560761859-S-XN

[9] Ameet Patra on CRM as Profit Center

LinkedIn. "CRM should be a profit center, not a cost center. If Salesforce only records activity, it is an expense. If it enforces revenue flow, it becomes an asset."

https://www.linkedin.com/posts/ameet-patra_a-different-perspective-on-servicenow-vs-activity-7411369617350225920-O6DF

[29] "Personalization at Scale: How AI Enabling Mass"

LinkedIn/Nicolas Babin. Analysis of how AI enables mass customization across industries—e-commerce, manufacturing, and healthcare delivering individualized experiences at scale.

https://www.linkedin.com/pulse/personalization-scale-how-ai-enabling-mass-nicolas-babin-tsuge

Primary Research

[5] "Claude Code Token Limits: A Guide for Engineering Leaders"

Faros.ai. RCT finding: experienced developers were 19% slower with AI tools but perceived themselves as 20% faster—a 40 percentage point perception gap.

https://www.faros.ai/blog/claude-code-token-limits

[6] "How to Customize Salesforce for Maximum Efficiency"

Revenue Grid. Source for the 34% selling time statistic—sales reps spend only 34% of their time actually selling.

https://revenuegrid.com/blog/enhance-your-experience-how-to-customize-salesforce-for-maximum-efficiency/

[8] "Why AI Systems of Action Replace Traditional CRM Workflows"

Gong. Documents the context loss problem: 6,000 words of conversation reduced to 30-60 words in CRM notes.

https://www.gong.io/blog/crm-system-of-record-vs-ai-system-of-action

[10] "18 Predictions for 2026"

Jakob Nielsen. Prediction that static interfaces will become obsolete, replaced by generative UI that assembles based on user intent and context.

https://jakobnielsenphd.substack.com/p/2026-predictions

[11] "Top 100 Developer Productivity Statistics with AI Coding Tools (2026)"

Index.dev. Source for the 10-30% average productivity gains from AI tools, with some studies showing up to 81% improvement among active users.

https://www.index.dev/blog/developer-productivity-statistics-with-ai-tools

[25] "Latency as UX: Why 200ms Matters for Perceived Intelligence"

Alex Smale. Research showing 200-300ms as the threshold for perceived responsiveness in interactive systems—the constraint that limits real-time AI capabilities.

https://www.alexsmale.com/latency-as-ux-why-200ms-matters-for-perceived-intelligence/

[28] "Test-Time Compute Analysis"

Hugging Face. Research demonstrating that small model + thinking time can outperform a model 14× larger with instant response—the foundation for why overnight batch processing unlocks AI capability.

https://huggingface.co/blog/Kseniase/testtimecompute

AI Design Patterns & UX

[12] "Guide to AI Integration Patterns for Enterprise-Scale Systems"

Transcenda. Documents the side-panel copilot pattern: AI lives in a collapsible panel providing space for history, reasoning, and deeper synthesis while leaving the main canvas intact.

https://www.transcenda.com/insights/guide-to-ai-integration-patterns-for-enterprise-scale-systems

[16] "AI Design Patterns Enterprise Dashboards"

AufaitUX. Modern dashboards should function as cognitive collaborators that summarize context and predict outcomes rather than just display data.

https://www.aufaitux.com/blog/ai-design-patterns-enterprise-dashboards/

[18] "Designing AI User Interfaces That Foster Trust and Transparency"

UXMatters. Interfaces can present context-sensitive explanations that show the reasoning behind AI outputs without overwhelming users with complex algorithmic details.

https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php

Academic Research on Human-AI Collaboration

[13] "Beyond Human-in-the-Loop"

ScienceDirect. AI offers unparalleled capabilities in processing large-scale data; humans contribute strategic thinking, contextual judgment, and ethical reasoning. Human-over-the-loop shifts humans to a supervisory role.

https://www.sciencedirect.com/science/article/pii/S2666188825007166

[14] "Human- versus Artificial Intelligence"

PMC/NIH. People are better suited than AI systems for a broader spectrum of cognitive and social tasks under unforeseen circumstances. AI excels at data processing while humans handle social-psychosocial interaction.

https://pmc.ncbi.nlm.nih.gov/articles/PMC8108480/

[15] "AI won't make the call: Why human judgment still drives innovation"

Harvard Business School. Human experience and judgment remain critical to making decisions because AI can't reliably distinguish good ideas from mediocre ones or guide long-term business strategies.

https://www.hbs.edu/bigs/artificial-intelligence-human-jugment-drives-innovation

[17] "Beyond Human-in-the-Loop" (HOTL pattern)

ScienceDirect. Human-over-the-loop (HOTL) shifts humans to a supervisory role, allowing AI to handle routine tasks while reserving human input for complex decisions.

https://www.sciencedirect.com/science/article/pii/S2666188825007166

[19] "Brain Cache: A Generative AI-powered Cognitive Exoskeleton"

MIT GenAI & HCI Conference. AI achieves cognitive augmentation through three mechanisms: externalizing biological memory, structuring knowledge, and activating insights.

https://generativeaiandhci.github.io/papers/2025/genaichi2025_51.pdf

[20] "Designing AI User Interfaces That Foster Trust and Transparency"

UXMatters. Context-sensitive explanations show reasoning behind AI outputs without overwhelming users with complex algorithmic details—the foundation of progressive disclosure in AI interfaces.

https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php

[21] "Building AI Trust: The Key Role of Explainability"

McKinsey. Research shows 40% of organizations identify explainability as a key AI risk, but only 17% actively work to mitigate it—evidence that opacity undermines trust.

https://www.mckinsey.com/capabilities/quantumblack/our-insights/building-ai-trust-the-key-role-of-explainability

[22] "Cognitive Load Theory"

John Sweller / ScienceDirect. The foundational research showing that human working memory has limited capacity—our brains can only handle so much at once before performance drops.

https://www.sciencedirect.com/topics/psychology/cognitive-load-theory

[23] "AI UX Design: Creating Transparency and Trust in AI Products"

Wildnet Edge. Documents how 2025 UI toolkits integrate explainability widgets that update dynamically, keeping users informed without demanding deep technical knowledge.

https://www.wildnetedge.com/blogs/ai-ux-design-creating-transparency-trust-in-ai-products

[27] "Mitigating AI Hallucinations: The Power of Multi-Model Approaches"

AI Sutra. Research showing that multi-model approaches—using several AI models together to check each other's work—reduce hallucinations and improve system trustworthiness.

https://aisutra.com/mitigating-ai-hallucinations-the-power-of-multi-model-approaches-2393a2ee109b

Sales Development & SDR Productivity

[31] "AI Agents for Outbound Sales: Complete Guide (2025)"

Smartlead.ai. Analysis showing SDRs spend 70-80% of time on repetitive research tasks, with only 20-30% on relationship-building and strategic thinking. Manual SDRs achieve 50-100 touches/day vs 200-300 for AI-assisted.

https://www.smartlead.ai/blog/ai-agents-for-outbound-sales

[32] "140+ Sales Statistics | 2026 Update"

Spotio. Sales reps spend only 30% of their time actively selling, with 70% consumed by administrative tasks, data entry, and research. Based on Salesforce State of Sales Report data.

https://spotio.com/blog/sales-statistics/

[33] "How to Enable Your SDRs with Data Enrichment: The Playbook"

Databar.ai. SDRs spend up to 40% of their workweek searching for contacts, with 10-15 minutes per prospect on manual research before personalization.

https://databar.ai/blog/article/how-to-enable-your-sdrs-with-data-enrichment-the-playbook

[34] "AI SDRs and Real-Time Lead Data Explained"

Salesforge.ai. Human SDRs spend only 35% of their time actively selling, with the rest consumed by research, data entry, and routine outreach. Real-time lead data improves lead scoring accuracy by up to 40%.

https://www.salesforge.ai/blog/ai-sdrs-real-time-lead-data-explained

[35] "The Best Sales Insights of 2025"

Gong. Analysis of 7.1 million opportunities showing sellers using AI generate 77% more revenue than those not using AI. AI enables personalized outreach at scale that previously required extensive manual research.

https://www.gong.io/blog/the-best-sales-insights-of-2025

[36] "AI Agents for Outbound Sales: Complete Guide"

Smartlead.ai. Comparison of manual vs AI-assisted SDR performance: manual SDRs achieve 50-100 touches/day at 5-10 minutes research per prospect; AI-assisted SDRs achieve 200-300 touches with research automated to seconds.

https://www.smartlead.ai/blog/ai-agents-for-outbound-sales

[37] "AI Sales Tools 2026: Transform Your Sales Performance"

MarketsandMarkets. Companies using AI report productivity jumped 30%, revenue grew 13-15%, and sales cycles became 68% shorter. 83% of sales teams with AI saw revenue growth vs 66% without.

https://www.marketsandmarkets.com/AI-sales/ai-sales-tools-whats-changing

[38] "What is AI Cold Calling? Tools, Strategies & Best Practices"

Outreach. 54% of sales teams now use AI for personalized outreach. Personalization that previously required 15-20 minutes of research per prospect now happens automatically.

https://www.outreach.io/resources/blog/ai-cold-calling

Customer Support & AI

[39] "Call Center Voice AI: How AI is Transforming Support Operations"

Assembled. Studies show agent after-call work dropped by 35% with AI; 81% of agents at companies without AI were overwhelmed by information during calls, compared to 53% at companies with AI deployed.

https://www.assembled.com/page/call-center-voice-ai

[40] "McKinsey: Seizing the Agentic AI Advantage"

McKinsey. In layered AI approaches, AI handles specific steps autonomously—classifies tickets, identifies root causes, resolves simple issues—delivering 20-40% savings in time and 30-50% backlog reduction for internal support operations.

https://www.mckinsey.com/capabilities/quantumblack/our-insights/seizing-the-agentic-ai-advantage

[41] "33 Crucial Customer Service Statistics (2026)"

Shopify/Katana. 49% of customers prefer talking to a live human over an AI chatbot for support; more than half say they'll switch to a competitor after just one bad experience.

https://www.shopify.com/blog/customer-service-statistics

[42] "10 Proven Ways AI Reduces Average Handle Time (2025)"

Dialzara. AI automation cuts documentation time by 50%; Observe.AI achieved 50% reduction in after-call work time; Zendesk's AI assistant improved first-call resolution by 18%.

https://dialzara.com/blog/10-ways-ai-reduces-average-handle-time-aht-in-2024

[43] "Boost SaaS Onboarding with AI-Powered Guidance"

Navigable AI (LinkedIn). Smart ticket classification cuts support time by 25-40%; ticket resolution time down 40% with AI-powered triage and routing.

https://www.linkedin.com/posts/navigable-ai_5-onboarding-questions-your-ai-assistant-activity-7411262970443481088-ECdO

[44] "AI Helpdesk vs Traditional IT Support: The Complete 2025 Comparison"

Dialzara. AI tools boost customer satisfaction by up to 66%; per-ticket costs drop from $17-22 to $3 with AI automation; Zendesk AI agents can automate up to 80% of customer interactions.

https://dialzara.com/blog/ai-help-desk-tools-vs-traditional-support-systems

[45] "How AI Changes CRM: What Businesses Need to Know"

Rapid Online. HubSpot's Customer Agent resolves over 50% of support tickets automatically with 40% less time closing tickets; Reddit cut support resolution times by 84% with Salesforce Agentforce.

https://www.rapidionline.com/blog/ai-crm

UI/UX Design Principles

[46] "Cognitive Load Theory"

John Sweller / ScienceDirect. Foundational research demonstrating that human working memory has limited capacity—excessive information causes performance degradation. Informs the "3-5 key signals" design principle.

https://www.sciencedirect.com/topics/psychology/cognitive-load-theory

[47] "Building AI Trust: The Key Role of Explainability"

McKinsey. Research showing 40% of organizations identify explainability as a key AI risk—hidden evidence undermines user trust, leading to override behavior.

https://www.mckinsey.com/capabilities/quantumblack/our-insights/building-ai-trust-the-key-role-of-explainability

[48] "Designing AI User Interfaces That Foster Trust and Transparency"

UXMatters. Interfaces presenting context-sensitive explanations that show reasoning behind AI outputs—without overwhelming users—form the foundation of progressive disclosure in AI interfaces.

https://www.uxmatters.com/mt/archives/2025/04/designing-ai-user-interfaces-that-foster-trust-and-transparency.php

Strategic Impact & Competitive Dynamics

[49] "The State of AI in 2025"

McKinsey. AI adoption leaders see performance improvements 3.8x higher than the bottom half of adopters, demonstrating the compounding advantage of early AI interface adoption.

https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai

[50] "Beyond Code Generation: How AI is Changing Tech Teams' Dynamics"

Stack Overflow. Traditional onboarding takes 3-6 months to full productivity; AI-enabled onboarding with standardized formats achieves productivity in 1-2 weeks.

https://stackoverflow.blog/2025/10/06/beyond-code-generation-how-ai-is-changing-tech-teams-dynamics/

[51] "The Widening AI Value Gap: Build for the Future 2025"

BCG. The gap between AI leaders and laggards is widening rapidly—early adopters gain compounding advantages that late movers struggle to close.

https://web-assets.bcg.com/e4/14/440d165648a29f02376854919b4d/indias-triple-ai-imperative.pdf

[52] "Unsticking Your AI Transformation"

Bain & Company. "You can't automate your way to transformation. You have to rethink the work itself." Process redesign—not technology—creates most of the transformation value.

https://www.bain.com/insights/unsticking-your-ai-transformation/

[53] "Agentic AI Predictions 2028"

Gartner. 15% of day-to-day work decisions will be made autonomously by 2028, up from near 0% in 2024—a fundamental shift in how knowledge work interfaces operate.

https://www.gartner.com/en/articles/intelligent-agent-in-ai

Note on Sources

This guide draws on a combination of primary research, industry reports, and practitioner observations from LinkedIn and other platforms. Where possible, we cite specific statistics with their original sources. Framework concepts (Two-Pane Cognition, Proposal Cards, Ralph Wiggum Protocol, etc.) are original to LeverageAI's methodology library unless otherwise attributed.

All sources were accessed in January 2026. URLs are provided for verification; content may have been updated since access.