The Intelligent RFP: Proposals That Show Their Work

Scott Farrell November 4, 2025 0 Comments

The Intelligent RFP

How Agentic Systems Orchestrate Smart RAG and Compliance Workflows to Transform Proposal Economics

TL;DR

  • Agentic RFP systems use token-powered loops to orchestrate proven RAG techniques (doc2query, HyDE, compliance matrices) at superhuman scale, enabling hypersprint iteration with full auditability
  • Smart RAG architecture shapes data for compliance: Q&A chunking transforms RFPs into question-shaped indexes and proposals into evidence libraries, enabling bipartite Q↔A graphs that detect gaps automatically
  • Agent receipts (citations, tool logs, eval scores, cost tracking) make systems faster AND more auditable than manual processes, resolving the speed-vs-governance trade-off

The Knowledge Evaporation Crisis

It’s Tuesday afternoon at a mid-sized construction engineering firm. Sarah, a principal engineer, opens her email to find the eighth request this week asking for “the same safety policy excerpt we used in the Melbourne Metro bid—but reformatted for this RFP’s section 3.2.1 requirements.”

She knows she wrote this six months ago. She knows it passed compliance review. She knows it won a $12M contract. But finding it? That’s 40 minutes of PDF searches, Slack thread archaeology, SharePoint diving, and hoping she saved the right version with the right edits.

By Friday afternoon, Sarah will have spent 12 hours answering compliance questions she’s answered before. The proposal will ship on time. The project will launch. And in three months, a colleague on a different bid will email her the exact same question.

This is the hidden tax of RFP workflows—not the initial authoring, but the knowledge evaporation, the constant rework, the institutional amnesia that makes every proposal feel like starting from scratch.

The Real Cost Breakdown

When executives calculate RFP costs, they typically count:

  • Bid manager salaries and overhead
  • External consultants for specialized sections
  • Design and production costs
  • Opportunity cost of senior staff time

What they miss is the compound drag of knowledge loss:

  • 60% of SME time is spent on questions they’ve already answered in previous proposals
  • Compliance matrices are manually maintained in spreadsheets, becoming outdated mid-process as requirements change
  • Pricing tables and BOQs (Bills of Quantities) can’t be effectively searched—they’re locked in Excel files named “Final_FINAL_v3_edited.xlsx”
  • Wins and losses don’t systematically feed into improvement loops—insights from client feedback evaporate
  • Scaling is linear: handling 20% more proposals requires 20% more headcount; there’s no efficiency curve

For a firm submitting 30-50 major proposals annually, this knowledge tax can represent 2,000-3,000 hours of wasted expert time per year. At $150-$300/hour blended rates, that’s $300K-$900K in pure rework costs.

Traditional “AI writing assistants” don’t solve this. Tools like Grammarly, ChatGPT, or even specialized proposal software help individuals write faster—but they don’t connect RFP requirements to past evidence. They don’t build institutional memory. They don’t compound learning.

What’s needed isn’t faster typing. It’s a fundamentally different architecture that treats knowledge as a strategic asset rather than disposable outputs.


Software 3.0: From Assistance to Autonomy

To understand what’s possible with agentic RFP systems, we need context on the broader technological shift happening right now.

Most business leaders jumped directly from traditional software (Software 1.0: humans write every line of code) to “AI assistants” (tools that help you work faster). They skipped an entire paradigm: Software 2.0 (large language models learning patterns from data) and are now missing the transition to Software 3.0 (autonomous agents operating in self-improving loops).

The Evolution in 60 Seconds

Software 1.0: Traditional programming. Humans write explicit instructions. Systems do exactly what you tell them, nothing more.

Software 2.0: Machine learning systems, particularly large language models (LLMs) running on GPUs. Systems learn patterns from data rather than following explicit rules. You train models that can generate, understand, and manipulate language at scale.

Software 3.0: Autonomous AI agents consuming tokens as computational fuel to think, act, and evolve. Agents don’t just assist—they build their own tools, write their own code, and form self-improving loops that operate 24/7 without human oversight.

The leap from 2.0 to 3.0 is profound: it’s the difference between “AI helps me write faster” and “AI orchestrates an entire compliance workflow while I sleep.”

The Triadic Engine

Software 3.0 systems are powered by what I call the Triadic Engine—three interdependent components that create autonomous intelligence:

1. Tokens (Computational Fuel)

Tokens are units of computation consumed when LLMs process input and generate output. They’re measured in API credits from providers like Anthropic (Claude), OpenAI (GPT-4), or Google (Gemini).

Here’s the crucial insight most CFOs miss: tokens aren’t costs—they’re R&D investments in intelligence generation.

Every token burned produces:

  • Analysis of RFP requirements
  • Retrieval and ranking of relevant evidence
  • Generation of compliance-mapped content
  • Evaluation of outputs against quality gates
  • Continuous learning and pattern refinement

The economic dynamic that changes everything:

  • LLM costs drop ~20% monthly (GPT-4 in 2025 costs a fraction of 2023 pricing)
  • Capabilities improve 10-20% monthly (better training, architectural improvements, scaling laws)
  • Tool integrations expand continuously (new APIs, better frameworks, more powerful capabilities)

This creates a compounding dynamic: organizations that established token-burning operations six months ago now have systems that are 50%+ more cost-efficient and significantly more capable than when they started—without changing a single line of code.

Meanwhile, companies that delayed “until costs come down” or “until we understand ROI better” face an insurmountable gap. Their competitors didn’t just get a head start. They climbed an exponential curve while hesitators stayed on linear ground.

2. Agency (Bounded Autonomy)

Agency means the system can pursue goals rather than just respond to prompts. Agentic systems make decisions, adapt strategies, and operate without continuous human oversight.

In RFP contexts, agency manifests as:

  • Planning: Agent analyzes RFP requirements and generates a task graph (which sections need evidence, what’s novel vs. reusable, which SMEs to query)
  • Retrieval: Agent searches past proposals using question-shaped queries, ranks evidence by compliance relevance, and identifies gaps
  • Generation: Agent drafts sections with inline citations to source documents
  • Evaluation: Agent runs quality checks (groundedness, citation accuracy, safety gates) before presenting to humans
  • Learning: Agent captures successful patterns, failed approaches, and SME feedback to improve next iteration

Critically, agency is bounded:

  • Token budgets prevent runaway costs (e.g., “spend max 100K tokens per RFP section”)
  • Evaluation gates block low-quality outputs from reaching SMEs
  • Human validation required before client-facing submission
  • Stop conditions defined (e.g., “halt if groundedness score < 0.85”)

This is autonomous operation with guardrails—freedom to explore within defined constraints.

3. Tools (Real-World Interfaces)

Agents aren’t abstract reasoning engines. They interact with the world through digital tools: function calls, API integrations, and system interfaces.

Essential tools for RFP agents include:

  • Information access: RAG retrieval over past proposals, web search for industry standards, documentation lookup
  • Structured queries: Database queries for vendor pricing, spreadsheet analysis for BOQs, table-QA for compliance matrices
  • Code execution: Running compliance checks, generating visualizations, reformatting documents
  • Communication: Generating SME punchlists, sending notifications, logging decisions
  • Analysis: Evaluating groundedness scores, calculating requirement coverage, detecting citation errors

The breakthrough: agents can build their own tools. If an agent repeatedly encounters a task its current toolset can’t handle efficiently—say, parsing a specific client’s RFP template format—it will construct a custom parser and add it to its toolkit.

This tool-building capability means agent systems become more powerful over time simply through exposure to diverse RFP scenarios. Your system in month 12 has capabilities that didn’t exist in month 1.

The Self-Sustaining Cycle

The Triadic Engine creates a self-sustaining cycle: Agents burn tokens to generate intelligence → Use tools to interact with systems → Evaluate progress toward goals → Adjust strategy and burn more tokens → Repeat continuously.

This isn’t automation (predefined workflows). This is systems that redesign their own workflows.


RAG Architecture for RFP Workflows: Beyond Generic Chunking

Here’s where most “AI for proposals” implementations fail catastrophically: they treat RFPs and proposals like blog posts and apply standard Retrieval-Augmented Generation (RAG) techniques.

Standard RAG workflow:

  1. Split documents into fixed-size chunks (e.g., 512 tokens)
  2. Embed chunks using a model like OpenAI text-embedding-ada-002
  3. Store embeddings in a vector database
  4. At query time, embed the question and retrieve top-k similar chunks
  5. Pass retrieved chunks to LLM for generation

This fails for RFP workflows because:

  • Fixed-size chunking breaks compliance logic: RFP requirement 3.2.1(b) might span 800 tokens across multiple paragraphs; arbitrary splits lose context
  • Keyword mismatch: Client asks for “WHS policy” but your proposal uses “Work Health & Safety framework”—semantic similarity isn’t enough
  • No requirement mapping: You retrieve “similar text” but can’t prove it addresses which specific RFP requirement
  • Compliance gaps invisible: No systematic way to detect which requirements lack evidence
  • Spreadsheets ignored: Pricing tables and BOQs are critical evidence but resist text-based embedding

The Solution: Compliance-First Data Shaping

Instead of generic RAG, we implement compliance-first data shaping: structuring RFPs and proposals specifically for requirement-evidence matching.

Q&A Chunking: The Core Innovation

The fundamental insight: treat RFPs as question corpora and proposals as answer corpora.

RFP Processing Pipeline

Step 1: Parse and Extract Requirements

Use layout-aware parsing (not naive text extraction) to:

  • Extract headings, numbered clauses, tables (submission checklists, scoring rubrics)
  • Classify fragments: Requirement (must/should clauses), Instruction (submission format), Evaluation criterion (scoring), Formality (page limits, fonts)
  • Assign stable IDs: “RFP_SEC3.2.1_REQ007”

→ This mirrors requirements traceability practice from systems engineering: every requirement has an ID linking to upstream sources and downstream artifacts (research: arXiv:2405.10845).

Step 2: Generate Question-Shaped Expansions

For every requirement, generate 1-3 “ask-shaped” paraphrases using doc2query technique:

Original requirement: “The Contractor shall demonstrate compliance with AS/NZS 4801 Occupational Health and Safety Management Systems.”

Generated questions:

  • “What evidence shows compliance with AS/NZS 4801?”
  • “How does the contractor meet occupational health and safety management standards?”
  • “Provide certification or documentation of AS/NZS 4801 compliance”

→ Research foundation: Nogueira & Lin (2019), “Document Expansion by Query Prediction,” cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf

This closes the vocabulary gap: even if your proposal doesn’t use the exact phrase “AS/NZS 4801,” question-shaped retrieval will match on “occupational health and safety management” and “certification.”

Step 3: Store with Rich Metadata

Each requirement becomes a node with attributes:

{
  "id": "RFP_SEC3.2_REQ007",
  "original_text": "The Contractor shall demonstrate...",
  "questions": ["What evidence shows compliance...", ...],
  "section": "3.2 Work Health & Safety",
  "priority": "must",
  "scoring_weight": 15,
  "submission_format": "Attach certification as Appendix C",
  "page_limit": 2,
  "due_date": "2025-11-15"
}

This metadata enables self-query retrieval: filter by Section=WHS, Priority=must before semantic search, dramatically improving precision.

→ Pattern documented: LangChain self-querying retriever (python.langchain.com/docs/how_to/self_query/)

Proposal Processing Pipeline

Step 1: Semantic Chunking (Not Fixed-Size)

Use layout-aware chunking that respects document structure:

  • Hierarchical chunking: Chapters → Sections → Subsections → Paragraphs
  • Semantic boundary detection: Don’t split mid-sentence or mid-clause
  • Parent-child relationships: Small chunks for retrieval precision, but link to parent sections for full context

→ Microsoft RAG guidance emphasizes semantically coherent chunks for policy-like texts (learn.microsoft.com/azure/architecture/ai-ml/guide/rag/rag-chunking-phase)

Why parent-child retrieval matters:

If you retrieve just a child chunk: “Our team holds current certifications including…” → context is lost.

If you return the parent section: “3.2 Work Health & Safety Compliance: Our team holds current certifications including AS/NZS 4801 (renewed 2024-03-15), supplemented by…” → full compliance story intact.

→ Implementation: LangChain ParentDocumentRetriever (python.langchain.com/docs/how_to/parent_document_retrieval/)

Step 2: Extract Claims and Link Evidence

Parse proposal content to identify:

  • Compliance claims: “We comply with AS/NZS 4801”
  • Evidence pointers: “See Appendix C for certification”
  • Exceptions/deviations: “We propose alternative compliance via ISO 45001”
  • Supporting artifacts: Links to certs, past project descriptions, staff CVs

These become your “answer candidates” for requirement matching.

Step 3: Create Answer-Shaped Index

Index with metadata:

{
  "chunk_id": "PROP_SEC3.2_PARA004",
  "text": "Our WHS framework aligns with AS/NZS 4801...",
  "parent_section": "PROP_SEC3.2 (full text)",
  "claim_type": "compliance",
  "evidence_refs": ["AppendixC_Cert_ASNZS4801.pdf"],
  "page": 47,
  "embedding": [0.023, -0.157, ...]
}

The Bipartite Q↔A Graph

Now the magic: index questions (from RFP) and answers (from proposals) separately, then create links.

Dual-index retrieval workflow:

  1. Query: “What evidence addresses RFP requirement 3.2.1(b) on WHS compliance?”
  2. Retrieve from Q-index: Find RFP requirement nodes matching query
  3. Retrieve from A-index: Use requirement’s generated questions to search proposal chunks
  4. Rerank with cross-encoder: Score top-k candidates for precision
  5. Return parent sections + citations: Full context with source page numbers

Store the bipartite links in a compliance graph:

RFP_REQ_3.2.1b --[answered_by]--> PROP_SEC3.2_PARA004
                --[confidence: 0.92]
                --[evidence: AppendixC_Cert_ASNZS4801.pdf]

This enables:

  • Automatic gap detection: Requirements with no high-confidence links → SME punchlist
  • Compliance matrix generation: Requirement → Status (covered/partial/gap) → Evidence → Owner
  • Bidirectional search: “Which RFP requirements does this proposal section satisfy?” (reverse lookup)

→ This extends GraphRAG (Microsoft): extract entities/relations and generate community summaries per section to reason over requirement networks, obligations, dependencies (microsoft.github.io/graphrag/)

Advanced Retrieval: HyDE, Reranking, Query Decomposition

On top of Q&A chunking, layer proven retrieval optimizations:

HyDE (Hypothetical Document Embeddings)

When a requirement is tersely phrased or uses unfamiliar terminology, generate a short “ideal answer” and retrieve passages near it.

RFP requirement: “Demonstrate WHS governance”

HyDE ideal answer: “Our WHS governance framework includes a dedicated safety officer, monthly audits, incident reporting protocols aligned with AS/NZS 4801, and executive accountability through KPIs tied to safety outcomes.”

Embed this ideal answer, retrieve real proposal chunks near it, then present those to LLM for actual generation.

→ Research: Gao et al. (2022), “Precise Zero-Shot Dense Retrieval without Relevance Labels,” arXiv:2212.10496

Cross-Encoder Reranking

After initial retrieval (which optimizes for recall), use a cross-encoder model (e.g., BAAI/bge-reranker-v2-m3) to re-score top-k candidates for precision.

Why this works: Initial embedding-based retrieval is fast but approximate. Cross-encoders jointly encode question + passage and produce fine-grained relevance scores.

→ Implementation: LangChain CrossEncoderReranker (python.langchain.com/docs/integrations/document_transformers/cross_encoder_reranker/), BGE reranker docs (bge-model.com)

Impact: Reranking typically improves precision@5 by 20-40% on long technical documents.

Query Decomposition

When RFP requirements are multi-part (“provide safety plan, past performance on similar projects, and risk register”), break into sub-queries:

  1. “What is our safety plan for this project?”
  2. “What past performance examples show similar project delivery?”
  3. “What risk register do we maintain?”

Retrieve evidence for each sub-query, then merge and rerank before generation.

→ Patterns: Haystack query decomposition (haystack.deepset.ai/blog/query-decomposition), NVIDIA RAG examples (github.com/NVIDIA/GenerativeAIExamples)

Spreadsheets as First-Class Evidence

Pricing tables, vendor quotes, and BOQs aren’t optional footnotes—they’re often the primary scoring criteria. We can’t ignore them just because they resist text-based RAG.

Approach 1: Table-QA (TAPAS-Style)

For structured tables where answers live in cells, use table-QA models that understand row/column semantics:

Query: “What’s the price for swing-stage hire for 16 weeks in CBD including delivery?”

Table-QA: Reasons over rows/columns → “$48,500 (line item 3.7, vendor: SafeScaff)”

→ Research: TAPAS (arXiv:2004.02349)

Approach 2: Row-as-Question

For each row in pricing table, generate a question-shaped chunk:

Row: Swing-stage hire | 16 weeks | CBD | Incl. delivery | $48,500 | SafeScaff

Question chunk: "Price for swing-stage hire, 16 weeks, CBD, including delivery?"

Answer: "$48,500 from SafeScaff (BOQ line 3.7)"

Now this row is retrievable via semantic search and can link to proposal text justifying the inclusion/exclusion.

→ This “row-as-question” approach combines ideas from text-to-SQL with RAG (arXiv:2410.01066v2) and self-query filters.


The Agent Loop: Plan → Retrieve → Act → Evaluate → Learn → Govern

Smart RAG architecture is the substrate. The Agent Loop is the operating system that runs on it.

This loop operationalizes the Triadic Engine: tokens fund exploration, agency applies bounded autonomy, tools ground impact—and receipts make it trustworthy.

You get hypersprints that are fast but also inspectable: every claim cites sources, every action leaves a trail, every improvement is measured.

Phase 1: PLAN

Input: RFP document + past proposal library + token budget + objectives (“Win this bid at ≤15% margin, full compliance, submit by Nov 15”)

Agent actions:

  1. Parse RFP into requirement graph (atomic requirements with IDs, priorities, scoring weights)
  2. Identify requirement types: novel (never seen before), standard (have past evidence), modified (similar but changed)
  3. Generate task DAG: which sections need drafting, which need SME input, dependencies, deadlines
  4. Allocate token budget across tasks (e.g., novel requirements get 3x tokens for exploration)

Output: Plan Receipt

{
  "plan_id": "RFP_2025_TransportVIC_7f2a",
  "objective": "Submit compliant proposal for TransportVIC project by 2025-11-15",
  "constraints": {
    "token_budget": 2000000,
    "quality_threshold": {"groundedness": 0.85, "citation_coverage": 0.95},
    "stop_conditions": ["deadline_reached", "budget_exhausted", "quality_failure"]
  },
  "task_dag": {
    "T001": {"desc": "Draft Sec 3.1 Project Approach", "deps": [], "tokens": 50000},
    "T002": {"desc": "Draft Sec 3.2 WHS Compliance", "deps": [], "tokens": 40000},
    "T003": {"desc": "SME review WHS", "deps": ["T002"], "tokens": 5000},
    ...
  },
  "timestamp": "2025-10-15T09:00:00+11:00"
}

Observability: Create root trace/span using OpenTelemetry GenAI semantic conventions, attach attributes for goal, risk level, budget.

→ Standard: OpenTelemetry GenAI semantics (opentelemetry.io/docs/specs/semconv/gen-ai/)

Phase 2: RETRIEVE (with Citations)

For each task in DAG:

  1. Query generation: Agent converts requirement into retrieval queries (using doc2query expansions + HyDE if needed)
  2. Filtered search: Apply metadata filters (Section, Priority, Project_Type) before semantic search
  3. Dual-index retrieval: Search both Q-index (RFP requirements) and A-index (past proposals)
  4. Rerank: Cross-encoder scores top-20 candidates, select top-5
  5. Return parents + citations: For each child chunk retrieved, return full parent section + source metadata (doc_id, page, lines, hash)

Output: Retrieval Receipt

{
  "task_id": "T002",
  "query": "Evidence for AS/NZS 4801 WHS compliance",
  "filters": {"section": "WHS", "priority": "must"},
  "retrieved": [
    {
      "chunk_id": "PROP_MelbMetro_SEC3.2_PARA004",
      "text": "Our WHS framework aligns with AS/NZS 4801...",
      "parent": "PROP_MelbMetro_SEC3.2 (full section text)",
      "source": {
        "doc_id": "Melbourne_Metro_2024_Proposal.pdf",
        "page": 47,
        "lines": "L340-L385",
        "hash": "sha256:a3f9b2..."
      },
      "rerank_score": 0.94,
      "why": "High-confidence match for WHS compliance evidence"
    },
    {...}
  ],
  "timestamp": "2025-10-15T09:15:23+11:00"
}

Why citations matter: Inline source IDs make every claim verifiable and reduce hallucinations. Most RAG frameworks (LangChain, LlamaIndex) now support citation patterns out-of-the-box.

→ LangChain QA with citations (js.langchain.com/docs/how_to/qa_citations/), Microsoft chunking guidance

Phase 3: ACT (Tool Use with Proof)

Agent takes actions via structured tool calls: draft content, query database, run compliance check, generate visualization, update compliance matrix.

Example tool calls for Task T002 (Draft WHS section):

  1. generate_section: Use retrieved evidence + requirement to draft 800-word WHS compliance section
  2. insert_citations: Add inline citation markers [1], [2] mapping to source docs
  3. check_formatting: Verify page limit (2 pages), font (Arial 11pt), margins
  4. update_compliance_matrix: Mark RFP_REQ_3.2.1b as “Addressed” with evidence link

Output: Action Receipt (per tool call)

{
  "action_id": "ACT_T002_001",
  "tool": "generate_section",
  "input": {
    "requirement_id": "RFP_REQ_3.2.1b",
    "retrieved_context": ["PROP_MelbMetro_SEC3.2_PARA004", ...],
    "length_target": 800,
    "tone": "formal_technical"
  },
  "output_summary": "Generated 847-word section with 3 inline citations",
  "affected_resources": ["draft_sec3.2.md"],
  "span_id": "span_7a3c",
  "duration_ms": 4200,
  "cost": {"prompt_tokens": 8340, "completion_tokens": 920},
  "timestamp": "2025-10-15T09:17:45+11:00"
}

Mechanism: Use function calling (OpenAI, Anthropic tool use) so every side effect is structured and replayable. Attach OpenTelemetry spans for each call.

→ OpenAI function calling (platform.openai.com/docs/guides/function-calling)

Phase 4: EVALUATE (Quality, Safety, Cost)

Every draft passes through evaluation gates before it’s considered “done.”

Quality Gates

  • Groundedness: Are claims supported by retrieved context? (Target: ≥0.85)
  • Context recall: Did we retrieve all relevant evidence? (Target: ≥0.80)
  • Answer relevance: Does output actually address the requirement? (Target: ≥0.90)
  • Citation coverage: Do all factual claims have source citations? (Target: 100%)

→ Metrics from RAGAS framework (docs.ragas.io/en/latest/concepts/metrics/)

Safety Gates

  • Prompt injection detection: Check for attempts to manipulate agent behavior (OWASP LLM01)
  • Sensitive data leakage: Ensure no client-confidential info from wrong projects
  • Hallucination detection: Flag statements not grounded in retrieved context

→ OWASP LLM Top 10 (genai.owasp.org/llmrisk/), Azure content safety guidance

Cost Gates

  • Token accounting: Track cumulative spend against budget per task
  • Efficiency metrics: Tokens per quality point (optimize for output value)
  • Budget alerts: Warn at 80% spend, halt at 100%

Output: Eval Receipt

{
  "artifact_id": "DRAFT_SEC3.2_v1",
  "eval_timestamp": "2025-10-15T09:20:12+11:00",
  "quality": {
    "groundedness": 0.92,
    "context_recall": 0.88,
    "answer_relevance": 0.95,
    "citation_coverage": 1.0
  },
  "safety": {
    "prompt_injection": "pass",
    "data_leakage": "pass",
    "hallucination_check": "pass"
  },
  "cost": {
    "tokens_used": 9260,
    "cumulative_task": 9260,
    "budget_remaining": 30740
  },
  "verdict": "APPROVED",
  "next_action": "route_to_SME_review"
}

Fail-fast behavior: If eval verdict is “REJECTED,” agent either:

  1. Retrieves additional context and regenerates (if context_recall low)
  2. Adjusts generation parameters (if groundedness low)
  3. Flags for SME intervention (if novel requirement, no past evidence)

Phase 5: LEARN (Institutional Memory)

After each task completes (successfully or with SME edits), capture learnings:

  • Successful patterns: “Requirement type X + retrieval strategy Y → groundedness 0.95” → promote to default
  • Failed approaches: “HyDE on pricing tables → poor results” → demote or flag
  • SME feedback: If SME edits agent output, capture diff + rationale → fine-tune retrieval/generation
  • New Q&A pairs: Novel requirements + approved responses → add to corpus for future bids

Store as configuration deltas and dataset versions (not raw prompts—preserve privacy):

{
  "learning_id": "L_2025_WHS_001",
  "trigger": "SME_edit",
  "pattern": "For WHS requirements, prioritize recent certifications over old policy docs",
  "config_change": {
    "retrieval_weights": {"recency": 0.3, "relevance": 0.7},
    "filter_defaults": {"doc_age_max_months": 24}
  },
  "dataset_version": "kb@a1c9",
  "policy_version": "v5",
  "timestamp": "2025-10-15T14:30:00+11:00"
}

Label risks and mitigations using NIST AI Risk Management Framework categories (Govern, Map, Measure, Manage).

→ NIST AI RMF (nist.gov/itl/ai-risk-management-framework)

Phase 6: GOVERN (End-to-End Audit Trail)

All receipts (plan, retrieval, action, eval, learning) roll up into a single trace that can be replayed:

Plan → Retrievals (+citations) → Tool calls → Evaluations → Decisions → SME edits → Final output

Modern observability stacks instrument:

  • Agent reasoning steps
  • Vector DB queries
  • Tool/function calls
  • LLM API calls (with token counts, latencies)
  • Evaluation scores

…all with shared telemetry and trace context, so audits and post-mortems are “one-click.”

→ Platforms: LangSmith (docs.langchain.com/langsmith/observability), Logfire, Arize Phoenix, Traceloop

The Agent Receipt (Complete Schema)

Every final artifact (section, table, appendix) carries a compact receipt:

{
  "artifact_id": "RFP_2025_TransportVIC:Sec3.2:Final",
  "objective": "Address RFP Section 3.2 WHS compliance requirements",
  "inputs": {
    "requirements": ["RFP_REQ_3.2.1a", "RFP_REQ_3.2.1b", "RFP_REQ_3.2.2"],
    "retrieval": [
      {
        "doc_id": "Melbourne_Metro_2024_Proposal.pdf#p47",
        "hash": "sha256:a3f9b2...",
        "why": "AS/NZS 4801 compliance evidence",
        "rerank_score": 0.94
      },
      {
        "doc_id": "PolicyWHS_v3.md#L120-188",
        "hash": "sha256:f8e3c1...",
        "why": "Current WHS framework description",
        "rerank_score": 0.89
      }
    ],
    "plan_id": "RFP_2025_TransportVIC_7f2a"
  },
  "actions": [
    {
      "tool": "generate_section",
      "span_id": "span_7a3c",
      "input_summary": "847 words, 3 citations",
      "result_ref": "draft_sec3.2_v1.md"
    },
    {
      "tool": "update_compliance_matrix",
      "span_id": "span_8b4d",
      "result_ref": "compliance_matrix.json"
    }
  ],
  "citations": [
    {"marker": "[1]", "doc_id": "Melbourne_Metro_2024_Proposal.pdf", "page": 47, "lines": "L340-L360"},
    {"marker": "[2]", "doc_id": "PolicyWHS_v3.md", "lines": "L140-L160"},
    {"marker": "[3]", "doc_id": "Cert_ASNZS4801_2024.pdf", "page": 1}
  ],
  "eval": {
    "quality": {
      "groundedness": 0.92,
      "context_recall": 0.88,
      "answer_relevance": 0.95,
      "citation_coverage": 1.0
    },
    "safety": ["LLM01:pass", "LLM02:pass", "data_leakage:pass"]
  },
  "sme_review": {
    "reviewer": "Sarah Chen, Principal Engineer",
    "status": "approved_with_edits",
    "edits": "Clarified cert renewal date, added reference to recent audit",
    "timestamp": "2025-10-15T16:20:00+11:00"
  },
  "cost": {
    "prompt_tokens": 18324,
    "completion_tokens": 4210,
    "total_cost_aud": 0.47
  },
  "timestamp": "2025-10-15T09:20:12+11:00",
  "version": {
    "model": "claude-sonnet-4-20250514",
    "kb_snapshot": "kb@a1c9",
    "policy": "v5"
  }
}

Why this matters for conservative industries:

  • Procurement audits can trace every claim back to source documents
  • Legal review can verify no hallucinations (groundedness scores + citations)
  • Finance can see exact cost per section (token accounting)
  • Quality teams can replay decision logic (full trace available)

This isn’t just “AI-assisted writing.” It’s governance-first autonomous intelligence.


What Changes: From Authors to Orchestrators

Understanding the architecture is one thing. Grasping what actually transforms in your organization is another.

For SMEs: Elevated from Authors to Validators

Before (Traditional Workflow):

Sarah receives email: “We need the WHS section for the TransportVIC bid. 800 words, due Friday, needs AS/NZS 4801 compliance evidence.”

She spends:

  • 30 mins searching for past examples in PDFs and SharePoint
  • 90 mins drafting from memory + copy-pasting + reformatting
  • 20 mins hunting for certification docs to attach
  • Total: 2.5 hours for one section (×15 sections = 37.5 hours per proposal)

After (Agentic Workflow):

Sarah receives notification: “Agent drafted WHS section for TransportVIC. Review receipt and approve.”

Receipt shows:

  • Section text (847 words) with inline citations [1], [2], [3]
  • Sources: Melbourne Metro proposal (p47), current WHS policy (L140-160), AS/NZS 4801 cert
  • Eval scores: Groundedness 0.92, Citation coverage 100%
  • Compliance matrix updated: RFP_REQ_3.2.1b marked “Addressed”

She spends:

  • 5 mins reading draft
  • 3 mins verifying citations point to correct sources
  • 7 mins editing: clarify cert renewal date, add reference to recent safety audit
  • Total: 15 mins for validation (×15 sections = 3.75 hours per proposal)

Result: 90% time reduction. Sarah shifts from “write everything” to “validate agent outputs and fill gaps the system flags.”

Her work becomes:

  • Reviewing agent-generated punchlists (novel requirements, low-confidence sections)
  • Providing targeted input where past evidence is weak
  • Approving outputs with full visibility into sources and reasoning
  • Strategic thinking about what wins bids (not formatting compliance text for the 30th time)

For Teams: Hypersprints and Compounding Knowledge

Hypersprints: Iteration at Machine Speed

Traditional proposal development operates on human timescales:

  • Week 1: Kickoff, assign sections, initial drafts
  • Week 2: First review, identify gaps, request SME input
  • Week 3: Revisions, integrate feedback, formatting
  • Week 4: Final review, compliance check, submit

During this month, the team explores maybe 5-7 different approaches to key sections (if they’re thorough).

Agentic hypersprints compress this:

Monday 9am: Kick off agent with RFP, token budget, objectives

Monday 9pm – Tuesday 6am (overnight): Agent performs 200+ iterations:

  • Try approach A for technical solution section, eval score 0.78 → iterate
  • Try approach A’ with different evidence sources, eval 0.84 → iterate
  • Try approach B (different framing), eval 0.91 → keep
  • Explore edge case: “What if client prioritizes sustainability over cost?” → generate alternative
  • Test formatting variations for readability
  • Optimize citation density (too many → cluttered; too few → unverifiable)

Tuesday 7am: SME arrives to:

  • Top 3 draft variations (ranked by eval scores)
  • Compliance matrix showing coverage of all requirements
  • Punchlist of 8 novel requirements needing SME input
  • Comparative analysis: “Approach B scores higher on technical depth but lower on cost-effectiveness; recommend B for tech-focused clients, A for cost-sensitive”

The team explores 40x more solution space in 9 hours (overnight) than they could in a month of human-paced iteration.

Compounding Institutional Knowledge

Traditional workflow: Each proposal is a discrete project. Knowledge evaporates after submission.

Agentic workflow: Each completed proposal leaves behind:

  1. Hardened Q&A pairs: RFP requirements + validated responses → added to RAG corpus
  2. Vetted evidence with citations: “For WHS compliance, use Melbourne Metro cert (recent) not Sydney Harbour cert (outdated)”
  3. SME decisions: When Sarah edits a draft, the diff + rationale is captured: “Always mention recent audits, not just certs”
  4. Evaluation data: Which retrieval strategies worked (HyDE on technical sections) vs. failed (HyDE on pricing tables)
  5. Win/loss insights: If proposal wins, mark all evidence as “high-value”; if loses, review eval scores for weak sections

This creates a compounding knowledge curve:

  • Proposal 1 (month 1): Agent has baseline RAG corpus, explores broadly, requires heavy SME input
  • Proposal 5 (month 3): Agent has 4 previous bids in corpus, retrieval precision improves, fewer SME questions
  • Proposal 20 (month 12): Agent has 19 bids, learned patterns cover 70% of requirements, SMEs focus on 30% novel/strategic sections

Knowledge doesn’t evaporate. It accumulates and improves the system automatically.

For Organizations: Token Budgets Replace Licensing Fees

Traditional Software Economics

Enterprise software costs are fixed and opaque:

  • Salesforce: $150-$300/user/month whether you use it or not
  • SAP: $2M-$10M implementation + $500K/year maintenance
  • Proposal software vendors: $50K-$200K/year licensing + $100K/year “professional services”

Costs don’t map to value. You pay for seats, not outcomes.

Token Economics

Agentic systems cost = computational work performed:

  • You spend tokens when agents think and act
  • Idle systems cost $0
  • Heavy usage reflects actual value generation

Example: TransportVIC proposal (40 sections, 2M token budget)

Section 3.1 (Novel technical approach): 150K tokens → $3.20
Section 3.2 (Standard WHS, high reuse): 40K tokens → $0.85
Section 3.3 (Pricing, mostly retrieval): 25K tokens → $0.53
...
Total: 1.8M tokens used → $38.40 for entire proposal

Compare to:

  • SME time savings: 35 hours × $200/hr = $7,000
  • Faster turnaround: 1 week vs. 4 weeks = 3 weeks competitive advantage
  • Higher quality: 200 iterations vs. 5 = better technical solutions

ROI is absurd: spend $40 in tokens to save $7K in labor and compress 4 weeks to 1 week.

More importantly: You control the systems. No vendor lock-in. No forced upgrades. No proprietary black boxes. Agents can rebuild, modify, or optimize any workflow through conversational interaction.

Competitive Dynamics: The Compounding Advantage

Month 1: Your competitor starts using agentic RFP workflows. You wait to “see if it works.”

Month 3: Their agents have processed 10 proposals. RAG corpus is richer. Retrieval precision improves. SME time per proposal drops 50%.

Month 6: They’ve completed 25 proposals. Agent has learned patterns covering 60% of common requirements. Token costs dropped 40% (LLM pricing improvements). Capabilities improved 30% (model updates). Their system is now 50%+ more cost-efficient AND more capable than their starting point.

Month 6 (You): You finally decide to start. Your agents begin with empty RAG corpus. You’re exploring patterns your competitor mastered in month 2. You’re paying full token costs they no longer pay. Your team is learning orchestration skills their team has been refining for 6 months.

Month 12 (Competitor): 60 proposals in corpus. Win rate improved 15% (better technical solutions through hypersprint exploration). Cost per proposal down 60%. Talent retention up (top SMEs love orchestration work vs. repetitive authoring). Industry reputation as “cutting-edge firm.”

Month 12 (You): 15 proposals in corpus (slower ramp). Struggling with agent orchestration culture change. Still paying higher token costs. Playing catch-up on capability curve.

The gap isn’t linear. It’s exponential because advantages compound monthly through:

  1. Growing RAG corpus (more past evidence = better retrieval)
  2. Dropping token costs (20%/month industry-wide)
  3. Improving model capabilities (10-20%/month)
  4. Team expertise in orchestration (learning curve advantage)

Early adopters don’t get a head start. They climb a curve that becomes steeper over time.


Building on Proven Foundations

It’s critical to emphasize: we’re not inventing speculative AI magic here.

We’re orchestrating proven best practices that RFP professionals, systems engineers, and AI researchers have established over decades:

Established Industry Practices

  • Compliance matrices are standard guidance from APMP (Association of Proposal Management Professionals). We’re automating their generation and maintenance, not inventing the concept.
  • Requirements traceability is fundamental to systems engineering and procurement (research literature, NIST standards). We’re making it systematic and tireless, not novel.
  • “RFP shredding” (breaking RFPs into atomic requirements) is literal industry terminology. We’re applying AI to do it faster and more consistently.

→ APMP RFP Shredding guidance: apmp-western.org/wp-content/uploads/2023/10/WRC2023-Conniff-Shred-For-Success.pdf

Proven AI Research Techniques

  • Doc2query (Nogueira & Lin, 2019) closed vocabulary gaps in academic information retrieval. We’re applying it to RFP-proposal alignment.
  • HyDE (Gao et al., 2022) showed hypothetical document embeddings improve zero-shot retrieval. We’re using it for terse requirements.
  • Parent-child retrieval prevents citation fragmentation (documented in LangChain, Microsoft RAG guidance). We’re using it for proposal evidence.
  • Cross-encoder reranking consistently lifts retrieval precision 20-40% (BGE, Cohere research). We’re applying it to compliance matching.
  • GraphRAG (Microsoft) enables reasoning over knowledge graphs. We’re using it for requirement dependency networks.

Governance and Safety Standards

  • NIST AI Risk Management Framework provides governance structure (Govern, Map, Measure, Manage). We’re aligning agent learning loops to it.
  • OWASP LLM Top 10 identifies security risks (prompt injection, data leakage). We’re implementing recommended mitigations.
  • OpenTelemetry GenAI semantics standardize observability for AI systems. We’re instrumenting agent loops with it.
  • RAGAS evaluation framework defines retrieval quality metrics (groundedness, context recall, faithfulness). We’re using these as eval gates.

→ Comprehensive reference list in article_prethink.md includes 20+ academic papers, industry standards, and platform docs

The innovation isn’t any single technique. It’s the synthesis: combining Software 3.0 thinking (tokens as fuel, agency as policy, tools as reach) with compliance-first RAG architecture (Q&A chunking, bipartite graphs, structured receipts) to create systems that are greater than the sum of their parts.

We’re not asking you to bet on unproven AI fantasies. We’re asking you to systematically orchestrate proven methods at machine speed with full auditability.


Implementation Roadmap: From Pilot to Production

Transitioning to agentic RFP workflows doesn’t require burning down existing processes. It requires developing new capabilities alongside current operations.

Phase 1: Two-Proposal Pilot (Weeks 1-4)

Goal: Validate Q&A chunking, measure baseline metrics, build team confidence

Step 1: Select two past RFP↔proposal pairs where you won the bid (known good outcomes)

Step 2: Implement basic RAG pipeline:

  • Parse RFPs into atomic requirements (use LLM-assisted extraction)
  • Chunk proposals with semantic boundaries (hierarchical chunking)
  • Generate doc2query expansions for requirements
  • Create embeddings and store in vector DB (Chroma, Pinecone, Weaviate)

Step 3: Measure baseline metrics:

  • Requirement coverage: What % of RFP clauses retrieve high-confidence evidence (rerank score ≥0.80)?
  • Retrieval precision@5: Of top-5 retrieved chunks, how many are actually relevant?
  • SME review time: How long does it take to validate retrieved evidence vs. hunting manually?
  • Gap detection accuracy: Does the system correctly identify which requirements lack evidence?

Step 4: Layer in optimizations and re-measure:

  • Add HyDE for terse requirements → measure precision improvement
  • Add cross-encoder reranking → measure precision improvement
  • Add parent-child retrieval → measure citation quality
  • Add metadata filters (section, priority) → measure precision improvement

Success criteria:

  • Requirement coverage ≥70%
  • Retrieval precision@5 ≥60%
  • SME review time reduced ≥40% vs. manual search
  • Zero false negatives on gap detection (better to over-flag than miss requirements)

Phase 2: Agent Loop Integration (Weeks 5-8)

Goal: Deploy full agent loop on one RFP section, validate receipts and governance

Step 1: Choose one RFP section (e.g., WHS compliance) for agent drafting

Step 2: Implement agent loop phases:

  • Plan: Agent generates task graph for section
  • Retrieve: Agent queries RAG with citations
  • Act: Agent drafts section with inline citation markers
  • Evaluate: Run groundedness, citation coverage, safety checks
  • Output: Generate receipt with all metadata

Step 3: SME validation workflow:

  • SME receives: draft text + receipt (sources, eval scores, cost)
  • SME validates citations point to correct evidence
  • SME edits for accuracy, tone, strategic positioning
  • System captures edits + rationale for learning phase

Step 4: Measure governance metrics:

  • Audit trail completeness: Can you trace every claim to source?
  • SME approval rate: What % of drafts approved with minor edits vs. major rewrites?
  • Time to review: How long does SME spend validating vs. authoring from scratch?

Success criteria:

  • 100% citation traceability (no orphaned claims)
  • ≥70% SME approval rate with minor edits only
  • ≥60% time reduction vs. manual authoring
  • Full observability: complete trace replay available

Phase 3: Full Proposal Automation (Weeks 9-16)

Goal: Scale to entire proposal, integrate spreadsheets, automate compliance matrix

Step 1: Extend agent loop to all proposal sections

Step 2: Implement spreadsheet handling:

  • Table-QA for structured pricing tables
  • Row-as-question for BOQs and vendor quotes
  • Link pricing evidence to proposal justification text

Step 3: Automate compliance matrix generation:

  • Requirement → Status (covered/partial/gap)
  • Evidence links (with citations)
  • Owner assignment (agent vs. SME)
  • Real-time updates as sections complete

Step 4: Deploy hypersprint workflow:

  • Kick off agent Monday 5pm with token budget
  • Overnight: agent explores 100-200 iterations
  • Tuesday 8am: SME reviews top-ranked drafts + punchlist
  • Iterate through week, final SME approval Friday

Success criteria:

  • ≥80% requirement coverage automated (agent-drafted with citations)
  • ≤20% requiring novel SME input (flagged via punchlist)
  • Proposal turnaround: 1-2 weeks vs. 4-6 weeks manual
  • Cost: $50-$200 in tokens vs. $10K-$30K in SME labor

Phase 4: Continuous Improvement (Ongoing)

Goal: Compound institutional knowledge, optimize retrieval, expand tool ecosystem

Activities:

  • After each proposal, capture learnings (successful patterns, SME edits, win/loss insights)
  • Quarterly: analyze eval metrics trends, tune chunking strategies, update retrieval weights
  • Expand tool ecosystem: custom parsers for client-specific RFP formats, integrations with CRM/ERP
  • Build feedback loops: link won/lost bids to proposal sections, identify high-value evidence patterns

Metrics to track:

  • Requirement coverage trend (should increase as corpus grows)
  • SME review time trend (should decrease as agent quality improves)
  • Win rate correlation (do agent-assisted proposals win more?)
  • Token efficiency (cost per quality point should improve monthly)

Ready to pilot agentic RFP workflows?

Start with two past proposals. Measure requirement coverage. Build orchestration expertise.

The compounding advantage starts today.


The Urgency: Cost of Delay

You might be thinking: “This sounds promising, but let’s wait 12 months until it matures, costs drop further, and best practices solidify.”

That’s the most expensive decision you could make.

Why Waiting Guarantees Irrelevance

The compounding dynamic:

Token costs drop 20%/month. Model capabilities improve 10-20%/month. Every month you operate agentic systems, you get:

  • Free efficiency gains: The same token budget buys exponentially more intelligence
  • Free capability upgrades: Models get smarter without you changing code
  • Institutional knowledge accumulation: Your RAG corpus grows, retrieval improves, patterns solidify
  • Team expertise: Orchestration skills develop through practice

Organizations that started 6 months ago now have systems that are 50%+ more cost-efficient and significantly more capable than their starting point.

But here’s the trap: token costs dropping benefits everyone equally. In 12 months, everyone will have access to cheap tokens. The competitive advantage doesn’t come from cheap tokens—it comes from:

  • Mature RAG corpus: 50+ proposals indexed with Q&A chunking
  • Proven agent patterns: 6-12 months of learning what retrieval strategies work for which requirement types
  • Orchestration culture: Teams comfortable with defining objectives, allocating token budgets, validating receipts
  • Tool ecosystem: Custom integrations, parsers, evaluators built over time

Competitors who start in 12 months will face:

  • Cheap tokens (same as you)
  • Empty RAG corpus (you have 50+ proposals)
  • Zero orchestration expertise (you have 12 months of pattern recognition)
  • Generic tooling (you have custom-built integrations)

The gap is unbridgeable because advantages compound monthly.

The Talent Dimension

Top proposal managers, SMEs, and technical leaders are watching this transition happen.

In 12-18 months, the employment landscape splits:

  • Agent-first firms: “Join us to orchestrate cutting-edge AI systems, explore solution spaces through hypersprints, build institutional knowledge that compounds”
  • Manual firms: “Join us to write compliance sections from scratch for the 40th time”

Which job posting attracts A-players?

Delaying adoption doesn’t just cost efficiency. It costs talent magnetism.

The Strategic Positioning Window

Early adopters don’t just gain technical advantages. They shape how the industry thinks about agentic workflows.

When your competitors see you:

  • Turning around proposals in 1 week vs. their 4 weeks
  • Submitting 30% more bids with same team size
  • Winning at higher rates through better technical solutions (hypersprint exploration)

…they scramble to copy. But by then you’re 12 months ahead on the learning curve.

Early adopters write the playbooks. Laggards follow them—at a disadvantage.


Conclusion: The Choice Ahead

We’re at an inflection point in professional services.

RFP workflows have been fundamentally unchanged for decades: human SMEs write proposals, knowledge evaporates after submission, costs scale linearly, institutional memory exists in people’s heads (and leaves when they do).

Agentic systems—powered by the Triadic Engine of tokens, agency, and tools—offer a different path:

  • Knowledge compounds rather than evaporates (Q&A chunked RAG stores)
  • Iteration happens at machine speed rather than human pace (hypersprints)
  • Governance becomes systematic rather than manual (agent receipts, audit trails)
  • Costs map to value rather than fixed licensing (token economics)
  • SMEs orchestrate rather than author from scratch (elevated roles)

This isn’t speculative AI magic. It’s the systematic orchestration of proven techniques—compliance matrices (APMP), requirements traceability (systems engineering), doc2query (academic IR), parent-child retrieval (Microsoft RAG guidance), evaluation frameworks (RAGAS), governance standards (NIST AI RMF)—woven together through autonomous agents that operate in token-powered loops.

Business leaders in proposal-heavy industries face a choice:

Path 1: Wait and See

Maintain traditional workflows. Hire more proposal managers. Invest in “AI writing assistants” that help individuals type faster. Optimize existing processes incrementally. This path feels safe. It guarantees competitive irrelevance.

Path 2: Embrace Agentic Workflows

Start with a two-proposal pilot. Measure requirement coverage. Implement the agent loop. Build orchestration expertise. Capture institutional knowledge systematically. Compound advantages monthly. This path feels risky. It’s the only rational strategy.

The uncomfortable truth: agentic advantages compound monthly through falling token costs, improving model capabilities, growing RAG corpuses, and team expertise. Organizations that start now will be exponentially ahead in 12 months. Those who delay will find the gap unbridgeable.

The question to ask yourself: What could your organization accomplish if proposal velocity wasn’t constrained by human working hours, and institutional knowledge didn’t evaporate after every bid?

The compounding advantage starts today—or six months behind your competitors.

Scott Farrell helps organizations transition to Software 3.0 through agent-first architectures and autonomous intelligence systems. He specializes in applying agentic workflows to knowledge-intensive professional services, with deep expertise in RAG architecture, compliance automation, and token economics.

To discuss implementing agentic RFP workflows in your organization, contact: [email protected]

Learn more at leverageai.com.au

Leave a Reply

Your email address will not be published. Required fields are marked *