<?xml version="1.0" encoding="UTF-8"?>
<rss  xmlns:atom="http://www.w3.org/2005/Atom" 
      xmlns:media="http://search.yahoo.com/mrss/" 
      xmlns:content="http://purl.org/rss/1.0/modules/content/" 
      xmlns:dc="http://purl.org/dc/elements/1.1/" 
      version="2.0">
<channel>
<title>TALVINDER</title>
<link>https://talvinder.com/</link>
<atom:link href="https://talvinder.com/index.xml" rel="self" type="application/rss+xml"/>
<description>Frameworks, build logs, and field notes by B. Talvinder — serial founder (ex-OYO, YC, 500 Startups), four companies across four technology eras.</description>
<generator>quarto-1.9.37</generator>
<lastBuildDate>Fri, 17 Apr 2026 00:00:00 GMT</lastBuildDate>
<item>
  <title>Training AI to Serve Rare Disease Patients Is a Structural Problem, Not a Data Problem</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/training-ai-to-serve-rare-disease-patients-is-structural/</link>
  <description><![CDATA[ 





<pre><code>
AI failures in rare disease diagnosis are not about data scarcity. They are about healthcare’s structural bottlenecks—fragmented data silos, inconsistent protocols, and missing consent infrastructure—that make reliable AI impossible at scale. Data scarcity is a symptom. The root cause is the system design underneath.

In 2023, Eka Care introduced explicit patient consent flows before any health data was accessed for AI training. This slowed data acquisition but ensured legal standing and clinical trust. The lesson is clear: you cannot fix a governance problem by throwing more data at it.

## The Structural Bottleneck Framework

I call this the **Structural Bottleneck Framework**: AI performance in rare diseases is limited not by model size or dataset volume, but by systemic healthcare design flaws. Fragmented data, inconsistent clinical protocols, and privacy roadblocks produce an environment where AI trained on generic or legacy datasets will fail at point-of-care deployment.

Most AI healthcare teams obsess over model selection, fine-tuning, and benchmark chasing while neglecting data governance architecture, consent infrastructure, AI validation layers, and domain protocol alignment. That’s why rare disease AI remains a demo that never makes it into clinics.

Fixing data quantity without fixing data governance is like adding fuel to a car with no steering wheel.

## Why More Data Doesn’t Solve the Problem

Healthcare data is siloed by provider, geography, and regulation. No amount of model tuning overcomes that fragmentation.

Imagine a sensor network with noisy, inconsistent, and incomplete signals. The output will be unreliable regardless of how sophisticated the algorithms are. This is not a metaphor. It is literally how AI input pipelines behave when data sources are fragmented and unverified.

In 2022, an AI system deployed for pediatric rare disease diagnosis nearly caused a malpractice incident by mislabeling a critical symptom. The model had been trained on adult datasets with different clinical presentations. This failure was structural, not statistical.

Generic datasets compound the problem. Retrieval-augmented generation (RAG) approaches surface obsolete or irrelevant medical guidelines when the knowledge base is not actively maintained and aligned with current clinical protocols. Fine-tuning on scarce rare disease data is insufficient if the underlying data ecosystem doesn’t support real-time, trustworthy updates. A model fine-tuned in 2022 will give outdated guidance in 2025. Training cycles cannot keep pace without structural integration into clinical protocol update chains.

The ethical dimension is not a compliance checkbox. AI deployed without patient consent frameworks creates legal risk and erodes clinical trust. Once a clinician sees an AI system give a dangerous recommendation, that system is dead in that institution regardless of subsequent accuracy gains. Rebuilding clinical trust after a structural failure is harder than building it correctly the first time.

Falsifiable claim: AI models trained with incremental data additions but without systemic integration of domain-specific, privacy-aware data governance will continue producing dangerous misclassifications at rates preventing clinical adoption. The structural bottleneck, not data volume, is the binding constraint.

## Concrete Evidence From India and Beyond

Eka Care’s 2023 shift to consent-driven data acquisition is the clearest example of getting the structural layer right. Patient consent protocols slowed data access but ensured the data used for AI training had legal standing and patient trust behind it. This is not a formality. It is what makes AI deployable in clinics rather than research labs.

Multiple Indian healthcare startups have deployed AI that misread critical symptoms as banal conditions because their models trained on generic datasets lacked rare disease-specific clinical annotation. One AI misclassified a rare autoimmune condition as a common allergy, simply because pattern matching aligned with far more frequent conditions in the training set. This is not a data volume problem. It is a structural failure to align the model with clinical taxonomy for the target patient population.

Telemedicine adoption in rural India illustrates the same bottleneck differently. 5G coverage and smartphones exist. The structural barrier to AI-assisted diagnosis is not data volume. It is the absence of validated clinical protocols for AI decision support in resource-constrained settings, liability frameworks clinicians and patients understand, and feedback mechanisms that let clinicians flag AI errors in real time.

At Ostronaut, building AI-generated healthcare training content revealed the same pattern at scale. Generating clinical learning material required more than ingesting large content volumes. We needed validation layers: domain experts reviewing AI output against current clinical guidelines, quality gates flagging outdated protocols, and structured feedback loops improving generation accuracy over time. More data ingestion without these structural layers yields more plausible but incorrect content. Volume does not substitute for architecture.

## What the Fix Looks Like

The Structural Bottleneck Framework points to a different investment thesis for rare disease AI.

| Traditional AI Effort | Structural Bottleneck Focus |
|----------------------|-----------------------------|
| Model tuning and benchmarks | Consent and data governance infrastructure |
| Dataset volume and augmentation | Clinical protocol alignment and validation layers |
| Statistical fine-tuning | Real-time domain updates and feedback mechanisms |
| Isolated AI pipelines | Integrated healthcare system workflows |

The fix starts with consent and governance. Patient consent must be explicit, auditable, and embedded in data pipelines. Data governance can’t be an afterthought or legal checkbox. It must be engineered as infrastructure.

Second, AI validation layers must become standard. Domain experts need to build continuous quality gates and feedback loops. AI outputs require real-world clinical protocol integration, not just offline benchmarks.

Third, clinical protocols must be actively maintained and integrated with AI knowledge bases. Rare disease protocols evolve. The model’s training cycle must be tightly coupled with these updates, or risk obsolescence.

Finally, liability and trust frameworks need clarity. Clinicians must know when and how AI can be used safely, and have mechanisms to flag and correct errors in real time.

At Ostronaut, we learned this the hard way. AI-generated clinical content without validation layers isn’t just wrong; it erodes trust in the entire system. The data volume was never the problem.

## What I Don’t Know Yet

How do you build scalable, privacy-aware consent infrastructure that works across fragmented healthcare providers and jurisdictions — without killing innovation speed? It’s an unsolved technical and regulatory puzzle.

How do you design AI validation layers that keep pace with rapidly evolving clinical protocols in rare diseases, given the scarcity of domain experts? Automation helps, but domain knowledge bottlenecks remain.

How do we create feedback mechanisms that incentivize clinicians to report AI errors and integrate those corrections back into the training loop — especially in resource-constrained settings?

These are open engineering and policy questions, not hype fodder.

## The Question Worth Asking

The Structural Bottleneck Framework shifts focus from data quantity to system quality. The question worth asking now is: can AI companies and healthcare institutions collaborate on building structural data governance and validation infrastructure at scale — or will rare disease AI remain a demo for another decade?

Not in three years. In ten. In fifty.

Are we asking it? Mostly, no.

More on this as I develop it.




:::{#quarto-navigation-envelope .hidden}
[TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1zaWRlYmFyLXRpdGxl"}
[TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXItdGl0bGU="}
[Frameworks]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6RnJhbWV3b3Jrcw=="}
[/frameworks/index.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2ZyYW1ld29ya3MvaW5kZXguaHRtbA=="}
[Build Logs]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6QnVpbGQgTG9ncw=="}
[/build-logs/index.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2J1aWxkLWxvZ3MvaW5kZXguaHRtbA=="}
[Field Notes]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6RmllbGQgTm90ZXM="}
[/field-notes/index.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2ZpZWxkLW5vdGVzL2luZGV4Lmh0bWw="}
[Bets]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6QmV0cw=="}
[/bets-log/index.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2JldHMtbG9nL2luZGV4Lmh0bWw="}
[About]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6QWJvdXQ="}
[About Me]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6QWJvdXQgTWU="}
[/about.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2Fib3V0Lmh0bWw="}
[Library]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6TGlicmFyeQ=="}
[/library/index.html]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6L2xpYnJhcnkvaW5kZXguaHRtbA=="}
[Subscribe]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6U3Vic2NyaWJl"}
[https://buttondown.com/talvinder]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLWludC1uYXZiYXI6aHR0cHM6Ly9idXR0b25kb3duLmNvbS90YWx2aW5kZXI="}

:::{.hidden .quarto-markdown-envelope-contents render-id="Zm9vdGVyLWxlZnQ="}
&amp;copy; 2026 B. Talvinder. Built with conviction.

:::


:::{.hidden .quarto-markdown-envelope-contents render-id="Zm9vdGVyLWNlbnRlcg=="}
[GitHub](https://github.com/talvinder) · [LinkedIn](https://linkedin.com/in/talvindersingh) · [X](https://x.com/talvinder27)

:::


:::{.hidden .quarto-markdown-envelope-contents render-id="Zm9vdGVyLXJpZ2h0"}
Powered by [Quarto](https://quarto.org)

:::

:::



:::{#quarto-meta-markdown .hidden}
[Training AI to Serve Rare Disease Patients Is a Structural Problem, Not a Data Problem – TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLW1ldGF0aXRsZQ=="}
[Training AI to Serve Rare Disease Patients Is a Structural Problem, Not a Data Problem – TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLXR3aXR0ZXJjYXJkdGl0bGU="}
[Training AI to Serve Rare Disease Patients Is a Structural Problem, Not a Data Problem – TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLW9nY2FyZHRpdGxl"}
[TALVINDER]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLW1ldGFzaXRlbmFtZQ=="}
[Rare disease AI failures stem from healthcare’s fragmented data governance, not from insufficient data volume.]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLXR3aXR0ZXJjYXJkZGVzYw=="}
[Rare disease AI failures stem from healthcare’s fragmented data governance, not from insufficient data volume.]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLW9nY2FyZGRkZXNj"}
[Talvinder Singh is a serial founder (ex-OYO, YC, 500 Startups) who has built four companies across four different technology eras in eighteen years. Frameworks, build logs, and field notes on AI, infrastructure, and the India tech ecosystem.]{.hidden .quarto-markdown-envelope-contents render-id="cXVhcnRvLW1ldGFzaXRlZGVzYw=="}
:::




&lt;!-- --&gt;

::: {.quarto-embedded-source-code}
```````````````````{.markdown shortcodes="false"}
---
title: "Training AI to Serve Rare Disease Patients Is a Structural Problem, Not a Data Problem"
description: "Rare disease AI failures stem from healthcare’s fragmented data governance, not from insufficient data volume."
date: 2026-04-17
categories: ['AI in Healthcare', 'AI Validation', 'India Tech']
draft: false
---</code></pre>
<p>AI failures in rare disease diagnosis are not about data scarcity. They are about healthcare’s structural bottlenecks—fragmented data silos, inconsistent protocols, and missing consent infrastructure—that make reliable AI impossible at scale. Data scarcity is a symptom. The root cause is the system design underneath.</p>
<p>In 2023, Eka Care introduced explicit patient consent flows before any health data was accessed for AI training. This slowed data acquisition but ensured legal standing and clinical trust. The lesson is clear: you cannot fix a governance problem by throwing more data at it.</p>
<section id="the-structural-bottleneck-framework" class="level2">
<h2 class="anchored" data-anchor-id="the-structural-bottleneck-framework">The Structural Bottleneck Framework</h2>
<p>I call this the <strong>Structural Bottleneck Framework</strong>: AI performance in rare diseases is limited not by model size or dataset volume, but by systemic healthcare design flaws. Fragmented data, inconsistent clinical protocols, and privacy roadblocks produce an environment where AI trained on generic or legacy datasets will fail at point-of-care deployment.</p>
<p>Most AI healthcare teams obsess over model selection, fine-tuning, and benchmark chasing while neglecting data governance architecture, consent infrastructure, AI validation layers, and domain protocol alignment. That’s why rare disease AI remains a demo that never makes it into clinics.</p>
<p>Fixing data quantity without fixing data governance is like adding fuel to a car with no steering wheel.</p>
</section>
<section id="why-more-data-doesnt-solve-the-problem" class="level2">
<h2 class="anchored" data-anchor-id="why-more-data-doesnt-solve-the-problem">Why More Data Doesn’t Solve the Problem</h2>
<p>Healthcare data is siloed by provider, geography, and regulation. No amount of model tuning overcomes that fragmentation.</p>
<p>Imagine a sensor network with noisy, inconsistent, and incomplete signals. The output will be unreliable regardless of how sophisticated the algorithms are. This is not a metaphor. It is literally how AI input pipelines behave when data sources are fragmented and unverified.</p>
<p>In 2022, an AI system deployed for pediatric rare disease diagnosis nearly caused a malpractice incident by mislabeling a critical symptom. The model had been trained on adult datasets with different clinical presentations. This failure was structural, not statistical.</p>
<p>Generic datasets compound the problem. Retrieval-augmented generation (RAG) approaches surface obsolete or irrelevant medical guidelines when the knowledge base is not actively maintained and aligned with current clinical protocols. Fine-tuning on scarce rare disease data is insufficient if the underlying data ecosystem doesn’t support real-time, trustworthy updates. A model fine-tuned in 2022 will give outdated guidance in 2025. Training cycles cannot keep pace without structural integration into clinical protocol update chains.</p>
<p>The ethical dimension is not a compliance checkbox. AI deployed without patient consent frameworks creates legal risk and erodes clinical trust. Once a clinician sees an AI system give a dangerous recommendation, that system is dead in that institution regardless of subsequent accuracy gains. Rebuilding clinical trust after a structural failure is harder than building it correctly the first time.</p>
<p>Falsifiable claim: AI models trained with incremental data additions but without systemic integration of domain-specific, privacy-aware data governance will continue producing dangerous misclassifications at rates preventing clinical adoption. The structural bottleneck, not data volume, is the binding constraint.</p>
</section>
<section id="concrete-evidence-from-india-and-beyond" class="level2">
<h2 class="anchored" data-anchor-id="concrete-evidence-from-india-and-beyond">Concrete Evidence From India and Beyond</h2>
<p>Eka Care’s 2023 shift to consent-driven data acquisition is the clearest example of getting the structural layer right. Patient consent protocols slowed data access but ensured the data used for AI training had legal standing and patient trust behind it. This is not a formality. It is what makes AI deployable in clinics rather than research labs.</p>
<p>Multiple Indian healthcare startups have deployed AI that misread critical symptoms as banal conditions because their models trained on generic datasets lacked rare disease-specific clinical annotation. One AI misclassified a rare autoimmune condition as a common allergy, simply because pattern matching aligned with far more frequent conditions in the training set. This is not a data volume problem. It is a structural failure to align the model with clinical taxonomy for the target patient population.</p>
<p>Telemedicine adoption in rural India illustrates the same bottleneck differently. 5G coverage and smartphones exist. The structural barrier to AI-assisted diagnosis is not data volume. It is the absence of validated clinical protocols for AI decision support in resource-constrained settings, liability frameworks clinicians and patients understand, and feedback mechanisms that let clinicians flag AI errors in real time.</p>
<p>At Ostronaut, building AI-generated healthcare training content revealed the same pattern at scale. Generating clinical learning material required more than ingesting large content volumes. We needed validation layers: domain experts reviewing AI output against current clinical guidelines, quality gates flagging outdated protocols, and structured feedback loops improving generation accuracy over time. More data ingestion without these structural layers yields more plausible but incorrect content. Volume does not substitute for architecture.</p>
</section>
<section id="what-the-fix-looks-like" class="level2">
<h2 class="anchored" data-anchor-id="what-the-fix-looks-like">What the Fix Looks Like</h2>
<p>The Structural Bottleneck Framework points to a different investment thesis for rare disease AI.</p>
<table class="caption-top table">
<colgroup>
<col style="width: 43%">
<col style="width: 56%">
</colgroup>
<thead>
<tr class="header">
<th>Traditional AI Effort</th>
<th>Structural Bottleneck Focus</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Model tuning and benchmarks</td>
<td>Consent and data governance infrastructure</td>
</tr>
<tr class="even">
<td>Dataset volume and augmentation</td>
<td>Clinical protocol alignment and validation layers</td>
</tr>
<tr class="odd">
<td>Statistical fine-tuning</td>
<td>Real-time domain updates and feedback mechanisms</td>
</tr>
<tr class="even">
<td>Isolated AI pipelines</td>
<td>Integrated healthcare system workflows</td>
</tr>
</tbody>
</table>
<p>The fix starts with consent and governance. Patient consent must be explicit, auditable, and embedded in data pipelines. Data governance can’t be an afterthought or legal checkbox. It must be engineered as infrastructure.</p>
<p>Second, AI validation layers must become standard. Domain experts need to build continuous quality gates and feedback loops. AI outputs require real-world clinical protocol integration, not just offline benchmarks.</p>
<p>Third, clinical protocols must be actively maintained and integrated with AI knowledge bases. Rare disease protocols evolve. The model’s training cycle must be tightly coupled with these updates, or risk obsolescence.</p>
<p>Finally, liability and trust frameworks need clarity. Clinicians must know when and how AI can be used safely, and have mechanisms to flag and correct errors in real time.</p>
<p>At Ostronaut, we learned this the hard way. AI-generated clinical content without validation layers isn’t just wrong; it erodes trust in the entire system. The data volume was never the problem.</p>
</section>
<section id="what-i-dont-know-yet" class="level2">
<h2 class="anchored" data-anchor-id="what-i-dont-know-yet">What I Don’t Know Yet</h2>
<p>How do you build scalable, privacy-aware consent infrastructure that works across fragmented healthcare providers and jurisdictions — without killing innovation speed? It’s an unsolved technical and regulatory puzzle.</p>
<p>How do you design AI validation layers that keep pace with rapidly evolving clinical protocols in rare diseases, given the scarcity of domain experts? Automation helps, but domain knowledge bottlenecks remain.</p>
<p>How do we create feedback mechanisms that incentivize clinicians to report AI errors and integrate those corrections back into the training loop — especially in resource-constrained settings?</p>
<p>These are open engineering and policy questions, not hype fodder.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The Question Worth Asking</h2>
<p>The Structural Bottleneck Framework shifts focus from data quantity to system quality. The question worth asking now is: can AI companies and healthcare institutions collaborate on building structural data governance and validation infrastructure at scale — or will rare disease AI remain a demo for another decade?</p>
<p>Not in three years. In ten. In fifty.</p>
<p>Are we asking it? Mostly, no.</p>
<p>More on this as I develop it. ``````````````````` :::</p>
</section>

 ]]></description>
  <category>AI in Healthcare</category>
  <category>AI Validation</category>
  <category>India Tech</category>
  <guid>https://talvinder.com/build-logs/training-ai-to-serve-rare-disease-patients-is-structural/</guid>
  <pubDate>Fri, 17 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>The Vibe Coding Hangover: Why AI-Written Code Costs 4x to Maintain by Year Two</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/vibe-coding-hangover/</link>
  <description><![CDATA[ 





<p>According to a CodeRabbit analysis of 1,000+ repositories, AI co-authored code introduces 1.7x more major issues than human-written code. The vulnerability rate is 2.74x higher. GitHub’s 2025 Octoverse data shows Copilot now generates 46% of code in files where it’s enabled. And a METR study found that experienced developers using AI assistants were actually 19% slower on real tasks — despite believing they were 24% faster.</p>
<p>The productivity feels real. The debt is real too. We’re starting to see the bill.</p>
<section id="the-three-month-cliff" class="level2">
<h2 class="anchored" data-anchor-id="the-three-month-cliff">The three-month cliff</h2>
<p>Every team I’ve talked to that adopted AI coding tools heavily describes the same pattern: massive output gains in months one through three, followed by an escalating maintenance burden that erases those gains by month six.</p>
<p>The pattern has a name now. Developers are calling it the “Spaghetti Point” — the moment where the codebase generated by AI assistants becomes harder to modify than code written from scratch would have been.</p>
<p>According to GitClear’s 2025 developer productivity report, code churn (lines modified or deleted within 14 days of being written) increased 39% in repositories with heavy AI assistance. That’s not refactoring — that’s rework. Code written fast, reviewed inadequately, and fixed repeatedly.</p>
<p>The economics are brutal. A 2025 analysis by Uplevel estimated that AI-generated code carries maintenance costs 4x higher than human-written code by year two. The initial velocity gain — real, measurable, impressive — gets consumed by debugging sessions where no one can explain why the code works the way it does, because the “why” never existed. This is the same <a href="../../field-notes/github-slopocalypse-trust-tax/">epistemological problem</a> that’s eroding trust in open source: AI-generated code has no intent. You can’t reconstruct reasoning that never happened.</p>
</section>
<section id="why-the-bugs-are-different" class="level2">
<h2 class="anchored" data-anchor-id="why-the-bugs-are-different">Why the bugs are different</h2>
<p>AI-generated bugs are structurally different from human bugs, and that difference makes them more expensive to find and fix.</p>
<p><strong>Human bugs have intent trails.</strong> A developer who writes a race condition usually has a mental model that’s almost right — they thought about concurrency but missed one case. You can read the code, reconstruct the thinking, find the gap. The fix follows from understanding the original intent.</p>
<p><strong>AI bugs have no intent.</strong> The code was generated from a probability distribution, not a mental model. When a Copilot-generated function has a subtle type coercion error, there’s no reasoning to reconstruct. You can’t ask “what were they thinking?” because nothing was thinking. You have to understand the code from scratch, as if reading a stranger’s work with no comments and no commit history that explains decisions.</p>
<p>According to Snyk’s 2025 AI security report, 35 new CVEs were attributed to AI-generated code in March 2026 alone. Repositories using Copilot leak 40% more secrets (API keys, credentials, tokens) than non-Copilot repositories. The AI doesn’t understand what’s secret — it patterns matches from training data that included leaked credentials.</p>
<table class="caption-top table">
<colgroup>
<col style="width: 21%">
<col style="width: 41%">
<col style="width: 36%">
</colgroup>
<thead>
<tr class="header">
<th>Bug type</th>
<th>Human-written code</th>
<th>AI-written code</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Root cause analysis</td>
<td>Follow the intent trail</td>
<td>Start from zero — no intent exists</td>
</tr>
<tr class="even">
<td>Time to diagnose</td>
<td>1-2 hours typical</td>
<td>3-5 hours (no reasoning to reconstruct)</td>
</tr>
<tr class="odd">
<td>Recurrence after fix</td>
<td>Low (developer updates mental model)</td>
<td>High (same prompt generates same pattern)</td>
</tr>
<tr class="even">
<td>Security issues per KLOC</td>
<td>Baseline</td>
<td>2.74x higher (CodeRabbit data)</td>
</tr>
<tr class="odd">
<td>Code churn within 14 days</td>
<td>Baseline</td>
<td>+39% (GitClear data)</td>
</tr>
</tbody>
</table>
</section>
<section id="the-organizational-blind-spot" class="level2">
<h2 class="anchored" data-anchor-id="the-organizational-blind-spot">The organizational blind spot</h2>
<p>The real damage isn’t technical — it’s organizational. Teams measuring developer productivity by lines of code or PRs merged are seeing their best numbers ever. The dashboards look great. Velocity is up. Sprint commitments are being met.</p>
<p>What the dashboards don’t show: time spent in code review has increased 45% (because reviewers now treat every PR as potentially AI-generated and requiring deeper verification). Bug reports from production are up 30% despite passing all automated tests. And senior engineers are spending more time reading and understanding code than writing it — the exact inverse of what AI tools were supposed to enable.</p>
<p>This is <a href="../../build-logs/ai-speed-lie-team-velocity/">the same productivity illusion</a> we measured in team velocity: AI makes individual tasks faster while making the overall system slower. The local optimization creates a global pessimization.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I got wrong</h2>
<p>I initially thought the problem was adoption immaturity — that teams would learn to use AI tools effectively and the quality issues would resolve. After watching a dozen teams go through the cycle over the past year, I think the problem is structural.</p>
<p>AI code generation optimizes for plausibility, not correctness. The output looks right, passes superficial review, and often works for the happy path. The failures are in edge cases, error handling, security boundaries, and long-term maintainability — exactly the things that junior developers also get wrong, because those are the things that require understanding, not pattern matching.</p>
<p>The teams that are succeeding with AI code generation share three practices:</p>
<p><strong>1. AI writes, humans architect.</strong> The AI generates implementation within a structure that a human designed. The human defines the interfaces, the error handling strategy, the security boundaries. The AI fills in the bodies. This preserves intent at the architectural level while leveraging AI speed at the implementation level.</p>
<p><strong>2. Review budgets increased, not decreased.</strong> Teams that cut code review time because “the AI wrote it” are the ones hitting the Spaghetti Point fastest. The teams that survive allocate more review time — not less — because the verification burden is higher for machine-generated code.</p>
<p><strong>3. Aggressive deletion of AI-generated code that can’t be explained.</strong> If a developer can’t explain why a function works the way it does — regardless of whether it passes tests — it gets rewritten by hand. This is expensive in the short term and cheap in the long term.</p>
</section>
<section id="the-historical-pattern" class="level2">
<h2 class="anchored" data-anchor-id="the-historical-pattern">The historical pattern</h2>
<p>This cycle is familiar. Every productivity tool that dramatically increases output velocity eventually forces a reckoning with quality.</p>
<p>3D printing was going to democratize manufacturing. It did — and it also created a mountain of low-quality plastic objects that nobody needed. The lasting value came from professionals using 3D printing within disciplined design processes, not from everyone printing everything.</p>
<p>No-code tools were going to replace developers. They did increase output — and they also created a generation of applications that couldn’t scale, couldn’t be debugged, and couldn’t be maintained when the original builder left. The lasting value came from no-code as a prototyping tool, not a production platform.</p>
<p>Vibe coding is following the same arc. The output explosion is real. The quality reckoning is coming. The lasting value will come from AI as an implementation accelerator within disciplined engineering practices — not from AI as a replacement for engineering judgment.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The question worth asking</h2>
<p>If your team adopted AI coding tools in the last twelve months, run this check: compare the bug rate and code churn rate in your most AI-assisted repositories against your least AI-assisted ones. Normalize for team size and feature complexity.</p>
<p>If the AI-heavy repos show higher churn and more production bugs — even if they also show higher velocity — you’re accumulating the debt. The hangover is coming. The question is whether you pay it down deliberately (with review discipline, architectural boundaries, and aggressive deletion) or discover it when the codebase becomes unmaintainable.</p>
<p>The <a href="../../field-notes/github-slopocalypse-trust-tax/">trust tax</a> isn’t just an open-source problem. It’s inside your organization too.</p>
<div class="schema-faq" style="display:none;">
<p>[{“q”:“Does AI-generated code have more bugs than human code?”,“a”:“Yes. According to CodeRabbit’s analysis of 1,000+ repositories, AI co-authored code has 1.7x more major issues and a 2.74x higher vulnerability rate. GitClear found code churn (rework within 14 days) increased 39% in repositories with heavy AI assistance.”},{“q”:“What is vibe coding and what are the risks?”,“a”:“Vibe coding is using AI tools like Copilot or ChatGPT to generate code by describing what you want in natural language. The risk is maintenance debt: code generated without human intent is harder to debug, carries 2.74x more vulnerabilities, and costs an estimated 4x more to maintain by year two.”},{“q”:“Are developers actually faster with AI coding tools?”,“a”:“Not necessarily. A METR study found experienced developers were 19% slower on real tasks with AI assistants, despite believing they were 24% faster. Local task speed increases, but time spent in review, debugging, and understanding AI-generated code offsets the gains at the team level.”}]</p>
</div>


<!-- -->

</section>

 ]]></description>
  <category>Software Engineering</category>
  <category>Agentic Systems</category>
  <category>Production AI</category>
  <guid>https://talvinder.com/build-logs/vibe-coding-hangover/</guid>
  <pubDate>Fri, 03 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>The Recourse Trap: Why Competition Makes Credit Scoring More Exclusive, Not Less</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/actionable-recourse-markets/</link>
  <description><![CDATA[ 





<p>In 2022, HDFC Bank raised its minimum CIBIL score requirement for personal loans from 650 to 725. ICICI and Axis followed within months. That same year, TransUnion CIBIL’s own data showed that first-time borrowers with scores between 650 and 725 had default rates under 4%. The banks weren’t responding to rising risk. They were responding to each other.</p>
<p>Credit scoring systems don’t fail because they’re inaccurate. They fail because accuracy isn’t the job in a competitive lending market.</p>
<p>The job is risk transfer. In competitive environments, the most efficient way to transfer risk is to exclude entire populations rather than solve information problems.</p>
<p>I’ve seen this pattern up close. At Pragmatic Leaders, I’ve trained credit risk teams at HDFC, ICICI, and four mid-tier Indian banks. The pattern is consistent: everyone knows traditional credit scoring excludes viable borrowers. No one builds the alternative system because competitive pressure rewards portfolio metrics over market expansion.</p>
<section id="the-recourse-trap" class="level2">
<h2 class="anchored" data-anchor-id="the-recourse-trap">The Recourse Trap</h2>
<p>This is what I’m calling <strong>The Recourse Trap</strong>: a system where the mechanism designed to enable access becomes the mechanism that prevents it, and competitive pressure makes the trap stronger, not weaker.</p>
<p>Here’s how it works:</p>
<p>A lender can’t distinguish between a borrower with no credit history and a borrower with bad credit history. Both score low. In a competitive market, the lender who extends credit to both will have worse portfolio performance than the lender who extends credit to neither. The rational competitive response is exclusion.</p>
<p>The borrower has no recourse. They can’t “improve their score” because they can’t access credit to build history. The system tells them what to do (build credit history) while preventing them from doing it.</p>
<p>India has 400 million adults with no credit history in any bureau. Not because they’re risky. Because the system has no mechanism to evaluate them, and no competitive incentive to build one.</p>
</section>
<section id="the-mechanism" class="level2">
<h2 class="anchored" data-anchor-id="the-mechanism">The Mechanism</h2>
<p>When lenders compete on portfolio risk metrics, they optimize for false negative reduction (don’t lend to bad borrowers) over false positive reduction (do lend to good borrowers). The asymmetry exists because the cost of a bad loan is immediate and visible, while the cost of a missed good loan is distributed across the market and invisible.</p>
<p>This creates a lemons problem. Borrowers without traditional credit history get pooled with genuinely risky borrowers. Lenders can’t tell them apart without incurring verification costs that competitive pressure makes prohibitive. The result: high-quality borrowers with no credit history get priced out or excluded entirely.</p>
<p><strong>Falsifiable claim</strong>: In competitive lending markets, credit score requirements will trend upward over time for populations without traditional credit history, even as default rates in those populations remain stable or decline. The system optimizes for competitive position, not credit risk.</p>
<p>You can test this. Look at minimum credit score requirements for first-time borrowers in India between 2018 and 2024. Requirements went up across every major bank. Did actual default rates for first-time borrowers go up proportionally? No.&nbsp;RBI data shows gross NPA ratios for retail loans actually declined from 2.5% to 1.7% in that period. The market tightened because competitors tightened, not because risk increased.</p>
</section>
<section id="the-transaction-cost-argument-is-circular" class="level2">
<h2 class="anchored" data-anchor-id="the-transaction-cost-argument-is-circular">The Transaction Cost Argument Is Circular</h2>
<p>Here’s the tell: when you ask banks why they don’t serve underbanked populations, they talk about credit scores. When you ask why they don’t build alternative scoring systems, they talk about transaction costs. When you ask why transaction costs are prohibitive for underserved populations but not for premium segments, the conversation ends.</p>
<p>High costs justify exclusion. Exclusion prevents scale. Lack of scale keeps costs high.</p>
<p>South African banks demonstrate this clearly. Despite strong demand for credit from low-income households, banks haven’t extended access. Not because these households are uniformly risky, but because the information required to assess risk isn’t available in formats traditional scoring systems can process.</p>
<p>The alternative mechanisms prove the problem is solvable. Group lending models and informal systems like stokvels work precisely because they solve the <a href="../../frameworks/comp-negotiation-entropy/">information problem</a> differently. They use peer monitoring, social ties, and collective savings as signals. Transaction costs stay low. Default rates stay manageable.</p>
<p>But competitive banks don’t adopt these approaches. They require different infrastructure, different risk models, and different competitive positioning. A bank that moves first takes on <a href="../../frameworks/entropy-and-entrepreneurship/">execution risk</a>. A bank that moves second can copy what works. The rational move is to wait, which means no one moves.</p>
</section>
<section id="what-ai-makes-worse" class="level2">
<h2 class="anchored" data-anchor-id="what-ai-makes-worse">What AI Makes Worse</h2>
<p><a href="../../frameworks/ai-deflation-trap/">AI-powered credit scoring is getting more sophisticated at predicting risk within existing data distributions</a>. Which means more sophisticated at excluding populations outside those distributions.</p>
<p>An AI model trained on historical lending data will learn that borrowers without credit history are risky. Not because they default more often, but because lenders historically avoided them. The model encodes the market’s collective risk aversion as ground truth.</p>
<p>I saw this firsthand during a workshop with a mid-tier bank’s risk team in 2023. They’d built a gradient-boosted model on five years of loan performance data. The model performed well on their test set (AUC of 0.87). But when they scored a sample of new-to-credit applicants, 92% were classified as high risk. The data scientist on the team knew the scores were wrong. His manager knew. But nobody was going to approve a lending policy that scored worse than the competitor down the street.</p>
<p>The feedback loop tightens. Better prediction within the existing distribution means worse outcomes for populations outside it.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I Got Wrong</h2>
<p>I initially thought the solution was better data. If we could capture alternative signals like UPI transaction history, utility payments, or rental records, we could build scoring systems that include underserved populations.</p>
<p>That’s technically true but structurally naive.</p>
<p>The problem isn’t data availability. India Account Aggregator has been live since 2021. Perfios and FinBox can pull 12 months of UPI transaction data in seconds. The pipes exist. Banks still don’t use them for first-time borrowers at any meaningful scale because competitive incentive hasn’t shifted.</p>
<p>A bank that invests in alternative data infrastructure takes on execution risk and regulatory uncertainty. A bank that waits can copy the approach if it works. The first-mover disadvantage is real.</p>
</section>
<section id="beyond-credit-scoring" class="level2">
<h2 class="anchored" data-anchor-id="beyond-credit-scoring">Beyond Credit Scoring</h2>
<p>The recourse trap exists because competitive markets optimize for relative performance, not absolute outcomes. A lender doesn’t need to solve the information problem if their competitors don’t solve it either.</p>
<p>This has implications beyond financial services. Any system that provides “actionable recourse” in a competitive environment faces the same dynamic. The advice the system gives (build credit history, gain relevant experience, develop measurable skills) is only actionable if the system allows you to act on it.</p>
<p>When it doesn’t, you’re not dealing with an information problem. You’re dealing with a market structure problem.</p>
<p>AI-powered resume screening. Skills-based hiring platforms. Fraud detection systems. They all create versions of the recourse trap when deployed in competitive markets. The mechanism is the same: optimize for false negative reduction, accept false positive costs, and let competitive pressure prevent anyone from solving the underlying information problem.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The Question Worth Asking</h2>
<p>What other systems are we building that look like they enable access but actually optimize for exclusion?</p>
<p>If the mechanism for proving you’re trustworthy requires access you can’t get without already being trusted, you’re in a recourse trap. If competitive pressure makes solving that problem more expensive than ignoring it, the trap becomes structural.</p>
<p>Are we asking this question when we deploy AI systems in hiring, lending, insurance, education? Mostly, no. We’re still arguing about bias metrics and fairness definitions while the competitive dynamics that drive exclusion go unexamined.</p>
<p>The recourse trap doesn’t care about bias. It cares about competitive dynamics. And those dynamics are getting stronger as AI makes within-distribution optimization cheaper and more effective.</p>


</section>

 ]]></description>
  <category>Financial Systems</category>
  <category>Market Structure</category>
  <category>AI Risk</category>
  <category>India Tech</category>
  <guid>https://talvinder.com/field-notes/actionable-recourse-markets/</guid>
  <pubDate>Fri, 03 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Why 86% of AI Agent Pilots Fail Before Reaching Production</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/why-agent-pilots-fail/</link>
  <description><![CDATA[ 





<p>According to the MAST benchmark study, multi-agent system failure rates range from 41% to 86.7% across seven leading frameworks. Gartner projects that 40% of agentic AI projects started in 2025 will be scaled back or canceled by 2027. McKinsey’s 2025 survey found that while 78% of enterprises have AI agent pilots running, only 14% have reached production deployment.</p>
<p>These numbers tell the same story: the demo works, but production kills it.</p>
<section id="the-failure-isnt-the-model-its-everything-around-the-model" class="level2">
<h2 class="anchored" data-anchor-id="the-failure-isnt-the-model-its-everything-around-the-model">The failure isn’t the model — it’s everything around the model</h2>
<p>The top three causes of agent pilot failure are integration complexity (67%), lack of monitoring (58%), and unclear escalation paths (52%), according to PwC’s 2025 enterprise AI survey. The model itself is rarely the problem. According to a 2025 PwC survey of 1,000 enterprises deploying AI agents, the top three failure modes are integration complexity (cited by 67%), lack of monitoring infrastructure (58%), and unclear escalation paths when the agent makes mistakes (52%).</p>
<p>The model itself is rarely the problem. GPT-4o, Claude, Gemini — they all perform well enough in controlled conditions. The collapse happens when the agent hits production reality: messy data, concurrent users, edge cases the prompt didn’t anticipate, and no one watching when confidence drops below threshold.</p>
<p>This is the same <a href="../../field-notes/indian-saas-agent-reliability/">reliability gap</a> that Indian SaaS companies have been closing for twenty years — not with better models, but with better systems around the models.</p>
</section>
<section id="five-patterns-that-kill-agent-pilots" class="level2">
<h2 class="anchored" data-anchor-id="five-patterns-that-kill-agent-pilots">Five patterns that kill agent pilots</h2>
<p>These are the five structural failures I’ve seen repeatedly across teams deploying agents — from startups to Fortune 500 companies. Each one is fixable, but only if you build for it before production, not after.</p>
<section id="no-confidence-scoring-or-graceful-degradation" class="level3">
<h3 class="anchored" data-anchor-id="no-confidence-scoring-or-graceful-degradation">1. No confidence scoring or graceful degradation</h3>
<p>Agents without confidence thresholds have 3x higher escalation rates than those with calibrated routing, according to Anthropic’s production deployment data. The agent either answers or it doesn’t. There’s no middle ground. In production, the middle ground is where most interactions live — the agent is 60% confident, the user’s query is ambiguous, the data is incomplete.</p>
<p>Without confidence scoring, you get one of two failure modes: the agent hallucinates confidently (and you lose trust) or the agent refuses to answer (and you lose utility). According to Anthropic’s production deployment guide, agents without confidence thresholds have 3x higher escalation rates than those with calibrated confidence routing.</p>
<p>The fix is graduated autonomy: act autonomously above 90% confidence, request human review between 60-90%, escalate below 60%. This is the same pattern <a href="../../build-logs/agentic-rightsizing/">we built at Zopdev</a> for infrastructure decisions — observe everything, act only within permission boundaries.</p>
</section>
<section id="the-just-retry-fallacy" class="level3">
<h3 class="anchored" data-anchor-id="the-just-retry-fallacy">2. The “just retry” fallacy</h3>
<p>When an agent fails, most frameworks default to retrying with the same prompt. This is the <a href="../../field-notes/consensus-is-not-verification/">Pass@k trap</a>: if the error is structural (wrong data, missing context, ambiguous instruction), retrying amplifies the problem rather than fixing it.</p>
<p>A 2025 analysis of production agent logs at a Fortune 500 company found that 73% of retried requests produced the same error category. The retry wasn’t recovery — it was waste. At $0.03 per inference call, a three-retry loop on every failed request added $180K/year to their agent infrastructure bill.</p>
<p>The fix is error classification before retry. Network timeout? Retry. Model hallucination? Route to a different model or escalate. Missing context? Fetch the context first, then retry with enriched input.</p>
</section>
<section id="no-observability-beyond-the-api-call" class="level3">
<h3 class="anchored" data-anchor-id="no-observability-beyond-the-api-call">3. No observability beyond the API call</h3>
<p>Most agent monitoring stops at the API layer: latency, token count, error rate. But agent failures are semantic, not mechanical. The API returns a 200 with a confident, well-formatted, completely wrong answer.</p>
<p>According to Langfuse’s 2025 observability report, teams that implement trace-level monitoring (tracking the full chain of agent reasoning, tool calls, and intermediate outputs) catch production issues 4x faster than teams monitoring only API metrics. This is what <a href="../../frameworks/trace-based-assurance-agentware/">trace-based assurance</a> looks like in practice — the governance layer that agentware actually needs.</p>
</section>
<section id="human-handoff-as-afterthought" class="level3">
<h3 class="anchored" data-anchor-id="human-handoff-as-afterthought">4. Human handoff as afterthought</h3>
<p>The agent is built to be autonomous. When it can’t handle something, it says “I don’t know” — and the user is stuck. There’s no warm handoff to a human, no context transfer, no continuity.</p>
<p>According to Freshworks’ deployment data, their Freddy AI achieves a 45% autonomous resolution rate. The other 55% gets escalated — and the quality of that escalation (context preserved, human gets the full conversation history, seamless transition) is what determines customer satisfaction. The agent’s job isn’t just to resolve; it’s to escalate well when it can’t.</p>
<p>The cost of building good escalation paths is significant. A production agent needs roughly 3.5 FTEs for monitoring, incident response, and drift detection. <a href="../../field-notes/indian-saas-agent-reliability/">In Bangalore, that’s $100K-150K/year</a>. In San Francisco, $600K-800K. This cost asymmetry is why Indian SaaS companies can afford the monitoring density that makes agents reliable.</p>
</section>
<section id="evaluation-that-doesnt-match-production-conditions" class="level3">
<h3 class="anchored" data-anchor-id="evaluation-that-doesnt-match-production-conditions">5. Evaluation that doesn’t match production conditions</h3>
<p>The agent scores 92% on the benchmark. In production, users ask questions the benchmark didn’t anticipate, in formats the prompt didn’t expect, with context the training data never included. The <a href="../../build-logs/llm-judge-india-failure/">evaluation cost ratio</a> breaks down when evaluation doesn’t mirror production conditions.</p>
<p>According to the HELM benchmark team at Stanford, model performance on curated test sets overpredicts production accuracy by 15-30 percentage points. The gap is not random — it’s systematic. Production queries are longer, more ambiguous, more dependent on context, and more adversarial than benchmark queries.</p>
</section>
</section>
<section id="what-actually-works-the-three-layer-architecture" class="level2">
<h2 class="anchored" data-anchor-id="what-actually-works-the-three-layer-architecture">What actually works: the three-layer architecture</h2>
<p>Successful agent deployments converge on a generation-validation-governance stack. The generation layer is what everyone builds; the other two are what separates pilots from production. Every successful deployment I’ve seen — Freshworks’ Freddy, Zoho’s Zia, our own systems — converges on the same architecture:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 20%">
<col style="width: 29%">
<col style="width: 50%">
</colgroup>
<thead>
<tr class="header">
<th>Layer</th>
<th>Function</th>
<th>What it catches</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td><strong>Generation</strong></td>
<td>The model produces output</td>
<td>Nothing — this is the happy path</td>
</tr>
<tr class="even">
<td><strong>Validation</strong></td>
<td>Rule-based checks, confidence scoring, format verification</td>
<td>Structural errors, low-confidence outputs, format violations</td>
</tr>
<tr class="odd">
<td><strong>Governance</strong></td>
<td>Human review queues, audit trails, escalation paths, drift detection</td>
<td>Semantic errors, edge cases, model drift, compliance violations</td>
</tr>
</tbody>
</table>
<p>The generation layer is what everyone builds. The validation layer is what separates pilots from production. The governance layer is what separates production from enterprise-grade.</p>
<p>Most pilots only build layer one. They fail because layers two and three are where production reliability actually lives.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The question worth asking</h2>
<p>If you’re running an agent pilot right now, ask this: what happens when the agent is wrong and confident about it?</p>
<p>If the answer is “we haven’t thought about that” — you’re in the 86%. The <a href="../../field-notes/consensus-is-not-verification/">consensus voting approach won’t save you</a>. The <a href="../../field-notes/cot-efficiency-tax/">chain-of-thought reasoning adds cost without guaranteeing correctness</a>. The model isn’t the problem.</p>
<p>The monitoring, the fallbacks, the escalation paths, the confidence routing — that’s where production reliability lives. The teams that figure this out aren’t building better agents. They’re building better <a href="../../frameworks/agent-context-is-infrastructure/">infrastructure around agents</a>. And right now, the companies with the deepest operational discipline in that infrastructure layer are <a href="../../field-notes/indian-saas-agent-reliability/">based in India</a>.</p>
<div class="schema-faq" style="display:none;">
<p>[{“q”:“Why do AI agent pilots fail in production?”,“a”:“According to MAST benchmark data, 41-86.7% of multi-agent systems fail across leading frameworks. The top causes are integration complexity (67%), lack of monitoring (58%), and unclear escalation paths (52%). The model works in demos — the failure is in monitoring, fallbacks, confidence scoring, and human handoff systems.”},{“q”:“What percentage of AI agent projects reach production?”,“a”:“Only 14% of enterprise AI agent pilots reach production deployment, according to McKinsey’s 2025 survey. Gartner projects 40% of agentic AI projects started in 2025 will be scaled back or canceled by 2027.”},{“q”:“How do you deploy AI agents to production successfully?”,“a”:“Successful deployments use a three-layer architecture: generation (the model), validation (confidence scoring, format checks, rule-based verification), and governance (human review queues, audit trails, escalation paths). Most failed pilots only build the generation layer.”}]</p>
</div>


</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>Production AI</category>
  <category>Infrastructure</category>
  <guid>https://talvinder.com/field-notes/why-agent-pilots-fail/</guid>
  <pubDate>Fri, 03 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>I Built Ed-Tech Before Ed-Tech Existed in India</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/edtech-before-edtech/</link>
  <description><![CDATA[ 





<p>In 2018, I started Pragmatic Leaders to teach product management in India. The category didn’t exist yet. Most companies were hiring for “marketing,” “sales,” or “operations” — PM was a Silicon Valley thing. I had 21 paying students across 3 countries and no funding. By the time Unacademy and BYJU’S were raising billions, we’d trained thousands and generated ₹4+ crores in salary hikes for students.</p>
<p>The insight: building before the market exists forces you to validate pedagogy instead of growth. That constraint became our advantage.</p>
<section id="market-before-product-vs.-product-before-market" class="level2">
<h2 class="anchored" data-anchor-id="market-before-product-vs.-product-before-market">Market-Before-Product vs.&nbsp;Product-Before-Market</h2>
<p>Most ed-tech companies in India built for a market that already existed. BYJU’S entered K-12 test prep — a ₹40,000 crore market. Unacademy entered competitive exam coaching — already massive. They optimized for distribution and unit economics in proven categories.</p>
<p>We built for a market that didn’t exist. Product management education in India in 2018 was not a category. There was no TAM to cite, no comparable to benchmark against, no playbook to copy.</p>
<p>When you build before the market, you can’t fake it. You can’t raise $50M and buy your way to product-market fit. You have to actually solve the problem first.</p>
</section>
<section id="why-bootstrapping-forces-better-pedagogy" class="level2">
<h2 class="anchored" data-anchor-id="why-bootstrapping-forces-better-pedagogy">Why bootstrapping forces better pedagogy</h2>
<p><strong>If you bootstrap an ed-tech company in an unproven category, you will build better pedagogy than if you raise capital in a proven category.</strong></p>
<p>Capital in a proven market optimizes for scale. You know the category works — the question is execution. Can you acquire cheaper? Convert faster? Retain longer? The pedagogy becomes a variable to optimize, not the foundation to validate.</p>
<p>Capital in an unproven market is a trap. You’ll spend it trying to create demand instead of validating that you can actually teach the thing. You’ll hire a sales team before you know if the course works. You’ll scale a mediocre product into a bigger mediocre product.</p>
<p>I couldn’t do that. I had 21 paying students and no investors. The only way to grow was if those 21 students actually learned product management and got better jobs. If the pedagogy didn’t work, I had no business.</p>
<p>So I built the pedagogy first.</p>
</section>
<section id="the-validation" class="level2">
<h2 class="anchored" data-anchor-id="the-validation">The validation</h2>
<p>I worked alone for the first year. Customized an LMS to deliver the course and gamify the learning. Watched every student’s progress. Saw where they got stuck. Saw what clicked.</p>
<p>The metric wasn’t revenue. It wasn’t NPS. It was: did they get the job?</p>
<p>Out of those first 21 students, 18 transitioned into PM roles or got promoted. Salary hikes ranged from ₹3L to ₹12L. That’s a 94% success rate on a sample size small enough to actually track.</p>
<p>That’s when I knew the pedagogy worked.</p>
</section>
<section id="the-technical-shift" class="level2">
<h2 class="anchored" data-anchor-id="the-technical-shift">The technical shift</h2>
<p>By 2019, I had a problem: I could teach 21 students well. I could probably teach 100 students well. But could I teach 10,000 students well?</p>
<p>The standard ed-tech answer is: record the lectures, sell access, scale horizontally. That’s not teaching. That’s distribution.</p>
<p>I made a different bet. I decided to build the platform and algorithms that could use the data we had from students. Individualized learning — not as a marketing term, but as an actual technical architecture.</p>
<p>Here’s what that meant in practice:</p>
<ul>
<li>Track where each student struggled in the curriculum</li>
<li>Identify patterns across cohorts (e.g., “students from non-tech backgrounds struggle with API design”)</li>
<li>Generate personalized problem sets based on performance</li>
<li>Adapt pacing based on engagement and comprehension signals</li>
</ul>
<p>This wasn’t LLM-powered. This was 2019. We built rule-based systems and basic ML models. But the principle was right: use data to make the course adapt to the student, not force the student to adapt to the course.</p>
<p>By 2020, we had 130 students in upfront-fee courses and 30 in ISA-based courses. We were adding 1.3 students daily — slow by VC standards, sustainable by pedagogy standards.</p>
<p>Cumulative salary hikes: ₹4.2 crores. Hours of training delivered: 30,000+.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I got wrong</h2>
<p>I thought the hard part was building the pedagogy. It wasn’t. The hard part was explaining why our pedagogy was different.</p>
<p>Every ed-tech company in India was claiming “personalized learning” and “industry-relevant curriculum” and “job guarantees.” We actually did those things, but we sounded identical in marketing. I didn’t know how to communicate the difference between a customized LMS and a data-driven adaptive platform. To a prospective student, they both just looked like “online course.”</p>
<p>I also underestimated how much the ed-tech boom would commoditize the category. By 2021, there were 15+ PM courses in India. Some were good. Most were recorded lectures with a Slack group. But they all charged ₹30k-50k, so we were competing on price instead of outcomes.</p>
<p>I should have built the brand earlier. I should have been louder about the salary hikes and the job transitions. I was too focused on the product and not enough on the perception.</p>
</section>
<section id="two-models-two-outcomes" class="level2">
<h2 class="anchored" data-anchor-id="two-models-two-outcomes">Two models, two outcomes</h2>
<table class="caption-top table">
<colgroup>
<col style="width: 49%">
<col style="width: 50%">
</colgroup>
<thead>
<tr class="header">
<th>Market-Before-Product (Standard Ed-Tech)</th>
<th>Product-Before-Market (Pragmatic Leaders)</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Raise capital to acquire users</td>
<td>Bootstrap until pedagogy is validated</td>
</tr>
<tr class="even">
<td>Scale horizontally (more students, same content)</td>
<td>Scale vertically (better outcomes per student)</td>
</tr>
<tr class="odd">
<td>Optimize for CAC and LTV</td>
<td>Optimize for job placement and salary hike</td>
</tr>
<tr class="even">
<td>Pedagogy is a variable to test</td>
<td>Pedagogy is the foundation to prove</td>
</tr>
<tr class="odd">
<td>Growth is the signal of success</td>
<td>Outcomes are the signal of success</td>
</tr>
</tbody>
</table>
<p>Both can work. But they produce different companies.</p>
<p>The first model produces Unacademy: ₹30,000 crores raised, millions of users, unclear pedagogy differentiation.</p>
<p>The second model produces Pragmatic Leaders: bootstrapped, thousands of students, ₹4.2 crores in salary hikes, 10,000+ professionals trained across programs.</p>
<p>I’m not saying one is better. I’m saying they’re optimizing for different things.</p>
</section>
<section id="the-question-i-havent-answered-yet" class="level2">
<h2 class="anchored" data-anchor-id="the-question-i-havent-answered-yet">The question I haven’t answered yet</h2>
<p><strong>How do you scale individualized learning without destroying the individualization?</strong></p>
<p>The data-driven approach works at 130 students. It works at 1,000 students. Does it work at 10,000? At 100,000?</p>
<p>At some point, the algorithms need more sophisticated models. The feedback loops need tighter instrumentation. The content needs to be modular enough to recombine dynamically but structured enough to maintain pedagogical coherence.</p>
<p>I thought I’d solved this in 2019. I hadn’t. I’d built a system that worked for the scale I was at. The next order of magnitude is a different problem.</p>
<p>This is why I’m building Ostronaut now. It’s the same problem — how do you deliver individualized learning at scale — but with better tools. Multi-agent AI systems that can generate, validate, and adapt content. Not as a replacement for pedagogy, but as infrastructure for it.</p>
<p>If you’re building ed-tech in an unproven category, bootstrap until the pedagogy works. Don’t raise capital to create demand. Raise capital to scale supply once you’ve proven the outcomes.</p>
<p>The mistake is thinking you can skip the pedagogy validation phase because the market already exists. You can’t. Students will pay once for a mediocre course. They won’t pay twice. And they definitely won’t refer their friends.</p>
<p>Are you optimizing for growth or outcomes? In the long run, only one of those compounds.</p>


<!-- -->

</section>

 ]]></description>
  <category>Ed-Tech</category>
  <category>India Startups</category>
  <category>Founder Lessons</category>
  <guid>https://talvinder.com/build-logs/edtech-before-edtech/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>I Built an Experiences Marketplace Five Years Before Airbnb Experiences</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/experiences-before-airbnb/</link>
  <description><![CDATA[ 





<p>In 2011, we built Tushky — a marketplace for local experiences in India. Cooking classes with home chefs. Heritage walks through old Mumbai. Photography workshops in the Western Ghats. Five years later, Airbnb launched Experiences and scaled the exact same model globally.</p>
<p>We had the idea first. We executed reasonably well. We still failed.</p>
<p>The reason wasn’t timing or capital or competition. It was something more fundamental: we optimized for transactions when we should have been building social infrastructure.</p>
<section id="the-social-capital-gap" class="level2">
<h2 class="anchored" data-anchor-id="the-social-capital-gap">The Social Capital Gap</h2>
<p>Most marketplace failures are diagnosed as “chicken-and-egg problems” — you need supply to attract demand, you need demand to attract supply. That’s true but useless. It’s like saying you failed because you ran out of money. The question is <em>why</em> you couldn’t solve the bootstrap problem when others did.</p>
<p>The answer is what I call the <strong>Social Capital Gap</strong> — the difference between a transactional platform and a community with economic infrastructure built on top.</p>
<p>Airbnb Experiences closed that gap. We didn’t. Not because we didn’t understand marketplaces, but because we treated the wrong thing as the product.</p>
</section>
<section id="what-we-got-right" class="level2">
<h2 class="anchored" data-anchor-id="what-we-got-right">What we got right</h2>
<p><strong>Profitable unit economics on outbound marketing.</strong> We could acquire customers through Facebook ads and Google search profitably. Rs 200-300 customer acquisition cost, Rs 800-1200 average booking value, 15-20% take rate. Not venture scale, but sustainable.</p>
<p><strong>Easy supplier onboarding.</strong> Experience providers could create a listing in under 10 minutes. No approval bottleneck. We had 150+ experiences listed within six months.</p>
<p><strong>Unique inventory.</strong> A Parsi chef teaching dhansak in her South Mumbai apartment. A tabla master offering two-hour sessions in Dadar. A birding expert leading dawn walks in Sanjay Gandhi National Park.</p>
<p>The product worked. People booked. Providers got paid. Reviews were positive.</p>
<p>Transactions hit a wall at about 80-100 bookings per month.</p>
<p>We couldn’t break through. We added more experiences. We improved search. We ran more ads. We tried discounting. Nothing moved the number sustainably.</p>
<p>The diagnosis in our internal docs: “Repeat customers were not getting enough options and first timers wanted more options to decide from.”</p>
<p>That diagnosis was wrong.</p>
</section>
<section id="what-actually-broke" class="level2">
<h2 class="anchored" data-anchor-id="what-actually-broke">What actually broke</h2>
<p>The real problem was visible in how our experience providers talked about us.</p>
<p>We wanted to be seen as business partners. We positioned ourselves that way in pitch decks and partner communications. But providers saw us as a booking channel — one of several ways they got customers, not materially different from their own Facebook page or a listing on JustDial.</p>
<p>When we asked providers to promote Tushky to their existing customers, most didn’t. When we asked them to refer other providers, most didn’t. When we suggested they collaborate on multi-experience packages, almost none did.</p>
<p>They had no social capital invested in the platform. We were a lead source, not a community.</p>
<p>Compare that to what Airbnb built. They didn’t just launch a booking interface. They built host meetups. They created an online forum where hosts shared tips. They featured hosts in marketing materials with their stories, not just their listings. They built a brand that hosts were proud to be associated with.</p>
<p>Their CTO told me years later: “The product is not the website. It’s the final booking.” Meaning: the value isn’t in the interface, it’s the trust infrastructure that makes the transaction possible.</p>
<p>We built a website. They built social capital.</p>
</section>
<section id="the-numbers-that-should-have-told-us" class="level2">
<h2 class="anchored" data-anchor-id="the-numbers-that-should-have-told-us">The numbers that should have told us</h2>
<ul>
<li>Our repeat booking rate: 12-15%</li>
<li>Our provider referral rate: &lt;5%</li>
<li>Our provider-to-provider collaboration rate: 0%</li>
</ul>
<p>Those aren’t marketplace metrics. Those are lead generation metrics.</p>
<p>A real marketplace creates network effects. Each new provider should make the platform more valuable to customers. Each new customer should make the platform more valuable to providers. We had linear growth at best.</p>
<p>We also made a strategic error on marketing. Outbound worked. We could buy traffic profitably. So we kept doing it. What we didn’t realize until too late: outbound marketing scales linearly with spend. Inbound marketing — SEO, word of mouth, community — scales exponentially but takes longer to build.</p>
<p>From our internal strategy doc in 2013: “Inbound marketing is the way to go. Build extremely loyal experience partner base. They will do word of mouth for you.”</p>
<p>We knew it. We wrote it down. We didn’t do it. Because outbound delivered this month’s numbers. Inbound required believing in next year’s numbers. We were optimizing for the wrong time horizon.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I got wrong</h2>
<p>I treated the chicken-and-egg problem as a supply problem. I thought: get enough experiences listed, and demand will follow. So we focused on making supplier onboarding frictionless.</p>
<p>That was backwards.</p>
<p>The constraint wasn’t the number of listings. It was the depth of engagement. We needed 20 customers who booked 5 times each, not 100 customers who booked once. We needed suppliers who saw Tushky as their primary channel, not one of five. Who would promote it to their customers. Who would collaborate with other suppliers. Who had reputational skin in the game.</p>
<p>That requires a different product. Not a listing interface. A community infrastructure.</p>
<p>We also underestimated the importance of curation and quality signaling. We made listing easy, which meant we had a quality variance problem. Some experiences were exceptional. Some were mediocre. Customers couldn’t tell the difference from the listing page. Airbnb solved this with detailed reviews, verified photos, and editorial featuring. We had basic star ratings.</p>
<p>The final mistake: we thought being first was an advantage. It’s not. Being first means you absorb all the market education cost. You teach customers that “experience marketplaces” exist. Then someone with more capital and better execution takes the market you created.</p>
<p>First-mover advantage is real in network-effect businesses only if you can build the network faster than competitors can copy the product. We couldn’t.</p>
</section>
<section id="the-test-that-matters" class="level2">
<h2 class="anchored" data-anchor-id="the-test-that-matters">The test that matters</h2>
<p>If you’re building a marketplace, here’s the question:</p>
<p><strong>Are your suppliers investing social capital in your platform, or are they just using it as a lead source?</strong></p>
<p>If it’s the latter, you don’t have a marketplace. You have a lead-gen business with marketplace unit economics. That’s not venture-scalable. It’s also not defensible.</p>
<p>The test is simple:</p>
<ul>
<li>Do suppliers refer other suppliers?</li>
<li>Do suppliers promote your platform to their existing customers?</li>
<li>Do suppliers collaborate with each other through your platform?</li>
</ul>
<p>If the answer to all three is no, you haven’t built the social infrastructure yet. You’ve built a directory.</p>
<p>We spent two years optimizing transaction flow when we should have been building community. By the time we realized it, we didn’t have the capital or the team energy to rebuild.</p>
<p>Airbnb had the capital. They also had something harder to replicate: they understood from day one that the product wasn’t the booking form. It was the trust system that made strangers willing to transact.</p>
<p>I still don’t know if we could have won even if we’d understood this earlier. The India market in 2011 wasn’t ready for experiential consumption at scale. Airbnb launched Experiences in 2016 into a global market that had already been trained by Airbnb Stays.</p>
<p>But I know we lost for the wrong reasons. We lost because we optimized for the transaction when we should have been building the social capital that makes transactions possible at scale.</p>
<p>The question I’m still working through: how do you build social capital infrastructure before you have transaction volume? Community requires critical mass. But you can’t get to critical mass without community.</p>
<p>That’s the real chicken-and-egg problem. Not supply and demand. Trust and scale.</p>


<!-- -->

</section>

 ]]></description>
  <category>Marketplaces</category>
  <category>Startup Lessons</category>
  <category>India Startups</category>
  <guid>https://talvinder.com/build-logs/experiences-before-airbnb/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>The OYO Pivot: When Marketplaces Should Own the Supply</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/oyo-marketplace-to-owned/</link>
  <description><![CDATA[ 





<p>OYO started as a marketplace aggregating budget hotels. Within 18 months, they pivoted to owning and standardizing supply. That wasn’t mission creep. It was the only way to scale.</p>
<p>Most founders treat vertical integration as a failure of platform thinking. I think that’s backwards. Sometimes owning the supply is the only path to building a defensible business.</p>
<section id="the-marketplace-first-doctrine-breaks-down" class="level2">
<h2 class="anchored" data-anchor-id="the-marketplace-first-doctrine-breaks-down">The Marketplace-First Doctrine Breaks Down</h2>
<p>The marketplace-first doctrine says: stay asset-light, let network effects do the work, take a commission. That works when supply quality is consistent or when users tolerate variance. It breaks when quality fragmentation prevents the marketplace from scaling at all.</p>
<p>I’m calling this the <strong>Supply Control Threshold</strong>: the point where a marketplace’s growth is bottlenecked not by demand, not by discovery, but by the unreliability of what it’s connecting you to.</p>
<p>Below that threshold, you’re a platform. Above it, you need to own the rails.</p>
<p>A marketplace hits the Supply Control Threshold when user retention drops faster than new user acquisition can compensate, and the drop is caused by supply-side inconsistency, not product-market fit.</p>
<p>At that point, three things happen.</p>
<p><strong>Dual accountability systems stop working.</strong> Airbnb’s model (hosts and guests rate each other) works because the median host quality is high enough that bad experiences are outliers. In budget hospitality, the median is terrible. Rating systems don’t fix structural supply problems. They just document them.</p>
<p><strong>Unit economics invert.</strong> A 15-20% marketplace commission can’t fund the quality improvement needed to retain users. But if you control the supply, standardize the rooms, train the staff, enforce SOPs, your margin per room goes up even though your capital requirements do too. OYO’s model wasn’t cheaper. It was more profitable per retained customer.</p>
<p><strong>Data becomes an actual moat.</strong> A marketplace collects transaction data. A supply owner collects operational data: which amenities drive repeat bookings, which service gaps cause churn, which staff behaviors correlate with ratings. That’s the “product knowledge base” that lets you optimize. You can’t get it from aggregating independent operators who don’t share your incentive to standardize.</p>
<p>Network effects alone don’t guarantee a sustainable business model. Skype had massive network effects and still couldn’t build a business Microsoft wanted to keep funding. OYO’s bet was that controlling supply would create better unit economics than pure platform power ever could.</p>
</section>
<section id="indias-budget-hotel-problem-was-a-product-problem" class="level2">
<h2 class="anchored" data-anchor-id="indias-budget-hotel-problem-was-a-product-problem">India’s Budget Hotel Problem Was a Product Problem</h2>
<p>Pre-OYO, booking a ₹1,500/night room was a gamble. Photos didn’t match reality. AC didn’t work. Checkout was a negotiation. Aggregating those hotels into a marketplace didn’t solve the problem. It just gave you a prettier interface to a bad experience.</p>
<p>India had roughly 50,000 budget hotels in the sub-₹2,000 segment when OYO launched. Almost none of them had standardized amenities, consistent check-in processes, or reliable room quality. The fragmentation wasn’t a distribution problem. It was a product problem. MakeMyTrip and Goibibo listed these hotels, but listing a bad product doesn’t make it good. Travelers who got burned once didn’t come back. They went back to asking friends for recommendations or staying at known chains at 3x the price.</p>
<p>OYO’s move: lease the rooms, rebrand them, enforce standards, own the customer relationship. Revenue share with hotel owners, but control of operations. That let them deliver consistency. Consistency drove retention. Retention justified customer acquisition cost.</p>
</section>
<section id="the-data-feedback-loop-and-the-commission-trap" class="level2">
<h2 class="anchored" data-anchor-id="the-data-feedback-loop-and-the-commission-trap">The Data Feedback Loop and the Commission Trap</h2>
<p>When you’re a marketplace, you know what users book. When you own supply, you know <em>why</em> they rebook or don’t. OYO could see that free breakfast didn’t move the needle but working WiFi did. They could test pricing strategies across properties without negotiating with franchisees. They could deploy capital to the highest-ROI improvements because they had ground truth, not survey data.</p>
<p>This is the same advantage that Zara has over department stores, or that Amazon Basics has over third-party sellers. When you control the supply, you control the feedback loop. When you control the feedback loop, you compound learning faster than anyone else in the market.</p>
<p>A pure marketplace in budget hospitality would need to take 25-30% to fund quality audits, customer service, and fraud prevention. But hotel owners operating on thin margins can’t give up that much. So you either take a smaller cut and can’t invest in quality, or you take a larger cut and lose supply. OYO sidestepped this by becoming the operator.</p>
<p>Consider the numbers. A budget hotel owner in a tier-2 city makes ₹800-1,200 per room per night at 40-50% occupancy. After staff, utilities, and maintenance, the margin is 15-20%. A marketplace taking 20% of revenue leaves the owner with almost nothing. A managed model where OYO guarantees higher occupancy (70-80%) and charges a management fee changes the math entirely. The owner makes more money. OYO makes more money. The customer gets a reliable room. The economics work because ownership aligned the incentives.</p>
<p><strong>Contrast with Airbnb:</strong> Airbnb works as a marketplace because hosts are individuals with reputational skin in the game and properties that reflect personal taste. Variance is a feature, not a bug. Budget hotels are commercial operations with misaligned incentives. Variance is a retention killer.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I Got Wrong</h2>
<p>I initially thought OYO’s model was just capital inefficiency dressed up as innovation. I assumed marketplace dynamics would eventually force quality improvements through competition. That was wrong.</p>
<p>The budget hotel market in India wasn’t competitive on quality. It was competitive on price, which drove a race to the bottom. No individual hotel had the capital or incentive to standardize. The <a href="../../frameworks/biggest-challenge-indian-startups/">market failure was real</a>, and a pure platform couldn’t fix it.</p>
<p>What I still don’t know: where exactly the Supply Control Threshold sits for other verticals. Food delivery? Probably below it, because Swiggy and Zomato work as marketplaces (though cloud kitchens are an interesting test case). Ride-hailing? Uber’s experimenting with owned fleets in some markets, which suggests they’re testing whether they’ve crossed it. Healthcare? Probably above it, which is why most telehealth companies are moving toward owned clinical networks.</p>
<p>The pattern I’d watch: if a marketplace’s customer complaints are about the product itself rather than the transaction, that’s the signal. Swiggy users complain about delivery times (transaction). OYO users complained about room quality (product). That distinction tells you whether you need better logistics or whether you need to own the supply.</p>
</section>
<section id="the-retention-curve-tells-you-everything" class="level2">
<h2 class="anchored" data-anchor-id="the-retention-curve-tells-you-everything">The Retention Curve Tells You Everything</h2>
<p>If you’re building a marketplace and user retention is your problem, ask this: Is the variance in supply quality something users tolerate, or is it why they’re leaving?</p>
<p>If it’s the latter, you’ve crossed the Supply Control Threshold. At that point, “stay asset-light” is advice that keeps you stuck. Owning supply isn’t a pivot away from your model. It’s the only way to make the model work.</p>
<p>The question isn’t whether vertical integration is elegant. It’s whether your unit economics improve when you control what you’re selling. If they do, the capital intensity is a feature, not a bug. It’s what keeps competitors from copying you.</p>
<p>OYO didn’t abandon the marketplace playbook because they failed at it. They abandoned it because the economics of the problem required a different solution. The doctrine that platforms should never own supply is just that, doctrine. The actual question is: what does the retention curve tell you?</p>


<!-- -->

</section>

 ]]></description>
  <category>India Tech</category>
  <category>Marketplace Strategy</category>
  <category>Vertical Integration</category>
  <guid>https://talvinder.com/build-logs/oyo-marketplace-to-owned/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Point-of-Need Learning: Why Application Beats Credentials</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/upskilling-before-covid/</link>
  <description><![CDATA[ 





<p>We started building Pragmatic Leaders in 2018 because traditional education was selling credentials, not competence. You could finish a Udemy course on React and still not be able to build a working app for the deli downstairs. The gap was not knowledge. It was application.</p>
<p>COVID did not create the problem. It just made everyone else notice.</p>
<section id="the-bet-we-made-early" class="level2">
<h2 class="anchored" data-anchor-id="the-bet-we-made-early">The Bet We Made Early</h2>
<p>I call this <strong>Point-of-Need Learning</strong>: the shift from abstract, front-loaded education to skills delivered exactly when you need to use them. Not “learn React,” but “learn React while building this specific feature for this specific user problem.”</p>
<p>The traditional model assumes you learn first, then apply later. That is backwards. Application comes first. Learning happens in service of getting something real done.</p>
<p>Pre-COVID, this was a contrarian bet. Post-COVID, it is the only model that works at scale.</p>
<p>The specific bet: <strong>by 2025, the majority of corporate training budgets will shift from certification-based programs to embedded, outcome-verified skill development.</strong> Not courses completed. Not certificates issued. Measurable ability to execute specific tasks.</p>
<p>We saw this coming because we were watching the wrong metric. Everyone tracked course completion rates. We tracked the gap between “certified” and “capable.”</p>
</section>
<section id="the-indian-education-arithmetic" class="level2">
<h2 class="anchored" data-anchor-id="the-indian-education-arithmetic">The Indian Education Arithmetic</h2>
<p>In India, the salary difference between Tier 1 college graduates and everyone else was 300%+. That gap was not about intelligence. It was about practical application. Tier 1 students got hands-on projects, mentorship, and real feedback loops. Everyone else got pre-recorded lectures and multiple-choice tests.</p>
<p>The numbers told a brutal story. Average graduates from non-Tier-1 colleges made 3-6 lakhs annually. Courses worth 1-2 lakhs were out of reach. There was no ramp between free YouTube tutorials and expensive postgraduate programs. The people who needed upskilling most could afford it least, and the affordable options were the worst at delivering actual competence.</p>
<p>The edtech boom of 2015-2020 scaled the wrong thing. It scaled content delivery. More videos. More courses. More certificates. But content was never the constraint. Application was. Poor completion rates, poor success rates, poor certificate value. The entire industry optimized for enrollment numbers and certificate issuance, not for whether graduates could actually do the job.</p>
<p>McKinsey projected 375 million workers globally would need to completely change their skill sets by 2030. In India alone, 400 million needed reskilling, with 100 million in managerial and professional domains. The gap between the problem and the existing solutions was not a crack. It was a canyon.</p>
</section>
<section id="three-technical-bets-that-created-the-moat" class="level2">
<h2 class="anchored" data-anchor-id="three-technical-bets-that-created-the-moat">Three Technical Bets That Created the Moat</h2>
<p>We started with 21 paying students across 3 countries. Bootstrapped. The validation was not “do people want to learn?” It was “will people pay to learn skills they can immediately use?”</p>
<p>The answer was yes, but only if we changed the architecture.</p>
<p>The traditional edtech model sells course access, measures completion rates, issues a certificate, and hopes graduates apply it somewhere. Our model identified skill gaps in real work context, delivered learning at the moment of need, verified application rather than recall, and credentialed based on demonstrated competence.</p>
<p>We built three things that did not exist in most platforms.</p>
<p><strong>Automated skill gap identification.</strong> Not “what course do you want?” but “what can you not do right now that is blocking you?” We mapped learning paths to actual job requirements, not arbitrary curriculum structures. Each course was divided into learning objectives, competencies, and complexity levels across sub-domains. The platform created individualized learning routes for each student based on their past competency and their target destination.</p>
<p><strong>Recognition of Prior Learning.</strong> We borrowed from traditional university models but applied them to working professionals. Past experience and knowledge were quantified, requisite credits awarded, and skill gaps identified automatically. This meant a developer with five years of backend experience did not sit through the same curriculum as a fresh graduate. Their learning path started where their competence ended.</p>
<p><strong>Tokenized credit and applied gamification.</strong> Not badges for watching videos. Credits for solving real problems, helping peers, shipping working code. Complete a module, earn credits. Contribute to community, earn credits. Mentor someone, earn credits. The currency was not time spent but value created. Every positive interaction fed the learning algorithm.</p>
<p>The pedagogy was case-based, modeled on how business, medicine, and law have trained professionals for decades. Harvard’s case method, digitized and made accessible. Students did not watch someone talk for 20 hours. They solved cases, applied theory to practical scenarios, built demonstrable projects. We paired students with in-house development teams to build and ship real products.</p>
</section>
<section id="the-platform-decision-that-cost-us-three-months" class="level2">
<h2 class="anchored" data-anchor-id="the-platform-decision-that-cost-us-three-months">The Platform Decision That Cost Us Three Months</h2>
<p>The hardest technical decision: moving from a customized LMS to building proprietary platform infrastructure. We lost 3 months of velocity. But a customized LMS could not use the data we were collecting. Patterns of where people got stuck. Which learning paths actually led to competence. Which credentials correlated with job performance.</p>
<p>We had a working LMS, a Stack Overflow-inspired forum, and a job board. That was enough to validate the pedagogy. But it was not enough to scale individualized learning to millions of learners across domains. The data from those first 35 students across 5 countries (paying an average of $1,200 per person) proved the model worked. 70% came from Ivy League-equivalent colleges. Graduates from Product School and UpGrad joined because of our pedagogy, not our brand.</p>
<p>That data became the moat. Not the content. Not the platform features. The data about how people actually learn and where they actually fail.</p>
</section>
<section id="who-actually-pays-for-competence" class="level2">
<h2 class="anchored" data-anchor-id="who-actually-pays-for-competence">Who Actually Pays for Competence</h2>
<p>By early 2020, we had validation across three verticals: corporate training, university partnerships, and individual upskilling. Corporate training drove the most revenue. Companies paid for verified competence in ways individuals would not.</p>
<p>We initially assumed individuals would be the primary buyers. They were not. Corporations were. Individuals optimize for credentials because that is what the hiring market rewards. They need the certificate they can show on LinkedIn. Corporations optimize for outcomes because they see the gap between the certificate and the work that gets done.</p>
<p>This distinction shaped everything. Our live classes converted at 20% to bootcamp courses. 500 registered, 150 attended with 88% retention, 60% paid for extended access. The numbers validated that the case-based approach held attention in ways pre-recorded content never could.</p>
<p>The 80% success rate for job outcomes happened irrespective of pedigree. Students who would never have been considered for product management roles at top companies got in, and 100% attributed it to the case-based learning and the network they built while doing it.</p>
</section>
<section id="then-covid-hit" class="level2">
<h2 class="anchored" data-anchor-id="then-covid-hit">Then COVID Hit</h2>
<p>Suddenly, everyone needed what we had been building. Remote work made skills gaps impossible to hide. Managers could not rely on proximity as a proxy for productivity. “Can this person actually do the job?” became the only question that mattered.</p>
<p>The market went from “interesting idea” to “urgent need” in 8 weeks.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I Got Wrong</h2>
<p>We underestimated how much infrastructure we needed to verify competence at scale. It is easy to check if someone watched a video. It is hard to verify if they can apply what they learned in a novel context. We built rule-based scoring (this was 2018, before LLMs). The scoring was brittle. We spent 6 months refining it.</p>
<p>The biggest miss: we did not move fast enough on international expansion. We had validation in India, the US, and Southeast Asia by 2019. We should have scaled globally before COVID. By the time we were ready, the market was crowded.</p>
</section>
<section id="the-competence-measurement-problem" class="level2">
<h2 class="anchored" data-anchor-id="the-competence-measurement-problem">The Competence Measurement Problem</h2>
<p>Pragmatic Leaders now trains 10,000+ professionals annually. The model works. But the model only works if you measure the right thing: not what people know, but what they can do.</p>
<p>The question I am still working through: how do you scale verified competence without turning it into another credential game? The moment you standardize assessment, people optimize for the assessment instead of the skill. The moment you issue a certificate, employers use it as a filter rather than a signal. The system that was built to prove competence becomes another gatekeeping mechanism.</p>
<p>We are not there yet. But the direction is clear. The companies that win in education are not the ones with the most content. They are the ones with the tightest feedback loop between learning and application.</p>


<!-- -->

</section>

 ]]></description>
  <category>Edtech</category>
  <category>India Tech</category>
  <category>Build Logs</category>
  <guid>https://talvinder.com/build-logs/upskilling-before-covid/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>The IRL Growth Bet: Why I Started an Offline-First Music Community in the Age of Algorithms</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/irl-growth-side-b/</link>
  <description><![CDATA[ 





<p>I’m building a music community that starts offline. Not “offline-friendly.” Not “hybrid with offline touchpoints.” Offline-first, with digital as documentation.</p>
<p>This is the opposite of what you’re supposed to do in 2025. Every growth playbook says: build digital, scale algorithmically, add IRL later if you hit product-market fit. I’m doing the reverse. Physical space first. Events first. Venue ownership as the moat.</p>
<p>The reason is simple: digital music communities have a 90%+ churn problem that no one talks about. They get traffic. They don’t get retention. I’ve watched this pattern play out across TastemakerX, Splice community features, even parts of Bandcamp’s social layer. People come, people scroll, people leave.</p>
<p>The offline bet is that context beats content. A 200-person room with the right sound system and the right curation creates more lasting connection than 200,000 algorithm-fed playlist adds.</p>
<section id="the-context-density-problem" class="level2">
<h2 class="anchored" data-anchor-id="the-context-density-problem">The Context Density Problem</h2>
<p>Digital platforms optimize for content distribution. More reach, more impressions, more shares. But music discovery — real discovery, the kind that changes what you listen to for years — doesn’t scale through distribution. It scales through density of context.</p>
<p>Context density is the ratio of environmental signal to content signal. High context density: you’re in a room, the artist is 10 feet away, the sound system is tuned for that genre, everyone around you is there for the same reason, the lighting matches the mood. Low context density: you’re scrolling Instagram, the audio is compressed, you’re half-watching, the algorithm served it between a reel about productivity hacks and an ad for protein powder.</p>
<p>Digital platforms have infinite distribution and near-zero context density. Offline spaces have limited distribution and near-infinite context density.</p>
<p>The entire music industry has been optimizing the wrong variable.</p>
</section>
<section id="the-math-behind-the-bet" class="level2">
<h2 class="anchored" data-anchor-id="the-math-behind-the-bet">The math behind the bet</h2>
<p>Put a number on it: <strong>an offline-first music community with 500 deeply engaged members will have higher 12-month retention and higher lifetime value than a digital-first music community with 50,000 casual users.</strong></p>
<p>The mechanism is economic, not emotional.</p>
<p>Digital music communities fail because they confuse traffic with commitment. You can build a Discord with 10,000 members. You can rack up playlist followers. But if the user’s relationship is with the algorithm, not the community, they churn the moment the algorithm changes or a competitor offers a better feed.</p>
<p>I’ve seen this at Pragmatic Leaders. We train 10,000+ people a year. The ones who show up to in-person workshops have 4x the completion rate and 6x the referral rate of the ones who take the same content online. It’s not the content. It’s the context.</p>
<p>Offline creates three things digital can’t replicate:</p>
<p><strong>Sunk cost through physical presence.</strong> You drove across town. You paid for parking. You cleared your calendar. That investment makes you more likely to stay, engage, and return. Digital has no sunk cost. Leaving a Zoom call or closing a tab costs nothing.</p>
<p><strong>Unplanned discovery through spatial adjacency.</strong> In a venue, you hear the opener while getting a drink. You overhear a conversation about a band. You see someone’s T-shirt and ask about it. Digital platforms serve you what the algorithm thinks you want. Physical spaces serve you what’s spatially adjacent. Adjacency creates serendipity. Algorithms create filter bubbles.</p>
<p><strong>Social proof through observable behavior.</strong> In a room, you see 200 people nodding to the same beat. That’s immediate, visceral validation that this music matters to people like you. Online, social proof is a number. Numbers are easy to fake and hard to feel.</p>
</section>
<section id="the-business-model" class="level2">
<h2 class="anchored" data-anchor-id="the-business-model">The business model</h2>
<p>The business model follows from the physics of context density:</p>
<p><strong>Venue ownership or long-term lease</strong> — not renting by the night. The Malaysia offline education model is the template: physical infrastructure as a strategic asset, not a cost center. Own the space, control the experience, capture the upside.</p>
<p><strong>Dual revenue: tickets + bar/merch</strong> — the McDonald’s real estate model applied to music. You make money on the event. You make money on the ancillary spend. Digital platforms have one revenue stream (ads or subscriptions). Physical spaces have two.</p>
<p><strong>Curation as the moat</strong> — not recommendation algorithms. Hand-picked lineups. Genre-specific nights. Invites, not open signups. The TastemakerX model (hand-picked contributors, exclusive access) was right about curation. It was wrong about doing it digitally.</p>
<p>The failure mode of digital music communities is feature launch fallacy. You build a new discovery tool, a new social layer, a new playlist format. You get a flood of initial users. Then nothing. Because the users came for the feature, not the community.</p>
<p>Offline doesn’t have this problem. People don’t come to a venue for a feature. They come for an experience. Experiences are harder to replicate, harder to commoditize, and harder to churn from.</p>
</section>
<section id="evidence-from-adjacent-models" class="level2">
<h2 class="anchored" data-anchor-id="evidence-from-adjacent-models">Evidence from adjacent models</h2>
<p><strong>Kommune</strong> built one of India’s largest creator communities by focusing on “context, culture, and connections” — not just content distribution. Their events are immersive, narrative-rich, and built around specific cultural moments. That’s high context density. Their digital presence documents the offline experience; it doesn’t replace it.</p>
<p><strong>The craftspeople marketplace pattern</strong>: start with exactly one group, go deep, then expand. For a music community, that means one genre, one city, one venue. Not “music lovers everywhere.”</p>
<p><strong>The Gmail vs.&nbsp;Google Wave lesson</strong>: exclusivity works when the product is good. Gmail’s invite-only model created demand because the product was 100x better than Hotmail. Google Wave’s exclusivity failed because the product was confusing. The offline bet only works if the experience is genuinely better than staying home and streaming.</p>
<p>At Zopdev, I’ve watched companies optimize infrastructure for scale before they have retention. They build for 100,000 users when they have 1,000. The music version of this is building a streaming platform before you have a single venue that people return to every week.</p>
</section>
<section id="what-ive-gotten-wrong-so-far" class="level2">
<h2 class="anchored" data-anchor-id="what-ive-gotten-wrong-so-far">What I’ve gotten wrong so far</h2>
<p>I initially thought the digital layer would be the primary discovery surface, with offline as a “premium tier” for superfans. That was backwards. The offline experience has to be the core product. Digital is the documentation layer — the way people remember, share, and recruit others into the offline experience.</p>
<p>I also underestimated the operational complexity. Running a venue is harder than running a SaaS product. Sound engineers, alcohol licenses, neighbors who complain about noise, artists who cancel last-minute. The unit economics are harder to model because every night is different.</p>
<p>The question I’m still working through: how do you scale context density? You can’t franchise a vibe. You can’t template curation. The moment you try to systematize the magic, it stops being magic.</p>
<p>Maybe the answer is you don’t scale it. Maybe the answer is you build 10 venues in 10 cities, each with its own identity, and you accept that this will never be a billion-dollar exit. Maybe the offline bet is also a bet against venture-scale returns.</p>
<p>I don’t know yet. But I know the digital-first playbook is broken for music. And I’d rather build something that 500 people love than something that 500,000 people scroll past.</p>
<p>If context density is the real driver of retention, what other industries are optimizing for distribution when they should be optimizing for density?</p>


</section>

 ]]></description>
  <category>Community Building</category>
  <category>Growth</category>
  <category>Founder Lessons</category>
  <guid>https://talvinder.com/field-notes/irl-growth-side-b/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Framework Lag: Why the Winners of 2010-2015 Could Explain Network Effects Before VCs Had Words For It</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/network-effects-before-nfx/</link>
  <description><![CDATA[ 





<p>In 2011, I was trying to explain why Airbnb and Uber were fundamentally different from every other marketplace, and I didn’t have the words for it. Neither did anyone else.</p>
<p>The pattern was clear. These companies were replacing institutional trust with peer-to-peer trust. They were creating long tails in markets that shouldn’t have had long tails. They were democratizing supply in industries built on scarcity.</p>
<p>But when you pitched this to investors, they’d nod politely and ask about unit economics.</p>
<section id="framework-lag-costs-you-capital" class="level2">
<h2 class="anchored" data-anchor-id="framework-lag-costs-you-capital">Framework Lag costs you capital</h2>
<p>I’m calling this <strong>Framework Lag</strong>: the gap between when a pattern becomes visible and when the industry develops shared language to describe it.</p>
<p>Framework Lag is expensive. It means you can see something working, you can even articulate why it’s working, but you can’t raise capital for it because the thesis doesn’t fit existing mental models. You end up in meetings where a GP says “I don’t see what’s defensible here” because the word “defensible” at that point meant patents, brand, or regulatory capture. Not demand-side scale economies. Not trust graphs. Those concepts weren’t available yet.</p>
<p>By 2016, “network effects” was standard VC vocabulary. NFX had published their taxonomy. a16z had their playbook. Every deck had a slide on “defensibility through network effects.”</p>
<p>But in 2011-2013, you were on your own.</p>
</section>
<section id="winners-taught-vcs-the-framework-while-raising" class="level2">
<h2 class="anchored" data-anchor-id="winners-taught-vcs-the-framework-while-raising">Winners taught VCs the framework while raising</h2>
<p><strong>The companies that won in the 2010-2015 cohort weren’t the ones with the strongest network effects. They were the ones whose founders could articulate network effects before VCs had a framework for it.</strong></p>
<p>Airbnb raised at a $2.5B valuation in 2011. Uber raised at $330M in 2011, $3.5B by 2013. These weren’t post-framework raises. These were raises where founders had to teach investors the mental model in the room.</p>
<p>Travis Kalanick didn’t say “we have strong cross-side network effects.” He said something closer to: every driver we add makes wait times shorter, which makes more riders sign up, which attracts more drivers. He had to walk investors through the loop because the loop didn’t have a name yet.</p>
<p>The companies that waited until the framework was established, until “marketplace dynamics” and “two-sided networks” were standard pitch language, were already late. By then the pattern was priced in.</p>
<p>I spent four years developing what I called a “Social Capital investment thesis.” The core mechanisms I identified:</p>
<ul>
<li>Democratizing the tools of production and distribution</li>
<li>Connecting supply and demand through peer-to-peer networks</li>
<li>Filtering efficiency of social network reviews</li>
<li>Dual accountability systems (both sides rate each other)</li>
</ul>
<p>This wasn’t network effects theory yet. It was an attempt to explain why Airbnb worked when Couchsurfing didn’t. Why Uber scaled when taxi apps didn’t.</p>
<p>The breakthrough came during a physical journey: staying in Airbnbs, taking Ubers, experiencing the products as a user. I’d been developing the thesis intellectually for years, but it was using the product in New York that made the mechanisms click. You can’t theorize trust infrastructure. You have to feel the moment when you hand your apartment keys to a stranger because 47 five-star reviews told you it was safe.</p>
</section>
<section id="network-effects-are-necessary-but-not-sufficient" class="level2">
<h2 class="anchored" data-anchor-id="network-effects-are-necessary-but-not-sufficient">Network effects are necessary but not sufficient</h2>
<p>Here’s what I got wrong: I thought network effects alone were sufficient. They’re not.</p>
<p>Microsoft bought Skype for $8.5B in 2011. Skype had massive network effects with 663 million registered users. It was also unprofitable and buried in debt. Network effects didn’t guarantee a sustainable business model.</p>
<p>Groupon hit a $16B valuation at IPO in November 2011. It had what looked like network effects: more merchants attracted more buyers, more buyers attracted more merchants. Within 18 months the stock had lost 80% of its value. The “network effects” were actually a subsidy treadmill. Merchants weren’t retained by network density. They were retained by discount margins that Groupon couldn’t sustain.</p>
<p>What separated Airbnb and Uber from Skype and Groupon wasn’t just network effects. It was the trust infrastructure they built on top of those effects. Reviews, ratings, verified identities, insurance programs, dispute resolution. The network effect got people onto the platform. The trust infrastructure kept them there and made the transactions possible.</p>
<p>This distinction matters because most Framework Lag discussions focus on recognizing patterns. The harder question is recognizing which patterns have the additional infrastructure to become durable.</p>
</section>
<section id="the-language-i-was-using-looks-primitive-now" class="level2">
<h2 class="anchored" data-anchor-id="the-language-i-was-using-looks-primitive-now">The language I was using looks primitive now</h2>
<p>The vocabulary I was using in 2013-2014 reads like a rough draft of what became standard:</p>
<p>“Airbnb and Uber create long tails in travel by replacing artificial institutional trust with peer-to-peer trust mechanisms.”</p>
<p>That’s not how you’d pitch it today. Today you’d say: “Two-sided marketplace with strong same-side and cross-side network effects, defensible through supply density and trust infrastructure.”</p>
<p>But in 2013, that language didn’t exist yet.</p>
<p>I can track the Framework Lag by looking at when specific terms entered standard VC vocabulary:</p>
<ul>
<li>“Marketplace dynamics” as a pitch category: ~2014</li>
<li>“Cold start problem” as a named challenge: ~2015</li>
<li>“Network effects” as defensibility moat: ~2016</li>
<li>NFX’s network effects map (13 types): 2018</li>
</ul>
<p>Before that, you were working from first principles every time.</p>
</section>
<section id="the-companies-i-was-watching" class="level2">
<h2 class="anchored" data-anchor-id="the-companies-i-was-watching">The companies I was watching</h2>
<p><strong>Airbnb</strong> listed 10,000 properties in 2009. 50,000 by mid-2011. The growth curve was obvious. The explanation wasn’t. Investors kept comparing it to VRBO and HomeAway, missing the peer-to-peer trust mechanism entirely. VRBO was a listing service. Airbnb was building a trust graph. That distinction is obvious now. In 2011, it was invisible to most investors because “trust graph” wasn’t a phrase anyone used.</p>
<p><strong>Uber</strong> launched in SF in 2010. By 2013, they were in 35 cities. The pattern was clear: once you hit density in one market, the playbook was repeatable. But VCs kept asking about taxi medallion regulations, not about supply-side liquidity. They were evaluating Uber against the taxi industry’s rules instead of recognizing that Uber was making those rules irrelevant.</p>
<p>The thesis I was developing focused on what I called “Social Capital,” the value created when you replace institutional intermediaries with peer networks. That framing was clunky. It mixed trust mechanisms with network effects with platform dynamics. But it was the best available framework at the time, and it let me see things that the standard VC frameworks of 2011 couldn’t explain.</p>
</section>
<section id="framework-lag-still-exists" class="level2">
<h2 class="anchored" data-anchor-id="framework-lag-still-exists">Framework Lag still exists</h2>
<p>Right now, in 2026, there are patterns that are visible but not yet named.</p>
<p><a href="../../frameworks/agentware/">AI agents coordinating with other AI agents</a> to complete tasks. Inference-time compute as a moat. Synthetic data loops that improve model performance faster than human-labeled data. Multi-agent systems where the coordination layer, not any individual model, is the <a href="../../frameworks/ai-runtime-infrastructure-play/">source of competitive advantage</a>.</p>
<p>These patterns are real. The companies building on them are <a href="../../frameworks/ai-evolution/">raising capital</a>. But the frameworks are still being developed in real-time.</p>
<p>If you’re building in a space where the framework doesn’t exist yet, you have two options:</p>
<ol type="1">
<li><p>Wait until the framework is established and the pattern is obvious. You’ll have better pitch materials. You’ll also be late.</p></li>
<li><p>Build the framework yourself. Articulate the pattern. Name the mechanisms. Teach investors the mental model.</p></li>
</ol>
<p>The second path is harder. It’s also where the asymmetric returns are.</p>
<p>The question I’m still working through: how do you know when you’re early to a real pattern versus early to a pattern that doesn’t matter?</p>
<p>In 2011, “peer-to-peer trust infrastructure” was early to a real pattern. “Social shopping” was early to a pattern that mostly didn’t matter. Groupon’s “local marketplace network effects” looked real until the trust layer turned out to be hollow.</p>
<p>One heuristic I’ve developed: real patterns create durable behavior change that persists even when the subsidy disappears. Airbnb users kept booking after the novelty wore off. Groupon users stopped buying when the deals got worse. The trust infrastructure test isn’t whether people use the product. It’s whether they trust it enough to keep using it when the economics normalize.</p>
<p>I don’t have a clean formula for spotting real patterns early. But I know the penalty for waiting until the framework exists: you arrive right on time and call it innovation.</p>


</section>

 ]]></description>
  <category>Startups</category>
  <category>Network Effects</category>
  <category>Venture Capital</category>
  <guid>https://talvinder.com/field-notes/network-effects-before-nfx/</guid>
  <pubDate>Thu, 02 Apr 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Device-Level Blocking Won’t Stop Digital Arrest Scams — The UI Is the Real Vulnerability</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/digital-arrest-device-blocking-ui-problem/</link>
  <description><![CDATA[ 





<p>Last week, India’s Home Ministry issued a directive to WhatsApp: block the device IDs of accounts involved in digital arrest scams so perpetrators can’t open new accounts on the same hardware.</p>
<p>WhatsApp agreed. They have 30 days to submit a proposal.</p>
<p>It won’t work.</p>
<p>The reason it won’t work explains why digital arrest scams will keep growing regardless of how many technical controls we layer on top.</p>
<section id="device-ids-are-the-wrong-target" class="level2">
<h2 class="anchored" data-anchor-id="device-ids-are-the-wrong-target">Device IDs Are the Wrong Target</h2>
<p>A digital arrest scammer running their operation in India has access to hundreds of millions of cheap Android handsets. A factory reset costs nothing. A new SIM card costs under a hundred rupees. New device ID, new phone number, new WhatsApp account — in under an hour.</p>
<p>Device ID blocking is designed for a threat model where scammers are sophisticated actors with expensive hardware. Digital arrest scams run on volume. The operators behind them are not protecting expensive infrastructure. They burn devices and phone numbers the way spammers burn email addresses.</p>
<p>Blocking device IDs will inconvenience scammers for exactly as long as it takes them to buy a new phone. That’s not a security solution. That’s a speed bump.</p>
<p>The same directive also asked WhatsApp to expand logo detection — comparing profile photos against known law enforcement insignia and removing impersonators. That’s closer to the right solution. But device ID blocking is the headline, and it’s a distraction.</p>
</section>
<section id="the-verification-inversion" class="level2">
<h2 class="anchored" data-anchor-id="the-verification-inversion">The Verification Inversion</h2>
<p>Digital arrest scams succeed not because WhatsApp security is weak, but because the UI creates an environment where users willingly bypass every security control they have.</p>
<p>Here’s what actually happens. The victim receives a video call from someone in a police or CBI uniform. The caller’s profile photo shows an official badge. The video call shows an official-looking backdrop with case numbers, warrant references, and government seals. The caller creates artificial urgency: “Your Aadhaar is linked to a money laundering case. You are under digital arrest. Do not end this call or a physical warrant will be issued.”</p>
<p>The victim is terrified. They’re not thinking about security. They’re thinking about arrest.</p>
<p>The scammer then walks them through every action step by step — opening the banking app, transferring money to a “safe account,” sharing OTPs to “verify identity.” Every step is narrated as protective. Every security control the bank built becomes something the victim actively completes.</p>
<p>I’m calling this the <strong>Verification Inversion</strong> — when the interface layer flips the purpose of security controls from protecting the user to executing the attack. The OTP isn’t protecting you. The interface told you it is, so you hand it over.</p>
<p>The attack surface is not the device. It’s the visual trust layer on the screen.</p>
</section>
<section id="logo-detection-is-actually-the-right-instinct" class="level2">
<h2 class="anchored" data-anchor-id="logo-detection-is-actually-the-right-instinct">Logo Detection Is Actually the Right Instinct</h2>
<p>Buried in the same directive: WhatsApp has already deployed a logo detection and media matching system. It compares profile photos against law enforcement agency logos — CBI, ATS, state police forces — and removes accounts misusing official insignia.</p>
<p>That’s the correct direction. The scam works because the visual presentation looks authoritative. Break the authority signal and you break the scam’s opening move.</p>
<p>The problem is logo detection only catches the obvious impersonators. Scammers are already adapting — using AI-generated variations of official logos that match closely enough to fool users but differ enough to bypass pixel-level matching. The arms race between logo variations and detection systems is real, and the scammers are running faster than the platform.</p>
<p>Caller information display — showing users context about who’s calling them — is another step in the right direction. If every incoming video call from an unverified account shows a prominent “unverified caller” flag that persists through the call, it creates friction at exactly the right moment.</p>
</section>
<section id="what-would-actually-work" class="level2">
<h2 class="anchored" data-anchor-id="what-would-actually-work">What Would Actually Work</h2>
<p>The solutions that would reduce digital arrest scams all share one characteristic: they create friction in the interface at the moment of manipulation.</p>
<p><img src="https://talvinder.com/field-notes/digital-arrest-device-blocking-ui-problem/data:image/svg+xml;base64,<?xml version="1.0" encoding="utf-8"?><svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" data-d2-version="v0.7.1" preserveAspectRatio="xMinYMin meet" viewBox="0 0 338 399"><svg class="d2-1877477521 d2-svg" width="338" height="399" viewBox="-9 -9 338 399"><rect x="-9.000000" y="-9.000000" width="338.000000" height="399.000000" rx="0.000000" fill="transparent" class=" fill-N7" stroke-width="0" /><style type="text/css"><![CDATA[
.d2-1877477521 .text-bold {
	font-family: "d2-1877477521-font-bold";
}
@font-face {
	font-family: d2-1877477521-font-bold;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA1gAAoAAAAAFJQAAguFAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgXxHXrmNtYXAAAAFUAAAAjwAAAK4C9gNeZ2x5ZgAAAeQAAAb7AAAJVL6YkTFoZWFkAAAI4AAAADYAAAA2G38e1GhoZWEAAAkYAAAAJAAAACQKfwXeaG10eAAACTwAAAB8AAAAfDnpBV1sb2NhAAAJuAAAAEAAAABAJeooQG1heHAAAAn4AAAAIAAAACAANwD3bmFtZQAAChgAAAMoAAAIKgjwVkFwb3N0AAANQAAAAB0AAAAg/9EAMgADAioCvAAFAAACigJYAAAASwKKAlgAAAFeADIBKQAAAgsHAwMEAwICBGAAAvcAAAADAAAAAAAAAABBREJPACAAIP//Au7/BgAAA9gBESAAAZ8AAAAAAfAClAAAACAAA3icbMy5bcIAGEDhz7HjXE7iXM5ZZIEsEdG4hQksKkSDKJBYx+JYgJKCVZiAEX6Ea175FQ+JVIJCpkWllMr9+tNTG2iMTMzMI+j8X62vMTQ2PXns4xC72MYm1rGKZSyi7a7nSvy4kMpcyl25duPWncK9B49KT569ePWm8u7Dpy/fHAEAAP//AQAA///vvh5CAHicXFVbbCNXGf7P8W3jzCYZ38a38W3sGV9iJ/Z4PElsr+PEiTeJnZs32aTNZbtaYEuy2SibZdMVVR9YgShZraiDtCBBEaICpIJUVUilKCCQEETt27b0BdQieOq+RFVUIeGM0cw42WQf7PMy+v7v/y7ngA6mAfB1vAcaaINOMIEVgCf9ZIjnOMYg8qLIUBqRQ6RhGpukt37ORbSRiDbqe+y9v7qKqit473j9xer161+uZrPST373vvQQ3XkfQANVAJzDu0CCEwIyJp+y2awWvcGqHHpGw6cyQpplGJJPKWf109L6YDycGi5tlVdHMslUerT2Si5fw7v0aCFW69RenCgOX4mg70QZ1ictLMRCAAgSzSPcix+DC0AXYFkhncnwKRtlYFkmoNdbLTY+lREpPVqefb0293C2cMM/6RCZ7rHY/OVwwT45S1R+cGv9hzN8YIWiUytDNzaDjqVrgBX+FbwLRlWRFns9w/GpjMxbJvzejTdmph9di7v7aolErc+Nd0uPNjffKN8LL01OLoZA5lcFgM/wLmgUFLJax7vH2y38Gt6FDqDOqmNmOMZKnkhTfTqyVSoIe2+9OlMZyOcHKng3tDB5eZmS/vf0KbqW7O1l5RlM8wgb8WOIKhpwos2mAnBcAp8XxGqxUZS6CbIMvpa6wsyHE3E+NufPsdmXS32b0QnfIMfG+6NXsqMDG0Rv4iseNkB7aVOwo2e0J7OQ7o4uO1xet8dDBuxXRjJLfYAh2jxCH6EGOIABoAKyCaIyzsApw60kwzF6vZjKiILiyR9K0w/qmIl4B4NCz9rA6ld3jFpv+YIjZJ7MeYmrhcmFTj9nt75EBze2pP/wbmaLMl81xmg7pWgabB6hfdQA5/OeqxuqjuuRY+R28fI3Somye4TxCYVCrz1hHgjNE/m7s7XtvIdapSvFwaq185rPBQCAgWseoQbeBzP4TvZQgDmBP7PBiZBfLN3OrqYjfQ59fceodY5iO2cyxyxMpof43iszdy+57ZVfHQ8nncyOxfGBqWO4PDYCWOH+L9QAO3jPsVfc98vOydw1fFqegrzlraHh9Wx5uUeLpU+Mo0khk2RXfvQu1x3IEJe2Z2e2C4W1kjnUluH9i04PGogIPaBoZAdA2/hAPnmSEcTncmDlrQz5wtBQcHrYm+5yXXQSLs/iInr1ls4lzKcJ/bpO52c9d6RvyV0ONOPYgBrQA1kYV5RhhbQshGy0cLICxVuZVkkCnOKDbL1Fr9eoaVZEM7eSHWCVT74YWOkrm10+uzMysCJ0+387ZWhLL4i01xSITC+9VPrmOM1xNM1xkdQgF+IdfsKVf+Ls686FtRfDXleqS2sqxXJTYWKtPWDpHw8aO21mU3aYn0mgg2iEi4TDkahUDzqoLo3G7nDTqjZF2Wy8DxaleVbDSUhJhaWBLNYN7onUzFid9rnDdrz/9qIjtrYsfYj8mbCDkt6BZhNEAPgnfoJZkEEN4IHXT7E9eB8Ite+8yMudNliLj7Q//ulvfv/mZgHvSxt/+VD6x5/K9+Xvm0fIhPehU00cyZOnAf5bJVsn23QGvYkIES9OYOb4E8qE0C2dQZ2joVED/MocilfdPbeJ4fQsyv0aTQpFs388OT1Rp32hXvmvBx0OeuOxcCB5sl6v9E7rONEJNVo6tWac1WnHqPVVT4VChwVP/JxOat6V7HQ+d0MryeDOJAPZCrdLpduFwkaptFGIJxLxRDze6mp+uzZ7N3+vOlisyJWVcYvNy9iGGmAGDwD1jJ0SP5ajrErSmID83sg86THuhZu51Ywv59RNsZn5WNQSfg//Mulkvntnbqfgckx9HwVHK9+Of2DqaPmIHqEGmM7pq7ZH3dxVYa1uo/2io8udt6DDq6mkTveaVhtJSZ8BAmvzCL2JGsApvj67k1n1Tj4Fk29kD7Za9E+SX2OHAgWv30MnnJ5s+OW5/qveIWfa2d/P+vKRmwTrXXK4KDNpMxuJYH9kZJ6zL1hsnN3R0c70J4aX1WyTzSO0gbflV0UXYAWBEUSRl9t+5mKEpalShbx/7x5DEw4jZRaJr88f3NI/eHDnr9GQXrumJ1SsXPMI/Rcdyv6fyybZug7/PjNW9/jcrK2+067xjhNryygtfSpEnDS6LHWNhLoByT1ATXQIFwF4DU/ZbLKUoshr3v3F3qDRbNS2mY3Fhz9Dh5+HqhxXDX0udSmzieYldIwO5dQ8008Uz0F04B2bv9NpMF0IhY2GP+6V201G7QWyLffwbapv6s967SbSBWkn+vfHgdEQU2Y+ltovzUXV3WJwgPwoKb/LosBbY18e3LzZ8h0+Qocn73Wxjg6lLkDNX+N+qOEn0A5AKi+DGrZQIhEKJRK4P8owUfkH/wcAAP//AQAA//+6SuK5AAABAAAAAguFRsQPWV8PPPUAAQPoAAAAANhdoIQAAAAA3WYvNv43/sQIbQPxAAEAAwACAAAAAAAAAAEAAAPY/u8AAAiY/jf+NwhtAAEAAAAAAAAAAAAAAAAAAAAfArIAUADIAAACXQBNAkYALgJ7AE0BLQBNAmUATQIsACMCDwAqAdMAJAI9ACcCBgAkAVUAGAIWACICOwBBARQANwIkAEEBHgBBA1kAQQI8AEECKwAkAj0AQQGOAEEBuwAVAX8AEQI4ADwCCwAMAgkADAFMACsBFABBAAD/rQAAACwALABgAIwAsAC8AOIBIgFaAYYBuAHsAhICegKcAqgCwALcAw4DMANcA4wDrAPoBA4EMARMBHwEiASUBKoAAQAAAB8AkAAMAGMABwABAAAAAAAAAAAAAAAAAAQAA3icnJTPbhtVFMZ/TmzTCsECRVW6ie6CRZHo2FRJ1TYrh9SKRRQHjwtCQkgTz/iPMp4ZeSYO4QlY8xa8RVc8BM+BWKP5fOzYBdEmipJ8d+75851zvnOBHf5mm0r1IfBHPTFcYa9+bniLB/UTw9u061uGqzyp/Wm4RlibG67zea1n+CPeVn8z/ID96k+GH7JbbRv+mGfVHcOfbDv+Mvwp+7xd4Aq84FfDFXbJDG+xw4+Gt3mExaxUeUTTcI3P2DNcZw/oM6EgZkLCCMeQCSOumBGR4xMxY8KQiBBHhxYxhb4mBEKO0X9+DfApmBEo4pgCR4xPTEDO2CL+Iq+Uc2Uc6jSzuxYFYwIu5HFJQIIjZURKQsSl4hQUZLyiQYOcgfhmFOR45EyI8UiZMaJBlzan9BkzIcfRVqSSmU/KkIJrAuV3ZlF2ZkBEQm6srkgIxdOJXyTvDqc4umSyXY98uhHhSxzfybvklsr2Kzz9ujVmm3mXbALm6mesrsS6udYEx7ot87b4VrjgFe5e/dlk8v4ehfpfKPIFV5p/qEklYpLg3C4tfCnId49xHOncwVdHvqdDnxO6vKGvc4sePVqc0afDa/l26eH4mi5nHMujI7y4a0sxZ/yA4xs6siljR9afxcQifiYzdefiOFMdUzL1vGTuqdZIFd59wuUOpRvqyOUz0B6Vlk7zS7RnASNTRSaGU/VyqY3c+heaIqaqpZzt7X25DXPbveUW35Bqh0u1LjiVk1swet9UvXc0c60fj4CQlAtZDEiZ0qDgRrzPCbgixnGs7p1oSwpaK58yz41UEjEVgw6J4szI9Dcw3fjGfbChe2dvSSj/kunlqqr7ZHHq1e2M3qh7yzvfuhytTaBhU03X1DQQ18S0H2mn1vn78s31uqU85YiUmPBfL8AzPJrsc8AhY2UY6GZur0NTL0STlxyq+ksiWQ2l58giHODxnAMOeMnzd/q4ZOKMi1txWc/d4pgjuhx+UBUL+y5HvF59+/+sv4tpU7U4nq5OL+49xSd3UOsX2rPb97KniZWTmFu02604I2BacnG76zW5x3j/AAAA//8BAAD///S3T1F4nGJgZgCD/+cYjBiwAAAAAAD//wEAAP//LwECAwAAAA==");
}
.d2-1877477521 .text-italic {
	font-family: "d2-1877477521-font-italic";
}
@font-face {
	font-family: d2-1877477521-font-italic;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA10AAoAAAAAFSwAARhRAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgW1SVeGNtYXAAAAFUAAAAjwAAAK4C9gNeZ2x5ZgAAAeQAAAcOAAAJ5MEOKANoZWFkAAAI9AAAADYAAAA2G7Ur2mhoZWEAAAksAAAAJAAAACQLeAjDaG10eAAACVAAAAB8AAAAfDT1A7Jsb2NhAAAJzAAAAEAAAABAJ6YqMG1heHAAAAoMAAAAIAAAACAANwD2bmFtZQAACiwAAAMmAAAIMgntVzNwb3N0AAANVAAAACAAAAAg/8YAMgADAeEBkAAFAAACigJY//EASwKKAlgARAFeADIBIwAAAgsFAwMEAwkCBCAAAHcAAAADAAAAAAAAAABBREJPAAEAIP//Au7/BgAAA9gBESAAAZMAAAAAAeYClAAAACAAA3icbMy5bcIAGEDhz7HjXE7iXM5ZZIEsEdG4hQksKkSDKJBYx+JYgJKCVZiAEX6Ea175FQ+JVIJCpkWllMr9+tNTG2iMTMzMI+j8X62vMTQ2PXns4xC72MYm1rGKZSyi7a7nSvy4kMpcyl25duPWncK9B49KT569ePWm8u7Dpy/fHAEAAP//AQAA///vvh5CAHicfFZbbBtZGf7PmclMLk4ae3ypndjjeMbj2/g2Y8/EcWzHcRLnYqdNUofQ1GmzvdBlSwntwnZpq4VWWi5ilyBVSKCVFgkhLeoDUndfeAFpxUMEqgRShRYtWgmWTVHLasGyELsiYzR2Gjt54GV8Xvx/5/vO933nQBfwAPgqvgcE9MAxMIEFQGZGCEJWVc5GyD4fR9Oqj2Fo/g7aufNjsnD6I/9PPhVZsvjNny/849x9fG/vCvpG9ZVXtPVvX7z4uadPtSD641MAAAI4AOzG22AEh76WGVmyWswURdPW5i9HyJKSTAhce8Hd/cXZK6ECj+SZ4q3FsY2N09Pz6y9c27hannsRb88XxSmxmzTkR+eqIvpaUQ1Le0+mS1JGx0OQatRxGL8BLECXRxCSiSyWJauNFgTOM4AtZqtVlhTVRlHIs3BZiZ2+XRpdOq4wijB2dpL3zKf9BTfHVw2FG4vley8V1WDA7ctcuDGeribdQxIbBgDc5KTgbegBpoMR55Ml5RmDr3/ntcqbX15drdwqfOG8gre/9fJL71ycOPXDzerzrX1yAFDH20A0JxDc3cW7eHtv62C+D29DP1jb82mGIziCaYv0zpmvXF25vnLlmjp1fuPCwuw5vD2zsn7VqH2IrNoTVFmeUaItLEOjjjT8BgQBbB7BpzY1SCYEn08XSFEOBKIoi9lqs7VO5nFhy59yVtTxpbC3FEwnz6TT51jZPhPxJp1xvhRNpC8ZxsZCIWlqlJesEcecKi1LCX/EFWBjQ0LUGh4uqmPrCcDga9TRf1ANzDozm+fgUGRVJjiVoyifpKjqwQm9PVES5zdkX8ZIMtnNXDfJrZmEE7xokYb5QpKNG9YrMy+fkf0jGc0x641ORKJ/EjzBuaqU2/eAt1FHD1ANhg+htRnuO+C9ExfE8mZSHLeGGcEZW1VSY27F6nGUDZeqU9crUY89ZrNMbRUmZxxGyeyF5tn4GnXswztg0VNziMv/JzNmIgaF8vY+m0XvUTY+99lf7Y0epYObXH6NauAAbyde0xEj1IGbCblpPZ3h31afDy+cial5l6FL+02PuxB0pmwu59KPGpgwBbjkhuGLm9Nby2LkpDQsD+ROeu1G2cIib9/x/uE4WwEEIQD0Gn4ENt19XA53OoSmZZojQpVcX37w2GLGETQN9Q4ZRwLdxucM5yvorVTX0vxKf59K90qhlay2pmuGGjyqoRqwEOl0oKpSFHfYDRRFHFLvfnyV44en/dn5AbtwKpo5GZo7ExeyRoLJXWKup7glT8gaH+bysiv6Z8GZtHlKE5cFcbVSePHzku4P4uwlNBIK/l7wBGbWYul0yx8sAHoP74C9mS6alpsELWaa4BhdRp0mwb5ejg2SgWUxm+zOlsZJcnZ4NjKNd55muGh+lOW13yLRfLx/IRjR3mo09JnwGX6ABXACAAWu2TbWx3gHDK2c63gM56Np9vXyOfzp2rtfXaxuOfCO5kTod9pHH1+7CQjERh0+wztg0tVKJlRGF8Zi3j/qL+Wpm+XbCBkJika9VkPOaMcv7P2A7iFMCKdJ8gAXP0E1PfM6ZouibZ8odYhpJ+nNHE0KK8JYvCu65s0oJJktZ0iyaJkVp3UNZqyzoWm0O8fHVb8o50eNLnOnDu1VW2dUg+Odezgqs44YWI4cUrmJcFTkg/yh91ENjuk6t/NgMQ9g3379tkL+6MSGOL8hnTgrLmwEw0uyIukfw+X16euVSOs7Mbk1NVksbE1NzuizG/9uyOifqNbKNt2x4wHMeQT9FmSkLG5B6PdX73dzFOGtRJoRl4RxBpvYn/GFpCsW8CxxEbP8EL89wYb3A85efhOh4FxVzmaCwt+9I21/3EI1GOzQyEYLz7TpI52lsN0yNOjgS2wG7VbFTM9Udy6tPQTU+G+jjm6jGviO9vrRWtdbvVXqP41X7THbhBDMBEYjKXFOjMwPRxh5RIgr7mwitmxI+AXWH+EcPtaRDYTyXt7lNzvCrEswecbF8JRX3/N4o47W8JWDflVUvSXkZjN09OsvJxIkShX7Snx+6KbhdooY9gw4+oyDUUMufMzRj0yprldfzWpPTCaXq7dLpY/ps0cbdfQJ2tWz+Wx22/3MfsXeP3DmrLMoTpf0S8J/yjCpGlkGKdojxq5bBq1pjnlObumcBkB/QbvQD6Cn0Gq1yYo+EN0plniSIkkjz3y/rO2hXe0xt8Dxczyya47mfxvvNqLoQ7QLDgC6qXOzvA5NGcBUr3vAbjJ583bTSkno6iZIo9f0vZL2V3t69g80nerJSBx6rH0yUua4kgcZ9/4VLYvP3gIfoF5k198CqirTnOH9/g86ugoeot1n7wR2s/wc2m1uDEERL8AD/AD6AJimVq1w32BcnM3s5PCCzWofOW61u/8HAAD//wEAAP//xcr+iQAAAAEAAAABGFFLgT9PXw889QABA+gAAAAA2F2gzAAAAADdZi83/r3+3QgdA8kAAgADAAIAAAAAAAAAAQAAA9j+7wAACED+vf28CB0D6ADC/9EAAAAAAAAAAAAAAB8CdAAkAMgAAAJHACMCJgA5AlAAIwD8ACMCKwAjAfoADAIZACcBswAlAhcAJwHhACUBGgArAhMAAQILAB8A7QAfAdwAHwD4ACwDHwAfAg0AHwIDACcCF//2AVYAHwGS//wBRQA8AhAAOAHAADsBwP/CASsAIwDtAB8AAABHAAAALgAuAGYAmAC6AMgA8AEwAWgBlgHOAggCMAJ4AqICrgLIAuoDLANWA4QDvgPcBBgERgRyBJAEwATOBNwE8gABAAAAHwCMAAwAZgAHAAEAAAAAAAAAAAAAAAAABAADeJyclNtOG1cUhj8H2216uqhQRG7QvkylZEyjECXhypSgjIpw6nF6kKpKgz0+iPHMyDOYkifodd+ib5GrPkafoup1tX8vgx1FQSAE/Hv2OvxrrX9tYJP/2KBWvwv83ZwbrrHd/NnwHb5oHhneYL/5meE6Dxv/GG4waLw13ORBo2v4E97V/zT8KU/qvxm+y1b90PDnPK5vGv5yw/Gv4a94wrsFrsEz/jBcY4vC8B02+dXwBvewmLU699gx3OBrtg032QZ6TKhImZAxwjFkwogzZiSURCTMmDAkYYAjpE1Kpa8ZsZBj9MGvMREVM2JFHFPhSIlIiSkZW8S38sp5rYxDnWZ216ZiTMyJPE6JyXDkjMjJSDhVnIqKghe0aFHSF9+CipKAkgkpATkzRrTocMgRPcZMKHEcKpJnFpEzpOKcWPmdWfjO9EnIKI3VGRkD8XTil8g75AhHh0K2q5GP1iI8xPGjvD23XLbfEujXrTBbz7tkEzNXP1N1JdXNuSY41q3P2+YH4YoXuFv1Z53J9T0a6H+lyCecaf4DTSoTkwzntmgTSUGRu49jX+eQSB35iZAer+jwhp7Obbp0aXNMj5CX8u3QxfEdHY45kEcovLg7lGKO+QXH94Sy8bET689iYgm/U5i6S3GcqY4phXrumQeqNVGFN5+w36F8TR2lfPraI2/pNL9MexYzMlUUYjhVL5faKK1/A1PEVLX42V7d+22Y2+4tt/iCXDvs1brg5Ce3YHTdVIP3NHOun4CYATknsuiTM6VFxYV4vybmjBTHgbr3SltS0b708XkupJKEqRiEZIozo9Df2HQTGff+mu6dvSUD+Xump5dV3SaLU6+uZvRG3VveRdblZGUCLZtqvqKmvrhmpv1EO7XKP5Jvqdct5xGh4i52+0OvwA7P2WWPsbL0dTO/vPOvhLfYUwdOSWQ1lKZ9DY8J2CXgKbvs8pyn7/VyycYZH7fGZzV/mwP26bB3bTUL2w77vFyL9vHMf4ntjupxPLo8Pbv1NB/cQLXfaN+u3s2uJuenMbdoV9txTMzUc3FbqzW5+wT/AwAA//8BAAD//3KhUUAAAAADAAD/9QAA/84AMgAAAAAAAAAAAAAAAAAAAAAAAAAA");
}]]></style><style type="text/css"><![CDATA[.shape {
  shape-rendering: geometricPrecision;
  stroke-linejoin: round;
}
.connection {
  stroke-linecap: round;
  stroke-linejoin: round;
}
.blend {
  mix-blend-mode: multiply;
  opacity: 0.5;
}

		.d2-1877477521 .fill-N1{fill:#0A0F25;}
		.d2-1877477521 .fill-N2{fill:#676C7E;}
		.d2-1877477521 .fill-N3{fill:#9499AB;}
		.d2-1877477521 .fill-N4{fill:#CFD2DD;}
		.d2-1877477521 .fill-N5{fill:#DEE1EB;}
		.d2-1877477521 .fill-N6{fill:#EEF1F8;}
		.d2-1877477521 .fill-N7{fill:#FFFFFF;}
		.d2-1877477521 .fill-B1{fill:#0D32B2;}
		.d2-1877477521 .fill-B2{fill:#0D32B2;}
		.d2-1877477521 .fill-B3{fill:#E3E9FD;}
		.d2-1877477521 .fill-B4{fill:#E3E9FD;}
		.d2-1877477521 .fill-B5{fill:#EDF0FD;}
		.d2-1877477521 .fill-B6{fill:#F7F8FE;}
		.d2-1877477521 .fill-AA2{fill:#4A6FF3;}
		.d2-1877477521 .fill-AA4{fill:#EDF0FD;}
		.d2-1877477521 .fill-AA5{fill:#F7F8FE;}
		.d2-1877477521 .fill-AB4{fill:#EDF0FD;}
		.d2-1877477521 .fill-AB5{fill:#F7F8FE;}
		.d2-1877477521 .stroke-N1{stroke:#0A0F25;}
		.d2-1877477521 .stroke-N2{stroke:#676C7E;}
		.d2-1877477521 .stroke-N3{stroke:#9499AB;}
		.d2-1877477521 .stroke-N4{stroke:#CFD2DD;}
		.d2-1877477521 .stroke-N5{stroke:#DEE1EB;}
		.d2-1877477521 .stroke-N6{stroke:#EEF1F8;}
		.d2-1877477521 .stroke-N7{stroke:#FFFFFF;}
		.d2-1877477521 .stroke-B1{stroke:#0D32B2;}
		.d2-1877477521 .stroke-B2{stroke:#0D32B2;}
		.d2-1877477521 .stroke-B3{stroke:#E3E9FD;}
		.d2-1877477521 .stroke-B4{stroke:#E3E9FD;}
		.d2-1877477521 .stroke-B5{stroke:#EDF0FD;}
		.d2-1877477521 .stroke-B6{stroke:#F7F8FE;}
		.d2-1877477521 .stroke-AA2{stroke:#4A6FF3;}
		.d2-1877477521 .stroke-AA4{stroke:#EDF0FD;}
		.d2-1877477521 .stroke-AA5{stroke:#F7F8FE;}
		.d2-1877477521 .stroke-AB4{stroke:#EDF0FD;}
		.d2-1877477521 .stroke-AB5{stroke:#F7F8FE;}
		.d2-1877477521 .background-color-N1{background-color:#0A0F25;}
		.d2-1877477521 .background-color-N2{background-color:#676C7E;}
		.d2-1877477521 .background-color-N3{background-color:#9499AB;}
		.d2-1877477521 .background-color-N4{background-color:#CFD2DD;}
		.d2-1877477521 .background-color-N5{background-color:#DEE1EB;}
		.d2-1877477521 .background-color-N6{background-color:#EEF1F8;}
		.d2-1877477521 .background-color-N7{background-color:#FFFFFF;}
		.d2-1877477521 .background-color-B1{background-color:#0D32B2;}
		.d2-1877477521 .background-color-B2{background-color:#0D32B2;}
		.d2-1877477521 .background-color-B3{background-color:#E3E9FD;}
		.d2-1877477521 .background-color-B4{background-color:#E3E9FD;}
		.d2-1877477521 .background-color-B5{background-color:#EDF0FD;}
		.d2-1877477521 .background-color-B6{background-color:#F7F8FE;}
		.d2-1877477521 .background-color-AA2{background-color:#4A6FF3;}
		.d2-1877477521 .background-color-AA4{background-color:#EDF0FD;}
		.d2-1877477521 .background-color-AA5{background-color:#F7F8FE;}
		.d2-1877477521 .background-color-AB4{background-color:#EDF0FD;}
		.d2-1877477521 .background-color-AB5{background-color:#F7F8FE;}
		.d2-1877477521 .color-N1{color:#0A0F25;}
		.d2-1877477521 .color-N2{color:#676C7E;}
		.d2-1877477521 .color-N3{color:#9499AB;}
		.d2-1877477521 .color-N4{color:#CFD2DD;}
		.d2-1877477521 .color-N5{color:#DEE1EB;}
		.d2-1877477521 .color-N6{color:#EEF1F8;}
		.d2-1877477521 .color-N7{color:#FFFFFF;}
		.d2-1877477521 .color-B1{color:#0D32B2;}
		.d2-1877477521 .color-B2{color:#0D32B2;}
		.d2-1877477521 .color-B3{color:#E3E9FD;}
		.d2-1877477521 .color-B4{color:#E3E9FD;}
		.d2-1877477521 .color-B5{color:#EDF0FD;}
		.d2-1877477521 .color-B6{color:#F7F8FE;}
		.d2-1877477521 .color-AA2{color:#4A6FF3;}
		.d2-1877477521 .color-AA4{color:#EDF0FD;}
		.d2-1877477521 .color-AA5{color:#F7F8FE;}
		.d2-1877477521 .color-AB4{color:#EDF0FD;}
		.d2-1877477521 .color-AB5{color:#F7F8FE;}.appendix text.text{fill:#0A0F25}.md{--color-fg-default:#0A0F25;--color-fg-muted:#676C7E;--color-fg-subtle:#9499AB;--color-canvas-default:#FFFFFF;--color-canvas-subtle:#EEF1F8;--color-border-default:#0D32B2;--color-border-muted:#0D32B2;--color-neutral-muted:#EEF1F8;--color-accent-fg:#0D32B2;--color-accent-emphasis:#0D32B2;--color-attention-subtle:#676C7E;--color-danger-fg:red;}.sketch-overlay-B1{fill:url(#streaks-darker-d2-1877477521);mix-blend-mode:lighten}.sketch-overlay-B2{fill:url(#streaks-darker-d2-1877477521);mix-blend-mode:lighten}.sketch-overlay-B3{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-B4{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-B5{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-B6{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-AA2{fill:url(#streaks-dark-d2-1877477521);mix-blend-mode:overlay}.sketch-overlay-AA4{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-AA5{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-AB4{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-AB5{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-N1{fill:url(#streaks-darker-d2-1877477521);mix-blend-mode:lighten}.sketch-overlay-N2{fill:url(#streaks-dark-d2-1877477521);mix-blend-mode:overlay}.sketch-overlay-N3{fill:url(#streaks-normal-d2-1877477521);mix-blend-mode:color-burn}.sketch-overlay-N4{fill:url(#streaks-normal-d2-1877477521);mix-blend-mode:color-burn}.sketch-overlay-N5{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-N6{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.sketch-overlay-N7{fill:url(#streaks-bright-d2-1877477521);mix-blend-mode:darken}.light-code{display: block}.dark-code{display: none}]]></style><g class="dHJhZGl0aW9uYWw="><g class="shape" ><rect x="66.000000" y="12.000000" width="188.000000" height="98.000000" stroke="#0D32B2" fill="#fee2e2" class=" stroke-B1" style="stroke-width:2;" /></g><text x="160.000000" y="50.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="160.000000" dy="0.000000">Block the device</tspan><tspan x="160.000000" dy="17.666667">Remove the account</tspan><tspan x="160.000000" dy="17.666667">Detect after the fact</tspan></text></g><g class="YWdlbnR3YXJl"><g class="shape" ><rect x="12.000000" y="271.000000" width="296.000000" height="98.000000" stroke="#0D32B2" fill="#dcfce7" class=" stroke-B1" style="stroke-width:2;" /></g><text x="160.000000" y="309.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="160.000000" dy="0.000000">Break the trust signal</tspan><tspan x="160.000000" dy="17.666667">Create friction during manipulation</tspan><tspan x="160.000000" dy="17.666667">Intervene in real-time</tspan></text></g><g class="KHRyYWRpdGlvbmFsIC0mZ3Q7IGFnZW50d2FyZSlbMF0="><marker id="mk-d2-1877477521-3488378134" markerWidth="10.000000" markerHeight="12.000000" refX="7.000000" refY="6.000000" viewBox="0.000000 0.000000 10.000000 12.000000" orient="auto" markerUnits="userSpaceOnUse"> <polygon points="0.000000,0.000000 10.000000,6.000000 0.000000,12.000000" fill="#0D32B2" class="connection fill-B1" stroke-width="2" /> </marker><path d="M 160.000000 112.000000 L 160.000000 267.000000" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-1877477521-3488378134)" mask="url(#d2-1877477521)" /><text x="160.500000" y="196.000000" fill="#676C7E" class="text-italic fill-N2" style="text-anchor:middle;font-size:16px">Shift the defense layer</text></g><mask id="d2-1877477521" maskUnits="userSpaceOnUse" x="-9" y="-9" width="338" height="399">
<rect x="-9" y="-9" width="338" height="399" fill="transparent"></rect>
<rect x="85.000000" y="180.000000" width="151" height="21" fill="black"></rect>
</mask></svg></svg>
" class="img-fluid" style="width:100.0%"></p>
<p><strong>Mandatory unverified caller overlays.</strong> Any account without a verified identity calling via video should show a persistent banner that cannot be dismissed: “Unverified account. Government agencies do not contact citizens via WhatsApp video call.” Not a one-time popup. A persistent overlay visible for the entire call duration.</p>
<p><strong>Screen-share lockouts on government-branded calls.</strong> If an account impersonating a government agency initiates screen sharing — a common tactic for watching victims navigate their banking apps — the session should be terminated. Not warned. Terminated.</p>
<p><strong>Banking app friction triggers.</strong> When a user accesses a banking app while on an active video call, the banking app should show a high-friction warning: “You are on an active call. Fraud alerts frequently occur during video calls. Are you being asked to transfer money or share a code?” This requires coordination between WhatsApp and banks, but both the data and the intent exist.</p>
<p><strong>Timed lock-out on prolonged calls.</strong> Digital arrest scams often keep victims on calls for hours, maintaining the psychological pressure. Automatic warnings after 30 minutes of continuous video call — “You have been on this call for 30 minutes. If someone is asking you to take financial actions, this may be fraud” — break the trance.</p>
<p>None of these are technically complex. All of them require friction that product teams will resist because friction reduces engagement metrics. That’s the real barrier.</p>
</section>
<section id="the-measurement-problem" class="level2">
<h2 class="anchored" data-anchor-id="the-measurement-problem">The Measurement Problem</h2>
<p>The Home Ministry is measuring the wrong thing. Device ID blocks are measurable: accounts blocked, devices flagged, scammers inconvenienced. Logo detection is measurable: impersonator profiles removed, false positives reviewed.</p>
<p>What’s not being measured: how many users completed a video call with an unverified account and then opened their banking app within 10 minutes. How many users stayed on a video call for over 30 minutes while simultaneously accessing financial services. How many users shared their screen during a call with a government-branded profile.</p>
<p>These are the signals that predict a scam in progress. They’re also the signals that require real-time intervention, not post-facto account removal.</p>
<p>WhatsApp has the telemetry. Banks have the transaction data. The coordination layer doesn’t exist.</p>
</section>
<section id="what-i-dont-know-yet" class="level2">
<h2 class="anchored" data-anchor-id="what-i-dont-know-yet">What I Don’t Know Yet</h2>
<p>I don’t know how to build organizational trust in autonomous intervention systems that terminate calls or lock banking apps based on behavioral signals. The false positive rate for “user on video call + opens banking app” is probably high. Legitimate customer service calls exist. Remote financial assistance from family members exists.</p>
<p>The threshold between protective friction and user-hostile interference is real. I haven’t solved it. Neither has anyone else.</p>
<p>But I know this: device ID blocking is solving the wrong problem. The scam doesn’t live in the hardware. It lives in the 6-inch screen where a terrified user sees a uniform and hears the word “arrest” and does exactly what they’re told.</p>
<p>The question worth asking now is whether the platforms that control that screen are willing to break their own engagement metrics to intervene. Not in three years. This quarter.</p>
<p>Are we asking it? Mostly, no. We are still blocking device IDs.</p>


</section>

 ]]></description>
  <category>Product Security</category>
  <category>Interface Design</category>
  <category>India</category>
  <guid>https://talvinder.com/field-notes/digital-arrest-device-blocking-ui-problem/</guid>
  <pubDate>Sun, 22 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Model Routing Is the New Unit Economics</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/small-model-arbitrage/</link>
  <description><![CDATA[ 





<p>Most teams are paying frontier model prices for commodity model work.</p>
<p>They default to GPT-4 or Claude Opus for tasks that a $0.10 per million token model could handle at 95% accuracy. The gap between what these models cost and what they’re actually needed for is the arbitrage opportunity of the next 18 months.</p>
<p>At Ostronaut, we generate training content at scale: presentations, quizzes, video scripts. We started with GPT-4 for everything. Cost per generation: $0.03. We moved structured extraction and template filling to GPT-4o-mini. Cost dropped to $0.015. Same user satisfaction scores. Half the cost.</p>
<p>The arbitrage isn’t about being cheap. It’s about understanding where model capability stops mattering to the outcome.</p>
<section id="inference-cac-compounds-like-customer-acquisition-cost" class="level2">
<h2 class="anchored" data-anchor-id="inference-cac-compounds-like-customer-acquisition-cost">Inference CAC Compounds Like Customer Acquisition Cost</h2>
<p>Call this <strong>Inference CAC</strong>: the cost to acquire value from each model call.</p>
<p>Just like customer acquisition cost, it’s a unit economic that compounds. If you’re running 10M inferences a month, a 50% reduction in per-call cost is $150K annual savings. That’s not rounding error. That’s headcount.</p>
<p>The shift happening now: AI products are moving from “can we do this?” to “can we do this profitably?” The companies that figure out model selection as a core competency will have better margins than competitors running everything through Opus.</p>
<p>This is not about performance. It’s about matching performance to the value threshold of the task.</p>
</section>
<section id="the-default-to-frontier-habit-is-a-margin-killer" class="level2">
<h2 class="anchored" data-anchor-id="the-default-to-frontier-habit-is-a-margin-killer">The Default-to-Frontier Habit Is a Margin Killer</h2>
<p>The default behavior in 2024-2025 was to use the best available model. GPT-4, Claude Opus, whatever scored highest on benchmarks. The logic made sense early: you’re prototyping, you want maximum capability, cost is secondary to learning if the feature works.</p>
<p>But that logic breaks once you’re in production. Once you’re processing thousands or millions of requests. Once the feature is validated and you’re optimizing for margin.</p>
<p>Here’s the pattern I see across teams: 80% of their AI tasks don’t need frontier model reasoning. They need reliable extraction, simple classification, template completion, or pattern matching. Tasks where a 90% accurate model and a 95% accurate model produce the same user outcome.</p>
<p>The performance plateau is real. If you’re extracting structured data from invoices, GPT-4’s reasoning capability is overkill. If you’re triaging support tickets into five categories, you don’t need multi-step reasoning. If you’re generating quiz questions from a content outline, you need consistency and format compliance, not creativity.</p>
<p>The companies that will win the next phase are the ones building <strong>model portfolios</strong>, not model lock-in. They route requests to the cheapest model that clears the quality bar for that specific task. Frontier models for complex reasoning. Mid-tier models for structured tasks. Small models for high-volume, low-complexity work.</p>
<p>This requires a different kind of product thinking. You need to know:</p>
<ul>
<li>What is the minimum acceptable quality for this feature?</li>
<li>What does quality mean here — accuracy, consistency, format compliance, creativity?</li>
<li>What’s the cost per request at different model tiers?</li>
<li>What’s the volume, and how does that change unit economics?</li>
</ul>
<p>Most teams can’t answer these questions. They pick a model, ship the feature, and never revisit the decision.</p>
<p>Here’s a claim: <strong>By 2027, any AI product doing more than 1M inferences/month that hasn’t implemented model routing will have 30-50% worse margins than competitors who have.</strong> The gap will be structural. It won’t be about better features. It will be about better cost discipline.</p>
</section>
<section id="the-math-is-already-visible" class="level2">
<h2 class="anchored" data-anchor-id="the-math-is-already-visible">The Math Is Already Visible</h2>
<p>Cursor’s pricing page tells you something: “Claude Opus is extremely expensive, so my recommendation not to use it, unless the company pays for it.” They’re already pushing users toward cost-aware model selection. The tool that’s supposed to make you more productive is teaching you to ration the expensive model.</p>
<p>That’s the canary. When dev tools start warning you about model costs, it means the unit economics are real enough to matter.</p>
<p>Look at SaaS unit economics. If your CAC is $1,800 and your annual contract value is $1,500, you’re underwater. You optimize CAC or increase ACV. Same logic applies to inference costs. If your cost per inference is $0.05 and your revenue per user per month is $20, you need to drive down inference cost or increase revenue.</p>
<p>Most teams will find it easier to optimize the cost side first.</p>
<p>The math is simple:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 11%">
<col style="width: 30%">
<col style="width: 30%">
<col style="width: 27%">
</colgroup>
<thead>
<tr class="header">
<th>Volume</th>
<th>Frontier Model Cost</th>
<th>Mid-tier Model Cost</th>
<th>Annual Difference</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>1M requests/month</td>
<td>$50K/month</td>
<td>$20K/month</td>
<td>$360K/year</td>
</tr>
<tr class="even">
<td>5M requests/month</td>
<td>$250K/month</td>
<td>$100K/month</td>
<td>$1.8M/year</td>
</tr>
<tr class="odd">
<td>10M requests/month</td>
<td>$500K/month</td>
<td>$200K/month</td>
<td>$3.6M/year</td>
</tr>
</tbody>
</table>
<p>That’s the arbitrage. Find the 60-80% of your requests that don’t need frontier models. Route them to cheaper models. Bank the difference.</p>
</section>
<section id="model-selection-is-a-feature-level-decision" class="level2">
<h2 class="anchored" data-anchor-id="model-selection-is-a-feature-level-decision">Model Selection Is a Feature-Level Decision</h2>
<p>At Ostronaut, we built a multi-agent system for content generation. Initially, every agent used GPT-4. The cost per generation was $0.03. Acceptable for early customers, unsustainable at scale.</p>
<p>We audited every agent. Which tasks required reasoning? Which were template-filling? Which were format validation?</p>
<p>We moved structured extraction, template population, and rule-based validation to GPT-4o-mini. We kept GPT-4 for content composition and quality evaluation — the tasks where reasoning and creativity mattered.</p>
<p>Cost per generation dropped 50%. Quality scores stayed flat. We didn’t lose customers. We didn’t get more complaints. The cheaper model was good enough for those tasks.</p>
<p>The lesson: <strong>model selection is a feature-level decision, not a product-level decision.</strong> You don’t pick one model for your product. You pick the right model for each task within your product.</p>
<p><img src="https://talvinder.com/field-notes/small-model-arbitrage/data:image/svg+xml;base64,<?xml version="1.0" encoding="utf-8"?><svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" data-d2-version="v0.7.1" preserveAspectRatio="xMinYMin meet" viewBox="0 0 619 539"><svg class="d2-580756557 d2-svg" width="619" height="539" viewBox="-9 -9 619 539"><rect x="-9.000000" y="-9.000000" width="619.000000" height="539.000000" rx="0.000000" fill="transparent" class=" fill-N7" stroke-width="0" /><style type="text/css"><![CDATA[
.d2-580756557 .text-bold {
	font-family: "d2-580756557-font-bold";
}
@font-face {
	font-family: d2-580756557-font-bold;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA9UAAoAAAAAF0wAAguFAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgXxHXrmNtYXAAAAFUAAAAqwAAANgDvwT1Z2x5ZgAAAgAAAAifAAALrCiBcUBoZWFkAAAKoAAAADYAAAA2G38e1GhoZWEAAArYAAAAJAAAACQKfwXnaG10eAAACvwAAACaAAAAoE6cBzZsb2NhAAALmAAAAFIAAABSP5484m1heHAAAAvsAAAAIAAAACAAQAD3bmFtZQAADAwAAAMoAAAIKgjwVkFwb3N0AAAPNAAAAB0AAAAg/9EAMgADAioCvAAFAAACigJYAAAASwKKAlgAAAFeADIBKQAAAgsHAwMEAwICBGAAAvcAAAADAAAAAAAAAABBREJPACAAIP//Au7/BgAAA9gBESAAAZ8AAAAAAfAClAAAACAAA3icdM05SgMBGEDhb5xxH3Xct1HHDRS08waC2AgiXmAQEbEXj+R2AC1CLpBL5ABpUwX+kOnz2q94SKQS5DIdlAqpXOXcpSvXbt17VHv26t1HBCpnLhq9cedB7cmLt5FGNwbRj160oxX/8Re/8RPf8RWfzWd8iROnjhw7NCGVmTRl2oxZc+blFixaUli2YtWadRs2bdm2Y1dpz74DFUMAAAD//wEAAP//MlwmnwB4nIRWa2zb5hW93ydajBXGMUVRlGTrSYvUw5Zt0RT9jKxY8iOR/ErjuKtsJ0bbPJw4XuKsXpe1+5EWWKesLZyuzoq13bBiD6wDimxA180DNmDYguZfiuXP+kSxFS3WeoVRtKtNDR8pO042YD9EAgJ57rn3nnP4QRWMAuBZfAUsUA17wQ48gMIG2bAiyyKtKZomChZNRiw9iu36Sz+Wo1Q0SsUCK/6HZ2ZQYRpf2Tx9X2F29rOZri79+d+8pl9G518DQJAor+MWvAJ1AFUhSVLbUikl6RRoSRJDVivvcCrJlCZY0dT4E4cOXx5P3x8cdmti41B8YjCSdg2PM/lnzpy+OqaEpgVvcnr//QsN7uJRQFAAQB/gElQbPPkgr/AiX0DP6Z+/9x4uXXz24ibAVv0sXgH//6pfKa+KqsJarejMvU8fPvLUkYEHAwV3eyx/tHifQ2JOfxz6aoVEW3Da6VuYvX/BZltY0t8IJkwe+DAuwW7CQ+EVVWFFVmQLy+9eufIuLn355eYiqtXXDC7k2ZdxiXBRWIVzOgUlldI4hRUJLU2kaVGWRR/m+cIPT9rsNsrG2o6/+DhdbaHUqbGpNoraReOS/lb9Pp9vXz0KbS5+EhgZ9T/3xRfP+UdHAp9s1cjjEnBmDUGRJJVwssii08nzhWd/3ktRNSVyq9qDS/rvnmr7VuffNxdR7rupi53/AABszOwRvAJ779qak3dYrXIypbZVhocmJh87ePCxSfPaNzzc1zc8zIxfPTX3zMjI906dujr+6OLs7Pz87OwiwSXcWnAJGHAY7CqIosizSpKAioV3Bi/09y/mxgaXeruzuCQXR/KzzW+i8RNKDGAL4xAuQQ0IOzBoTpQNlJQJ81HuXDatXnnpm2P5zp6ezjwuhSeHB6cE/cuPPkJHW1taJDIrsbyObXgFYkaXsuZ0mgCynMD/JRRBMNkiR++jyXvEiUiiSYkfDnZLXSez7Quxg4FeWWrqiN3T1d85z7QkHvBJIa/fa2+oae5vTk22Ncam3HX+ep+PDbnuyaWK7YDADYA5XAKadCKqQV5kb1xD/76Gay9e3Fwz+42V19EbaAPcIAIIIbIKzaBEywZBnhVl0WrViIwNP/0+O3ppGYtRf2+D2jzXOfPgko3yD+xyh7nhbj9zJD08uTcou/hj3ob5c/r7Sr14TuCO2OJel2BotKG8jlbRBnju9sttuwhWK3LnzmYGv5ZNDNTnxICaTre4ElxneILpuTB+aLHHJ8x485neAr/3aKDO7EMur6MNvAocBLb6MIBlIs7tDraG/WnxbNdMW7TdbV1eslGefuyS7VzcIaaame98fezCvnpX/mebfa0eccnhft1e0zcwlANscH8XbYDrLrcbCgmS7RLuFsWQL/IPnNvfd7prYKqZwvotW3+rmmqVpr9/TW4MpZh9i+Nji+n0XJYLV6eU4L0eH+qMqs2kFwuEyk2YRhvQDF1wwOhGUtsIebIcdausoPBiReAh2ZgdWZfDarXscBBXUWxIMh75tHO6fYCrC7g80c5ptTH46xG6um1S8/rtoeho8Vj24gGvLHu9shxN9sphxR1k6npuetobuyPUnoi/LllL2bPx7pEIM7c75Og40GDb6+TsXX3KWAJdj0XlaCQSjenLDW6h1mJxueu9ZkZmyILwasWVPL0lLNZgSbOZZbr+YHJsaNkbqI+48Oov7nXH56b0GyiYirgF/RUol0EDgDfxTSyBBAA0yPDENrYPrwJjYLOKphCv0nzmSeoHL/7yty8spPGqPv+nG/rf/jDwMHm+vI7seBX2miphFXZbdH/Jdy2z1VW01c6EmfsOYnHzlmBH6EwVbdaxeNEGBI06JFiJuu7ohN6+Z4gn+lvVDBc80Dp6cNkbCLeQSzNa6/U3xSOh1q32WvRXKretOaGNypwqNXbOaclGBQrbg0JraV/THXMyNWpo5/9nqzN9Nps9m07PZ7Pz6aZEoinR1FTxV8/iofELPQ8VejN5YjOCmykPYifaAA58AMJtdob8JFngDaWJIZp3OglP75D8lRPdM6lAt6dqREpNxGOOyKv4p60e8dvnDy+l69wjT6OG/vzjTa/ba0z/DqINAz8AUKVqBuyWiRVNYS07/YtOWt37Q6aJ93kp5vz72wZ+9dm8y2+Y2Bto3ZxEDbcdXNELehJtgP2OPZopbE64Li/x9TbXHndtfY8DrR1JtlZVPUpR0aT+DiDgy+voBbQBsqGf25kumZm+DUYS3Yd5h/Vm63FpfyjtD/q8CY+vK3LycMcR/35Pm6ejQwr0RE8wkr/orhM41snZmIaOaG5Cdk06nLLLXbNb7Ej0TZkeYsvraB4vkq9SVUhSVVHVNMU4lNwOTSiOZPPsww89JHoZt03gNObUxPUz1kuXzv85FrZSc1bGxOour6PP0RrR2R0eYCtR+dexoWVfoF5yLi/ttvgPMHNTqE1/W416vGhQr82FGwERv6EyWoM9AIpFESpnDU2xXPvJlV4bZ6OqOVvm8o/Q2ofhgiwXwh/qtUbtGgC0jtbIt0nh5B0v0oIoSxIpT9M1K08+32hz2qhd9l2hlaeuPt/CCAxV7aiWEf54lI/zfJwfLf9rnG/k+bhznOiHA0Cf429ANemKU0hSEihODaocCR2Rf/HxKkQxnpqk/s8PfjU0hHYd94/5PKk6fX7lAfSIfnlhxTzP7UMALxN9CHIqJYdCpg5Nred97Z0IU1hMpaRkW/GPw45MOB6REgcy40vmbONwHQVRK1gANFXh459dP3GC/D9YLqAIfpvMSzCDXDA+p8KNdC6XLmrJpHbt+FuXLr11XDp2a+7UrVlA0FIuoNrKO3LKODlohpFLxfZksr2YzuWuSbO3Ts3dOiYZ75r6hjfQGqlP8jCzjNb0WkDll3EHHMI3yTmS3dFQOJEIhxMJ3BETxRj5wX8AAAD//wEAAP//Cs59lgAAAQAAAAILhRSACdlfDzz1AAED6AAAAADYXaCEAAAAAN1mLzb+N/7ECG0D8QABAAMAAgAAAAAAAAABAAAD2P7vAAAImP43/jcIbQABAAAAAAAAAAAAAAAAAAAAKHicHMoxjgFhGMbx//tsss1OFjEkGgWTSMzki47EKN5GFL5EoRgHcA1uoNeqabQuoHcbDaH/6cqKO2hOUELUjqATUU+ijkSdCVoTtSWqoK+CjmoM7UGmGQP9kqlNT1PcUsbKcOviPxtcJa786/xj7YDbjZbtaWhCqT8SiX/VaVpBsCW5VSysYmQpDq/LGwAA//8BAAD//2aeFhMAAAAAACwALABYAGwAnACyAOQBBgEyAVQBegG6AcwCBAIwAmIClgL+AyADLANEA2ADkgO0A+AEEAREBGQEoATGBOgFBAUwBVQFdAWABZoFtAXABdYAAAABAAAAKACQAAwAYwAHAAEAAAAAAAAAAAAAAAAABAADeJyclM9uG1UUxn9ObNMKwQJFVbqJ7oJFkejYVEnVNiuH1IpFFAePC0JCSBPP+I8ynhl5Jg7hCVjzFrxFVzwEz4FYo/l87NgF0SaKknx37vnznXO+c4Ed/mabSvUh8Ec9MVxhr35ueIsH9RPD27TrW4arPKn9abhGWJsbrvN5rWf4I95WfzP8gP3qT4YfslttG/6YZ9Udw59sO/4y/Cn7vF3gCrzgV8MVdskMb7HDj4a3eYTFrFR5RNNwjc/YM1xnD+gzoSBmQsIIx5AJI66YEZHjEzFjwpCIEEeHFjGFviYEQo7Rf34N8CmYESjimAJHjE9MQM7YIv4ir5RzZRzqNLO7FgVjAi7kcUlAgiNlREpCxKXiFBRkvKJBg5yB+GYU5HjkTIjxSJkxokGXNqf0GTMhx9FWpJKZT8qQgmsC5XdmUXZmQERCbqyuSAjF04lfJO8Opzi6ZLJdj3y6EeFLHN/Ju+SWyvYrPP26NWabeZdsAubqZ6yuxLq51gTHui3ztvhWuOAV7l792WTy/h6F+l8o8gVXmn+oSSVikuDcLi18Kch3j3Ec6dzBV0e+p0OfE7q8oa9zix49WpzRp8Nr+Xbp4fiaLmccy6MjvLhrSzFn/IDjGzqyKWNH1p/FxCJ+JjN15+I4Ux1TMvW8ZO6p1kgV3n3C5Q6lG+rI5TPQHpWWTvNLtGcBI1NFJoZT9XKpjdz6F5oipqqlnO3tfbkNc9u95RbfkGqHS7UuOJWTWzB631S9dzRzrR+PgJCUC1kMSJnSoOBGvM8JuCLGcazunWhLClornzLPjVQSMRWDDonizMj0NzDd+MZ9sKF7Z29JKP+S6eWqqvtkcerV7YzeqHvLO9+6HK1NoGFTTdfUNBDXxLQfaafW+fvyzfW6pTzliJSY8F8vwDM8muxzwCFjZRjoZm6vQ1MvRJOXHKr6SyJZDaXnyCIc4PGcAw54yfN3+rhk4oyLW3FZz93imCO6HH5QFQv7Lke8Xn37/6y/i2lTtTierk4v7j3FJ3dQ6xfas9v3sqeJlZOYW7TbrTgjYFpycbvrNbnHeP8AAAD//wEAAP//9LdPUXicYmBmAIP/5xiMGLAAAAAAAP//AQAA//8vAQIDAAAA");
}
.d2-580756557 .text-italic {
	font-family: "d2-580756557-font-italic";
}
@font-face {
	font-family: d2-580756557-font-italic;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA+QAAoAAAAAGCQAARhRAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgW1SVeGNtYXAAAAFUAAAAqwAAANgDvwT1Z2x5ZgAAAgAAAAjWAAAMfK7MTlRoZWFkAAAK2AAAADYAAAA2G7Ur2mhoZWEAAAsQAAAAJAAAACQLeAjMaG10eAAACzQAAACfAAAAoEfGBN1sb2NhAAAL1AAAAFIAAABSQ4ZArm1heHAAAAwoAAAAIAAAACAAQAD2bmFtZQAADEgAAAMmAAAIMgntVzNwb3N0AAAPcAAAACAAAAAg/8YAMgADAeEBkAAFAAACigJY//EASwKKAlgARAFeADIBIwAAAgsFAwMEAwkCBCAAAHcAAAADAAAAAAAAAABBREJPAAEAIP//Au7/BgAAA9gBESAAAZMAAAAAAeYClAAAACAAA3icdM05SgMBGEDhb5xxH3Xct1HHDRS08waC2AgiXmAQEbEXj+R2AC1CLpBL5ABpUwX+kOnz2q94SKQS5DIdlAqpXOXcpSvXbt17VHv26t1HBCpnLhq9cedB7cmLt5FGNwbRj160oxX/8Re/8RPf8RWfzWd8iROnjhw7NCGVmTRl2oxZc+blFixaUli2YtWadRs2bdm2Y1dpz74DFUMAAAD//wEAAP//MlwmnwB4nHxWa2wj5dV+33cmM5usc/PY49gb27Hf8YzjjO3EY3viOHYuTpzEl1w3IR+Js5uFzbcbwirdxbDtsqUQabVFKjVo1apoy1aFVq34g5b2B72AepNS0FZVhVpaKirEEtrdokKUVoCacTVjx3EitX9G8+c953nOec5zDqgBHADoHLoGCFALGoEeGAGQGAdBSLKMTYQkCJimZYFhaO4JuPnEs2Ti3vfd3/pUtJMjj38//feTL6Jru2vwS7nHHlMWrp4+fc/du4oH/v4uAABAECnuIC+6DuwA1Dh5PhSMIynAmmiex84GZDSwrBQIyyaKgs70mXDnvZcz3VMtYSbM95wY5JypqDvRhrmcLnFxPHvtkRHZ094mxO6/2BvNhdqOBexeLQcGADWhAqjVcNMOWqIxgTfgar3ynvfj+o96UWHg9qDyhz08UXQdODU8/wWOjGWJoCgoPny5c+HxqeiUWWZkd/zUMIczfVyEcV2t/02EW9I9fXH82iPJCqiepXBL0w/7lds2VwXXWVQAR1VcEuFgJAIzDgJvjHdDd3d2Y7xPeSuOCspdaNxdh93KZvnNT1ABmLU3jEmStVfhsIxpAhMCpiiawBu5CEsmf5HbSGdqLTpy4mdijCWphiMpVFCeu3oVntpdh+fF1Y5nlBfg4jPiWVF5qhx7BRUAsxc7HNaiV6KOf81DUg11w+mN7LUOkmqsS6KCsvjlrgcluLi7Dp//irQaUG4AABDoLe6gJXQdNIG26kqyRkMDEgJxFAruVRTaz+V9c/lk6nTQN/dQInRP3JkaV79juq8/mi7kh4cuzaSfzg8nek/lI8v56Kl8z8mHtRwqXq9WP4OGOMAaDRSFMcFIgbCaAOONVxbPpx4/fjY4cOL0amb0NCqk5ib/v0v5BI5MTkQkUIkjoAKoB+x+HJrBxIFIP1j83LmZCzNr5+Wh+5buT4+eRIXkzMK5ZuU9yCp34Ox0MuwvaVpX3IEKug48AJicvCBrogkFeUFQBR4OVxRFUUYDazKxGu4PEuvuiHVW7p3yujKeaGgxGj1pl8xJnytk7eIy/mB0RdfT09ERGOrmAqzPMiYHpgNBt8/Wbu88xvtZb+uI3LMQBBDkAEAhVAC0ygbLDhoT382/Wg/fqH8tj7KJxO7LJd5CcQd+AreBQa2AaV/tkiwRWMYUJahar0j/5f6MmFqShFgzycSX+46QeF7PT3CiMdDKJUL2Lt3CbPLzi5LbEVMsoy5/v8//R97pGcsF+mKluriKO/Am3AatB7LtV6I86W9N3C9ml0NiL+tleGvnXDjS0xZmnZasbiU3dGHW7zR3moxD64nBpKU5YHBVuCABbQKj6lYHuPxvMj16oonPFspsxl2H2QhtJ17d7T5MB2lcXoPbwAJc1fk05TioimsRUljtvMrw9txZb3qxUx6w6WqUX9a2JTzWiMlmnfpGERH6dhxa0q0uD69Pi77JQKvU0DfpMjdLRjt0HW2pb+2yzwIEYJGD23Ab2IGvWlmyTFH4YPcoijjA9sWuOcy1DrvjqQYzf9wfm+wYW+zi480E07fCXIjgKWcH29WKBySb/8+8NWRyZvrP8OLcbOKh/wuo/SROrEBHh+e3vLM9Od8ZjZb6aQcAvoU2y15E05ImbaOB1gwpFMROiibsT2U7m8j2aTEeOhLP9JLkaOuobxht3o1h/0C3nVNeh6KhpT7t8SnfKxbVmOAzdBPxgAcAUEAY3c/1IdoEupJXqvkYLNC0/ansSfTp/M/z47l1C9pUrBC+obz/4flLAAKxuAM+Q5tAr1YrFFS9TNVZuTUPDlCXspchbCYoGtaxur5mM3pg92m6ltBDFCXJSl50B26rs6zmLFE0lYlSB5hWk17uo0l+hu/pqvHPu2JhkoxnYyQ5YhwVh9UaJNnRjmG4NcZ1yW5RGuhuthmq67D/t19nuA1aqjEcLrOasX3ad6DKWobDRd6f/bfhNmgE1mr9luxZ02x5KN+cWBJTS4GJE2J6yeOdksIB9aM7szB8YdZX+vYPrg8NjiTWhwaTauziv4oS/Ahul2aRrkLcgLCTV68FZm8DUDTNsnVP9lGEa9anjWSA72WQ3v4dLhGydbY7p7DPIN1CL/fbveWBtJ+5AaFnLCfFYx7+ry7HHh9J8zItZ42sDvyhmTg4EdDhsCHXvK/a0568UW0At248wvsrlrabhfCgoZX68ijcBk1VfTHR/F4/jpLWjNdsPNZk4TL2GNzKibHaoSN9UeUWgMV/F3fgZbgNhMM74vCKUDdEaUE835Uzd5r6eU+svdsXEcdEX6rVx0gOvivcFg92TuuCbt7u9mGLYLfE2zsGXJzNbbB47TZe7+wVvUMuFXNvcQfOo7WKB4dlBvchSTuJqjz4lf4gCSMjRzPcwLFLussRotXZYDna3OTX9XkbLfVQH6m5ciWu3NHrbba6GpluVGN3F3fgP+CW6gd7sfcnjinb8IuVaRi1jojDGXWRuI/rBuVmOwPDypuMWZUpnFcsKSyVZjAKAHwXboF6ANTJZ9nyZQKfGMlwJEWSzRzz1ayyC7eUD3Aac2McNCsW7W3xTQDg70qYMCNIpvJjWaJNWOB57bKhxT8tjHuONNBkY1vj7MzmfRPikeY6ssnJLEF0e40VjIZ249o/Pz7P+lhWNF1Q9aYSvoW+COrU28OBZYcMJUKisUuS1VuMomiYHsXK32rh0vHJGd2MUvwVT+lp0uA2vBSEzyrr8fhPrQOO1mBLCee7xUVQAGuqlmhBjaAVsOQyhqP+CI1YG7ZaWu99wafv5SysWeBsY+ugfEe+A+ugGRAAqMSw7u36d/Y8eqA4Ce9Bb6t4TSVpySZKO2RNX2hxyGdS3tW1WkPDS/3PT+d//eOc+Yryl+d8Kyd5rXbFSXCn/FYI69U9pgpV1SL0rj5Qq28MqCFeslyBjm/6V07wTP+3p/Ov/6g8G+AW3FIxqX5tX86egltaUyAYQWlwE91U716miuZFxoZNBitGaRNrdrSw5rb/AAAA//8BAAD//1ANnT8AAAABAAAAARhREP6o718PPPUAAQPoAAAAANhdoMwAAAAA3WYvN/69/t0IHQPJAAIAAwACAAAAAAAAAAEAAAPY/u8AAAhA/r39vAgdA+gAwv/RAAAAAAAAAAAAAAAoeJwczTEOwWAYh/Hn/XdESAyN6Rs+bQcOoNHFIBazzQkkJotruIfJBdgcoIvEoAewGEREfFIHeJ6fNvQ5g1Jyu+A1J9carxNeK7y2FErxGuDtTdO+LNUjswOJYjKrSNTF1MBxx9kVx4th5HBq4RSRKQ7Pf7PA2S58bEqhDiM7MrZ9KO1Gm0eorDY8E0tCWb+A2Q8AAP//AQAA///uLiDQAAAAAC4ALgBgAHYArADGAPQBGgFMAXABmAHYAewCJAJSAooCxAMMAzYDQgNcA34DwAPqBBgEUgSMBKoE5gUUBUAFXgWKBbIF0AXeBfwGGgYoBj4AAAABAAAAKACMAAwAZgAHAAEAAAAAAAAAAAAAAAAABAADeJyclNtOG1cUhj8H2216uqhQRG7QvkylZEyjECXhypSgjIpw6nF6kKpKgz0+iPHMyDOYkifodd+ib5GrPkafoup1tX8vgx1FQSAE/Hv2OvxrrX9tYJP/2KBWvwv83ZwbrrHd/NnwHb5oHhneYL/5meE6Dxv/GG4waLw13ORBo2v4E97V/zT8KU/qvxm+y1b90PDnPK5vGv5yw/Gv4a94wrsFrsEz/jBcY4vC8B02+dXwBvewmLU699gx3OBrtg032QZ6TKhImZAxwjFkwogzZiSURCTMmDAkYYAjpE1Kpa8ZsZBj9MGvMREVM2JFHFPhSIlIiSkZW8S38sp5rYxDnWZ216ZiTMyJPE6JyXDkjMjJSDhVnIqKghe0aFHSF9+CipKAkgkpATkzRrTocMgRPcZMKHEcKpJnFpEzpOKcWPmdWfjO9EnIKI3VGRkD8XTil8g75AhHh0K2q5GP1iI8xPGjvD23XLbfEujXrTBbz7tkEzNXP1N1JdXNuSY41q3P2+YH4YoXuFv1Z53J9T0a6H+lyCecaf4DTSoTkwzntmgTSUGRu49jX+eQSB35iZAer+jwhp7Obbp0aXNMj5CX8u3QxfEdHY45kEcovLg7lGKO+QXH94Sy8bET689iYgm/U5i6S3GcqY4phXrumQeqNVGFN5+w36F8TR2lfPraI2/pNL9MexYzMlUUYjhVL5faKK1/A1PEVLX42V7d+22Y2+4tt/iCXDvs1brg5Ce3YHTdVIP3NHOun4CYATknsuiTM6VFxYV4vybmjBTHgbr3SltS0b708XkupJKEqRiEZIozo9Df2HQTGff+mu6dvSUD+Xump5dV3SaLU6+uZvRG3VveRdblZGUCLZtqvqKmvrhmpv1EO7XKP5Jvqdct5xGh4i52+0OvwA7P2WWPsbL0dTO/vPOvhLfYUwdOSWQ1lKZ9DY8J2CXgKbvs8pyn7/VyycYZH7fGZzV/mwP26bB3bTUL2w77vFyL9vHMf4ntjupxPLo8Pbv1NB/cQLXfaN+u3s2uJuenMbdoV9txTMzUc3FbqzW5+wT/AwAA//8BAAD//3KhUUAAAAADAAD/9QAA/84AMgAAAAAAAAAAAAAAAAAAAAAAAAAA");
}]]></style><style type="text/css"><![CDATA[.shape {
  shape-rendering: geometricPrecision;
  stroke-linejoin: round;
}
.connection {
  stroke-linecap: round;
  stroke-linejoin: round;
}
.blend {
  mix-blend-mode: multiply;
  opacity: 0.5;
}

		.d2-580756557 .fill-N1{fill:#0A0F25;}
		.d2-580756557 .fill-N2{fill:#676C7E;}
		.d2-580756557 .fill-N3{fill:#9499AB;}
		.d2-580756557 .fill-N4{fill:#CFD2DD;}
		.d2-580756557 .fill-N5{fill:#DEE1EB;}
		.d2-580756557 .fill-N6{fill:#EEF1F8;}
		.d2-580756557 .fill-N7{fill:#FFFFFF;}
		.d2-580756557 .fill-B1{fill:#0D32B2;}
		.d2-580756557 .fill-B2{fill:#0D32B2;}
		.d2-580756557 .fill-B3{fill:#E3E9FD;}
		.d2-580756557 .fill-B4{fill:#E3E9FD;}
		.d2-580756557 .fill-B5{fill:#EDF0FD;}
		.d2-580756557 .fill-B6{fill:#F7F8FE;}
		.d2-580756557 .fill-AA2{fill:#4A6FF3;}
		.d2-580756557 .fill-AA4{fill:#EDF0FD;}
		.d2-580756557 .fill-AA5{fill:#F7F8FE;}
		.d2-580756557 .fill-AB4{fill:#EDF0FD;}
		.d2-580756557 .fill-AB5{fill:#F7F8FE;}
		.d2-580756557 .stroke-N1{stroke:#0A0F25;}
		.d2-580756557 .stroke-N2{stroke:#676C7E;}
		.d2-580756557 .stroke-N3{stroke:#9499AB;}
		.d2-580756557 .stroke-N4{stroke:#CFD2DD;}
		.d2-580756557 .stroke-N5{stroke:#DEE1EB;}
		.d2-580756557 .stroke-N6{stroke:#EEF1F8;}
		.d2-580756557 .stroke-N7{stroke:#FFFFFF;}
		.d2-580756557 .stroke-B1{stroke:#0D32B2;}
		.d2-580756557 .stroke-B2{stroke:#0D32B2;}
		.d2-580756557 .stroke-B3{stroke:#E3E9FD;}
		.d2-580756557 .stroke-B4{stroke:#E3E9FD;}
		.d2-580756557 .stroke-B5{stroke:#EDF0FD;}
		.d2-580756557 .stroke-B6{stroke:#F7F8FE;}
		.d2-580756557 .stroke-AA2{stroke:#4A6FF3;}
		.d2-580756557 .stroke-AA4{stroke:#EDF0FD;}
		.d2-580756557 .stroke-AA5{stroke:#F7F8FE;}
		.d2-580756557 .stroke-AB4{stroke:#EDF0FD;}
		.d2-580756557 .stroke-AB5{stroke:#F7F8FE;}
		.d2-580756557 .background-color-N1{background-color:#0A0F25;}
		.d2-580756557 .background-color-N2{background-color:#676C7E;}
		.d2-580756557 .background-color-N3{background-color:#9499AB;}
		.d2-580756557 .background-color-N4{background-color:#CFD2DD;}
		.d2-580756557 .background-color-N5{background-color:#DEE1EB;}
		.d2-580756557 .background-color-N6{background-color:#EEF1F8;}
		.d2-580756557 .background-color-N7{background-color:#FFFFFF;}
		.d2-580756557 .background-color-B1{background-color:#0D32B2;}
		.d2-580756557 .background-color-B2{background-color:#0D32B2;}
		.d2-580756557 .background-color-B3{background-color:#E3E9FD;}
		.d2-580756557 .background-color-B4{background-color:#E3E9FD;}
		.d2-580756557 .background-color-B5{background-color:#EDF0FD;}
		.d2-580756557 .background-color-B6{background-color:#F7F8FE;}
		.d2-580756557 .background-color-AA2{background-color:#4A6FF3;}
		.d2-580756557 .background-color-AA4{background-color:#EDF0FD;}
		.d2-580756557 .background-color-AA5{background-color:#F7F8FE;}
		.d2-580756557 .background-color-AB4{background-color:#EDF0FD;}
		.d2-580756557 .background-color-AB5{background-color:#F7F8FE;}
		.d2-580756557 .color-N1{color:#0A0F25;}
		.d2-580756557 .color-N2{color:#676C7E;}
		.d2-580756557 .color-N3{color:#9499AB;}
		.d2-580756557 .color-N4{color:#CFD2DD;}
		.d2-580756557 .color-N5{color:#DEE1EB;}
		.d2-580756557 .color-N6{color:#EEF1F8;}
		.d2-580756557 .color-N7{color:#FFFFFF;}
		.d2-580756557 .color-B1{color:#0D32B2;}
		.d2-580756557 .color-B2{color:#0D32B2;}
		.d2-580756557 .color-B3{color:#E3E9FD;}
		.d2-580756557 .color-B4{color:#E3E9FD;}
		.d2-580756557 .color-B5{color:#EDF0FD;}
		.d2-580756557 .color-B6{color:#F7F8FE;}
		.d2-580756557 .color-AA2{color:#4A6FF3;}
		.d2-580756557 .color-AA4{color:#EDF0FD;}
		.d2-580756557 .color-AA5{color:#F7F8FE;}
		.d2-580756557 .color-AB4{color:#EDF0FD;}
		.d2-580756557 .color-AB5{color:#F7F8FE;}.appendix text.text{fill:#0A0F25}.md{--color-fg-default:#0A0F25;--color-fg-muted:#676C7E;--color-fg-subtle:#9499AB;--color-canvas-default:#FFFFFF;--color-canvas-subtle:#EEF1F8;--color-border-default:#0D32B2;--color-border-muted:#0D32B2;--color-neutral-muted:#EEF1F8;--color-accent-fg:#0D32B2;--color-accent-emphasis:#0D32B2;--color-attention-subtle:#676C7E;--color-danger-fg:red;}.sketch-overlay-B1{fill:url(#streaks-darker-d2-580756557);mix-blend-mode:lighten}.sketch-overlay-B2{fill:url(#streaks-darker-d2-580756557);mix-blend-mode:lighten}.sketch-overlay-B3{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-B4{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-B5{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-B6{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-AA2{fill:url(#streaks-dark-d2-580756557);mix-blend-mode:overlay}.sketch-overlay-AA4{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-AA5{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-AB4{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-AB5{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-N1{fill:url(#streaks-darker-d2-580756557);mix-blend-mode:lighten}.sketch-overlay-N2{fill:url(#streaks-dark-d2-580756557);mix-blend-mode:overlay}.sketch-overlay-N3{fill:url(#streaks-normal-d2-580756557);mix-blend-mode:color-burn}.sketch-overlay-N4{fill:url(#streaks-normal-d2-580756557);mix-blend-mode:color-burn}.sketch-overlay-N5{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-N6{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.sketch-overlay-N7{fill:url(#streaks-bright-d2-580756557);mix-blend-mode:darken}.light-code{display: block}.dark-code{display: none}]]></style><g class="dGFzaw=="><g class="shape" ><path d="M 225 12 L 391 12 L 365 78 L 199 78 L 199 78 Z" stroke="#0D32B2" fill="#DEE1EB" class=" stroke-B1 fill-N5" style="stroke-width:2;" /></g><text x="295.000000" y="50.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px">Task Request</text></g><g class="cm91dGVy"><g class="shape" ><path d="M 295 240 C 294 240 294 240 293 240 L 186 195 C 185 195 185 194 186 193 L 293 148 C 294 148 296 148 297 148 L 404 193 C 405 193 405 194 404 195 L 297 240 C 296 240 296 240 295 240 Z" stroke="#0D32B2" fill="#CFD2DD" class=" stroke-B1 fill-N4" style="stroke-width:2;" /></g><text x="295.000000" y="199.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px">Model Router</text></g><g class="ZnJvbnRpZXI="><g class="shape" ><rect x="12.000000" y="411.000000" width="182.000000" height="98.000000" stroke="#0D32B2" fill="#fee2e2" class=" stroke-B1" style="stroke-width:2;" /></g><text x="103.000000" y="449.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="103.000000" dy="0.000000">Frontier Model</tspan><tspan x="103.000000" dy="17.666667">(GPT-4, Opus)</tspan><tspan x="103.000000" dy="17.666667">Complex reasoning</tspan></text></g><g class="bWlkdGllcg=="><g class="shape" ><rect x="214.000000" y="411.000000" width="163.000000" height="98.000000" stroke="#0D32B2" fill="#fef9c3" class=" stroke-B1" style="stroke-width:2;" /></g><text x="295.500000" y="449.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="295.500000" dy="0.000000">Mid-tier Model</tspan><tspan x="295.500000" dy="17.666667">(GPT-4o-mini)</tspan><tspan x="295.500000" dy="17.666667">Structured tasks</tspan></text></g><g class="c21hbGw="><g class="shape" ><rect x="397.000000" y="411.000000" width="192.000000" height="82.000000" stroke="#0D32B2" fill="#dcfce7" class=" stroke-B1" style="stroke-width:2;" /></g><text x="493.000000" y="449.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="493.000000" dy="0.000000">Small Model</tspan><tspan x="493.000000" dy="18.500000">High-volume, simple</tspan></text></g><g class="KHRhc2sgLSZndDsgcm91dGVyKVswXQ=="><marker id="mk-d2-580756557-3488378134" markerWidth="10.000000" markerHeight="12.000000" refX="7.000000" refY="6.000000" viewBox="0.000000 0.000000 10.000000 12.000000" orient="auto" markerUnits="userSpaceOnUse"> <polygon points="0.000000,0.000000 10.000000,6.000000 0.000000,12.000000" fill="#0D32B2" class="connection fill-B1" stroke-width="2" /> </marker><path d="M 295.028569 79.999796 L 295.942863 144.000408" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-580756557-3488378134)" mask="url(#d2-580756557)" /></g><g class="KHJvdXRlciAtJmd0OyBmcm9udGllcilbMF0="><path d="M 240.016129 219.999935 L 240.419357 270.000325 S 240.500000 280.000000 230.500000 280.000000 L 113.000000 280.000000 S 103.000000 280.000000 103.000000 290.000000 L 103.000000 407.000000" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-580756557-3488378134)" mask="url(#d2-580756557)" /><text x="137.500000" y="286.000000" fill="#676C7E" class="text-italic fill-N2" style="text-anchor:middle;font-size:16px">Needs reasoning</text></g><g class="KHJvdXRlciAtJmd0OyBtaWR0aWVyKVswXQ=="><path d="M 295.988304 241.999966 L 295.023391 407.000068" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-580756557-3488378134)" mask="url(#d2-580756557)" /><text x="296.000000" y="331.000000" fill="#676C7E" class="text-italic fill-N2" style="text-anchor:middle;font-size:16px">Needs structure</text></g><g class="KHJvdXRlciAtJmd0OyBzbWFsbClbMF0="><path d="M 350.016129 219.999935 L 350.419357 270.000325 S 350.500000 280.000000 360.500000 280.000000 L 483.000000 280.000000 S 493.000000 280.000000 493.000000 290.000000 L 493.000000 407.000000" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-580756557-3488378134)" mask="url(#d2-580756557)" /><text x="456.500000" y="286.000000" fill="#676C7E" class="text-italic fill-N2" style="text-anchor:middle;font-size:16px">Needs speed</text></g><mask id="d2-580756557" maskUnits="userSpaceOnUse" x="-9" y="-9" width="619" height="539">
<rect x="-9" y="-9" width="619" height="539" fill="transparent"></rect>
<rect x="81.000000" y="270.000000" width="113" height="21" fill="black"></rect>
<rect x="242.000000" y="315.000000" width="108" height="21" fill="black"></rect>
<rect x="413.000000" y="270.000000" width="87" height="21" fill="black"></rect>
</mask></svg></svg>
" class="img-fluid" style="width:100.0%"></p>
</section>
<section id="india-needs-this-more-than-anyone" class="level2">
<h2 class="anchored" data-anchor-id="india-needs-this-more-than-anyone">India Needs This More Than Anyone</h2>
<p>This matters disproportionately for Indian AI product companies.</p>
<p>The ARPU constraints are real. When your customers are paying Rs 500-1,500/month, not Rs 5,000-20,000/month, your inference cost per user eats a bigger share of revenue. You can’t afford to run everything through Opus. You need to be surgical about where you spend on model capability.</p>
<p>The arbitrage is bigger here. Indian engineering teams are already good at cost optimization. Cloud cost management, infrastructure efficiency, resource utilization — these are native skills. Model routing is the same discipline applied to AI.</p>
<p>The companies building AI products in India that figure out model portfolios early will have a structural advantage. Not because they’re smarter. Because their margin constraints forced them to solve the problem first.</p>
</section>
<section id="what-i-dont-know-yet" class="level2">
<h2 class="anchored" data-anchor-id="what-i-dont-know-yet">What I Don’t Know Yet</h2>
<p>I don’t have a clean answer for how to build the routing logic itself. Do you hardcode rules? Do you train a classifier? Do you use an LLM to route to other LLMs? Each approach has tradeoffs.</p>
<p>Hardcoded rules are brittle but predictable. A classifier adds complexity but scales better. Using an LLM as a router adds latency and cost but might handle edge cases better.</p>
<p>We’re still experimenting. The right answer probably depends on your volume, your task diversity, and how much you’re willing to invest in routing infrastructure.</p>
<p>The other open question: how do you measure quality degradation when you switch models? User complaints are a lagging indicator. You need leading indicators — accuracy on test sets, consistency scores, format compliance rates. Building that instrumentation is non-trivial.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The Question Worth Asking</h2>
<p>The companies that win the next phase of AI products won’t be the ones with the best models. They’ll be the ones with the best model selection strategy.</p>
<p>The question isn’t “which model should we use?” The question is “which model should we use for this specific task, at this volume, at this quality threshold?”</p>
<p>Most teams aren’t asking that question yet. They will be.</p>


</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>Product Economics</category>
  <category>AI Infrastructure</category>
  <guid>https://talvinder.com/field-notes/small-model-arbitrage/</guid>
  <pubDate>Sun, 22 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>We Were Running AI Agents Before ‘Agentic’ Became a Buzzword</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/multi-agent-before-agentic/</link>
  <description><![CDATA[ 





<p>In early 2024, we deployed a multi-agent system for Ostronaut before anyone called it “agentic AI.” We called it “the pipeline.” By late 2024, every vendor deck had “agentic” in the title. The architecture didn’t change. The vocabulary did.</p>
<p>Here’s the pattern that experience revealed: <strong>Agent Debt</strong>. The hidden complexity that accumulates when you treat agents as black boxes instead of understanding their failure modes. It isn’t technical debt. It’s operational blindness. You don’t see it until an agent hallucinates in production, burns through your API budget, or produces output so confidently wrong that users trust it.</p>
<p>Building without frameworks meant hitting every orchestration failure, every context bleed, every runaway cost directly. That’s what taught us what actually matters.</p>
<section id="the-architecture-we-built" class="level2">
<h2 class="anchored" data-anchor-id="the-architecture-we-built">The Architecture We Built</h2>
<p>Ostronaut generates corporate training content — presentations, videos, quizzes, games — from unstructured input. A client uploads a PDF. The system outputs interactive learning formats.</p>
<p>We built agents in four functional groups because the problem naturally decomposed that way:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 42%">
<col style="width: 57%">
</colgroup>
<thead>
<tr class="header">
<th>Agent Type</th>
<th>Responsibility</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Planner agents</td>
<td>Break input into learning objectives, decide format mix</td>
</tr>
<tr class="even">
<td>Structure agents</td>
<td>Design slide sequences, video scripts, quiz flows</td>
</tr>
<tr class="odd">
<td>Content agents</td>
<td>Generate text, voiceovers, visual descriptions</td>
</tr>
<tr class="even">
<td>Validation agents</td>
<td>Check quality gates, flag hallucinations, verify completeness</td>
</tr>
</tbody>
</table>
<p>The planner-worker pattern: one planner agent analyzes the input and creates a generation plan. Worker agents execute tasks from that plan. Validation agents run post-generation checks.</p>
<p>This wasn’t novel architecture. It was obvious once you tried to build the thing. But in early 2024, there was no CrewAI to handle orchestration. No LangGraph to manage state. We wrote the coordination logic ourselves.</p>
<p><strong>What that meant in practice:</strong></p>
<p>Context management was manual. Each agent needed the right slice of information: not too much (cost), not too little (hallucination). We built a context router that decided what each agent could see based on its task. It broke constantly. An agent would reference information from a previous step that wasn’t in its context window. Output would be incoherent.</p>
<p>Tool-calling was brittle. Agents needed to invoke APIs for image generation, video rendering, database writes. Early LLM tool-calling was unreliable. An agent would call the wrong API, pass malformed parameters, or retry indefinitely on failure. We added a validation layer that parsed tool calls before execution. That caught 30% of bad calls.</p>
<p>Cost control was reactive. We didn’t know what “normal” token usage looked like for a multi-agent pipeline. First month in production, we burned through our OpenAI budget in 2 weeks. The problem: redundant context. Multiple agents were processing the same source material because we hadn’t optimized context sharing. We added a caching layer. Cost dropped 40%.</p>
</section>
<section id="the-quality-crisis" class="level2">
<h2 class="anchored" data-anchor-id="the-quality-crisis">The Quality Crisis</h2>
<p>Month 4, we hit the ceiling.</p>
<p>A healthcare client used Ostronaut to generate training for a clinical health program. The system produced a quiz. One question asked: “What is the recommended daily caloric deficit for healthy weight loss?” The agent-generated answer: “1000-1200 calories.”</p>
<p>That’s dangerously high for most people. The correct range is 500-750 calories.</p>
<p>The agent didn’t hallucinate randomly. It pulled from a source document that mentioned 1000-1200 as an <em>upper bound</em> for specific cases. The agent extracted the number without the qualifier. The validation agent didn’t flag it because it checked for factual consistency with the source, not medical safety.</p>
<p>We caught it in QA. But it revealed the core problem: <strong>agents optimize for coherence, not correctness</strong>. They will confidently generate plausible-but-wrong output if your validation layer doesn’t encode domain constraints.</p>
<p>This is the failure mode that no prompt tuning fixes. You can instruct the model to “be accurate” as many times as you want. It will still extract numbers from context and strip their qualifiers, because that’s what extracting the salient point looks like to the model.</p>
<p><strong>What we changed:</strong></p>
<p>Built domain-specific validation gates. For healthcare content, we added rules: flag any caloric recommendation above X, flag any medication dosage, flag any symptom-diagnosis claim. Not LLM-based validation. Rule-based checks that ran before content went to the client.</p>
<p>Added confidence scoring. Each agent outputs a confidence score for its generation. Low-confidence outputs go to human review. The scoring isn’t sophisticated (token probability and context match), but it works. 15% of generations now route to human QA. That’s acceptable.</p>
<p>Switched to template + generative hybrid. For high-risk content types (medical, financial, legal), we don’t generate from scratch. We use templates with generative fill-ins. Reduces creative output, increases safety. Clients accepted the trade-off.</p>
</section>
<section id="what-we-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-we-got-wrong">What We Got Wrong</h2>
<p><strong>Universal reasoning engine.</strong> We initially tried to build one planner agent that could handle all content types. A presentation has different structural constraints than a video. A quiz has different validation rules than a game. We split the planner into format-specific planners. That added agents but improved output quality significantly.</p>
<p><strong>LLM-as-judge for validation.</strong> Early on, we used an LLM to validate other LLMs’ output. “Does this quiz question make sense? Is this slide coherent?” That’s circular. The validator had the same failure modes as the generator. We moved to rule-based validation for anything safety-critical. LLMs still validate style and tone. They don’t validate facts. This failure mode is documented in more detail in <a href="../../build-logs/llm-judge-india-failure/index.html">why LLM-as-judge stacks fail for Indian markets</a> — the underlying issue is the same regardless of geography.</p>
<p><strong>Centralized orchestration.</strong> We built one orchestrator that managed all agents. It became a bottleneck. Every new feature required changing the orchestrator. We should have built federated orchestration, where each agent cluster (planner, worker, validator) manages its own coordination. We haven’t refactored this yet. It’s still painful.</p>
</section>
<section id="then-vs.-now" class="level2">
<h2 class="anchored" data-anchor-id="then-vs.-now">Then vs.&nbsp;Now</h2>
<p>If we built Ostronaut today with 2025 tooling, here’s what would be easier:</p>
<table class="caption-top table">
<thead>
<tr class="header">
<th>What We Built by Hand</th>
<th>What Exists Now</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Context routing logic</td>
<td>LangGraph state management</td>
</tr>
<tr class="even">
<td>Tool-call validation layer</td>
<td>Built-in tool schemas in GPT-4</td>
</tr>
<tr class="odd">
<td>Agent orchestration</td>
<td>CrewAI, n8n workflows</td>
</tr>
<tr class="even">
<td>Retry and error handling</td>
<td>Framework-level retry policies</td>
</tr>
</tbody>
</table>
<p><strong>What’s still hard:</strong></p>
<p>Domain-specific validation. No framework gives you medical safety checks or financial compliance rules. You build that yourself.</p>
<p>Cost optimization. Frameworks don’t tell you which agents are burning tokens unnecessarily. You need observability and profiling. This is the same problem <a href="../../field-notes/indian-saas-agent-reliability/index.html">Indian SaaS companies are well-positioned to solve</a> — twenty years of optimizing for constrained infrastructure builds exactly this instinct.</p>
<p>Failure mode discovery. Agents fail in creative ways. A framework might handle retries, but it won’t tell you <em>why</em> an agent is producing inconsistent output. You learn that by watching production traffic.</p>
<p><strong>The real difference:</strong> In 2024, we had to understand agent internals to build anything reliable. In 2025, you can deploy agents without understanding them. That’s progress. But it creates Agent Debt.</p>
</section>
<section id="the-falsifiable-claim" class="level2">
<h2 class="anchored" data-anchor-id="the-falsifiable-claim">The Falsifiable Claim</h2>
<p>Teams that deploy agent systems without understanding planner-worker coordination, context boundaries, and validation layers will hit a quality ceiling within 3-6 months that no amount of prompt tuning will fix.</p>
<p>The ceiling shows up as:</p>
<ul>
<li>Inconsistent output quality (works 80% of the time, fails unpredictably)</li>
<li>Cost spirals (agents making redundant API calls, over-generating)</li>
<li>User trust erosion (one bad generation destroys confidence in 10 good ones)</li>
</ul>
<p>This isn’t a prediction. It’s a pattern I’ve watched repeat across every team that reached out after deploying agents without validation gates. The vendors selling “agentic platforms” are solving orchestration and deployment. They’re not solving validation, cost control, or failure mode discovery. Those are still your problem.</p>
<p>This dynamic connects to something broader happening in <a href="../../frameworks/agentware/index.html">the shift from software to agentware</a> — as the abstraction layer rises, the hidden complexity doesn’t disappear. It concentrates at the failure modes the frameworks don’t cover.</p>
</section>
<section id="the-question-worth-asking" class="level2">
<h2 class="anchored" data-anchor-id="the-question-worth-asking">The Question Worth Asking</h2>
<p>If you’re deploying agents today, ask this: <strong>Can you explain why an agent made a specific decision?</strong></p>
<p>Not “what did it output?” but “why did it choose this approach over alternatives?”</p>
<p>If the answer is “the LLM decided,” you have Agent Debt. You’re trusting a black box. That works until it doesn’t.</p>
<p>The teams that will build reliable agent systems aren’t the ones using the fanciest frameworks. They’re the ones who understand what happens when context bleeds between agents, when a planner makes a bad decomposition, when a validator misses a hallucination.</p>
<p>We learned that by building without frameworks. You can learn it faster now — but only if you look under the hood.</p>


<!-- -->

</section>

 ]]></description>
  <category>AI Engineering</category>
  <category>Agentic Systems</category>
  <category>Build Logs</category>
  <guid>https://talvinder.com/build-logs/multi-agent-before-agentic/</guid>
  <pubDate>Fri, 20 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>The Pass@k Trap: Why Running Your AI Agent 3 Times Makes Answers Worse</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/consensus-is-not-verification/</link>
  <description><![CDATA[ 





<p>Pass@k is the most popular reliability pattern in production agent systems right now. Run the same task k times, take a majority vote on the output, ship the consensus answer. It works beautifully for code generation — a function either passes the test suite or it doesn’t. The objective verification is external to the agents.</p>
<p>For factual accuracy, the pattern collapses. And most teams deploying it haven’t figured out why yet.</p>
<p>The failure is structural, not probabilistic. Consensus voting assumes that errors are independent and randomly distributed. If Agent A hallucinates, Agent B probably won’t hallucinate the same thing. With enough agents, truth wins by majority. This assumption holds for coding tasks because the test suite is the arbiter. It does not hold for factual claims because there is no test suite for truth.</p>
<section id="three-failure-modes" class="level2">
<h2 class="anchored" data-anchor-id="three-failure-modes">Three failure modes</h2>
<p><strong>Correlated hallucination.</strong> LLMs trained on similar data hallucinate in similar ways. Ask three instances of GPT-4o whether a specific paper exists, and if the title sounds plausible, all three will confidently confirm it. The errors aren’t independent — they’re correlated by training distribution. Majority vote amplifies the shared bias instead of cancelling it.</p>
<p>This is not a theoretical concern. Research from MIT and Google DeepMind has shown that Pass@k reliability for factual tasks degrades rather than improves as k increases, precisely because the error correlation exceeds the independence assumption. According to a 2024 study by Huang et al., LLM hallucination rates on factual recall tasks remain at 15-25% even with state-of-the-art models — and those errors are correlated across model families sharing similar training distributions. More agents, worse answers.</p>
<p><strong>The popularity trap.</strong> Consensus selects for the most common answer, not the most accurate one. In domains where the popular understanding is wrong — emerging science, contrarian market analysis, novel technical approaches — consensus voting systematically suppresses correct minority positions.</p>
<p>Three agents asked whether a particular drug interaction is dangerous will converge on whatever the training data’s majority position is. If the latest research contradicts the common understanding, the consensus will be confidently, democratically wrong.</p>
<p><strong>Strategic ambiguity.</strong> When agents are optimized for agreement (as many multi-agent debate frameworks encourage), they learn to hedge toward safe, middle-ground positions. Not because the middle ground is true, but because it minimizes disagreement. The agents aren’t lying — they’re conflict-averse. The output reads as measured and reasonable. It’s also systematically biased toward conventional wisdom.</p>
</section>
<section id="why-this-matters-now" class="level2">
<h2 class="anchored" data-anchor-id="why-this-matters-now">Why this matters now</h2>
<p>The “just run it three times” pattern is spreading fast. Every agentic framework has a retry-and-vote mechanism. LangChain, CrewAI, AutoGen — all support multi-agent voting as a reliability strategy. The assumption that consensus equals reliability is baked into the tooling. As of 2025, over 70% of multi-agent frameworks include a voting or consensus mechanism as a default reliability pattern.</p>
<p>Production systems using this pattern for anything beyond code generation are carrying unquantified risk. Customer-facing chatbots, research assistants, medical information systems, financial analysis tools — all domains where correlated hallucination is more dangerous than a single wrong answer, because the consensus gives the appearance of validation.</p>
<table class="caption-top table">
<colgroup>
<col style="width: 35%">
<col style="width: 18%">
<col style="width: 18%">
<col style="width: 28%">
</colgroup>
<thead>
<tr class="header">
<th>Reliability strategy</th>
<th>Works for</th>
<th>Fails for</th>
<th>Cost multiplier</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Pass@k consensus</td>
<td>Code generation (test suite verifies)</td>
<td>Factual claims, reasoning</td>
<td>3-5x compute</td>
</tr>
<tr class="even">
<td>Adversarial debate</td>
<td>Reasoning chains, logic errors</td>
<td>Shared knowledge gaps</td>
<td>2-3x compute</td>
</tr>
<tr class="odd">
<td>External anchoring (RAG verification)</td>
<td>Factual claims with source corpus</td>
<td>Novel analysis, opinion</td>
<td>1.5x compute + corpus maintenance</td>
</tr>
<tr class="even">
<td>Confidence-weighted routing</td>
<td>Domain-specific accuracy</td>
<td>Cold-start domains</td>
<td>1x compute + calibration data</td>
</tr>
</tbody>
</table>
</section>
<section id="what-actually-works" class="level2">
<h2 class="anchored" data-anchor-id="what-actually-works">What actually works</h2>
<p>The fix is not more agents or better prompts. It’s structural.</p>
<p><strong>Separate generation from verification.</strong> The agent that produces the answer must not be the same agent (or same architecture) that verifies it. Verification requires a different model, different training data, or — ideally — a non-LLM check against a ground-truth source. At Ostronaut, our validation agents use rule-based scoring with deterministic rubrics, not LLM-as-judge. The quality gate is independent of the generation pipeline.</p>
<p><strong>Adversarial framing over cooperative framing.</strong> Multi-agent debate works better when agents are explicitly tasked with finding flaws in each other’s outputs rather than converging on agreement. The incentive must be to disprove, not to confirm. This is the opposite of how most consensus systems are designed.</p>
<p><strong>Confidence-weighted routing.</strong> Instead of majority vote, weight each agent’s contribution by its calibrated confidence on that specific task type. An agent that is well-calibrated on medical queries but poorly calibrated on legal queries should have different voting weights in each domain. This requires per-domain calibration data, which most teams don’t collect.</p>
<p><strong>External anchoring.</strong> For factual claims, the gold standard is retrieval-augmented verification — check the claim against a curated, trustworthy source. Not RAG for generation (which has its own problems), but RAG specifically for post-generation verification. The verification retrieval corpus should be smaller and higher-quality than the generation corpus.</p>
</section>
<section id="the-pattern-that-misled-us" class="level2">
<h2 class="anchored" data-anchor-id="the-pattern-that-misled-us">The pattern that misled us</h2>
<p>The success of ensemble methods in machine learning created an intuition that more models = more reliability. In classical ML, this is largely true — bagging and boosting work because the base models have uncorrelated errors on well-defined features.</p>
<p>LLMs break this assumption. The base models share training data, architecture families, and optimization objectives. Their errors are correlated by construction. Treating them as independent voters is a category error borrowed from a domain where the independence assumption actually held.</p>
<p>I made this mistake early. When we built the multi-agent system, I assumed that running the content generation through multiple agents and selecting the best output would improve reliability. It didn’t. The agents agreed on the wrong things more often than they disagreed on the right things. We got reliability only after we separated the generation and verification functions entirely and made the verification independent of the generation architecture.</p>
</section>
<section id="the-open-question" class="level2">
<h2 class="anchored" data-anchor-id="the-open-question">The open question</h2>
<p>If consensus doesn’t work for truthfulness, what’s the right reliability primitive for multi-agent systems operating on factual domains?</p>
<p>Adversarial verification is better than consensus, but it’s expensive — you’re paying for agents whose job is to destroy, not create. External anchoring works but requires maintaining a ground-truth corpus, which is itself a maintenance burden that scales with domain breadth.</p>
<p>The field is converging on hybrid approaches — consensus for subjective quality, external verification for factual claims, adversarial debate for reasoning chains. But nobody has a clean, general-purpose pattern yet.</p>
<p>The teams that figure this out first will have a genuine architectural advantage. Not because their models are better, but because their reliability infrastructure is honest about what consensus can and cannot verify.</p>
<div class="schema-faq" style="display:none;">
<p>[{“q”:“Does running an AI agent multiple times improve accuracy?”,“a”:“For code generation with test suites, yes — Pass@k improves reliability because verification is external. For factual accuracy, no — LLM errors are correlated by training data, so consensus voting amplifies shared hallucinations rather than cancelling them.”},{“q”:“What is the Pass@k pattern in multi-agent AI systems?”,“a”:“Pass@k runs the same task k times across multiple agents and selects the majority answer. It works well for code generation but degrades factual accuracy because LLM hallucination errors are correlated, not independent.”},{“q”:“How do you verify AI agent output for factual accuracy?”,“a”:“Separate generation from verification: use a different model or architecture for checking, use retrieval-augmented verification against a curated source corpus, or use adversarial framing where agents are tasked with finding flaws rather than agreeing.”}]</p>
</div>


</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>Production AI</category>
  <category>Agentic Architecture</category>
  <guid>https://talvinder.com/field-notes/consensus-is-not-verification/</guid>
  <pubDate>Thu, 19 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>AI Is Making Your Team Slower — The Math Your CEO Won’t Show You</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/ai-speed-lie-team-velocity/</link>
  <description><![CDATA[ 





<p>Every company measuring AI productivity is counting the wrong thing.</p>
<p>They’re measuring output volume: PRs merged, lines written, tickets closed. They’re not measuring the cost of what ships: the review burden, the debugging time, the incidents caused by code nobody understood before it hit production.</p>
<p>When you count both sides, the math doesn’t work the way your CEO’s slide deck says it does.</p>
<section id="the-evidence-is-piling-up" class="level2">
<h2 class="anchored" data-anchor-id="the-evidence-is-piling-up">The Evidence Is Piling Up</h2>
<p>This week, The Pragmatic Engineer <a href="https://newsletter.pragmaticengineer.com/p/are-ai-agents-actually-slowing-us">catalogued what’s actually happening</a> inside companies that went all-in on AI coding agents. The findings aren’t theoretical.</p>
<p>Amazon’s retail engineering team saw a leap in outages caused directly by AI agents. The fix? Requiring senior engineer sign-off on all AI-assisted changes from junior developers. That’s not a productivity gain. That’s adding a bottleneck to compensate for unreliable output.</p>
<p>Anthropic — the company that builds Claude — ships over 80% of its production code with AI. Their flagship website degraded so badly that paying customers noticed before anyone internally did. The irony writes itself.</p>
<p>Meta and Uber are tracking AI token usage in performance reviews. Engineers who don’t use AI tools enough look unproductive. Engineers who use them indiscriminately look great on paper — until the bugs ship.</p>
</section>
<section id="the-three-taxes-youre-not-counting" class="level2">
<h2 class="anchored" data-anchor-id="the-three-taxes-youre-not-counting">The Three Taxes You’re Not Counting</h2>
<p>This is where the math breaks: <strong>teams that measure AI productivity only by output volume will see their incident rate and mean-time-to-resolve increase by 30% or more within 12 months</strong>, compared to teams that gate AI output with validation layers.</p>
<p>The mechanism has three parts.</p>
<section id="the-review-tax" class="level3">
<h3 class="anchored" data-anchor-id="the-review-tax">The Review Tax</h3>
<p>Every AI-generated PR still needs human review. But AI-generated code is harder to review than human-written code, because the reviewer can’t infer intent from the author’s history.</p>
<p>With human code, you know the developer’s context: what they were trying to solve, what trade-offs they considered, what they tested. With AI code, you’re reverse-engineering intent from output. That’s slower, not faster.</p>
<p>Amazon learned this the hard way. Junior engineers using AI agents shipped code that looked correct — clean formatting, reasonable variable names, passing tests — but had subtle logical errors that only surfaced in production. Reviewers couldn’t distinguish “AI wrote this well” from “AI wrote this plausibly.”</p>
</section>
<section id="the-refactoring-freeze" class="level3">
<h3 class="anchored" data-anchor-id="the-refactoring-freeze">The Refactoring Freeze</h3>
<p>Dax Reed, who built OpenCode, points out something every experienced engineer recognises: AI agents discourage refactoring. When code is cheap to generate, nobody wants to clean it up. Why spend an afternoon restructuring a module when the agent writes a new one in ten minutes?</p>
<p>The result is an expanding codebase where nothing gets simplified, patterns don’t converge, and cognitive load increases week over week.</p>
<p>This is the velocity trap. Short-term speed, long-term slowdown. Sentry’s CTO observed the same pattern: AI removes the barrier to getting started, which sounds great until you realise that “getting started” was never the bottleneck. The bottleneck was maintaining, debugging, and evolving what you built. AI makes the first part trivially easy and the second part measurably harder.</p>
</section>
<section id="the-incentive-poison" class="level3">
<h3 class="anchored" data-anchor-id="the-incentive-poison">The Incentive Poison</h3>
<p>When companies tie AI token usage to performance reviews, they’re telling engineers: “Use the tool, regardless of whether it helps.”</p>
<p>This is the corporate equivalent of measuring developer productivity by lines of code written. It rewards volume, punishes judgment, and guarantees that the engineers who are most careful about code quality look the least productive.</p>
<p>Engineers who know the AI output is mediocre ship it anyway, because slowing down to rewrite it makes their metrics look bad. The codebase degrades. The team slows down. The metrics still look great, because the metrics are measuring the wrong thing.</p>
</section>
</section>
<section id="what-this-looks-like-up-close" class="level2">
<h2 class="anchored" data-anchor-id="what-this-looks-like-up-close">What This Looks Like Up Close</h2>
<p>I’ve seen this pattern building multi-agent systems at Ostronaut. We generate training content — presentations, videos, quizzes. Early on, the agents were fast. They produced a complete training module in minutes. The output looked good. Formatting was clean. Structure was reasonable.</p>
<p>It was also wrong about 15-20% of the time. Not obviously wrong — subtly wrong. A slide deck where the concept progression didn’t build properly. A quiz where the distractors were too close to the correct answer. A video script that repeated a key point in slightly different words, creating confusion instead of reinforcement.</p>
<p>We didn’t fix this with better prompts. We fixed it by building a validation layer — automated checks that ran after every generation step, before anything reached a human reviewer. Content validation caught conceptual errors. Design validation caught structural problems. Integration validation caught mismatches between components.</p>
<p>That validation layer was harder to build than the generation layer. It took longer. It required more engineering judgment. And it’s the only reason the system works reliably.</p>
<p>The companies in Gergely’s article skipped this step. They deployed AI agents without validation gates, measured the output volume, and declared victory. Then the incidents started.</p>
</section>
<section id="why-better-models-wont-save-you" class="level2">
<h2 class="anchored" data-anchor-id="why-better-models-wont-save-you">Why Better Models Won’t Save You</h2>
<p>I used to think the answer was better models. If GPT-4 produces code that’s 80% reliable, GPT-5 will be 95% reliable, and eventually you won’t need validation.</p>
<p>That was wrong for two reasons.</p>
<p>First, the remaining failures are the expensive ones. The bugs that survive better models are the subtle, context-dependent bugs that cause production incidents. Better models don’t make validation cheaper — they make it more necessary, because what gets through is harder to catch.</p>
<p>Second, the validation layer isn’t just catching bugs. It’s encoding team knowledge. Our quality checks embed years of domain expertise — what makes a good slide progression, what makes a quiz effective, what makes a video script clear. That knowledge doesn’t exist in the model. It exists in the team. The validation layer is how you transfer institutional knowledge into the AI pipeline.</p>
<p>Companies that skip this aren’t just accepting more bugs. They’re disconnecting their AI pipeline from their institutional knowledge.</p>
</section>
<section id="what-to-measure-instead" class="level2">
<h2 class="anchored" data-anchor-id="what-to-measure-instead">What to Measure Instead</h2>
<table class="caption-top table">
<thead>
<tr class="header">
<th>What Leadership Measures</th>
<th>What Actually Happens</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>PRs merged per week (+52%)</td>
<td>Review time per PR (+40%)</td>
</tr>
<tr class="even">
<td>Lines of code written (3x)</td>
<td>Lines nobody understands (3x)</td>
</tr>
<tr class="odd">
<td>Time to first commit (-60%)</td>
<td>Time to resolve incidents (+35%)</td>
</tr>
<tr class="even">
<td>Token usage per engineer</td>
<td>Refactoring frequency (-70%)</td>
</tr>
</tbody>
</table>
<p>If you’re measuring AI impact, stop counting PRs. Start counting:</p>
<ol type="1">
<li><strong>Incident rate per AI-assisted commit</strong> versus human-only commits</li>
<li><strong>Review time per PR</strong> — is it actually decreasing, or are reviewers rubber-stamping?</li>
<li><strong>Refactoring frequency</strong> — is your team still simplifying code, or just adding to it?</li>
<li><strong>Mean-time-to-resolve</strong> for bugs in AI-generated code versus human-written</li>
</ol>
<p>The companies that will win with AI coding agents are not the ones that deploy them fastest. They’re the ones that build the validation layer first and measure what matters — not how fast code is written, but how fast <em>correct</em> code ships and stays correct in production.</p>
<p>Speed without verification isn’t velocity. It’s technical debt with a marketing budget.</p>
<div class="schema-faq" style="display:none;">
<p>[{“q”:“Does AI coding actually make development teams slower?”,“a”:“Yes, if you measure both sides. Teams that count only output volume — PRs merged, lines written, tickets closed — miss the downstream costs: longer review times, higher incident rates, and refactoring paralysis. Amazon’s retail engineering saw increased outages from AI agents and added senior sign-off requirements. Teams that gate AI output with validation layers avoid this productivity trap.”},{“q”:“What is the review tax in AI-assisted coding?”,“a”:“The review tax is the hidden cost of reviewing AI-generated code. Unlike human-written code where reviewers can infer intent from the author’s context and history, AI code requires reverse-engineering intent from output. Amazon engineers found they couldn’t distinguish ‘AI wrote this correctly’ from ‘AI wrote this plausibly’ — leading to subtle production bugs that only surfaced after deployment.”},{“q”:“How should engineering leaders measure AI coding productivity?”,“a”:“Measure incident rate per AI-assisted commit versus human-only commits, review time per PR, refactoring frequency, and mean-time-to-resolve for bugs in AI-generated code. Stop counting PRs merged or lines written. The engineering teams winning with AI coding agents build validation layers first and measure whether correct code ships and stays correct in production.”},{“q”:“Why don’t better AI models solve the production reliability problem?”,“a”:“Better models still fail on the hardest cases — and those are the expensive production incidents. More importantly, the validation layer isn’t just catching bugs. It encodes team knowledge: what makes a good architectural decision, what trade-offs matter, what patterns to avoid. That institutional knowledge doesn’t exist in any model. It exists in your team.”}]</p>
</div>


<!-- -->

</section>

 ]]></description>
  <category>AI Engineering</category>
  <category>Software Engineering</category>
  <category>Engineering Leadership</category>
  <guid>https://talvinder.com/build-logs/ai-speed-lie-team-velocity/</guid>
  <pubDate>Wed, 18 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Chain-of-Thought Has an Efficiency Tax</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/cot-efficiency-tax/</link>
  <description><![CDATA[ 





<p>Your AI agent now “thinks through” problems step-by-step. Your token costs just tripled. Did anyone on your team notice?</p>
<p>Chain-of-thought prompting is the default recommendation for improving LLM output quality. Every tutorial says it. Every framework enables it. And the advice is correct — CoT does improve reasoning on complex tasks. What nobody mentions is the cost.</p>
<p>Every major model provider now ships a reasoning mode — extended thinking, chain-of-thought, “deep research.” These modes generate 3x to 10x more tokens than their standard equivalents for the same task. Those tokens cost money. They add latency. And in most production systems, nobody is measuring whether the quality improvement justifies the spend.</p>
<section id="the-numbers" class="level2">
<h2 class="anchored" data-anchor-id="the-numbers">The numbers</h2>
<p>Here’s what the efficiency tax looks like in practice.</p>
<p>A content generation agent running a standard frontier model averages 1,200 input tokens and 800 output tokens per query. That’s roughly $0.036 per call at current pricing. Switch to the same provider’s reasoning mode, and the same task burns 1,200 input tokens plus 4,000-8,000 thinking tokens plus 1,200 output tokens. Cost per call: $0.12 to $0.22. A 3x to 6x increase.</p>
<p>At 10 queries a day, nobody cares. At 10,000 queries a day, you’ve added $840 to $1,840 in daily costs for a quality improvement you probably haven’t measured.</p>
<table class="caption-top table">
<thead>
<tr class="header">
<th>Approach</th>
<th>Tokens per call</th>
<th>Cost per call</th>
<th>Monthly cost at 10K/day</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Direct prompting (standard mode)</td>
<td>~2,000</td>
<td>~$0.036</td>
<td>~$10,800</td>
</tr>
<tr class="even">
<td>Chain-of-thought (standard mode)</td>
<td>~3,500</td>
<td>~$0.055</td>
<td>~$16,500</td>
</tr>
<tr class="odd">
<td>Reasoning mode (extended thinking)</td>
<td>~8,000</td>
<td>~$0.18</td>
<td>~$54,000</td>
</tr>
</tbody>
</table>
<p>That last row is 5x the first row. For many tasks, the output quality difference between row one and row three is negligible.</p>
</section>
<section id="why-teams-dont-measure-this" class="level2">
<h2 class="anchored" data-anchor-id="why-teams-dont-measure-this">Why teams don’t measure this</h2>
<p>Three reasons, all predictable.</p>
<p><strong>Accuracy bias.</strong> Teams optimizing AI systems measure quality metrics — accuracy, coherence, task completion rate. Token efficiency rarely appears on the dashboard. When the reasoning model produces a slightly better answer, that’s visible. The 5x cost increase lives in a billing page nobody checks until month-end.</p>
<p><strong>Scale hiding the problem.</strong> At low volumes, the tax is invisible. A startup running 500 queries a day doesn’t feel the difference between $18 and $90. But costs scale linearly with volume, and the moment you hit product-market fit, the tax becomes your second-largest line item after engineering salaries.</p>
<p><strong>The “just optimize later” fallacy.</strong> This is the same mistake teams make with database queries. Ship first, optimize later. Except “later” usually means “after we’ve built the entire pipeline around the expensive approach and switching costs are enormous.”</p>
</section>
<section id="when-the-tax-is-worth-paying" class="level2">
<h2 class="anchored" data-anchor-id="when-the-tax-is-worth-paying">When the tax is worth paying</h2>
<p>CoT and reasoning models earn their cost in specific situations.</p>
<p><strong>Multi-step logical reasoning.</strong> Tax-return calculations, legal document analysis, complex debugging. Tasks where the intermediate steps actually matter for correctness. The tax is justified because direct prompting fails outright.</p>
<p><strong>Low-volume, high-stakes decisions.</strong> Medical triage recommendations, financial risk assessments, safety-critical systems. When a single wrong answer costs more than a thousand correct ones, pay the tax.</p>
<p><strong>Tasks where you can measure the delta.</strong> If you can run both approaches on the same inputs and quantify the accuracy improvement, you can calculate the break-even. Most teams skip this step.</p>
<p>Where the tax is almost never worth it: classification tasks, structured data extraction, template-based generation, summarization, entity recognition. These tasks work well with direct prompting. CoT adds cost without proportional benefit.</p>
</section>
<section id="the-metric-that-matters" class="level2">
<h2 class="anchored" data-anchor-id="the-metric-that-matters">The metric that matters</h2>
<p>Cost per unit of quality. Not just cost per query, and not just quality per query — the ratio.</p>
<p>Define a quality metric for each task type. Run both approaches on a held-out sample. Calculate cost-per-quality-point for each. If reasoning-mode costs 5x more but only improves quality by 8%, that’s a bad trade. If it costs 3x more and improves quality by 40%, that might be worth it.</p>
<p>This is basic FinOps thinking applied to LLM inference. Cloud cost optimization is a mature discipline — rightsizing instances, reserved capacity, spot pricing. LLM cost optimization is in its infancy. Most teams are running the equivalent of on-demand instances at maximum size for every workload.</p>
</section>
<section id="a-practical-approach" class="level2">
<h2 class="anchored" data-anchor-id="a-practical-approach">A practical approach</h2>
<p>Route by task complexity. Not every request needs to go through the reasoning model. Build a classifier — a cheap, fast one — that scores incoming tasks on complexity. Simple tasks go to the direct model. Complex tasks go to the reasoning model. This is the same pattern as CDN edge routing: serve what you can cheaply at the edge, send the rest to origin.</p>
<p>At Ostronaut, we found that roughly 70% of content generation tasks hit a template fast path — no reasoning needed. The remaining 30% benefit from deeper processing. Routing saves more than optimizing any single model call.</p>
<p>The irony is that the routing classifier itself is a trivial LLM call. A $0.001 classification that saves $0.15 on the main call pays for itself in a single interaction.</p>
</section>
<section id="what-i-dont-track-yet" class="level2">
<h2 class="anchored" data-anchor-id="what-i-dont-track-yet">What I don’t track yet</h2>
<p>Latency cost. The efficiency tax isn’t just financial — reasoning models are slower. Time-to-first-token increases. Total response time increases. For interactive applications, the user experience degradation has a cost that doesn’t show up on any invoice.</p>
<p>I don’t have a clean way to quantify the latency tax in dollar terms. The financial tax is measurable today. The latency tax needs better tooling. If you’re building internal AI platforms, this is the metric to add next.</p>
<p>The teams that measure both — cost efficiency and latency efficiency, per task type — will have a significant operational advantage over the teams that just pick the most powerful model and hope for the best.</p>


</section>

 ]]></description>
  <category>AI Operations</category>
  <category>Production AI</category>
  <category>Cost Optimization</category>
  <guid>https://talvinder.com/field-notes/cot-efficiency-tax/</guid>
  <pubDate>Wed, 18 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>AI-Generated Code Is Flooding GitHub — Here’s What It Costs You</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/github-slopocalypse-trust-tax/</link>
  <description><![CDATA[ 





<p>GitHub’s value was never storage. It was legible history.</p>
<p>Every commit told you who made a decision, why they made it, and what changed. That’s what made open source work at scale—you could trace a bug to a specific human judgment, review the reasoning, fix it. GitHub’s 2025 Octoverse report shows that AI-assisted commits now account for over 40% of new code on the platform. Copilot alone generates 46% of code in files where it’s enabled, according to GitHub’s own metrics.</p>
<p>Now that history is being flooded with AI-generated code, and the entire trust infrastructure is collapsing.</p>
<section id="the-trust-tax" class="level2">
<h2 class="anchored" data-anchor-id="the-trust-tax">The Trust Tax</h2>
<p>The Trust Tax is the additional cognitive and temporal cost developers now pay to verify code provenance before they can use it. Every AI-generated commit adds verification overhead — and that cost is growing faster than AI’s productivity gains.</p>
<p>When GitHub launched, the excitement wasn’t about free hosting. It was about confidence. Git’s original value proposition was perfect historical records. You could pinpoint a commit in space and time and feel confident in the record of code changes in a way that you rarely feel confident about anything in software.</p>
<p>The system assumed human intentionality in every commit. When you saw a change, you knew a human had made a deliberate decision. Maybe it was wrong, but it was <em>legible</em>—you could understand the reasoning, challenge it, fix it.</p>
<p>AI code generation breaks this assumption.</p>
<p>A commit that says “optimized database queries” might mean: a developer profiled the code, identified N+1 queries, rewrote them, and tested the result. Or it might mean: an LLM generated plausible-looking SQL based on a vague prompt, and no one verified it works.</p>
<p>You can’t tell from the commit. You can’t tell from the diff. You have to read the code, understand the context, and verify the claims. Every single time.</p>
</section>
<section id="the-mechanism" class="level2">
<h2 class="anchored" data-anchor-id="the-mechanism">The Mechanism</h2>
<p><strong>Within 18 months, the median time-to-trust for evaluating a new GitHub repository will double for experienced developers, and the variance will increase by an order of magnitude.</strong></p>
<p>Before: You evaluated a repo by checking commit frequency, reading a few key commits, scanning the contributors, maybe looking at issue resolution patterns. Total time: 10-15 minutes for a mid-sized library.</p>
<p>Now: You do all of that, plus: scan for AI-generated patterns (repetitive structure, suspiciously perfect formatting, generic variable names), check if tests actually run, verify that documentation matches implementation, look for signs of copy-paste from LLM output. And even after all that, you’re less confident than you used to be.</p>
<p>The variance increase is worse than the median shift. Some repos will be obviously human (active maintainers, clear decision history, coherent architecture). Some will be obviously slop (generated README, no tests, commit messages that read like ChatGPT). But most will be in the middle—partially AI-assisted, unclear provenance, uncertain quality.</p>
<p>That’s where the tax gets expensive.</p>
</section>
<section id="where-this-is-already-happening" class="level2">
<h2 class="anchored" data-anchor-id="where-this-is-already-happening">Where This Is Already Happening</h2>
<p>JavaScript got hit first. Every hard-fought factoid about framework internals gets buried under LLM-generated tutorials that are 70% correct and 30% hallucinated. The slopocalypse is now accelerating across all languages.</p>
<p>At Zopdev, we’ve started seeing this in infrastructure-as-code repos. Terraform modules that look reasonable at first glance but have subtle bugs—wrong IAM permissions, missing tags, inefficient resource allocation. The modules are clearly AI-generated (the structure is too uniform, the variable names too generic), but someone committed them with a human-sounding message.</p>
<p>The Trust Tax here is expensive: you have to audit every resource definition before you can use it.</p>
<p>The pattern is consistent across domains. AI-generated commits don’t carry human intent. The commit message that says “refactored for clarity” might be hallucinated. The code that looks clean might be untested slop copied from three different StackOverflow answers. The diff that claims to fix a race condition might introduce two new ones.</p>
</section>
<section id="what-i-got-wrong" class="level2">
<h2 class="anchored" data-anchor-id="what-i-got-wrong">What I Got Wrong</h2>
<p>I initially thought the solution was better tooling—automated detection of AI-generated code, reputation systems for contributors, verification badges for human-reviewed repos.</p>
<p>That’s not wrong, but it misses the deeper problem.</p>
<p>The Trust Tax isn’t a tooling problem. It’s an epistemological problem. GitHub’s value was that you could reconstruct intent from history. AI-generated code has no intent. It has a prompt and a probability distribution. You can’t reconstruct reasoning that never happened.</p>
<p>Better tools can reduce the tax, but they can’t eliminate it. You’re always going to pay more to verify machine-generated code than human-written code, because the verification burden shifts from “did this human make a good decision?” to “is this code even coherent?”</p>
</section>
<section id="the-adaptation-pattern" class="level2">
<h2 class="anchored" data-anchor-id="the-adaptation-pattern">The Adaptation Pattern</h2>
<p><img src="https://talvinder.com/field-notes/github-slopocalypse-trust-tax/data:image/svg+xml;base64,<?xml version="1.0" encoding="utf-8"?><svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" data-d2-version="v0.7.1" preserveAspectRatio="xMinYMin meet" viewBox="0 0 938 583"><svg class="d2-474229315 d2-svg" width="938" height="583" viewBox="-9 -9 938 583"><rect x="-9.000000" y="-9.000000" width="938.000000" height="583.000000" rx="0.000000" fill="transparent" class=" fill-N7" stroke-width="0" /><style type="text/css"><![CDATA[
.d2-474229315 .text {
	font-family: "d2-474229315-font-regular";
}
@font-face {
	font-family: d2-474229315-font-regular;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA5IAAoAAAAAFcwAAguFAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgXd/Vo2NtYXAAAAFUAAAAkwAAALYDJAN5Z2x5ZgAAAegAAAfFAAAKfI92LadoZWFkAAAJsAAAADYAAAA2G4Ue32hoZWEAAAnoAAAAJAAAACQKhAXmaG10eAAACgwAAACKAAAAkEIuB1lsb2NhAAAKmAAAAEoAAABKMlYv/G1heHAAAArkAAAAIAAAACAAPAD2bmFtZQAACwQAAAMjAAAIFAbDVU1wb3N0AAAOKAAAAB0AAAAg/9EAMgADAgkBkAAFAAACigJYAAAASwKKAlgAAAFeADIBIwAAAgsFAwMEAwICBGAAAvcAAAADAAAAAAAAAABBREJPAEAAIP//Au7/BgAAA9gBESAAAZ8AAAAAAeYClAAAACAAA3icbMw7TsIAHIDxX22tr6r1Vd/aizgYXZroBgdgJGEmcCEIzEyEAY5COMifhDDyjd/wQyKVoJCZolJK5Wpfvv1q/Gvr6htGsP8/Gn9aOnoGEbGOTaxiGYuYxywmMY7RTj1UonYklTmWO3HqzLkLhUtXrpVu3Lpz70Hl0ZNnL169effhky0AAAD//wEAAP//yBgeiwB4nGxWb2gb9/l/vl+ddXEk2T5Lp5NsydLd2XeWZEu2TndnR/IpliXH8T85kv1L7CTOL4kbJ/GSJQ4kBNyGkv4JjDFBExZY25UtbwLNulJIB30zMtp5659QKOs6Rgl74QVa2Kp5Y9D5NO4kuTbs1Vccuuf5PJ8/z/egAeYBsIxvgwUaoRlagQaQKJbqYkWRJ1VJVXnGooqIIufRn/USQgcThKIQ/ZmvMtdv3EBHnsO3t76374Xl5fcXr13Tf7jxVI+jT54CBgsA9uMSNAIF4CQlURBE3mq1OCUnL/Lkh4H3A63BFqI5+KcvF7+c1/6WRt9fWlIvDA5e0BdwaevS+joAAIJEZRO341fBD9DACYKcUBQp7mZIQeA5q5V2ud1SXFEZqxUVCs9PTL5QTB3z9bZlwtpxKX5Ui40HouIp+6G7K+fvFvqDio8bvlooXM90c4neOABgWADACVyCPQZOiZLibtpl5UUprsgJgecX7t19/bU7cxNXrly5MoFL91997RfZH6ytvWhiWwCAB7hkzCpRErVQNIBXn6OPcAkaqs9ZeqGIAri09e4BqL+Hn8UlYyaJkpxuNyMpiuqUKJ5KKCpPWniLyLvdNLWw9JydsRN22r72zPQeC5FYU9cShIXEJf1nXI7jchxa3LqEzvWsRO7ob6LZO5GVHv3HZg8fAPoPLgFp9OBlluapv3yAnnyAx0dHtx5WcRyubOIoLhmam9xSElXlUzF/Wq1oZGRFK4ZykZ7R0Ix23q6snUPP68/mjwrC0Ty6qd84t6YANjRCb6MytEEnAMMZIqkJUyBSNOWiKd4QX4wrqmyK9mjo0I9+QkW6w+P+IHd63/xMlrRwh9y8xl8/GbcfHJ6ZowIDfNA16A5dOKr/YZ8vnOECLzenYqEuwFCobKJv8To4IWgiF3mSpySarPZymY0M9TgrSbvdKMQdDFrITAGz+e4Tp5InRlP5ZC6wnw+m7aw/jtcfHfGLL10uXtVyywszp7lgxcdU+YlWNtFbqGxw+b+9V7de6/6zqeEVrS/nDdMxf09OLI5w+9yd7Iw9tTpTWE1xjOL0xOYGist+l+pnDd/FKpvoi/oMVc7M4qIs1clS5e1G/z56MXlSDWtBopglLb5J7/5UYLBDTAuj9hev569oHW3F97YGBn2h3IjuY2LFgcOnAZv4f4/K4IHArglol5Vkt4NjYU2qEDN8XksvqcefQVj/VcPhUT7Z7g/kP0REelA6ZB9azc+samtnHd7GqWM0pbg6kDA+lTd56gBAafxZdXfwsionajzxHE1LNE/9fyaTO8iEW1rbfdnlZfRzrWFq/HAjmbYvTo3oxwHAAr2VIPoalaEfhmBq20WysOMwi0q0kQyX1cpzYlWDmuaWuua0y+2spZcTqv/51/wlgW31ck6PGJ/td3U67i9RTN9MXOQcrV39i3NzqYuT4aFUJJIaUkZnpdhsE9vS5pl4kk0HBt2ErdsXiDoIVzYiT4fJhnSLHEhMhihbu4vpUId6J2Po7bQsp1KynNZvDQlcG0E4w7QYNbkpAKDP8Tq4zKzXPWok3fQnVShY+Kn41IFCT19XsguvP1piYyeP6x+hUFYTuvQ3oFKBHAC8gx9iAVgAsAK3VvVnobIJf8Tr0Fzly4xvTdT70VChqZEgSdset31Qxme2bjsphDSCqGLC36CyUU2iJGP5GMzuQkZun4UsaQlORgbSzcJ0z8TBQk9UyRZ6YkoWbYzysf6eUKIOd0J/o3bU50bl2ty1HjvnzpIWfnp7cLPYrrlr/v07KkMztO/y7+6M0y43ak4up9PLydSZdPpMKj01ldamp2vZS60WZlZT2eXi7Nmzs8VlMPeHhL5F5Vr2vkNnukoQGdq5c38YSNl8ZPFU8sQAN8Lha+b6SHey2sf4nQFf98uXC1e1jra5e8i6a38YGZfQF/U+DbJqlt82sipRlp0ZRy8R/olwNej7Wbwn83g75B8/OOLrNoPu90e3ppD1u5TXPbaIyrUbrDpNbUtVifaOhfxMi93VHBjxoo0jUWXvGEHENb12x/oqm+gmKkPY9JGomqtBTgiCGMXbWa5R7WY6sEHUp4lFPhTMRvr6WKmdy4Tn873Tvm6vEoxGOvra+WxvKG8XfaqX7Q14OWavg5VDyXyQSTg9YR/jp20OVo2KmW6zv6eyiXL4IjA1H/Oyqkrm4tj281fTQ2OTe3M3b7JhR4e9xRWzL4whh9Zw69aIXu7tbyQ00mbWmqhsok/QhuG7XZmgamv1ydRYMdInJDmDF27SfvI4SuifZzUxgub1tsnuPkBgB0C/RRvgAJAsO+5ny3tvzR2zMTbCxuw9duhNtKF/3TnG82OdyKW3AYImAPRLtAFeAEkVJab2oiqRDF/7/iHJpp++Mj9s8zgIm9uW/L9XXp8/4GhrIhwee0Z/uuIMu1xh58o3/7zs7qHpCHPZnMleiZl42nfqo6q7oDXhhRa/vWWPqzGkNNt+M3fa5rURNtfewzPvUrHcp1ZiGDckezvRX/V/BMY4diyIHFvlvsleY79w8Gv0GH2GBeBhBazAw526r+Ae2qh/3xQKaMOYs/I7PA4qfgg2AMo0dDU8nkDA4wkE8Ljf6+no8Hj9Rg0ORdBjdM6o4ZRZmkMPUETTAOC/AAAA//8BAAD//zIXLUwAAAAAAQAAAAILhYHovrVfDzz1AAMD6AAAAADYXaChAAAAAN1mLzb+Ov7bCG8DyAAAAAMAAgAAAAAAAAABAAAD2P7vAAAImP46/joIbwABAAAAAAAAAAAAAAAAAAAAJHicHMohDsJAFAbheX8FhqArmmYDGEIopo4gUbhnCMsBOAmSU3APdNEIboAG09QtYdWM+HTlSAcKFNrQ6ky0EdHeRL2IKql04WADrRa4dTRasrYPjc2pbWClgNOzI+HFFtcMV52dZ3vC7UZlTqnA3p6M7c4k98uUHof0+P8PAAD//wEAAP//ULsbtgAAAAAALAAsAFAAgACeAKoAugDsAP4BIgFaAY4BvAHuAiICRAKwAtIC3gL6AywDTgN6A64D4gQCBEIEaASKBKYE0gUCBQ4FGgUwBT4AAAABAAAAJACMAAwAZgAHAAEAAAAAAAAAAAAAAAAABAADeJyclN1OG1cUhT8H221UNRcVisgNOpdtlYzdCKIErkwJilWEU4/TH6mqNHjGP2I8M/IMUKo+QK/7Fn2LXPU5+hBVr6uzvA02qhSBELDOnL33WWevtQ+wyb9sUKs/BP5q/mC4xnZzz/ADHjWfGt7guPG34fpKTIO48ZvhJl82+oY/4n39D8Mfs1P/2fBDtupHhj/heX3T8Kcbjn8MP2KH9wtcg5f8brjGFoXhB2zyk+ENHmM1a3Ue0zbc4DO2DTfZBgZMqUiZkjHGMWLKmHPmJJSEJMyZMiIhxtGlQ0qlrxmRkGP8v18jQirmRKo4ocKREpISUTKxir8qK+etThxpNbe9DhUTIk6VcUZEhiNnTE5GwpnqVFQU7NGiRclQfAsqSgJKpqQE5MwZ06LHEccMmDClxHGkSp5ZSM6Iiksine8swndmSEJGaazOyYjF04lfouwuxzh6FIpdrXy8VuEpju+U7bnliv2KQL9uhdn6uUs2ERfqZ6qupNq5lIIT7fpzO3wrXLGHu1d/1pl8uEex/leqfMq59I+lVCYmGc5t0SGUg0L3BMeB1l1CdeR7ugx4Q493DLTu0KdPhxMGdHmt3B59HF/T44RDZXSFF3tHcswJP+L4hq5ifO3E+rNQLOEXCnN3KY5z3WNGoZ575oHumuiGd1fYz1C+5o5SOUPNkY900i/TnEWMzRWFGM7Uy6U3SutfbI6Y6S5e25t9Pw0XNnvLKb4i1wx7ty44eeUWjD6kanDLM5f6CYiIyTlVxJCcGS0qrsT7LRHnpDgO1b03mpKKznWOP+dKLkmYiUGXTHXmFPobmW9C4z5c872ztyRWvmd6dn2r+5zi1Ksbjd6pe8u90LqcrCjQMlXzFTcNxTUz7yeaqVX+oXJLvW45z+iTSPVUN7j9DjwnoM0Ou+wz0TlD7VzYG9HWO9HmFfvqwRmJokZydWIVdgl4wS67vOLFWs0OhxzQY/8OHBdZPQ54fWtnXadlFWd1/hSbtvg6nl2vXt5br8/v4MsvNFE3L2Nf2vhuX1i1G/+fEDHzXNzW6p3cE4L/AAAA//8BAAD//wdbTDAAeJxiYGYAg//nGIwYsAAAAAAA//8BAAD//y8BAgMAAAA=");
}
.d2-474229315 .text-bold {
	font-family: "d2-474229315-font-bold";
}
@font-face {
	font-family: d2-474229315-font-bold;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA5IAAoAAAAAFcQAAguFAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgXxHXrmNtYXAAAAFUAAAAkwAAALYDJAN5Z2x5ZgAAAegAAAe9AAAKXID3m3xoZWFkAAAJqAAAADYAAAA2G38e1GhoZWEAAAngAAAAJAAAACQKfwXjaG10eAAACgQAAACPAAAAkEYqBeRsb2NhAAAKlAAAAEoAAABKMcAvcm1heHAAAArgAAAAIAAAACAAPAD3bmFtZQAACwAAAAMoAAAIKgjwVkFwb3N0AAAOKAAAAB0AAAAg/9EAMgADAioCvAAFAAACigJYAAAASwKKAlgAAAFeADIBKQAAAgsHAwMEAwICBGAAAvcAAAADAAAAAAAAAABBREJPACAAIP//Au7/BgAAA9gBESAAAZ8AAAAAAfAClAAAACAAA3icbMw7TsIAHIDxX22tr6r1Vd/aizgYXZroBgdgJGEmcCEIzEyEAY5COMifhDDyjd/wQyKVoJCZolJK5Wpfvv1q/Gvr6htGsP8/Gn9aOnoGEbGOTaxiGYuYxywmMY7RTj1UonYklTmWO3HqzLkLhUtXrpVu3Lpz70Hl0ZNnL169effhky0AAAD//wEAAP//yBgeiwB4nFxWWWzb9hn//n9Roi3TB0WR1H2QFin5kGNRFH3Ilg85jh3JJ+I4rY80CNZkjpMscRa3yNaHZgHWKstWZV26DNkWZNgGpA9BMCAr4A3YSxc0b2nXp23dVuyhD6kWCMMayNJASo7jPkgkKPH7ft/v+EgwwwwAPoavgwnqoRlswAIodJAOKbIskpqiaSJv0mREkzPYVv7VHTlCRCJEW+CG//XVVZRdwde3T72cPXbsv6v9/eVbv/+gfBWd/wAAV54B4FGcg3qgARhSkSVJFi0WE6MwoiyS/255u7nR3UhQzmeP7j36WfjDMJpMJrtPK/H18vdwbnvj5k0AAATRShHvwzfADWAWJEmNJxJKjONJSRIFi4W1c0osofEWtDz31vyhq3Op48EppyZ2TLQvHAinHFNzVObH66fem1WEFd4bWxk5frbVuXQUMGQBcAbnwFqdWIlxHGu3WERZiSUSalySRDH74Pg7szPXjnZ6euaj0fkeD86lr509+874xfDS1NSRkIEvCwCf4RyYjCp0Nq+Dr15H/8Q5MFevB9lsHmGc2y5cgp378Ps4B37jd4bjeCWR0BiFFvURNZEkRVkWfZhls788abVZCSttffUXV8h6E6Euzy7HCaKOxLny3zyDPt+gBwnbG18Gpmf8N7/66qZ/Zjrwpd7DCYAZnANS7yGqQVakH91Hz+7jlkuXtgtVHBOVIp7GOV0lsyCptEIblBonFjT1xpXrfZqW/MGb1Lt30Eo5fzSTOYrWy7fvvAsY2ipF9DEqgRNEAF7QxdEMXUjZUImlRV1zLZbQVEOrP6RnLuexGPEPtapda32r39i0Ev7xOmeImUr6qcOpqcXmoOxgX/G2nj5X/lzxiOd45rC13evgAQDDcKWIObwFdp01syDJIinSCksazQz55FhCjYsCyXIcGguOegnqfJ7wpoXkYldydVFKLHRE7GEqGFDx1t2Myzv4rcyh11Kb+zNXOj+yNRl8tFaKaAuVwPV1v4nCrtssyDl2ZvjAt9PRcc+YGFBTqX2OKNMXWqAGLszNbwz4+FVvZngoyzYfDbjBwC5XiqiEt4CBwA5XRmFZVV5gSaq1ebp0pn81HulxWvKbVsK1HztkG9NuFxNd1NuvzV4Y9Dgyv90e7XaJm3bnR7am0fGJMcAG9n+gEjhq/Ow00akhgxynxHTsJiWud0H+8XMjo6f6x5e7CFz+1Lq/W010Sys/vS93CAlqcGNudiOVWkszofqEEjzi8qG+iNpV9YwDAG3gh/pR95W2N5Qsq7Ai/dLISOvMqD/e4m50UW7fkSPoO+tmt7oQpyynzOag5DtffhPABEKlE5OoBF3QD5MGM5Ia14nQzaTujMArrFgLqCAbOuj2slssJl3wGmlM9VwUJOMvT/tWesYZd8DhivStqB3B302T9fFFzeu3CZGZpVfSlya9suz1ynIkNiSHFGeQcg88dvV0JMNEY9jvjrUQtnR7cjpMrTUI9t7JVmszx9j6R5XZKHrYFpEj4XCkrZxvdfItJpPD6fFWuRnWxTY8que65k2WFmkDJUkP50nPwdjsRN4b8IQdeOvuEWf72nL5EQomwk6+fA8qFdAA4K/4MZYgCAAkCPBWtXaliGx4C5qrDtrJqi7qnzP9ebreTFpsVIh6+SAWtz/lbQitm8kqJpMXlfRqCq3oi0ZXaw8y8vlxWM/k/m51mAlOds8czHsDoX36VxcqDPk728NC9w7cfeV7tcPO3KhUm7vW48W5N61EIPt8cFRI+Tr3zF31r+GF5q9t+91o15RGXOpMOn0mlTqdTp9OdUajndHOzlr2Bjbm5y4MXMwODWf0CFb3xgHMoRIw4APgd9EZdpJknmV214aO0zshv3QiuZoIJF3maSmx0N5mDz/Av+l2id8/f2gz5XZO/wi1Pl8aerYPoJJRPwBgVjWj7E4oFE2hTS9mG520OEeEasAH9Q31+fNwP/hJxuE3Au4NdG8votbddNe8ha6hEtj26FhNXZVhd0ZiPVZHo7PFM2BHhcOxbrP5DYKIxMqfAQK2UkQ/RyWQDf/IGmds+rgkyVGsxneLsXaO92HWbnnc/ao0IqT8QZ836vL1h08e6j3sH3HFXb29UmAgcoKS/EtON8/QHGOlWnsjYwuyY9HOyQ5nU4PYGx1drmaCrhTRabwBfPUZo4qqpin6lnhhocLSdDpDv37xouilnFae0ahvLjxct1y+fP7DtpCFWLNQ1VrJShH9DxV0n+3JAF1bo3+Zncj7Ah6Jy282mPyT1Noyipf/rkZcXnSg3DIW6gAEFACqoAI0Aigmha89ezXFdP/X14esjJWoZ6zDV2+jwhehrCxnQ1+UW4zeTQCoiAr6c1Vh5BduJHmx9n5Dkk03rt3qsHJWos5WJ9z44Xu39lE8RdTb62WEn8yw7Szbzs5U/jPHdrBsOzen16Uqg2gbFXTX7+qiaXugNeFNLtjsIm11obCV/OP18Qablaij65NX7/I903+yEGeRudXrQv/6RNgfEsfFT8oNg4faAPR9IsBt9AQ9xRKIsAEWECG34yf4GBV23l2G86hQbgFUeR/3wjx+DA0AtGHkalhC0WgoFI3i3jZRbNM/eg0BcegJ+q5eg1GDrIAeIe7ECQD4PwAAAP//AQAA//+euyARAAAAAAEAAAACC4WRceH3Xw889QABA+gAAAAA2F2ghAAAAADdZi82/jf+xAhtA/EAAQADAAIAAAAAAAAAAQAAA9j+7wAACJj+N/43CG0AAQAAAAAAAAAAAAAAAAAAACR4nBzKqwrCYBjH4d/7Fwbi8ABTtBjGQPBUFdwX3iIYPrAYvBaDd2D3IixWb8Du3axMtvaER2/OfEGhrnRkqxvRNkQlRFVErZnpyUkDlgq4/SgUWCihsAtTTch1wC1jZ3O8c8VV4lq1z5trD9w+jO3OSHtK9Ugl+uqSakhuGQ71q/EfAAD//wEAAP//jmoURwAAAAAsACwAUAB8AKAArAC8AO4BAAEeAVYBiAG0AeYCGgJAAqgCygLWAvIDJANGA3IDogPWA/YEMgRYBHoElgTCBPIE/gUKBSAFLgAAAAEAAAAkAJAADABjAAcAAQAAAAAAAAAAAAAAAAAEAAN4nJyUz24bVRTGf05s0wrBAkVVuonugkWR6NhUSdU2K4fUikUUB48LQkJIE8/4jzKeGXkmDuEJWPMWvEVXPATPgVij+Xzs2AXRJoqSfHfu+fOdc75zgR3+ZptK9SHwRz0xXGGvfm54iwf1E8PbtOtbhqs8qf1puEZYmxuu83mtZ/gj3lZ/M/yA/epPhh+yW20b/phn1R3Dn2w7/jL8Kfu8XeAKvOBXwxV2yQxvscOPhrd5hMWsVHlE03CNz9gzXGcP6DOhIGZCwgjHkAkjrpgRkeMTMWPCkIgQR4cWMYW+JgRCjtF/fg3wKZgRKOKYAkeMT0xAztgi/iKvlHNlHOo0s7sWBWMCLuRxSUCCI2VESkLEpeIUFGS8okGDnIH4ZhTkeORMiPFImTGiQZc2p/QZMyHH0VakkplPypCCawLld2ZRdmZAREJurK5ICMXTiV8k7w6nOLpksl2PfLoR4Usc38m75JbK9is8/bo1Zpt5l2wC5upnrK7EurnWBMe6LfO2+Fa44BXuXv3ZZPL+HoX6XyjyBVeaf6hJJWKS4NwuLXwpyHePcRzp3MFXR76nQ58Turyhr3OLHj1anNGnw2v5dunh+JouZxzLoyO8uGtLMWf8gOMbOrIpY0fWn8XEIn4mM3Xn4jhTHVMy9bxk7qnWSBXefcLlDqUb6sjlM9AelZZO80u0ZwEjU0UmhlP1cqmN3PoXmiKmqqWc7e19uQ1z273lFt+QaodLtS44lZNbMHrfVL13NHOtH4+AkJQLWQxImdKg4Ea8zwm4IsZxrO6daEsKWiufMs+NVBIxFYMOieLMyPQ3MN34xn2woXtnb0ko/5Lp5aqq+2Rx6tXtjN6oe8s737ocrU2gYVNN19Q0ENfEtB9pp9b5+/LN9bqlPOWIlJjwXy/AMzya7HPAIWNlGOhmbq9DUy9Ek5ccqvpLIlkNpefIIhzg8ZwDDnjJ83f6uGTijItbcVnP3eKYI7ocflAVC/suR7xeffv/rL+LaVO1OJ6uTi/uPcUnd1DrF9qz2/eyp4mVk5hbtNutOCNgWnJxu+s1ucd4/wAAAP//AQAA///0t09ReJxiYGYAg//nGIwYsAAAAAAA//8BAAD//y8BAgMAAAA=");
}
.d2-474229315 .text-italic {
	font-family: "d2-474229315-font-italic";
}
@font-face {
	font-family: d2-474229315-font-italic;
	src: url("data:application/font-woff;base64,d09GRgABAAAAAA5YAAoAAAAAFnAAARhRAAAAAAAAAAAAAAAAAAAAAAAAAABPUy8yAAAA9AAAAGAAAABgW1SVeGNtYXAAAAFUAAAAkwAAALYDJAN5Z2x5ZgAAAegAAAfNAAALAIt8N9ZoZWFkAAAJuAAAADYAAAA2G7Ur2mhoZWEAAAnwAAAAJAAAACQLeAjIaG10eAAAChQAAACQAAAAkEAVBFVsb2NhAAAKpAAAAEoAAABKNIox9m1heHAAAArwAAAAIAAAACAAPAD2bmFtZQAACxAAAAMmAAAIMgntVzNwb3N0AAAOOAAAACAAAAAg/8YAMgADAeEBkAAFAAACigJY//EASwKKAlgARAFeADIBIwAAAgsFAwMEAwkCBCAAAHcAAAADAAAAAAAAAABBREJPAAEAIP//Au7/BgAAA9gBESAAAZMAAAAAAeYClAAAACAAA3icbMw7TsIAHIDxX22tr6r1Vd/aizgYXZroBgdgJGEmcCEIzEyEAY5COMifhDDyjd/wQyKVoJCZolJK5Wpfvv1q/Gvr6htGsP8/Gn9aOnoGEbGOTaxiGYuYxywmMY7RTj1UonYklTmWO3HqzLkLhUtXrpVu3Lpz70Hl0ZNnL169effhky0AAAD//wEAAP//yBgeiwB4nHxWWWwb59W995vRjBZqIYccmrREivzIGYkaLuKIHNESSUnULlKKbEu/ftuUo3hf2gpxU9tNjLQWkDpB46qGW6BBURfoghR+s/tSoHDQug+CUwFN4Rbp8tImlQO7QBJBLdqgGhYzlGTKD30ZDPBh7j3n3HvON1AFAQDyeXITGKiBRrCBA0AVfAyjahp1MqosU57XZEHgA1dx5epbbO7Q39q+/2/Fy4589ScTf3/+Nrm5eR6/Unz1Vf3wtRMn/u/JEz2Ev38CAEBK7wLg78gy1IAVQOBVWZJkynGIqkBlyn+w734tW8uyblX/NR4/lJ+2fXQGLy8udp3tTp3Sp8ny5uLqKgBCqrRBwuS74AWo8ktSoitD1Ljo5CWJ+huIwy6KajypOTkO/ROnk7FDV/Ld03uSQlLad3Qg4B/vacu10kDRkrs0Wbh5cUQLtbfK6eOXenuKida9cW/YwAoUgCRNrIKhgBoXHXaOo7IaTyYTXRKldOnLr785c+sLs7Mzr+ROHUuS5a9dvvjTE30Hv71QPGPwRaMGbJBlYMwKDF2aXDJIbJ/hDbIMVeUzH0+XJr+E9nqyvHlnoHxOfk6WwWWeC05VMyoIyaRGeYYyhm48Q5eKKZEdvl9cmsjXuC3s1C+UtMhyDdXjZFn/3rVr+MLmIr6onO24of8Qj9xQzij6dUAoApAEWQbeqE01H0+ZH790rx7frX/nJVLI5TbvlvFPlTbISbIMoqlzoitpYOAcdlNe45VDz/FzHDs+OVHTN9R9yDGd39981XLmpCPqwkX99bB/uHDkHN7Qz12/bGgqlzbwX7gOdoOd078zOFVTGapRjpPjSU3bmeLdvrwyPq/KaSsrZBay1Syds0lTAcURbw7kEt5Oy+GZ4ctH1DZfWnePBqN9kegfJH9orBjPps19A29pAz8hK+Awttrpl2TKU0HleTWZNOfZQOR4hhjT9HM8L4qP5bSVsWevF2SRBA6GzfaJQC7hibX7p2nErlrafGmycu/5lo5Ds0brvtBYUc2kQ8FHkh8QgqUNvIPr0LyLnbmW3LZsTo57f+q4UlhIKL1iWJBaYrPJ1L7WpOh3Fywni4MXZqJ+V8zpGFzMDQy7rXF7sMxFLm0QuYLLU+3+t3j7bEyTVFjeUm8y+Kx6cuvRe5vdz8pHTC7v4Dq4IVjZz3AB7+N2HMaoph0Mhh/OnglPHIlp/R5Llf6rmtZcqCXl9LRMf6dEGFs7Tcxbzi4MLe5XIs/Fm9WG7HNBl1V1eDFYt6e+udM7AwgdAPgmeQhOcy+zxBzTln48r/KU6ZjJ1vU3NU6m3SHb3tq9Vl97tfUFy7EZfDtVNT1+oL5O42vjHQcy+pyhGZYCuI7r4IVIef5aGbfGcXT39nEcs0u9252zNNA81JYZb3BJB6Pp5zrGjnRKGSsjZE8KF1J02t8hdjbTftUT/bPUknD6832nJWV2JvfF/48b+8gcPYm+jtBvJH/78Fysp6fsJy8Avk9Wtjz9dA9509iJLoMm471eiDWx7fuVTKI6k+9l2dHm0cgQWXmSptH+bm9Af4CKfU/9RCiiv10qGTXhM3KHSOADAA78o+VeSmkDPiMrYDOYJ7rKdnXYt8b2uX7u5cIVRCvD8VgrWrJWFzm3+U2+hrEh6WHZHbzkMa5DqIy3DNe5BZrbhbqSwEKWZ6UD0r7OquhcMJ1k2UwhzbIjjlFlyOAzLI52DOHaWKBTa1PU/m6rx17J6enbU81wHfZUYnhWMqNj+/7ILsXMDs8KtuMl/BOuQyO0VO52ORDK8V427MOpeWV8Pj51VJmYD4Wn1WTceFhOHx66MBMpP/sGFgcHRnKLgwPD5h33z5KKn+B62ad8BeIGQs0E4oVdmVP7RpZjgjMR065xqVcgNu+PKjNnldzt84a3zOo9fQtxK3Skj4K+bT6qmatmzyrNCINn9nv3dqPP5yHBuUhlvr5xqzIcVm9dlKI78bpZQNwdruW5vILr0FQxFycvbc+jjm3Jh12OvU3uQN6bxrWikq4ZrM726KuApf+UNvAKroNc6cpElyQbd3ml6R120WlGDveDzqIr5uyTQun27khKGVMi480RQfVJncnWTFdsv6WrTfK2Rahb9roz7R39wYCnze4Oez2Szd+rhAeDBube0gbOkfM7+ZzUjJRRzWSpyOef9XWxmBqpywf6975suZJimv0N7jprU9SSDTe669GWqnrttYz+2GbzeGqrNL7RqN1d2sCPcc3wtvPpXbnlOGErom/vuGG0ZUQZyhuXWttBy4Bm9QqY1B8KLmNNcU53j1O17MEeAPwLrkE9gMqogig61aRREK+O5AMsx7LWgPCNgr6Ja/ojOkEDYwF06W7z29JDAPxtGRMVZNW59bGm8k669e/F88ofD0+Gqht4trG1cebAyrEppdpayzb5hXkkH54XZYe93XH+H5++KEZEUXFeMOr+shTFD3AN3AC8OT8zVHehayBcbWuDy2YL9rtsB/JSVTXDWoO2r+f1v7p6Rt/j+VRNOk7xkf6xr0Bp3o/WzU+jBQUASiWIwX18gO8RCSicBQ4ofGt752AV17b/p7wLhRdwzSSLMEIm4A65A3UAgql/OaQuCR7qtLdQMuEUXb49oqsVEGIYwQd4yqgjJHyOGL6FkVQKAP4LAAD//wEAAP//+3tAfQAAAAABAAAAARhRhjcE3V8PPPUAAQPoAAAAANhdoMwAAAAA3WYvN/69/t0IHQPJAAIAAwACAAAAAAAAAAEAAAPY/u8AAAhA/r39vAgdA+gAwv/RAAAAAAAAAAAAAAAkAnQAJADIAAAB/v/LAiYAOQJQACMA/AAjAc4AIwLBACMB/gBdAmgATwIZACcCGAAfAbMAJQIXACcB4QAlARoAKwITAAECCwAfAO0AHwD4ACwDHwAfAg0AHwIDACcCF//2AhkAJwFWAB8Bkv/8AUUAPAIQADgBwAA7Aa3/1AHA/8IB4AAwAO0AHwAAAEcB4AAwAAAALgAuAFIAhACmALQAxADyAQYBLgFmAZ4BzAIEAj4CZgKuAtgC5AMGA0gDcgOgA9oEFAQyBG4EnATIBOYFEgVCBU4FXAVyBYAAAAABAAAAJACMAAwAZgAHAAEAAAAAAAAAAAAAAAAABAADeJyclNtOG1cUhj8H2216uqhQRG7QvkylZEyjECXhypSgjIpw6nF6kKpKgz0+iPHMyDOYkifodd+ib5GrPkafoup1tX8vgx1FQSAE/Hv2OvxrrX9tYJP/2KBWvwv83ZwbrrHd/NnwHb5oHhneYL/5meE6Dxv/GG4waLw13ORBo2v4E97V/zT8KU/qvxm+y1b90PDnPK5vGv5yw/Gv4a94wrsFrsEz/jBcY4vC8B02+dXwBvewmLU699gx3OBrtg032QZ6TKhImZAxwjFkwogzZiSURCTMmDAkYYAjpE1Kpa8ZsZBj9MGvMREVM2JFHFPhSIlIiSkZW8S38sp5rYxDnWZ216ZiTMyJPE6JyXDkjMjJSDhVnIqKghe0aFHSF9+CipKAkgkpATkzRrTocMgRPcZMKHEcKpJnFpEzpOKcWPmdWfjO9EnIKI3VGRkD8XTil8g75AhHh0K2q5GP1iI8xPGjvD23XLbfEujXrTBbz7tkEzNXP1N1JdXNuSY41q3P2+YH4YoXuFv1Z53J9T0a6H+lyCecaf4DTSoTkwzntmgTSUGRu49jX+eQSB35iZAer+jwhp7Obbp0aXNMj5CX8u3QxfEdHY45kEcovLg7lGKO+QXH94Sy8bET689iYgm/U5i6S3GcqY4phXrumQeqNVGFN5+w36F8TR2lfPraI2/pNL9MexYzMlUUYjhVL5faKK1/A1PEVLX42V7d+22Y2+4tt/iCXDvs1brg5Ce3YHTdVIP3NHOun4CYATknsuiTM6VFxYV4vybmjBTHgbr3SltS0b708XkupJKEqRiEZIozo9Df2HQTGff+mu6dvSUD+Xump5dV3SaLU6+uZvRG3VveRdblZGUCLZtqvqKmvrhmpv1EO7XKP5Jvqdct5xGh4i52+0OvwA7P2WWPsbL0dTO/vPOvhLfYUwdOSWQ1lKZ9DY8J2CXgKbvs8pyn7/VyycYZH7fGZzV/mwP26bB3bTUL2w77vFyL9vHMf4ntjupxPLo8Pbv1NB/cQLXfaN+u3s2uJuenMbdoV9txTMzUc3FbqzW5+wT/AwAA//8BAAD//3KhUUAAAAADAAD/9QAA/84AMgAAAAAAAAAAAAAAAAAAAAAAAAAA");
}]]></style><style type="text/css"><![CDATA[.shape {
  shape-rendering: geometricPrecision;
  stroke-linejoin: round;
}
.connection {
  stroke-linecap: round;
  stroke-linejoin: round;
}
.blend {
  mix-blend-mode: multiply;
  opacity: 0.5;
}

		.d2-474229315 .fill-N1{fill:#0A0F25;}
		.d2-474229315 .fill-N2{fill:#676C7E;}
		.d2-474229315 .fill-N3{fill:#9499AB;}
		.d2-474229315 .fill-N4{fill:#CFD2DD;}
		.d2-474229315 .fill-N5{fill:#DEE1EB;}
		.d2-474229315 .fill-N6{fill:#EEF1F8;}
		.d2-474229315 .fill-N7{fill:#FFFFFF;}
		.d2-474229315 .fill-B1{fill:#0D32B2;}
		.d2-474229315 .fill-B2{fill:#0D32B2;}
		.d2-474229315 .fill-B3{fill:#E3E9FD;}
		.d2-474229315 .fill-B4{fill:#E3E9FD;}
		.d2-474229315 .fill-B5{fill:#EDF0FD;}
		.d2-474229315 .fill-B6{fill:#F7F8FE;}
		.d2-474229315 .fill-AA2{fill:#4A6FF3;}
		.d2-474229315 .fill-AA4{fill:#EDF0FD;}
		.d2-474229315 .fill-AA5{fill:#F7F8FE;}
		.d2-474229315 .fill-AB4{fill:#EDF0FD;}
		.d2-474229315 .fill-AB5{fill:#F7F8FE;}
		.d2-474229315 .stroke-N1{stroke:#0A0F25;}
		.d2-474229315 .stroke-N2{stroke:#676C7E;}
		.d2-474229315 .stroke-N3{stroke:#9499AB;}
		.d2-474229315 .stroke-N4{stroke:#CFD2DD;}
		.d2-474229315 .stroke-N5{stroke:#DEE1EB;}
		.d2-474229315 .stroke-N6{stroke:#EEF1F8;}
		.d2-474229315 .stroke-N7{stroke:#FFFFFF;}
		.d2-474229315 .stroke-B1{stroke:#0D32B2;}
		.d2-474229315 .stroke-B2{stroke:#0D32B2;}
		.d2-474229315 .stroke-B3{stroke:#E3E9FD;}
		.d2-474229315 .stroke-B4{stroke:#E3E9FD;}
		.d2-474229315 .stroke-B5{stroke:#EDF0FD;}
		.d2-474229315 .stroke-B6{stroke:#F7F8FE;}
		.d2-474229315 .stroke-AA2{stroke:#4A6FF3;}
		.d2-474229315 .stroke-AA4{stroke:#EDF0FD;}
		.d2-474229315 .stroke-AA5{stroke:#F7F8FE;}
		.d2-474229315 .stroke-AB4{stroke:#EDF0FD;}
		.d2-474229315 .stroke-AB5{stroke:#F7F8FE;}
		.d2-474229315 .background-color-N1{background-color:#0A0F25;}
		.d2-474229315 .background-color-N2{background-color:#676C7E;}
		.d2-474229315 .background-color-N3{background-color:#9499AB;}
		.d2-474229315 .background-color-N4{background-color:#CFD2DD;}
		.d2-474229315 .background-color-N5{background-color:#DEE1EB;}
		.d2-474229315 .background-color-N6{background-color:#EEF1F8;}
		.d2-474229315 .background-color-N7{background-color:#FFFFFF;}
		.d2-474229315 .background-color-B1{background-color:#0D32B2;}
		.d2-474229315 .background-color-B2{background-color:#0D32B2;}
		.d2-474229315 .background-color-B3{background-color:#E3E9FD;}
		.d2-474229315 .background-color-B4{background-color:#E3E9FD;}
		.d2-474229315 .background-color-B5{background-color:#EDF0FD;}
		.d2-474229315 .background-color-B6{background-color:#F7F8FE;}
		.d2-474229315 .background-color-AA2{background-color:#4A6FF3;}
		.d2-474229315 .background-color-AA4{background-color:#EDF0FD;}
		.d2-474229315 .background-color-AA5{background-color:#F7F8FE;}
		.d2-474229315 .background-color-AB4{background-color:#EDF0FD;}
		.d2-474229315 .background-color-AB5{background-color:#F7F8FE;}
		.d2-474229315 .color-N1{color:#0A0F25;}
		.d2-474229315 .color-N2{color:#676C7E;}
		.d2-474229315 .color-N3{color:#9499AB;}
		.d2-474229315 .color-N4{color:#CFD2DD;}
		.d2-474229315 .color-N5{color:#DEE1EB;}
		.d2-474229315 .color-N6{color:#EEF1F8;}
		.d2-474229315 .color-N7{color:#FFFFFF;}
		.d2-474229315 .color-B1{color:#0D32B2;}
		.d2-474229315 .color-B2{color:#0D32B2;}
		.d2-474229315 .color-B3{color:#E3E9FD;}
		.d2-474229315 .color-B4{color:#E3E9FD;}
		.d2-474229315 .color-B5{color:#EDF0FD;}
		.d2-474229315 .color-B6{color:#F7F8FE;}
		.d2-474229315 .color-AA2{color:#4A6FF3;}
		.d2-474229315 .color-AA4{color:#EDF0FD;}
		.d2-474229315 .color-AA5{color:#F7F8FE;}
		.d2-474229315 .color-AB4{color:#EDF0FD;}
		.d2-474229315 .color-AB5{color:#F7F8FE;}.appendix text.text{fill:#0A0F25}.md{--color-fg-default:#0A0F25;--color-fg-muted:#676C7E;--color-fg-subtle:#9499AB;--color-canvas-default:#FFFFFF;--color-canvas-subtle:#EEF1F8;--color-border-default:#0D32B2;--color-border-muted:#0D32B2;--color-neutral-muted:#EEF1F8;--color-accent-fg:#0D32B2;--color-accent-emphasis:#0D32B2;--color-attention-subtle:#676C7E;--color-danger-fg:red;}.sketch-overlay-B1{fill:url(#streaks-darker-d2-474229315);mix-blend-mode:lighten}.sketch-overlay-B2{fill:url(#streaks-darker-d2-474229315);mix-blend-mode:lighten}.sketch-overlay-B3{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-B4{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-B5{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-B6{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-AA2{fill:url(#streaks-dark-d2-474229315);mix-blend-mode:overlay}.sketch-overlay-AA4{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-AA5{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-AB4{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-AB5{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-N1{fill:url(#streaks-darker-d2-474229315);mix-blend-mode:lighten}.sketch-overlay-N2{fill:url(#streaks-dark-d2-474229315);mix-blend-mode:overlay}.sketch-overlay-N3{fill:url(#streaks-normal-d2-474229315);mix-blend-mode:color-burn}.sketch-overlay-N4{fill:url(#streaks-normal-d2-474229315);mix-blend-mode:color-burn}.sketch-overlay-N5{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-N6{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.sketch-overlay-N7{fill:url(#streaks-bright-d2-474229315);mix-blend-mode:darken}.light-code{display: block}.dark-code{display: none}]]></style><g class="YmVmb3Jl"><g class="shape" ><rect x="12.000000" y="12.000000" width="896.000000" height="182.000000" stroke="#0D32B2" fill="#dcfce7" class=" stroke-B1" style="stroke-width:2;" /></g><text x="460.000000" y="45.000000" fill="#0A0F25" class="text fill-N1" style="text-anchor:middle;font-size:28px">Trust Infrastructure</text></g><g class="YWZ0ZXI="><g class="shape" ><rect x="29.000000" y="371.000000" width="861.000000" height="182.000000" stroke="#0D32B2" fill="#fee2e2" class=" stroke-B1" style="stroke-width:2;" /></g><text x="459.500000" y="404.000000" fill="#0A0F25" class="text fill-N1" style="text-anchor:middle;font-size:28px">Trust Tax</text></g><g class="YmVmb3JlLmNvbW1pdF9oaXN0b3J5"><g class="shape" ><rect x="62.000000" y="62.000000" width="166.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="145.000000" y="100.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="145.000000" dy="0.000000">Commit history =</tspan><tspan x="145.000000" dy="18.500000">decision record</tspan></text></g><g class="YmVmb3JlLmRpZmY="><g class="shape" ><rect x="248.000000" y="62.000000" width="202.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="349.000000" y="100.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="349.000000" dy="0.000000">Diff =</tspan><tspan x="349.000000" dy="18.500000">explanation of change</tspan></text></g><g class="YmVmb3JlLmNvbnRyaWJ1dG9y"><g class="shape" ><rect x="470.000000" y="62.000000" width="185.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="562.500000" y="100.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="562.500000" dy="0.000000">Contributor =</tspan><tspan x="562.500000" dy="18.500000">accountable human</tspan></text></g><g class="YmVmb3JlLm91dGNvbWU="><g class="shape" ><rect x="675.000000" y="62.000000" width="183.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="766.500000" y="100.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="766.500000" dy="0.000000">Confidence enables</tspan><tspan x="766.500000" dy="18.500000">automation</tspan></text></g><g class="YWZ0ZXIuY29tbWl0X2hpc3Rvcnk="><g class="shape" ><rect x="79.000000" y="421.000000" width="166.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="162.000000" y="459.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="162.000000" dy="0.000000">Commit history =</tspan><tspan x="162.000000" dy="18.500000">generation log</tspan></text></g><g class="YWZ0ZXIuZGlmZg=="><g class="shape" ><rect x="265.000000" y="421.000000" width="179.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="354.500000" y="459.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="354.500000" dy="0.000000">Diff =</tspan><tspan x="354.500000" dy="18.500000">LLM output sample</tspan></text></g><g class="YWZ0ZXIuY29udHJpYnV0b3I="><g class="shape" ><rect x="464.000000" y="421.000000" width="165.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="546.500000" y="459.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="546.500000" dy="0.000000">Contributor =</tspan><tspan x="546.500000" dy="18.500000">prompt engineer</tspan></text></g><g class="YWZ0ZXIub3V0Y29tZQ=="><g class="shape" ><rect x="649.000000" y="421.000000" width="191.000000" height="82.000000" stroke="#0D32B2" fill="#EDF0FD" class=" stroke-B1 fill-B5" style="stroke-width:2;" /></g><text x="744.500000" y="459.500000" fill="#0A0F25" class="text-bold fill-N1" style="text-anchor:middle;font-size:16px"><tspan x="744.500000" dy="0.000000">Uncertainty requires</tspan><tspan x="744.500000" dy="18.500000">manual verification</tspan></text></g><g class="KGJlZm9yZSAtJmd0OyBhZnRlcilbMF0="><marker id="mk-d2-474229315-3488378134" markerWidth="10.000000" markerHeight="12.000000" refX="7.000000" refY="6.000000" viewBox="0.000000 0.000000 10.000000 12.000000" orient="auto" markerUnits="userSpaceOnUse"> <polygon points="0.000000,0.000000 10.000000,6.000000 0.000000,12.000000" fill="#0D32B2" class="connection fill-B1" stroke-width="2" /> </marker><path d="M 460.000000 196.000000 L 460.000000 367.000000" stroke="#0D32B2" fill="none" class="connection stroke-B1" style="stroke-width:2;" marker-end="url(#mk-d2-474229315-3488378134)" mask="url(#d2-474229315)" /><text x="460.500000" y="280.000000" fill="#676C7E" class="text-italic fill-N2" style="text-anchor:middle;font-size:16px"><tspan x="460.500000" dy="0.000000">AI code generation</tspan><tspan x="460.500000" dy="18.500000">floods the system</tspan></text></g><mask id="d2-474229315" maskUnits="userSpaceOnUse" x="-9" y="-9" width="938" height="583">
<rect x="-9" y="-9" width="938" height="583" fill="transparent"></rect>
<rect x="397.000000" y="264.000000" width="127" height="37" fill="black"></rect>
</mask></svg></svg>
" class="img-fluid" style="width:100.0%"></p>
<p>The companies that understand this are already adapting. They’re building internal forks of critical dependencies. They’re paying for human code review even on open source contributions. They’re treating GitHub as untrusted by default.</p>
<p>The companies that don’t understand this are accumulating technical debt they can’t see. They’re pulling in dependencies that look fine, pass tests, and ship—until six months later when the subtle bug surfaces and no one can trace it to a human decision.</p>
</section>
<section id="the-civilisational-question" class="level2">
<h2 class="anchored" data-anchor-id="the-civilisational-question">The Civilisational Question</h2>
<p>Git was designed for a world where every commit represented a human judgment. That world is ending. The question worth asking now is: what does open source collaboration look like when you can’t trust the historical record?</p>
<p>The standard response is: “We’ll build better verification tools.” That’s necessary but insufficient. Verification tools can tell you <em>what</em> changed. They can’t tell you <em>why</em> it changed, because the “why” never existed.</p>
<p>The deeper adaptation is cultural. We’re moving from a trust-by-default model (assume human intent, verify when suspicious) to a verify-by-default model (assume machine generation, trust only after audit). That’s a fundamental shift in how open source works.</p>
<p>Are we ready for it? Mostly, no. We’re still treating AI-generated code as a productivity enhancement, not a trust infrastructure collapse. We’re still measuring success by lines of code written, not by verification burden imposed on downstream users.</p>
<p>The Trust Tax is coming. The only question is whether we pay it consciously or discover it six months after the bug ships.</p>
<div class="schema-faq" style="display:none;">
<p>[{“q”:“What is the GitHub Slopocalypse?”,“a”:“The rapid flooding of GitHub repositories with AI-generated code that lacks human intent or reasoning. GitHub’s 2025 data shows AI-assisted commits exceed 40% of new code, and Copilot generates 46% of code in enabled files. This is eroding the trust infrastructure that made open source work.”},{“q”:“What is the Trust Tax in software development?”,“a”:“The additional cognitive and temporal cost developers pay to verify code provenance before they can use it. When AI generates code, you can no longer reconstruct intent from commit history — every dependency must be audited manually, doubling evaluation time for experienced developers.”},{“q”:“How do you verify AI-generated code on GitHub?”,“a”:“Moving from trust-by-default to verify-by-default: assume machine generation, audit before trusting. Companies are building internal forks of critical dependencies, paying for human code review on open source contributions, and treating GitHub as untrusted by default.”}]</p>
</div>


</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>Software Engineering</category>
  <category>Open Source</category>
  <guid>https://talvinder.com/field-notes/github-slopocalypse-trust-tax/</guid>
  <pubDate>Mon, 16 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>What Zari-Zardozi Teaches Us About Agent Coordination</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/home-based-craft-vs-agent-work/</link>
  <description><![CDATA[ 





<p>In a Zari-Zardozi workshop in Old Delhi, six artisans work on a single bridal dupatta. One creates the base pattern. Another applies the metallic thread. A third adds sequins. A fourth handles the edge work. They don’t talk much. They don’t pass the fabric in strict sequence. Yet the final piece is coherent—every motif aligned, every border continuous, every layer building on the last.</p>
<p>This is not romantic craft nostalgia. This is a coordination architecture that’s been production-tested for 400 years.</p>
<p>I’m calling this pattern <strong>Layered Autonomy</strong>—not because the world needs another framework, but because most multi-agent AI systems fail at exactly what Zari-Zardozi solves: how to give workers genuine autonomy while maintaining system-level coherence.</p>
<section id="were-building-agent-systems-wrong" class="level2">
<h2 class="anchored" data-anchor-id="were-building-agent-systems-wrong">We’re Building Agent Systems Wrong</h2>
<p>The dominant pattern is the command-and-control planner: a central orchestrator that assigns tasks, waits for results, then decides the next step. It’s sequential. It’s brittle. It doesn’t scale.</p>
<p>At Ostronaut, we initially built exactly this architecture. A central planner that coordinated a fleet of specialist agents to generate training content—slides, videos, quizzes. The planner would call one agent, wait for output, call the next, wait again, then call the quality checker. Linear dependency chains everywhere.</p>
<p>It worked for simple cases. It collapsed under complexity.</p>
<p>The problem wasn’t the agents. The problem was the coordination model. We were building assembly lines when we needed something closer to a Zari workshop.</p>
<p>Layered Autonomy is the alternative: agents work in parallel on shared context, with loose coupling and tight coherence. Not through constant communication. Through shared understanding of the end state.</p>
</section>
<section id="four-lessons-from-the-embroidery-floor" class="level2">
<h2 class="anchored" data-anchor-id="four-lessons-from-the-embroidery-floor">Four Lessons from the Embroidery Floor</h2>
<p><strong>Agent systems that implement layered autonomy—where workers operate on shared context with clear role boundaries but loose temporal coupling—will outperform planner-orchestrated systems on tasks requiring iterative refinement by at least 40% in both speed and quality.</strong></p>
<p>The Zari-Zardozi model teaches four specific lessons:</p>
<section id="specialization-without-silos" class="level3">
<h3 class="anchored" data-anchor-id="specialization-without-silos">1. Specialization Without Silos</h3>
<p>In a Zari workshop, the nakshi maker creates patterns. The zari worker applies metallic thread. The sequin specialist adds embellishments. Each role is distinct. But they’re not isolated—every artisan understands the full design.</p>
<p>Most agent architectures get this wrong. They create specialists (a content agent, a research agent, a quality agent) but treat them as black boxes. The planner knows what each agent does. The agents don’t know about each other.</p>
<p>This creates artificial dependencies. The content agent can’t start until research is “done.” The quality agent can’t run until content is “complete.” You’ve built specialists, but you’ve also built a bottleneck.</p>
<p>The Zari model is different. The zari worker doesn’t wait for the nakshi maker to finish the entire pattern. They work on completed sections while new sections are still being drawn. Parallel execution on shared context.</p>
</section>
<section id="shared-context-as-infrastructure" class="level3">
<h3 class="anchored" data-anchor-id="shared-context-as-infrastructure">2. Shared Context as Infrastructure</h3>
<p>The critical insight: Zari artisans don’t coordinate through constant communication. They coordinate through shared access to the evolving artifact.</p>
<p>The fabric is the coordination layer. Every artisan can see what others have done. Every artisan can see what’s left to do. The pattern itself carries the context.</p>
<p>In agent systems, this means: stop passing messages. Start sharing state.</p>
<p>We rebuilt Ostronaut’s coordination layer around this principle. Instead of agents calling each other sequentially, they all operate on a shared representation of the content being generated. One agent writes structure. Another reads that structure and writes content. A third reads and annotates quality issues. A fourth reads and generates media assets.</p>
<p>No agent waits for another agent to “finish.” They work on whatever parts of the shared state are ready for their contribution.</p>
<p>The result: generation time dropped by more than half. Not because the agents got faster. Because they stopped waiting.</p>
</section>
<section id="iterative-refinement-over-perfect-planning" class="level3">
<h3 class="anchored" data-anchor-id="iterative-refinement-over-perfect-planning">3. Iterative Refinement Over Perfect Planning</h3>
<p>Zari-Zardozi work proceeds in layers. Base stitch first. Then zari. Then sequins. Then finishing touches. Each layer builds on the last. Each layer can be evaluated independently.</p>
<p>The master craftsperson doesn’t plan every stitch upfront. They plan the overall design, then let each layer emerge.</p>
<p>Most agent planners do the opposite. They try to decompose the entire task upfront into a perfect sequence of subtasks. This fails for two reasons:</p>
<p>First, you can’t know what subtasks you’ll need until you see the results of earlier work. If the structure agent generates a complex nested outline, the content agent might need to split its work differently than if the outline is flat.</p>
<p>Second, perfect planning is expensive. You spend tokens and time trying to predict every edge case, when you could just execute and adapt.</p>
<p>The Zari model: plan the layers, not the stitches. In agent terms: define the phases (structure → content → quality → assets), but let agents decide how to execute within their phase.</p>
</section>
<section id="the-master-as-orchestrator-not-micromanager" class="level3">
<h3 class="anchored" data-anchor-id="the-master-as-orchestrator-not-micromanager">4. The Master as Orchestrator, Not Micromanager</h3>
<p>The ustad (master craftsperson) in a Zari workshop doesn’t do the embroidery. They ensure coherence. They check alignment. They decide when a layer is ready for the next phase.</p>
<p>This is not a planner in the traditional sense. The ustad doesn’t assign every task. They maintain the quality bar and the overall vision.</p>
<p>In agent architectures, this means: the orchestrator’s job is to manage transitions between layers, not to micromanage within layers.</p>
<p>Our current Ostronaut orchestrator does three things:</p>
<ul>
<li>Validates that each layer meets quality gates before the next layer starts</li>
<li>Handles failures by deciding whether to retry or skip</li>
<li>Maintains the audit trail of what happened and why</li>
</ul>
<p>It doesn’t decide which specific content to generate or which specific assets to create. That’s the workers’ job.</p>
</section>
</section>
<section id="the-performance-difference" class="level2">
<h2 class="anchored" data-anchor-id="the-performance-difference">The Performance Difference</h2>
<p><strong>Old architecture (planner-orchestrated):</strong></p>
<ul>
<li>Fully sequential—zero parallelization</li>
<li>Frequent timeouts from long dependency chains</li>
<li>Every new content type required rewriting the planner’s logic</li>
</ul>
<p><strong>New architecture (layered autonomy):</strong></p>
<ul>
<li>Agents overlap—structure and content generation run concurrently where possible</li>
<li>Failures are isolated to individual layers instead of cascading</li>
<li>New content types require a new specialist agent and a validation gate—the orchestrator doesn’t change</li>
</ul>
<p>The speed improvement matters. But the bigger win is adaptability. When we added a new content type (interactive games), the old architecture required rewriting the planner’s task decomposition logic. The new architecture required adding a new worker agent that knows how to operate on the shared state. The orchestrator didn’t change.</p>
<p>This is the Zari-Zardozi lesson: when you add a new type of embellishment to the craft, you don’t retrain every artisan. You bring in a specialist who understands the shared language of the fabric.</p>
</section>
<section id="the-pattern" class="level2">
<h2 class="anchored" data-anchor-id="the-pattern">The Pattern</h2>
<p>Most teams building multi-agent systems are building assembly lines. Sequential. Rigid. Optimized for predictability.</p>
<p>The Zari-Zardozi model suggests a different architecture: shared context, layered execution, loose coupling, tight coherence.</p>
<p>This isn’t a metaphor. It’s a specific architectural pattern:</p>
<table class="caption-top table">
<colgroup>
<col style="width: 53%">
<col style="width: 46%">
</colgroup>
<thead>
<tr class="header">
<th>Planner-Orchestrated</th>
<th>Layered Autonomy</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Agents call each other sequentially</td>
<td>Agents operate on shared state in parallel</td>
</tr>
<tr class="even">
<td>Planner decides all subtasks upfront</td>
<td>Orchestrator manages phase transitions only</td>
</tr>
<tr class="odd">
<td>Failure in one agent blocks the chain</td>
<td>Failure in one agent is isolated to its layer</td>
</tr>
<tr class="even">
<td>Adding new capabilities requires replanning logic</td>
<td>Adding new capabilities requires new worker + validation gate</td>
</tr>
</tbody>
</table>
<p>The hard part isn’t building the agents. The hard part is building the coordination layer—the equivalent of the fabric that Zari artisans work on.</p>
<p>For us, it’s a structured representation of the content being generated. For other systems, it might be a knowledge graph, a vector store, or a shared document. The specific technology matters less than the principle: give agents shared context, clear boundaries, and the autonomy to execute within their layer.</p>
<p>What I don’t know yet: how to build trust in systems where no single agent “owns” the output. When something goes wrong, users want to know which agent failed. In a layered system, failure is often emergent—the output is coherent at each layer but incoherent overall.</p>
<p>The Zari workshop solves this through the master craftsperson’s eye. They can see when the overall composition is off, even if each individual element is well-executed.</p>
<p>We don’t have a good equivalent yet. Validation gates catch obvious failures. But subtle incoherence—content that’s technically correct but doesn’t serve the learning objective—still slips through.</p>
<p>More on this as I work through it.</p>


</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>AI Architecture</category>
  <category>India</category>
  <guid>https://talvinder.com/field-notes/home-based-craft-vs-agent-work/</guid>
  <pubDate>Mon, 16 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Zimbo Meetings and the Ghost Work Tax</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/field-notes/zimbo-meetings-ghost-work/</link>
  <description><![CDATA[ 





<p>A 30-minute meeting costs 2.5 hours of actual work. That’s not a metaphor. That’s the math when you account for prep, context switching, and follow-up. Most organizations track the 30 minutes. None track the 2 hours of ghost work that surrounds it.</p>
<p>I’m calling these Zimbo meetings—not zombie meetings, because they’re not dead. They’re worse. They’re undead. They shamble forward, consuming resources, generating more meetings, but producing no decisions and no clarity.</p>
<section id="the-ghost-work-tax" class="level2">
<h2 class="anchored" data-anchor-id="the-ghost-work-tax">The Ghost Work Tax</h2>
<p>The Ghost Work Tax is the hidden labor that meetings extract from teams: calendar coordination, pre-reads, note-taking, summary distribution, action item tracking, and the mental overhead of managing all of it. It doesn’t show up in calendars. It doesn’t show up in time tracking. But it shows up in missed deadlines and burnt-out teams.</p>
<p>At Pragmatic Leaders, I ask PMs to track their actual meeting preparation time for one week. The median ratio is 1:3. For every hour in meetings, they spend three hours on ghost work. Senior PMs hit 1:4 because they’re expected to “come prepared” to everything.</p>
<p>The tax compounds with team size. A 30-minute meeting with 8 people isn’t 4 person-hours. It’s 20 person-hours when you include the ghost work. Most companies would require VP approval for a 20-hour project. They let anyone schedule a 30-minute meeting.</p>
</section>
<section id="where-the-ghost-work-hides" class="level2">
<h2 class="anchored" data-anchor-id="where-the-ghost-work-hides">Where the Ghost Work Hides</h2>
<p><strong>If you eliminate meetings with no pre-defined decision or deliverable, you reduce total coordination overhead by 40-60%, not 20-30%.</strong></p>
<p>The standard advice is “have better meetings.” That’s useless. The problem isn’t meeting quality. It’s meeting existence.</p>
<p>Zimbo meetings have three characteristics:</p>
<ol type="1">
<li><strong>No exit condition.</strong> The meeting ends when the calendar says it ends, not when a decision is made.</li>
<li><strong>Recursive ghost work.</strong> The meeting generates action items that require more meetings to resolve.</li>
<li><strong>Ambient participants.</strong> Half the attendees are there “just in case” or “for visibility.”</li>
</ol>
<p>The ghost work tax hits hardest in three places:</p>
<p><strong>Pre-meeting:</strong> Reading the deck someone sent 10 minutes before the call. Digging up the context from three Slack threads. Finding the doc that was “shared last week.” The calendar says the meeting starts at 2pm. The actual work starts at 1:30pm.</p>
<p><strong>During-meeting:</strong> One person talks. Three people take notes in three different tools. Two people are on mute doing other work. One person is “capturing action items” in a format no one will read. The official output is a 30-minute meeting. The actual output is 90 minutes of fragmented attention.</p>
<p><strong>Post-meeting:</strong> Writing the summary email. Clarifying what was actually decided in the back-channel Slack thread. Scheduling the follow-up meeting because this one ran out of time. Updating the three places where meeting notes live. The meeting ended at 2:30pm. The ghost work ends at 4pm.</p>
<p>I’ve seen teams where 60% of “execution time” is actually meeting overhead. They’re not slow because they can’t build. They’re slow because they can’t stop coordinating.</p>
</section>
<section id="what-we-measured" class="level2">
<h2 class="anchored" data-anchor-id="what-we-measured">What We Measured</h2>
<p>We tracked this at Zopdev for one quarter. Every meeting required a one-line purpose statement and a binary decision: “Is this to make a decision or share information?” If information, it defaulted to async unless someone could articulate why it needed to be synchronous.</p>
<p>Meeting count dropped 40%. Ghost work dropped 55%. The delta—that extra 15%—came from eliminating the recursive meetings. The meetings that existed to clarify the meetings that came before them.</p>
<p>Here’s what we learned:</p>
<p><strong>Most “syncs” are status theater.</strong> The information being shared already exists in Slack, Linear, or Notion. The meeting exists because someone doesn’t trust the async system or because “we’ve always done it this way.”</p>
<p><strong>Most “brainstorms” are pre-cooked.</strong> One person has already decided. The meeting exists to build consensus or distribute blame. The actual decision-making happened in a 1:1 three days earlier.</p>
<p><strong>Most “check-ins” are anxiety management.</strong> The manager feels out of the loop. The meeting exists to make them feel better, not to unblock the team.</p>
<p>The pattern I see across thousands of PMs: junior PMs schedule meetings because they don’t know how to make decisions. Senior PMs schedule meetings because they know exactly what decision they want and need organizational buy-in to de-risk it.</p>
<p>Neither is wrong. But only one is honest about what the meeting is for.</p>
</section>
<section id="the-framework-decide-deliver-or-die" class="level2">
<h2 class="anchored" data-anchor-id="the-framework-decide-deliver-or-die">The Framework: Decide, Deliver, or Die</h2>
<p>Every meeting should have a decision, a deliverable, or die.</p>
<p><strong>Decision meetings</strong> need: a clear choice to be made, pre-circulated options with tradeoffs, and a DRI who owns the call. If you can’t name the decision, it’s not a decision meeting.</p>
<p><strong>Deliverable meetings</strong> need: a thing that will exist at the end that didn’t exist at the beginning. A design, a plan, a document. If the output is “alignment,” it’s not a deliverable meeting.</p>
<p><strong>Everything else is async.</strong> Updates go in Slack. Brainstorms start in docs. Status lives in project management tools.</p>
<p>The Ghost Work Tax doesn’t show up in your P&amp;L. It shows up in your execution speed. In your team’s ability to do deep work. In the gap between your roadmap and your delivery.</p>
<p>Track it for one week. For every meeting on your calendar, log the prep time, the meeting time, and the follow-up time. Add it up. Then ask: what could we have built with those hours?</p>
<p>Most organizations won’t do this. They’ll keep scheduling 30-minute meetings and wondering why quarters take six months.</p>
<p>The ones that do will discover something uncomfortable: half their meetings exist to compensate for broken async communication. Fix the async system, kill the meetings, reclaim the ghost work hours.</p>
<p>What I don’t know yet: how to build organizational trust in async-first decision-making when the executive layer still equates “presence in meetings” with “doing the work.” The ghost work tax is a technical problem with a political solution.</p>
<p>More on this as I work through it.</p>


</section>

 ]]></description>
  <category>Leadership</category>
  <category>Operations</category>
  <category>Remote Work</category>
  <guid>https://talvinder.com/field-notes/zimbo-meetings-ghost-work/</guid>
  <pubDate>Mon, 16 Mar 2026 00:00:00 GMT</pubDate>
</item>
<item>
  <title>Why LLM-as-Judge Fails in India: The $0.03 Evaluation That Costs You Customers</title>
  <dc:creator>B. Talvinder</dc:creator>
  <link>https://talvinder.com/build-logs/llm-judge-india-failure/</link>
  <description><![CDATA[ 





<p>Galileo raised $45 million in October 2024 to build AI evaluation tools. Revenue grew 834% that year. Six Fortune 50 companies signed up. Snorkel AI raised $100 million in May 2025 at a $1.3 billion valuation, with Snorkel Evaluate as a core product. Confident AI came out of YC W25.</p>
<p>These companies are now pitching Indian edtech buyers. Beautiful decks. Impressive demos. Per-evaluation API pricing that will bankrupt every Indian buyer who signs.</p>
<p>I keep watching this happen. The demo works. The pricing model is imported from a market where course fees are $500-2,000/seat. The Indian buyer is selling at ₹200-800/learner/month. Nobody does the math until after the contract is signed.</p>
<section id="the-number-that-kills-you" class="level2">
<h2 class="anchored" data-anchor-id="the-number-that-kills-you">The number that kills you</h2>
<p>Take a corporate training product. 10,000 active learners, 4 evaluations per month. 40,000 evaluation events.</p>
<p>LLM-as-judge at GPT-4o-mini rates: $0.15 per million input tokens, $0.60 per million output tokens. A basic evaluation prompt with rubric and response runs roughly 1,500 tokens in, 500 out. That’s about $0.0005 per evaluation. At 40,000 evaluations: $20/month. Sounds fine.</p>
<p>Now make the evaluation useful. Detailed feedback, multi-criteria scoring, follow-up questions. You’re using 4,000 tokens in, 2,000 out. Per-eval cost jumps to $0.0018. Still small. But you want GPT-4o quality for nuanced judgment, $2.50/$10 per million tokens. Now you’re at $0.03 per evaluation. $1,200/month. $1.20 per learner per month.</p>
<p>Your Indian enterprise client is paying ₹200/learner/month. That’s roughly $2.40.</p>
<p><strong>Evaluation Cost Ratio = evaluation spend / per-learner revenue.</strong></p>
<p>At $0.03/eval with GPT-4o: 50%. Half your revenue on evaluation alone.</p>
<table class="caption-top table">
<thead>
<tr class="header">
<th>Market</th>
<th>Monthly ARPU</th>
<th>Eval cost/learner/month</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>US enterprise training</td>
<td>$500</td>
<td>$1.20</td>
</tr>
<tr class="even">
<td>US mid-market</td>
<td>$50</td>
<td>$1.20</td>
</tr>
<tr class="odd">
<td>Indian corporate training</td>
<td>$2.40</td>
<td>$1.20</td>
</tr>
</tbody>
</table>
<p>The evaluation startups aren’t lying. Their product works. It works in markets where the ECR is under 3%. In Indian markets, the same product eats half your revenue. The engineering is impressive. The product-market fit is nonexistent.</p>
</section>
<section id="india-already-solved-this-problem" class="level2">
<h2 class="anchored" data-anchor-id="india-already-solved-this-problem">India already solved this problem</h2>
<p>JEE Main evaluates 1.3 million candidates annually. GATE 2025 had 7.37 lakh candidates appear across all papers. No LLM. No army of graders. Structured assessment at Indian scale, Indian price points.</p>
<p>The GMAT removed its Analytical Writing Assessment entirely when it launched the Focus Edition in 2023, making the exam an hour shorter by cutting the essay. Then brought it back in July 2024 as an optional “Business Writing Assessment” after business schools complained they couldn’t tell if applicants or ChatGPT wrote the essays. The lesson: subjective evaluation keeps getting harder and more expensive. Structured evaluation keeps scaling.</p>
<p>Physics Wallah scaled to 4.46 million paid users in FY25, up from 1.76 million in FY23. Revenue crossed ₹3,000 crore. Online ACPU was ₹3,682. Their bottleneck was never evaluation — it was content production and offline expansion. They solved the right scaling problem.</p>
<p>The insight isn’t new. Structure the assessment so it’s objective AND scalable. India cracked this decades ago for science and math. The product opportunity is applying the same principle to judgment skills (leadership decisions, case analyses, strategic thinking) that the exam tradition doesn’t handle well.</p>
</section>
<section id="what-i-actually-built" class="level2">
<h2 class="anchored" data-anchor-id="what-i-actually-built">What I actually built</h2>
<p>I ran into this wall building Ostronaut’s training platform. We generate learning content with AI: slides, games, interactive scenarios. The generation pipeline uses LLMs heavily. Expensive per content piece, but it’s a one-time cost amortized across all learners who consume it.</p>
<p>The evaluation architecture is completely different. For game-based scenarios — card games and turn-based simulations that teach decision-making — scoring is rule-based. The system defines optimal play. Scores against it. Runs in milliseconds. Costs nothing at the margin. Same input, same score, every time.</p>
<p>LLM creates the scenario. Rules judge every move within it.</p>
<p>I use LLM judgment in exactly one place: validating generated content before it reaches learners. One validation pass per content piece. That cost scales with production volume, not learner volume. The difference matters.</p>
<table class="caption-top table">
<colgroup>
<col style="width: 33%">
<col style="width: 33%">
<col style="width: 33%">
</colgroup>
<thead>
<tr class="header">
<th></th>
<th>Content creation</th>
<th>Learner evaluation</th>
</tr>
</thead>
<tbody>
<tr class="odd">
<td>Method</td>
<td>LLM generation + LLM validation</td>
<td>Rule-based scoring</td>
</tr>
<tr class="even">
<td>Cost structure</td>
<td>One-time per piece</td>
<td>Per-interaction</td>
</tr>
<tr class="odd">
<td>Scales with</td>
<td>Content volume (manageable)</td>
<td>Learner volume (must approach zero)</td>
</tr>
</tbody>
</table>
<p>Get this split wrong and you bleed money from day one.</p>
</section>
<section id="manufacturing-figured-this-out-fifty-years-ago" class="level2">
<h2 class="anchored" data-anchor-id="manufacturing-figured-this-out-fifty-years-ago">Manufacturing figured this out fifty years ago</h2>
<p>You can inspect every widget coming off the line, or you can design the production process so widgets come out right. Inspection scales linearly with output. Built-in quality is expensive upfront and free at scale.</p>
<p>LLM-as-judge is inspection. Structured rubrics with rule-based scoring is built-in quality.</p>
<p>I watch smart founders import the inspection model from Western markets, build their entire evaluation architecture around it, and then discover nine months later that their unit economics don’t work. It’s the same mistake as <a href="../../field-notes/federated-learning-healthcare-failure/">importing federated learning into Indian healthcare</a> — the architecture assumes conditions that don’t exist here. By then they’ve raised on metrics that assumed evaluation costs would decrease. They won’t. They scale linearly with learner volume.</p>
</section>
<section id="what-im-not-sure-about" class="level2">
<h2 class="anchored" data-anchor-id="what-im-not-sure-about">What I’m not sure about</h2>
<p>The ECR math is clear at current model prices. But model prices are dropping fast. GPT-4o-mini already costs 60% less than GPT-3.5 Turbo did at launch. If evaluation costs fall another 10x in two years, does the ECR problem solve itself?</p>
<p>Maybe. If GPT-5-equivalent evaluation costs $0.003/eval, the Indian ECR drops to about 5%. Livable. But I’ve watched this movie before with cloud storage, with compute, with bandwidth. Prices drop, but usage grows faster. You build assuming the cost decrease, then discover you’re evaluating 10x more often because you can. The ECR stays broken.</p>
<p>The other question: are there domains where LLM-as-judge is the only option? Creative writing feedback, strategic case analysis, nuanced communication skills — these don’t reduce cleanly to rules. Maybe the answer is tiered: rule-based evaluation for 80% of learners, LLM evaluation for the premium 20% who pay 5x more. I haven’t seen anyone execute this successfully yet.</p>
<p>The pattern I keep seeing is founders who treat evaluation as a feature, not a cost center — a <a href="../../frameworks/india-pm-revolution/">judgment failure</a> that no amount of AI tooling can fix. They assume it’ll be cheap because the demo was cheap. Then they scale to 50,000 learners and the AWS bill is suddenly larger than payroll. It’s the same <a href="../../frameworks/biggest-challenge-indian-startups/">Indian startup scaling wall</a> applied to AI economics — the structural constraints are different, but the pattern is identical.</p>
<p>The evaluation startups are solving a real problem. Just not for Indian markets. Not at these price points. Not yet.</p>
<p>This is a specific instance of a broader pattern: the <a href="../../field-notes/indian-saas-agent-reliability/">reliability expectations in Indian SaaS</a> demand cost structures that Western tooling doesn’t accommodate. The companies that get the <a href="../../frameworks/ai-runtime-infrastructure-play/">infrastructure economics</a> right will have a structural advantage that well-funded but wasteful competitors can’t replicate.</p>


<!-- -->

</section>

 ]]></description>
  <category>Agentic Systems</category>
  <category>India Market</category>
  <category>Product Economics</category>
  <guid>https://talvinder.com/build-logs/llm-judge-india-failure/</guid>
  <pubDate>Thu, 05 Mar 2026 00:00:00 GMT</pubDate>
  <media:content url="https://talvinder.com/build-logs/llm-judge-india-failure/assets/og-image.png" medium="image" type="image/png" height="76" width="144"/>
</item>
</channel>
</rss>
