Portfolio About Team Blog Contact Get in Touch
LLM Trends

Context Window Economics:
A New Moat in LLMs

Two years ago, 4,096 tokens was a hard ceiling for most production LLM applications. Developers built elaborate chunking and retrieval pipelines to stay under the limit. Today, frontier models support context windows north of a million tokens. That's not a linear improvement in a feature — it's a structural change in what's possible to build, and it's creating new competitive dynamics that most investors haven't fully mapped.

We've been thinking hard about the second-order effects. The first-order story is obvious: longer context means you can give a model more information. The second-order story is about cost, architecture, and where defensible positions form.

The Cost Curve Is Not Intuitive

Attention in transformer models is quadratic in sequence length. Double the context, quadruple the compute — roughly speaking. Models have gotten much more efficient at long context through architectural improvements like sliding window attention, linear attention variants, and better memory management during prefill. But the fundamental cost structure still puts pressure on unit economics at long context.

What that creates is a tiered economics problem. For a use case that genuinely needs 500k tokens of context — say, a legal firm reviewing a full merger agreement — the cost is high but the value delivered is even higher. For a use case that accidentally stuffs 500k tokens into context because the application was poorly architected, the cost is high and the value could have been delivered with a smarter retrieval approach using 10k tokens.

The companies building tooling around context management — smart caching, context compression, hybrid retrieval strategies that decide what actually needs to be in context versus what can be retrieved on demand — are solving a problem with direct economic impact for their customers. That's a different conversation than "our tool improves developer experience."

KV Cache as Strategic Infrastructure

One specific angle worth understanding: the KV cache is increasingly where cost optimization happens in long-context serving. For repeated context — a system prompt, a large document that appears in many queries, a code repository — caching the key-value pairs computed during prefill means you don't pay for reprocessing them on every request. Prompt caching can reduce costs by 60-90% for use cases with shared context.

The interesting investment angle here is not cache management as a standalone product — it's KV cache management integrated into the broader serving layer. Organizations running high-volume applications with shared context blocks need cache invalidation logic, cross-request cache sharing, and cost attribution that shows which requests benefited from cache hits. None of that exists as a mature product today.

Retrieval vs. Stuffing: A Framework

Long context hasn't killed retrieval-augmented architectures — it's changed when you use each approach. The decision isn't "use RAG or use full context." It's more nuanced:

  • Dense retrieval (RAG): Low latency, cost-efficient, works well when the relevant information is a small fraction of a large corpus. Requires good embedding quality and retrieval precision.
  • Full context stuffing: Higher cost and latency, but better for tasks requiring cross-reference across many parts of a document, or when you don't know in advance which parts are relevant.
  • Hybrid: Retrieve candidate chunks, then stuff a subset into long context for reasoning. More complex to implement, but often the right answer for knowledge-intensive tasks.

The teams building the abstraction layer over these three strategies — routing intelligently based on query type, document structure, and cost budget — are working on genuinely hard problems. And the market for that abstraction is real: any organization building knowledge-intensive AI applications needs to make this tradeoff, and most are making it badly right now because they don't have the tooling.

Where the Moats Are Forming

From where we sit, three positions look defensible in the context economics space:

First, the organizations that accumulate proprietary long-context training data. Instruction-following at long context is harder to train than short context, and the models that handle long context well did work to get there. Datasets and synthetic data pipelines specifically designed for long-context reasoning are an under-discussed competitive input.

Second, the serving infrastructure optimized for long-context use cases. This is architecturally different from serving optimized for short, high-volume requests. Memory bandwidth, KV cache management, and prefill optimization matter in ways they don't for short context. Companies building serving stacks with this in mind aren't competing on the same axis as general serving solutions.

Third, the application layer for specific domains where long context genuinely changes the quality of output. Legal review, code analysis across large repositories, scientific literature synthesis — these are use cases where the difference between 32k and 1M context is the difference between a demo and a real product. The first teams to build production-grade applications in each of these verticals, with serving infrastructure that makes the economics work, will be hard to displace.

Our Current View

Context window expansion is compressing several years of architectural decisions into a short window. Teams that adapted their application design for long context early are accruing advantages in data, infrastructure optimization, and product maturity. The window to build that lead is open now, but it won't stay open indefinitely as the rest of the market catches up.

We're actively looking at companies in this space. If you're building in context-aware serving, long-context evaluation, or application infrastructure for knowledge-intensive domains, we'd like to hear from you early.

Working on long-context infrastructure or retrieval architecture? Let's talk.