← Obsta Labs

Your Token Bill Is a Decision Receipt

AI didn't eliminate engineering cost. It moved it from execution to discovery.

March 2026 · Obsta Labs

Your AI token bill is not a compute cost. It is the price of discovering the correct solution.

AI collapsed the cost of writing code. It did not collapse the cost of knowing what to write. Once a model understands the solution, producing it is cheap. The expensive part is the exploration, the dead ends, the architectural pivots, the 40-turn investigation that narrows six approaches down to one.

That's not compute. That's decision-making. And for the first time, we can measure it.

14.2M reasoning tokens → 4,712 lines of code3,021 tokens per line — the cost of discovering the solution, not typing it

Most of those tokens were not spent writing code. They were spent comparing architectures, testing approaches, rejecting dead ends, and converging on a final design. Once the solution was understood, producing the code took a fraction of the total.

Three eras of engineering economics

Engineering has always measured whatever resource was scarce.

Era 1: Hardware. The scarce resource was physical. Engineering questions sounded like: "Can we run this on 64MB?" "How many nodes do we need?" The budget tracked machines.

Era 2: Cloud. Abstracted the hardware, kept the same model. Dashboards tracked CPU%, memory, latency, cost per request. The budget still tracked compute — just rented instead of owned.

Era 3: Decisions. LLMs collapsed the cost of execution. Code production became cheap. The scarce resource became reasoning: exploring the design space, testing hypotheses, iterating architecture. The budget now tracks decisions.

The inversion

For most of software history, ideas were cheap and implementation was expensive. A whiteboard sketch took 20 minutes. Building it took six months.

AI reversed this. Implementation is increasingly cheap. What's expensive is figuring out the right thing to build. The reasoning tokens your model burns while circling an architecture decision, comparing tradeoffs, invalidating approaches — that's the real cost. The code it writes afterward is a rounding error.

The most expensive tokens in a project are the early ones.The first 5M tokens produced almost no code — but created the compression that enabled everything after.

Early exploration often looks inefficient. It produces little code while the system probes the design space. But those tokens are not waste — they are the cost of narrowing many possibilities down to one viable architecture.

The snowball

Wasted exploration is worse than it looks. Every token in an AI session's context is re-read by the model on every subsequent turn. A dead-end investigation that produces 50,000 tokens of noise at turn 10 is not a 50,000-token mistake. If the session continues for another 100 turns, that noise is re-processed 100 times. The real cost is 5 million tokens.

This makes the cost curve of unmanaged sessions quadratic, not linear. Dead exploration introduced early in a session is catastrophically more expensive than dead exploration introduced late, because more turns remain to compound the re-read cost.

47.4M tokens stripped from sessions. 2.9B re-reads avoided.Each removed token would have been re-read an average of 61 more times.

The implication: cleanup is not housekeeping. Removing dead exploration from a live session has a multiplied return. Stripping 10,000 tokens with 50 turns remaining saves 500,000 tokens of future re-reads. The earlier the cleanup, the larger the multiplier.

This also explains why long AI sessions get expensive so fast. It is not just that they run longer. The accumulated noise from abandoned approaches participates in every API call, each re-read costing real money. The session is paying for its own dead ends on every turn.

The system has its own response to this: compaction. When context grows past a threshold, the model compresses the conversation into a summary and continues from there. This stops the snowball, but destructively. Compaction discards exploration and signal alike. The decisions, the constraints, the architectural reasoning that took thousands of tokens to reach are flattened into a paragraph. The snowball stops, but the session's reasoning quality resets with it.

Surgical cleanup is the alternative. Remove the dead exploration while preserving the decisions. This delays or avoids compaction, keeping reasoning quality intact while eliminating the re-read tax on noise. The value is not just the tokens saved. It is the compaction event avoided.

Agent friction

Not all token waste comes from reasoning. A second source is agent mechanics: failed file reads retried against wrong paths, parallel operations that cancel each other when one target is invalid, oversized reads that hit limits and force repeated attempts, and low-value narration between tool calls. These tokens do not buy exploration or decisions. They are coordination failures.

But once they enter a live session, they snowball exactly like abandoned reasoning branches — re-read on every subsequent turn, compounding the same way. A bad architectural hypothesis and a bad file-reading strategy have the same economic property: both become future context tax.

1.2M friction tokens recovered over 326 cleanup cycles.None of it reasoning. All of it re-read on every turn until removed.

This matters because it breaks the simplistic framing that expensive tokens equal deep thinking. They don't. Some fraction of every token bill is process overhead from poorly disciplined execution. The cost of AI-assisted engineering is shaped not only by reasoning quality, but by execution discipline in the agent loop.

Tokens are a decision proxy

Tokens are not a perfect measure of thinking. But they are the first measurable proxy we have for exploration, hypothesis testing, and architectural iteration during AI-assisted development. Design and exploration always existed in software engineering. What changed is that AI systems expose the reasoning process in measurable units.

Traditional metrics measure artifacts: LOC, complexity, latency. They tell you what was built and how it performs. They say nothing about the cost of arriving at the idea.

Tokens don't measure the intrinsic difficulty of a problem. They measure the cost of discovering a solution within a given system of humans, tools, and models. New metrics make this visible: tokens per LOC (decision density), cost per decision (price of each commitment), exploration ratio (discovery vs. execution).

The search space map

If tokens trace reasoning, then a session's token trail is a map of the design space it traversed. Not every path leads to a commitment. Most are probes — directions explored, evaluated, and abandoned before the session converges.

This can be quantified. Multiply the decisions a session made by how broadly it searched (branch factor) and how much it re-explored due to context damage (re-exploration multiplier). The result is estimated path probes — the number of meaningful design directions the session evaluated before arriving at its commitments.

27 decisions. ~204 paths explored. 7.5:1 exploration-to-commit ratio.The session evaluated roughly 200 design directions to arrive at 27 commitments.

The re-exploration multiplier is the hidden cost. A session that explored 100 paths but re-explored 50 of them because compaction destroyed the reasoning trace didn't traverse 150 paths. It traversed 100 and wasted 50. The multiplier separates productive exploration from damage-induced repetition.

This turns a token bill into something more useful than a cost report: a search space map showing how many directions were tried, how many survived, and how much work was repeated. Two sessions can spend the same tokens and traverse completely different search spaces. One explored broadly and converged. The other circled the same territory because it kept forgetting where it had been.

Decision budgets

If tokens measure reasoning, then projects now have a new budget line: decisions.

Decision Budget: 50M tokens. Exploration: 18M. Implementation: 25M. Refinement: 7M.

This is not speculation. It is what token-level session instrumentation already shows. Projects have distinct reasoning phases — exploration, stabilization, execution — with measurably different token economics. Exploration burns tokens fast with low code output. Execution converts tokens to code efficiently. The transition between them is where tooling matters.

What changes

When you stop framing AI costs as compute and start framing them as decisions, different questions appear:

Not "how do we reduce token usage?" but "how do we reach decisions faster?"

Not "how do we make the model cheaper?" but "how do we reduce wasted exploration?"

Not "how big should the context window be?" but "how do we preserve decision quality across sessions?"

The tooling that follows from this framing looks like decision infrastructure: serializing intent before execution, maintaining reasoning stability during long sessions, tracking which tokens contributed to commitments and which were noise.

That's not prompt engineering. It's the beginning of decision economics — optimizing how humans and AI reach the correct solution together.

Update, 10 March 2026. The numbers above were measured across 252 sessions over the project’s first months. Current state with continuous watch mode running 24/7: 12M tokens removed per day, 503M cache-read tokens avoided (snowball), $377 in cost savings — in a single 24-hour window. The patterns described here hold; the scale has increased.