a builder's codex
codex · operators · Eugene Yan · ins_mine-transcripts-to-promote-config

Close the feedback loop by mining session transcripts for patterns to promote into config

By Eugene Yan · Senior Applied Scientist; writes at eugeneyan.com on AI/ML systems · 2026-05-03 · essay · Closing the feedback loop — mining session transcripts for promotable patterns

Tier A · TL;DR
Close the feedback loop by mining session transcripts for patterns to promote into config

Claim

The fastest practical route from "we ship LLM features" to "the system gets better on its own cadence" is to treat session transcripts as the corpus from which improvements get harvested. Each session is a recorded interaction; each correction, retry, or visibly better path is a candidate pattern. The closing move is to harvest those patterns into config — prompts, tool definitions, retrieval rules, evaluator rubrics — so the next session inherits what the last session figured out. The corpus is the gym; the transcripts are the reps; the config promotions are the strength gain that compounds.

Mechanism

The cost of running an LLM call is now low. The cost of human attention reviewing a transcript is the binding constraint. Two design moves break the constraint: (a) transcripts are stored in a queryable shape (not free-text logs), so an evaluator agent can pre-filter the most-promising candidates; (b) "promote into config" is a real, tracked PR-shaped operation — not a one-off prompt edit — so promotions are reviewable, attributable, reversible. Once both are in place, the feedback loop runs continuously: agent runs → transcript stored → evaluator surfaces candidates → human or higher-tier agent reviews → config diff merges → next run inherits. Without the queryable shape, candidates drown in noise. Without promotion-as-PR, every improvement is local and forgotten on the next prompt edit.

Conditions

Holds when:

Fails when:

Evidence

"Close the feedback loop — mine session transcripts for patterns to promote into config."

— Eugene Yan, eugeneyan.com, 2026-05-03 (synthesised from his publication that week; see A trace alone teaches nothing; learning requires feedback attached to the trace for the parallel framing from Harrison Chase the same week).

The same idea appeared in Karpathy's Sequoia Ascent talk a few days earlier — the bottleneck is verification, not intelligence. Yan operationalises it: verification produces the labels; promotion turns labels into compounding system improvement.

Signals

Counter-evidence

Cross-references

Open the interactive view → View original source → Markdown source →