Perspectives

AI Decisions Without Evidence Are Just Opinions

Piper Cyterski
March 30, 2026
7 min read

Imagine your VP of Product asks: "Why are we using Claude instead of GPT-4o for the summarization feature?" And the honest answer is: someone tested it in a notebook three months ago, said it was better, and the team went with it. There's no record of what "better" meant. No rubric. No dataset. No scores. Just a decision that got made and a rationale that got lost.

Now imagine a different version: you pull up the experiment. Here's the dataset -- 200 real support tickets. Here are the criteria -- factual accuracy, format compliance, latency. Here are the scores, side by side, pinned to specific prompt versions and model configurations. Claude won on accuracy by 12 points. GPT-4o won on format compliance and was 40% cheaper. The team chose Claude because accuracy was the binding constraint for this use case, and the format issues were solvable with a structured output wrapper. The reasoning is right there, attached to the evidence.

That's the difference between a decision and a defensible decision. And right now, almost every AI team is operating entirely in the first mode.

The Defensibility Problem

This isn't about bureaucracy or process for its own sake. It's about building AI systems that you can stand behind when someone asks hard questions.

A customer wants to know why the AI gave them a wrong answer. Your legal team wants to know what testing was done before a feature launched. A new model comes out and the team needs to decide whether to switch, but the basis for the original decision is gone. An engineer leaves, and their judgment was the only record of why things were built the way they were.

In every one of these situations, the team without evidence is guessing. They're reverse-engineering their own rationale, re-running experiments they already ran, or -- worst case -- defending decisions they can't actually explain. That's not a minor inconvenience. It's a credibility problem that gets worse with every decision that goes unrecorded.

The teams building AI features that affect revenue, trust, and safety need to be able to answer "why?" with something more substantive than "it felt right in the playground."

Rigor Isn't Slower. Ad Hoc Is.

There's a persistent misconception that rigor slows you down. That structured experimentation is something you do when you have the luxury of time, and in the early days you just need to move fast and iterate.

The opposite is true. Ad hoc is slow. It just doesn't feel slow because the costs are distributed and invisible.

When there's no record of past experiments, people re-derive conclusions the team already reached. When there's no shared criteria, three engineers evaluate the same outputs and come to three different conclusions, then spend an hour in a meeting arguing about which model is "better" without agreeing on what "better" means. When there's no versioned lineage, a quality regression takes days to diagnose because nobody can identify which of the five things that changed last week caused the drop.

Rigor feels like overhead in the moment. Over a quarter, it's a multiplier. Teams with structured evidence make decisions in minutes that ad hoc teams debate for days -- not because they're smarter, but because they have data and a shared framework for interpreting it.

Collaboration Requires Shared Ground

This is the part that gets underestimated: AI evaluation is a team activity, not a solo one.

On a team of five engineers, each person has a slightly different intuition about what "good" means. One person cares most about factual accuracy. Another is focused on latency. A third thinks tone is the most important dimension. Without shared, explicit criteria, every evaluation is shaped by whoever ran it and whatever they happened to prioritize. The results aren't wrong, exactly -- they're just not comparable. Each person is measuring with a different ruler.

Shared criteria solve this. When the team agrees on a rubric -- factual accuracy scored on this scale, format compliance defined by these rules, tone evaluated against these anchoring examples -- everyone is working from the same definition of quality. Code review has shared standards. Design has shared systems. AI evaluation should too.

But shared criteria only work if they're accessible, versioned, and used consistently. If the rubric lives in someone's notebook, it's not shared -- it's one person's rubric that they might tell you about if you ask. A system of record makes the criteria, the datasets, and the results visible to the whole team by default. That's what turns evaluation from a solo exercise into a collaborative discipline.

What Gets Lost Without a Record

The specific things that disappear when experiment results aren't captured systematically:

Negative results. Positive results get remembered -- "we switched to Claude and quality improved." Negative results vanish. Nobody writes up "we tried few-shot prompting with 8 examples and it was worse than 3 examples for our use case." Six months later, someone tries 8-shot prompting again because the idea sounds reasonable and there's no record showing it already failed.

Decision reasoning. "We use GPT-4o for this task" persists as institutional knowledge. "We use GPT-4o because Claude scored higher on accuracy but 30% lower on format compliance with our structured output schema, and format compliance was the binding constraint for this integration" does not persist. Without the reasoning, the decision can't be re-evaluated when circumstances change. You just carry forward a frozen conclusion whose justification has evaporated.

Baselines. Your summarization quality score is 3.8 today. Is that good? Compared to what? Without a recorded baseline, every score is an isolated number. With one, it's a trajectory -- you can see whether you're improving, regressing, or flat, and correlate those changes with specific configuration changes.

Institutional knowledge. When an engineer leaves, everything they learned through months of experimentation leaves with them -- unless the system captured it along the way. The difference between "Sarah knows why we do it this way" and "the experiment history shows why we do it this way" is the difference between fragile and durable knowledge.

What a System of Record Actually Looks Like

It means that every experiment produces structured, queryable data as a side effect of being run. Not because someone exported results and uploaded them somewhere. Not because someone wrote a doc after the fact. Because the act of running the experiment is the act of recording it.

Concretely:

  • Every evaluation run is tied to a specific dataset version, specific criteria versions, and specific model configuration.
  • Results are immutable. You can run new experiments, but you can't retroactively edit old scores. Tuesday's results still say what they said on Tuesday.
  • Lineage is automatic: any result can be traced back to the exact inputs, criteria, and configuration that produced it.
  • Everything is queryable: "show me every experiment that used Claude on this dataset, sorted by accuracy score" is a filter, not a research project.
  • Criteria are shared artifacts, versioned alongside the experiments that use them.

The key property is that record-keeping costs zero marginal effort. If it requires a separate step, it won't happen consistently. If it's a byproduct of doing the work, it always happens.

The Compound Return

Evidence compounds in a way that intuition can't.

Each recorded experiment becomes a baseline for the next. Each refined criterion makes the measuring stick more accurate. Each decision backed by data becomes a node in a growing graph of institutional knowledge that any team member -- including people who haven't been hired yet -- can navigate and build on.

After six months, a team with a system of record doesn't just have better individual experiments. They have a body of evidence -- a searchable, versioned history of what they've tried, what worked, what didn't, and why. New team members ramp up in days because the reasoning is in the system, not locked in someone's head. Model migration decisions take hours instead of weeks because the evaluation infrastructure is already built and the baselines already exist.

The teams building this discipline now are creating an advantage that gets wider every quarter. Not because they're working harder, but because every experiment they run makes the next one faster, sharper, and better informed. That's not something you can catch up on by "being more rigorous" later. It's a compounding lead.