Why this matters more than the model selection
AI models produce plausible-looking outputs even when they are wrong. That property, fluency without grounding, is what makes them feel useful and what makes them dangerous to deploy without validation.
When an AI tool gets a fact wrong, it does not announce it. It does not slow down. It does not flag uncertainty in a way most users will read. It produces the wrong answer in the same confident voice it produces correct ones. That is a property of how the technology works, not a bug to be patched.
Validation is the only mechanism that prevents this property from quietly damaging your operations.
The hallucination problem is structural, not transient
Researchers have measured the rate at which large language models produce false but plausible outputs across many tasks. Even at the frontier of model capability, hallucination rates are non-zero on factual tasks, and they vary substantially by domain and prompt structure. The Stanford AI Index has tracked these rates across model generations and benchmarks for several years.
This matters because two business habits are dangerous when combined:
- Treating LLM output as if it were a search result.
- Skipping validation because "the model has gotten so good."
The second statement is partly true and entirely beside the point. Even if hallucination rates drop substantially, you cannot tell from any individual output whether that output is the wrong one. You need a process that catches the wrong ones, regardless of the rate.
Validation does not depend on knowing how often the model is wrong. It depends on having a mechanism that works even when you do not know which output is the wrong one.
What validation actually looks like in practice
Validation is not an oversight committee. It is a small set of checks attached to a specific workflow, run consistently. There are five components.
1. Acceptance criteria written before the prototype runs
Before any AI output is allowed to touch the workflow, you write down what acceptable output looks like. Concretely. "Reply tone matches our brand voice" is not an acceptance criterion. "Reply addresses the customer by name, references the specific product they asked about, includes our standard disclosure for refund requests, and does not include claims about pricing" is.
If you cannot write the acceptance criteria, the workflow is not yet ready to automate. That is a useful conclusion, not a failure.
2. Comparison against twenty known-good cases
Before the AI runs in production, you collect twenty real examples of the workflow that have already happened: twenty inbound leads, twenty quote requests, twenty support emails, along with what the correct human output was for each.
Run the AI against the same inputs. Compare its output to the human output using the acceptance criteria you wrote. Score it. If the AI fails on one in five, you know the failure rate is at minimum 20 percent. That is the floor. The ceiling is somewhere north of that, and you do not know yet how far north.
Twenty cases is not statistically rigorous. It does not need to be. It is a forcing function. Most teams skip this step entirely. The teams that do it spend an hour and surface failure modes that would have taken a quarter to discover in production.
3. Documented human-approval points
Customer-facing or revenue-affecting outputs route through a human approver during validation. This is not "we will spot-check sometimes." It is every output, reviewed by a named person, recorded. Not because the AI cannot be trusted forever, but because the only way to know whether it can be trusted is to compare its output to a human decision repeatedly.
After accuracy thresholds are consistently met for a defined window, approval moves to sampling, for example every tenth output, or any output flagged by a deterministic rule.
4. Sampling for a defined window after launch
Validation does not end when the AI goes live. A randomly sampled set of outputs is reviewed for a defined post-launch window. The findings update the acceptance criteria and trigger re-validation if accuracy drifts.
Drift is a real phenomenon. Inputs change. Customer language changes. The model itself may change if the vendor pushes an update. Sampling is the mechanism by which you notice drift before it produces complaints.
5. A measurable metric tied to the workflow
There is a quantitative metric that tells you the workflow is still working a month later. Time saved. Response time. Error rate. Revenue per workflow run. Pick one. Report it weekly. No metric, no scale. If you cannot measure it, you cannot tell whether continuing to run it is creating value or destroying it.
Borrowed directly from safety-critical engineering
I spent twelve-plus years building software certified under DO-178C, the standard used in commercial aircraft. That work demands defined workflows, written acceptance criteria, validated outputs, and measurement before deployment. It also imposes a discipline that does not show up in most software work: you must show how the system is allowed to fail.
Avionics software is not allowed to fail unpredictably. Engineers define how it is allowed to fail, what triggers human intervention, and how long the system is allowed to run before re-validation. That mental model maps directly onto AI inside a business, not because a small business needs aircraft-grade certification, but because the underlying property is the same: an output that downstream people will trust and act on.
The translation for SMBs is small but important.
- Define the workflow's allowed failure modes. What kinds of wrong outputs are tolerable? Which are not?
- Define what triggers human intervention. Specific keywords, output formats, confidence signals, or randomly sampled review.
- Define a re-validation interval. Quarterly is a reasonable default for most SMB workflows. More often if the inputs are changing fast.
This is not a certification process. It is a one-page document for each workflow.
A practical validation routine for SMBs
Here is what we put in place during a typical AI Workflow Sprint. It fits on one page.
- Acceptance criteria document: what acceptable output looks like, written before the prototype runs.
- Twenty-case validation report: the AI's output on twenty real cases, scored against acceptance criteria, with failures categorized.
- Approval procedure: who approves, what they look at, what triggers escalation.
- Metric definition: the quantitative metric, the source of truth for the data, and the reporting cadence.
- Re-validation trigger: when the workflow is checked again, by whom.
Five documents. None of them are long. All of them are required.
Common objections and answers
"We do not have time for this."
You have time for the rework that ungoverned AI output produces. You have time to apologize to customers when the model invents a refund policy. The validation routine takes less time than either of those, and it produces an artifact you can use to onboard the next team that owns this workflow.
"Our competitors are moving faster."
Many of them are also abandoning AI projects within a quarter. Speed without validation is not a competitive advantage. It is an unfunded liability that shows up later as customer complaints, regulatory exposure, or quietly broken processes nobody trusts anymore.
"The model is good enough that this seems like overkill."
The model's average accuracy is irrelevant for any individual output. You cannot tell from a single response whether the model was right. Validation does not depend on the model being bad. It depends on you needing to know which outputs are wrong before they reach a customer.
"We will validate after we launch."
Then you are committing to a validation routine you have not designed, on outputs you have not specified, against criteria you have not written. The most common version of "we will validate after launch" is "we will not validate." Be honest about which one you are choosing.
What this is not
Validation is not an oversight committee, a steering group, or a quarterly executive review. It is a small set of checks attached to one workflow at a time, owned by the same person who owns the workflow itself.
If your validation routine produces a lot of meetings, it is the wrong shape. If it produces five short documents and a weekly review of a metric, it is the right shape.
Where Geaux Digital Media fits
This validation discipline is the reason the practice exists. Most AI consulting markets a tool. We market a process: define the workflow, constrain the AI, validate the output, measure the result, and scale only after it works.
If you have an AI tool you have already deployed and you are not sure it is producing reliable output, an AI Workflow Review can include a retrospective validation pass against twenty real cases. That is often the highest-leverage hour you will spend on AI this quarter.
Further reading
Brent Dorsey is the founder of Geaux Digital Media and a Senior Systems & Software Engineer with 20+ years across Marine Corps technical systems and DO-178C avionics software for Boeing, GE Aviation, BAE Systems, and RTX. Geaux Digital Media helps Louisiana small businesses implement AI workflows that are defined, validated, and measured before they scale. Request an AI Workflow Review →