Link Margin: Why Scale Won't Save You
Abstract
Large language models improve with scale, but scale increases capability—not reliability. In high-stakes domains, the dominant failure mode is not hallucination but unresolved ambiguity: models lack access to authoritative variant spaces, resolution rules, and consequence thresholds at inference time. This produces outputs that are plausible yet unverifiable.
This essay argues that reliability requires architectural link margin, not larger models. It presents a deterministic pattern in which models propose mappings, a separate authorization layer verifies them against authoritative sources, and only validated claims are allowed to exist. The result is not smarter models, but systems that can prove correctness, fail safely when context is insufficient, and be audited.
I. Link Margin as an Engineering Concept
In RF engineering, there's a concept called link margin.
It's the buffer between "signal received" and "signal lost." If you're designing a satellite link, you calculate everything: transmit power, antenna gain, atmospheric absorption, receiver sensitivity, and so on. You end up with a number. Say you need −100 dBm to decode the signal, and you're receiving −90 dBm. That's 10 dB of margin. It means conditions can get ten times worse—rain, interference, pointing errors—and you still get through.
If your margin is zero, you're on the edge. Any noise kills the link.
Engineers don't hope for good weather. They build margin.
II. Scale as an Article of Faith
The AI industry is hoping for good weather.
The prevailing belief is that scale solves everything. More parameters, more data, more compute, more RLHF, and the model will get it right. GPT-5 will be better than GPT-4. GPT-6 better still. Somewhere on that curve, the thinking goes, we cross a threshold where the model stops making mistakes that matter.
This is not an engineering argument. It's an article of faith.
Over the last six months, I measured what actually happens when AI systems are deployed into real production codebases—not benchmarks, not demos, but outputs evaluated against concrete requirements. I tracked a failure mode I call the fiction rate: outputs that are syntactically valid and reviewable but functionally wrong. Roughly 5,500 AI-assisted commits in those systems had a fiction rate of 87% (internal measurement across ~5,500 AI-assisted commits over six months).
The code compiled. It often looked reasonable. It just didn't do what it was supposed to do.
This isn't a complaint. It's a measurement.
III. Variant Resolution Is the Problem
Let me make this concrete.
Ask an AI to calculate the nutrition for "chicken breast." Simple request. The model will confidently return an answer—probably something like 165 calories and 31 grams of protein.
The problem is that USDA FoodData Central contains over 100 distinct chicken breast variants: raw versus cooked, skin-on versus skin-off, bone-in versus boneless, fried versus grilled versus roasted. Each variant has different macros—about 120 kcal for raw skinless, over 220 kcal for fried with skin. FDC explicitly distinguishes these as separate entries—the data exists. The problem is identity, not a missing source.
The model has no reliable way to bind your request to the correct variant. The text representation alone does not encode which preparation, cut, or state applies, and no amount of scale guarantees that binding without explicit context and rules.
So the model pattern-matches to something. It returns a chicken breast. Not the right chicken breast.
For a diabetic calculating insulin, the wrong variant means the wrong dose. That's not a hallucination problem. It's a resolution problem.
And scale doesn't fix it.
A superintelligent model could know everything ever written about chicken—poultry science, Maillard reactions, nutritional biochemistry. It still wouldn't know which of the 100+ USDA records applies to your dinner, because that information isn't in its weights. It's in a database that the model doesn't have authoritative access to at inference time.
Scale builds capability. It does not build margin.
IV. Why the Usual Fixes Fail
The usual responses don't work either.
RAG retrieves documents and stuffs them into context. But retrieving text is not the same as resolving which authoritative record applies. You can retrieve all 100 chicken variants; the model still has to choose one. On what basis? Hope.
RLHF optimizes for human approval. But humans cannot verify variant selection at scale. When the model says "chicken breast, 165 kcal," the rater thinks "that sounds right" and clicks thumbs up. The system learns plausibility, not correctness.
Guardrails filter outputs after generation. They are reactive, not constitutive. A guardrail can check that the output is valid JSON. It cannot reliably verify that the selected database record is the correct one.
JSON mode enforces structure. Required fields present. Types correct. None of that guarantees the content is right. You can have a perfectly structured lie.
None of these approaches, by themselves, manufactures link margin. They do not provide the model with the authoritative variant space, the resolution rules, or the domain tolerances. They just make the guessing more elaborate.
V. Intelligence and Reliability Are Orthogonal
The problem isn't that models are stupid. They're not.
The problem is that intelligence and reliability are orthogonal.
A model can be arbitrarily intelligent and still lack, at inference time:
- The variant space — the live database of possibilities: food records, drug formulations, legal filings.
- The resolution rules — the logic that ties context to a specific variant. "Grilled" means cooked. "Skinless" means without skin. These rules are domain-specific and change over time.
- The domain tolerances — how much error is acceptable? Ten percent might be fine for a recipe blog. Five percent is dangerous for diabetic management. One-tenth of a percent can be lethal for drug dosing.
- The consequence thresholds — where error becomes harm. These are found in regulations, clinical guidelines, and case law, not in training data.
Intelligence is probabilistic. It generates plausible outputs from learned distributions.
Safety is deterministic. It requires knowing the exact right answer and proving it.
These are different problems. Solving one does not solve the other.
A system that is 99% accurate is still wrong 1% of the time. In high-stakes domains, 1% is catastrophic.
VI. An Architecture That Holds
So what does work?
After months of iteration, I arrived at an architecture that holds—not because it's clever, but because it's strict. This is the pattern that became Ontic: authority lives outside the model, gates enforce it, and oracles provide ground truth.
Principle 1: Models propose. The authorization gate decides.
No model output—regardless of confidence, format, or source—becomes a claim unless it passes through a deterministic authorization gate. The model proposes "chicken breast, FDC ID 171077." The gate verifies: Does the record exist? Does the description match? Are the required resolution dimensions satisfied? Is confidence above threshold?
If any check fails, the claim does not exist. Concretely: if FDC has no entry for the requested ingredient state, the workflow returns unknown and fails closed instead of interpolating.
Principle 2: Mappers, not calculators.
Models select identifiers. They do not perform arithmetic. "2 cloves garlic" maps to an FDC ID, unit "clove," quantity "2." Deterministic code does the math: look up grams per clove, multiply, calculate nutrients.
Models hallucinate arithmetic. Code does not.
Principle 3: Resolution dimensions are explicit.
For each claim type, define what context is required. Chicken breast requires cooking state, skin state, and bone state. If the input is "grilled chicken breast," the cooking state is inferred as cooked. If the input is just "chicken breast," the system either rejects (strict mode) or defaults with disclosure ("assumed raw").
The system knows what it does not know.
Principle 4: Confidence is computed, not claimed.
Models do not estimate their own confidence. That is circular. The authorization gate computes confidence deterministically from multiple factors:
- Identity score (string alignment)
- Resolution quality (supplied vs inferred vs defaulted context)
- Evidence integrity (source fetch success and validity)
All factors must meet minimums. High identity does not compensate for missing resolution.
Principle 5: Provenance is cryptographically sealed.
Every authorized claim includes a signed audit trail: what the database returned, which rules were applied, which calculations ran, and which inferences were made.
Years later, you can prove exactly what happened, not just assert it.
VII. Why Nutrition Comes First
The first domain this architecture targets is nutrition.
Nutrition is unusually well-suited for this work. The authoritative oracle (USDA FoodData Central) is public. Regulatory tolerances are explicit. Known consequence thresholds include diabetic management, renal limits, and allergen exposure.
If you can solve the chicken breast problem with verifiable accuracy, you've proven something transferable.
And it transfers cleanly:
| Domain | Oracle | Same Architecture |
|---|---|---|
| Nutrition | USDA FDC | ✓ |
| Drug dosing | FDA Orange Book | ✓ |
| Compliance | SEC EDGAR | ✓ |
| Legal research | Case law databases | ✓ |
Swap the oracle, keep the chassis. The problem is always the same: map ambiguous input to an authoritative record, verify the mapping, compute deterministically, and prove the chain.
VIII. Why This Generalizes
This matters beyond recipes.
Enterprises are sitting on billions of dollars of AI capability they cannot deploy. Models are good enough. The bottleneck is trust.
A bank cannot ship AI-generated loan summaries if it cannot prove correctness. A hospital cannot accept dosage suggestions that cannot be audited. A law firm cannot rely on citations that might be fabricated.
What's missing is not intelligence. It's verification infrastructure—systems that can take model outputs and prove, not claim, that they are grounded in authoritative sources.
That is what link margin means in this context: the distance between "the model said it" and "we can defend it."
IX. Build Margin, Not Hope
Engineers don't hope for good weather. They build margin.
The AI industry has been hoping—hoping that scale will save us, that the next model will finally be reliable enough, that guardrails will catch what matters.
Hope is not an engineering strategy.
Link margin is.
The question is not whether AI will transform every industry. It will. The question is whether that transformation happens responsibly or recklessly—whether we build systems we can verify, or systems we merely hope are right.
That choice is still ours.