← All writing
Featured Essay May 2026 · ~8 min read

What 25 Years in Large-Scale Systems Taught Me About AI That Most Companies Miss

Most enterprise AI strategies are built on a quiet assumption: that intelligence is the hard part. After two and a half decades engineering systems for national laboratories, scientific computing, and Fortune-scale operations, I'm convinced it isn't. The hard part is everything around the model.

When I joined the world of large-scale computing in the early 2000s, the dominant problem was scale. We were wiring together thousands of cores to simulate climate, decode genomes, model physical systems no single machine could hold. The lessons of that era — about reliability, observability, data lineage, and the brittle interface between research code and production infrastructure — quietly became the foundation of everything I do in AI today.

Most companies skip those lessons. They treat AI as a layer to bolt on. They are about to learn, expensively, why that doesn't work.

1. The model is the smallest part of the system

In a high-performance computing environment, the simulation kernel is maybe 5% of the codebase. The other 95% is data movement, scheduling, fault tolerance, validation, and the long tail of operational glue that lets scientists actually trust a result. Enterprise AI is the same shape, but most teams don't see it yet. The LLM is the kernel. Production-grade AI is the 95%.

2. Reliability is a research problem before it is an operations problem

The reliability of an AI system is not something you bolt on at deployment. It has to be designed into how data is collected, how features are versioned, how prompts and retrieval indices evolve, and how human feedback loops close. Companies that hand their AI initiative to a platform team after the model is built are repeating the mistake scientific computing made — and corrected — twenty years ago.

3. Trust is engineered, not declared

In national-lab work, the answer to "is this result trustworthy?" is never a single number. It is a chain: provenance of inputs, reproducibility of method, calibration of uncertainty, and an audit trail a peer can challenge. Trustworthy AI in the enterprise needs the same chain — and almost no one is building it. Governance frameworks and AI policies are necessary, but they are not sufficient. Trust lives in the pipeline, not the policy document.

4. The bottleneck is almost always the human-system interface

The most successful AI deployments I have led did not win on model quality. They won because we paid obsessive attention to the moment a human had to act on a system's output — the latency, the framing, the explanation, the recourse. AI that is technically correct and operationally useless is the default outcome. Avoiding it is a design discipline, not a model-selection problem.

5. The organization is the architecture

Conway's Law applies to AI with a vengeance. The shape of your AI capability will mirror the shape of your engineering, data, and decision-making organizations. If those are siloed, your AI will be siloed. If they are slow, your AI will be slow. The transformation work is rarely about technology. It is almost always about how teams are wired to each other.

What this means for leaders

If you are leading an AI program right now, the temptation is to optimize for visible wins: a chatbot, a copilot, a demo that lands well in the boardroom. Those have their place. But the durable advantage — the kind that compounds over years — comes from the unglamorous work of building the system around the model. Data infrastructure that earns trust. Evaluation harnesses that catch regressions. Governance that operates at the speed of engineering. Teams that ship and learn faster than the model itself improves.

That is the lesson 25 years of large-scale systems work keeps teaching me, in every new domain, with every new generation of technology: the model is never the moat. The system is.

Continue the conversation

Discuss this essay on LinkedIn

I publish each piece as a LinkedIn article so the discussion happens where the network already lives. Comments, counter-arguments, and case examples welcome.

Discuss on LinkedIn →
More writing
Trustworthy AI in Regulated Environments: A Practitioner's Field Guide Coming soon Why Digital Transformation Programs Stall — and What Actually Restarts Them Coming soon