DEV Community

Cover image for What Running AI for the Air Force Taught Me About Enterprise AI
Jamie Thompson
Jamie Thompson

Posted on • Originally published at sprinklenet.com

What Running AI for the Air Force Taught Me About Enterprise AI

I spent the better part of a year as Principal Investigator on an Air Force basic research program focused on trust measurement, multi-source signal analysis, and uncertainty quantification. Before that, I contributed to AI research efforts for the Air Force Research Laboratory through SBIR programs. And before any of that, I spent nearly two decades building AI products across industries.

None of it looked like the AI demos you see at conferences.

There were no polished chatbots on stage. No live demos where everything magically works. The actual work was methodical, sometimes tedious, and deeply focused on whether the outputs could be trusted, explained, and acted on by real people in real situations.

That experience changed how I think about enterprise AI. Not because government AI is so different from commercial AI, but because it strips away the hype and forces you to confront what actually matters. Here are five lessons I took away that apply to any organization trying to make AI work at scale.

1. Nobody needs another AI demo. They need production systems.

The defense and intelligence communities are drowning in AI prototypes. Every contractor, startup, and research lab has a demo. Most of them are impressive for about fifteen minutes. Then someone asks, "How does this connect to our existing systems?" or "What happens when the data format changes?" and the conversation gets quiet.

What I learned working with Air Force programs is that the gap between a working demo and a production system is enormous. It is not a gap you can close with more funding or more data scientists. It requires thinking about integration from day one. It requires designing for the unglamorous realities of authentication, access control, data pipelines, and system monitoring.

The organizations that succeed with AI are not the ones with the most sophisticated models. They are the ones that treat AI deployment the way they would treat any critical infrastructure project: with proper engineering, testing, and operational planning.

If your AI only works in a demo environment, you do not have an AI capability. You have a science project.

2. Trust is not a feature. It is the architecture.

My research focused specifically on trust and influence measurement frameworks. One of the clearest takeaways was that trust in AI systems cannot be bolted on after the fact. You cannot build a black-box system and then add an "explainability module" later. That approach fails every time.

Trustworthy AI requires architectural decisions made at the foundation. That means audit trails that log every decision and every data source consulted. It means source citations so a human can verify why the system produced a given output. It means uncertainty quantification so users know not just what the system thinks, but how confident it is.

In government contexts, this is not optional. An analyst cannot act on AI output they cannot explain to their leadership. A program manager cannot defend a recommendation to Congress if the reasoning is opaque.

But this applies equally in the commercial world. A CFO will not trust an AI-generated forecast if they cannot trace it back to source data. A compliance officer will not sign off on an AI system that cannot explain its decisions. A board of directors will not accept "the model said so" as justification for a strategic pivot.

If your AI system does not have built-in explainability, source attribution, and audit trails, you do not have a trustworthy system. You have a prototype with good marketing.

3. Multi-source data is where the real value lives.

The hardest part of AI is not the model. It never has been. The hardest part is connecting to the messy, fragmented, inconsistent data sources that organizations actually rely on.

During my Air Force research, the signal analysis work involved synthesizing information from multiple sources, each with different formats, different levels of reliability, and different update cadences. The model was almost the easy part. The data integration was where the real engineering happened.

This is true in every enterprise I have worked with. The data lives in SharePoint, in legacy databases, in email threads, in PDFs that were scanned in 2014, in Slack channels, and in the heads of subject matter experts who are three years from retirement. Getting AI to work means building connectors to all of it, normalizing it, handling conflicts between sources, and doing so in a way that is maintainable over time.

This is exactly why I built Knowledge Spaces as a multi-source RAG platform. The value is not in choosing the right LLM. The value is in connecting the right data, from the right sources, with the right context, so the AI can actually be useful.

Any vendor who tells you their AI solution "just works" without a serious conversation about data integration is selling you a fantasy.

4. Small teams move faster, and speed is everything right now.

I have seen AI programs run by teams of five deliver results in months that teams of fifty could not deliver in years. This is not an exaggeration. It is a pattern I observed repeatedly across SBIR programs, research labs, and commercial engagements.

Large defense contractors have enormous resources, deep relationships, and decades of institutional knowledge. But when the technology is evolving as fast as AI is evolving right now, those advantages can become liabilities. Big teams mean more coordination overhead. Legacy processes mean slower iteration. Risk aversion means waiting for someone else to prove the concept before committing.

Small, focused teams can prototype, test, deploy, learn, and iterate in the time it takes a large program to complete its requirements gathering phase. When the underlying models are improving every few months, the ability to adapt quickly is not just nice to have. It is the difference between deploying something useful and deploying something obsolete.

This does not mean large organizations should only work with small companies. It means they should structure their AI initiatives to preserve agility. Use small, empowered teams. Reduce approval layers. Accept that the first version will not be perfect and plan to iterate. The organizations that move fastest will learn fastest, and the ones that learn fastest will win.

5. The fractional Chief AI Officer model works.

Not every organization needs a full-time Chief AI Officer. Most do not. What they need is someone who has built AI systems in production, who understands both the technology and the organizational dynamics, and who can set the right architecture and strategy without the overhead of a permanent C-suite hire.

I have been operating as a fractional Chief AI Officer for multiple organizations, and the model works for a simple reason: the critical decisions in enterprise AI are architectural and strategic, not operational. You need experienced judgment to decide which problems AI should solve, which data sources to prioritize, which vendors to trust, and which hype to ignore. You do not need that person sitting in every standup meeting.

A good fractional CAIO sets the foundation, builds the evaluation frameworks, trains the team to execute, and then stays engaged enough to course-correct when the landscape shifts. They bring pattern recognition from working across multiple organizations and industries, which is something no internal hire can replicate.

The organizations getting the most value from AI right now are not the ones who hired the most impressive AI team. They are the ones who found someone who has done this before, who can tell them directly what AI can and cannot do for their specific situation, and who can keep the team focused on outcomes instead of technology for its own sake.

The Bottom Line

AI in the enterprise is not a technology problem. It is an integration problem, a trust problem, and a leadership problem. The technology is mature enough. The models are good enough. What most organizations lack is the experience to deploy AI in a way that is reliable, explainable, and connected to the data and systems that actually matter.

That is what working with the Air Force taught me. Not how to build better models, but how to build AI systems that people can actually trust and use.

If your organization is navigating these same challenges, I am always happy to compare notes.

Jamie Thompson is the Founder and CEO of Sprinklenet AI, where he builds enterprise AI platforms for government and commercial clients. He writes weekly at newsletter.sprinklenet.com.

Top comments (0)