In a recent interview, Alexandr Wang, the 28-year-old CEO & co-founder1 of Scale AI, offered one of the clearest articulations to date of a national AI doctrine. His thesis is deceptively simple: if the United States wants to win the AI race, it must out-deploy its adversaries. In other words, speed of integration beats speed of innovation.
For Wang, the decisive factor isn’t who trains the biggest model or accumulates the most FLOPs. It’s who operationalizes AI faster, more safely, and in more domains. That belief puts Scale AI in a very particular place in the AI stack: not in the apps layer, not in the compute layer, but as a middleware complement to the training and inference layer. It is an enabler, validator, and optimizer of the models themselves.
In a field dominated by algorithmic fetishism and GPU worship, Wang’s view is not only contrarian, but likely correct.
The Real AI Stack
The AI stack is frequently misrepresented as a vague hierarchy of “models, compute, and apps.” In reality, it’s more physical, more capital-intensive, and more bottlenecked than people think. The actual structure looks like this:
Physical Infrastructure —> Power, land, cooling, water, substations, fiber, data centers. Constraints: siting, interconnection, megawatts.
Compute Infrastructure (training & inference) —> GPU clusters, orchestration, model labs (OpenAI, Anthropic, Meta). Constraints: chip supply, GPU scheduling, orchestration efficiency, capital. Scale AI complements this layer.
Applications/APIs/Wrappers —> Vertical apps copilots, agentic front ends. Constraints: distribution, defensibility, UI/UX differentiation, the models’ agglomerative nature.
Most investors overindex on Layer 3 (apps) or Layer 2 (compute). Wang is betting on a complement to Layer 2: the middleware that connects raw models to real-world workflows, especially in high-stakes, high-integrity environments like defense and autonomy.
Scale AI: The Data Foundry
Wang’s key metaphor is instructive. Just as Nvidia requires foundries like TSMC to turn its chip designs into GPUs, AI requires what Wang calls a “data foundry”: an industrial process that refines raw data into structured, targeted, safety-tested inputs suitable for model training and deployment.
This isn’t mere labeling. It involves:
Fusion of multimodal inputs (e.g., LiDAR + radar + GPS)
Edge-case synthesis
Fine-tuning and RLHF
Red-teaming and adversarial scenario testing
Continuous post-deployment feedback loops
That work is increasingly essential. For the last decade, deep learning advances were driven by algorithmic novelty. Today, the field is defined by scaling laws: more data, more compute, same architecture. But we’re hitting diminishing returns on “just more tokens.” The next leaps will come from high-relevance, high-quality, hard-to-generate data, especially in edge domains.
That’s what Scale buids. Not the model, not the compute layer, but the epistemic operations needed to make a model perform safely in reality.
Agentic Warfare
Wang’s strategic clarity is most evident when he talks about defense. He doesn’t romanticize AI-powered drones or killer robots. Instead, he focuses on planning loops: the slow, manual, error-prone processes that underpin modern command and control.
His thesis: LLM-based agents trained on decades of military doctrine, logistics, geography, and adversary behavior can compress planning cycles from days to minutes. Scale’s Thunderforge program, developed with the Department of Defense and INDOPACOM, is designed for this: embedding agents into real-world operational planning, simulation, and wargaming.
Wang calls this shift “agentic warfare”. This isn’t automation; it’s augmentation. Not replacing generals, but giving them cognitive infrastructure with superhuman recall and speed.
And the implications are serious. Whoever closes the OODA loop fastest in future conflicts will win. It’s that simple.
The China Clock
Wang is sober about China’s trajectory. He doesn’t call for decoupling or moralizing. He simply points to evidence: Chinese labs like DeepSeek and Alibaba’s Qwen are producing world-class models with far less compute than U.S. leaders.
What explains that? Wang’s answer: data orchestration and national deployment speed. China isn’t just training models. It’s deploying them in military, surveillance, and enterprise use cases at scale, with coordinated feedback loops and policy alignment. America, by contrast, is fragmented and procedural.
This is why Wang’s concept of data dominance matters. He’s not saying “whoever has the most tokens wins.” He’s saying whoever can generate, refine, and integrate task-relevant data into model loops faster will have the epistemic edge. And in high-stakes domains, that edge compounds fast.
Infrastructure Revisited
Wang argues the U.S. is building AI atop fragile scaffolding. He calls for the creation of enterprise-wide infrastructure within government and defense: shared cloud, standardized data frameworks, and national-scale testing environments.
This isn’t infrastructure in the physical sense (that’s Layer 1). It’s digital and procedural infrastructure. It is the glue that lets models be evaluated, integrated, iterated, and trusted.
This is where Scale AI positions itself: not as a model lab, not as a defense contractor, but as the cognitive operating layer for applied AI. The validator. The enabler. The bridge from sandbox to sovereign-scale deployment.
Conclusion: The Right Bet
Wang has made a clear bet. While others chase app-layer virality or model-layer fame, Scale AI is building the hard, unglamorous, necessary middle: the feedback, evaluation, and data scaffolding that lets AI leave the lab and enter the real world.
He’s not trying to outcompete OpenAI. He’s trying to make OpenAI’s models deployable, auditable, and effective at the edge, whether that edge is a battlefield, an enterprise, or an autonomous vehicle navigating fog and chaos. If he’s right, the most important company in the AI value chain won’t be the one that trains the biggest model. It will be the one that makes those models real.
Love your based analysis on AI !
My latest that references some of your work.
https://open.substack.com/pub/pramodhmallipatna/p/agi-meets-the-data-wall