3 Pilgrim LLC
Version 1.0 · February 5,2026
Click here for full PDF of paper
1) Why This Paper Exists
Over the last several years, AI capability has improved sublinearly while cost has grown superlinearly. Larger clusters consume more power, generate more heat, move more bits across longer fabrics, and require increasingly elaborate coordination—yet benchmark gains continue to flatten.
Industry narratives initially treated this as a temporary engineering lag: better chips, denser interconnects, improved cooling, or more data would restore prior scaling slopes. Instead, each local improvement shifted pressure elsewhere in the system.
This paper argues that the observed flattening is not accidental, cyclical, or purely economic. It is the natural result of multiple physical and informational constraints coupling into a single limiting surface.
We call that surface the Compute Efficiency Frontier (CEF).
The CEF is not a wall you hit in one dimension. It is a multidimensional boundary beyond which marginal capability per unit of cost, power, data, or scale asymptotically approaches zero. Past this frontier, additional investment produces entropy, coordination loss, and stranded capital—not intelligence.
2) What the Paper Says (Plain Language)
The idea in one line
Modern AI systems operate within a constrained region of
possibility—the Compute Efficiency Frontier—where adding more
resources yields diminishing returns, and beyond which returns
collapse.
What defines the frontier
The CEF is formed by the interaction of six independent but
coupled constraints:
Compute – finite switching efficiency, error correction overhead
Power – delivery limits, grid availability, conversion losses
Heat – thermal density, removal rates, material limits
Data – finite novelty, signal dilution, contamination
Parallelism – synchronization costs, Amdahl-type limits
Transmission – latency, bandwidth, finite signal speed
Each constraint is rooted in a different physical or informational law. None alone explains the slowdown. Together, they define a convex efficiency boundary.
Why improvements don’t break through
Hardware advances, faster optics, higher TDP accelerators, and
better cooling improve local efficiency, but they do not introduce
new independent degrees of freedom. They slide systems
along the frontier rather than moving it
outward.
Why scale stops helping
As systems grow, coordination, synchronization, and movement costs
rise faster than useful computation. Capability per joule, per bit
moved, and per unit time converges toward zero at the
frontier—even as total expenditure explodes.
3) What Distinguishes This Framework
Topology, not tactics
This paper does not catalog cooling techniques, networking
upgrades, or facility designs. It presents the geometry that
explains why all such tactics encounter the same diminishing
returns. The CEF is a systems-level object, not a data center
checklist.
Physics and information theory explain the economics
The flattening of ROI curves is not primarily a market failure or
management error. Capital efficiency mirrors physical and
informational limits. The economics are downstream of the
physics.
A portfolio-level decision rule
The CEF reframes investment decisions: beyond the frontier,
additional capital reliably converts into heat, latency, and idle
silicon. This is not a matter of execution quality; it is a
structural boundary.
4) Theoretical Implications
(Assuming the Framework Is Correct)
Scaling saturation is structural
As clusters grow, marginal capability per unit of compute, power,
or data trends toward zero at the CEF. Alleviating one constraint
tightens others, preserving the frontier.
No single wall breaks the curve—independence does
There is no thermal fix, network fix, or hardware fix that
restores old scaling slopes. Only new independent axes of
freedom—true dimensional expansion—can move the frontier.
Densifying existing axes cannot.
This aligns directly with the earlier semiotic correction: most scaling today increases correlated capacity, not independent degrees of freedom.
The optimization target shifts
Progress shifts from raw capacity to coherence: intelligence per
joule, per bit transmitted, per meter of distance, per unit
latency. Total FLOPs purchased becomes a secondary
metric.
5) Potential Implications
(Downstream, Not Predictions)
A) Strategy & Economics
From magnitude to efficiency
The “bigger-is-better” era gives way to an efficiency regime.
Outside a narrow operating band, enlarging clusters produces
negative marginal returns. Smaller, specialized, modular systems
become economically dominant.
Capex converts into opex pressure
Rising device power densities and facility constraints force
superlinear spending on power delivery and cooling to support
sublinear capability gains—an inversion that eventually constrains
deployment regardless of demand.
B) Infrastructure & Operations
Cooling becomes architecture, not plumbing
Thermal management moves from an implementation detail to a
first-order design constraint. Even so, improved cooling manages
the Heat wall—it does not remove it.
Bandwidth grows; latency persists
Faster optics and photonics reduce electrical loss and raise
throughput, but finite signal speed and coordination overhead
preserve the Transmission and Parallelism constraints.
Power and siting dominate timelines
“Time to power” and grid access become binding constraints.
On-site generation mitigates delay but does not bypass the
underlying limits.
C) Data, Training, and Evaluation
The Data wall hardens
As useful signal becomes scarcer, synthetic feedback and
model-on-model training raise the cost of novelty. Value shifts
toward domain-specific data and closed-loop generation with
measurable independence.
Scaling laws evolve into efficiency laws
The industry’s growing emphasis on test-time compute, curriculum
design, and algorithmic efficiency reflects implicit recognition
of the CEF—even when not named as such.