Back Original

After "AI": Anticipating a post-LLM science & technology revolution

After "AI": Anticipating a post-LLM science & technology revolution

Published: 2025-12-29 Updated: 2025-12-29

I, for one, welcome the coming age of the post-LLM-datacenter-overinvestment-bust-fueled backyard GPU supercomputer revolution.


No expert can escape the logarithmic growth curve of expertise. Not Lt. Cdr. Geordi La Forge (foreground), not Dr. Leah Brahms (middle), and certainly not the Ship's Computer (background). Image source: screenrant, wikipedia

(Recently riffed out in Linkedinese, and somewhat revised below, for ownership and sanity's sake.)

Lt. Cdr. Geordi La Forge of the Starship Enterprise (pictured above) is the archetypical Gen AI-using Engineer and Executive. He uses the Ship's antimatter-powered LLM all the time, not infrequently to solve novel one-off problems (sometimes under life-and-death pressure). However, he is such a successful user only because he can spot incredibly subtle errors and compensate/augment with his own elite training combined with years and years of hands-on field experience.

Essentially, he has hard-earned specialist tacit knowledge, which is entirely absent from any static data corpus, no matter how detailed the corpus is. Taste, judgement, and intuition are not born, they are bred. Through diligent training, deliberate practice, and frequent feedback from violent contact with reality. Frequent violent contact.

However, Human and Computer expertise grows along logarithmic growth curves. Gains grow quickly in the beginning, for a short while, then slower but still perceptible for a bit more, and then, suddenly, imperceptibly. No amount of additional input produces any meaningful additional output.

Said another way…

Expert systems need expert users.

I dub this The Geordi LaForge Paradox.

The larger the language model, the more sophisticated the user's world model must be to. Only a trained radiologist's mental model can reasonably examine the computer model's classification. Only Geordi can challenge the opinions of both, the Ship's Computer, and Dr. Brahms's deeply informed opinions, because only Geordi LaForge has learned the hard things the only way there is to learn them; the long, hard way.

And this science fact has far-reaching implications.

But it is not where people are looking for the proverbial keys, viz. only under the white-hot lamp-post of nuclear-powered abstract LLM machines. Moths to a flame; locked in a self-created gravity well of sunk costs… both emotional/psychological attention and on-ground action.

My optimistic #LLM tech bet is that…

As of 2030-ish we will have accrued some non-trivial direct benefits (of transformer architecture, scaling + compression):

  • Likely a bunch of commodity super-specialist pattern matching expert systems (astronomical / radiological imaging / computer vision / multi-media production tools etc.)
  • And some commodity generalist language translation engines (literally the Babelfish from the Hitchiker's guide - Azmat Yusuf's Talk Machine and ilk).

However, we stand to gain far greater indirect benefits, as and when GPU investments transmute into serving today's severely under-served but world-reforming science/industry application areas…

Think "on-campus GPU supercomputers too cheap to meter".

I dearly hope we get a version of what happened after the boom years of railroads, telecoms, and/or cloud computing…

  • Massive capital investments fuelling totally unexpected industrial and economic and socio-political phenomena, by way of infrastructure ownership re-allocations through write-offs, fire sales, and bankruptcy style M&A.
  • Perhaps (I hope) we get a disintermediation of datacenters. People ship out containers to private industry and universities and so forth — stacks of supercomputers in your backyard.

A whole new breed of Oxide Computer Company companies.

Could a "GPUs too cheap to meter" phase—say, about a decade, up to 2040—remarkably speed up cycle times of traditional deterministic modeling / simulation type workloads.

Think:

  • Commodity whole genome sequencing in seconds, and high-fidelity scans and imaging studies (think supercomputer powered MRI machines).
  • n of one precision medicine breakthroughs - collapse expert-led differential diagnosis into days or hours, instead of years and months
  • Which in turn accelerate vaccine response to epidemics (hours to first trials, days to validation, because perhaps we don't need large pools of people and lots of time to make statistically significant observations)
  • Whole rocket system design/simulation/testing in computers (no more finding limits primarily from "unplanned mid-flight disassemblies").
  • Full-brain (maybe even full-body) simulations for mouse models. Or at least nematode models.
  • Multiphysics and material science innovation (with or without LLM assist).
  • etc. etc. etc.

Whatever is a stupid-expensive computing investment today.

Society has been building massively parallel compute hardware for… something… GPUs and Memory became commodity units, primarily through consumer applications (not the needs of heavy industry / defence etc.).

  • First, in service of gaming and video streaming applications
    • -> fed into classical ML/AI
      • -> together fed into distributed ledgers
        • -> now repurposed / commandeered into scaling-style AI

Except, scaling-style AI is hitting the same energy and physics walls as crypto mining, only faster, and at orders-magnitude more capex and opex.

In the words of Arvind Krishna, CEO of IBM, in conversation with Nilay Patel, on the Decoder podcast. (Emphasis mine).

So, let’s ground this in today’s costs because anything in the future is speculative.

It takes about $80 billion to fill up a one-gigawatt data center.

That’s today’s number.

If one company is going to commit 20-30 gigawatts, that’s $1.5 trillion of CapEx. To the point we just made, you’ve got to use it all in five years because at that point, you’ve got to throw it away and refill it.

Then, if I look at the total commits in the world in this space, in chasing AGI, it seems to be like 100 gigawatts with these announcements. That’s $8 trillion of CapEx.

It’s my view that there’s no way you’re going to get a return on that because $8 trillion of CapEx means you need roughly $800 billion of profit just to pay for the interest.

LLM enthusiasts will easily presume IBM are salty about missing the boat. However, IBM have also been around longer than anybody else. Long enough to know a thing or two about extremely leveraged CapEx.

Short of cheaper-than-solar fusion power plants, or financial engineering miracles—maybe there are great debt restructuring wizards out there, maybe quantitative easers will ease irrationality longer than I can remain solvent or salient—I bet the napkin math will sort itself out into…

The Big Question is…

Who is cultivating the option to snap up and repurpose vapourised datacenter investments at fire sale prices, soon as the "datacenter debt" cometh calling?