Coffee Break: The Abundance Bros Notice the AI Bubble but What Is Their Narrative Leaving Out?

The AI bubble has gotten so big even the Abundance Bros are noticing, but maybe there is more to the AI narrative than meets the eye.

Maybe Sam Altman and company are playing a completely different game than can be understood using conventional economic analysis.

I’ve been writing about AI mania and how it’s driving bad decision making at Meta, Elon Musk’s web of companies, OpenAI, Oracle, Anduril, and Google (and I’m hoping to get around to similar pieces on Amazon, Microsoft, Palantir, and Anthropic).

In each of those pieces I’ve followed Ed Zitron’s skepticism of LLM business models and Gary Marcus’ skepticism of the potential of LLMs to reach AGI (Artificial General Intelligence).

Now it seems that Zitron’s and Marcus’ AI bearishness is hitting the mainstream:

And if something is getting mainstream attention, it’s going to be commented on by the clique of centrist pundits that some call the Abundance Bros and who I think of as keepers of the conventional wisdom.

The Atlantic Monthly’s Charlie Warzel has a piece called “AI is a Mass-Delusion Event” that manages to both split the baby and be quite bearish about LLMs as we know them:

What if generative AI isn’t God in the machine or vaporware? What if it’s just good enough, useful to many without being revolutionary? Right now, the models don’t think—they predict and arrange tokens of language to provide plausible responses to queries. There is little compelling evidence that they will evolve without some kind of quantum research leap. What if they never stop hallucinating and never develop the kind of creative ingenuity that powers actual human intelligence?

The models being good enough doesn’t mean that the industry collapses overnight or that the technology is useless (though it could). The technology may still do an excellent job of making our educational system irrelevant, leaving a generation reliant on getting answers from a chatbot instead of thinking for themselves, without the promised advantage of a sentient bot that invents cancer cures.

Good enough has been keeping me up at night. Because good enough would likely mean that not enough people recognize what’s really being built—and what’s being sacrificed—until it’s too late. What if the real doomer scenario is that we pollute the internet and the planet, reorient our economy and leverage ourselves, outsource big chunks of our minds, realign our geopolitics and culture, and fight endlessly over a technology that never comes close to delivering on its grandest promises? What if we spend so much time waiting and arguing that we fail to marshal our energy toward addressing the problems that exist here and now? That would be a tragedy—the product of a mass delusion. What scares me the most about this scenario is that it’s the only one that doesn’t sound all that insane.

Noah Smith is flat out asking “Will data centers crash the economy? This time let’s think about a financial crisis before it happens.

The U.S. economic data for the last few months is looking decidedly meh. The latest employment numbers were so bad that Trump actually fired the head of the Bureau of Labor Statistics, accusing her of manipulating the numbers to make him look bad. But there’s one huge bright spot amid the gloom: an incredible AI data center building boom.

AI technology is advancing fast, threatening (promising?) to upend many sectors of the economy. Nobody knows yet exactly who will profit from this boom, but one thing that’s certain is that it’s going to take a lot of computing power (or “compute”, as they say).

Whoever provides this compute is going to make a huge amount of revenue. Whether that means they’ll make a lot of profit is another question, but let’s table that for right now; you can’t make profit if you don’t make revenue. So right now, tech companies have the choice to either sit out of the boom entirely, or spend big and hope they can figure out how to make a profit.

Roughly speaking, Apple is choosing the former, while the big software companies — Google, Meta, Microsoft, and Amazon — are choosing the latter. These spending numbers are pretty incredible:

For Microsoft and Meta, this capital expenditure is now more than a third of their total sales.

I think it’s important to look at the telecom boom of the 1990s rather than the one in the 2010s, because the former led to a gigantic crash. The railroad boom led to a gigantic crash too, in 1873 (before the investment peak on Kedrosky’s chart). In both cases, companies built too much infrastructure, outrunning growth in demand for that infrastructure, and suffered a devastating bust as expectations reset and loans couldn’t be paid back.

In both cases, though, the big capex spenders weren’t wrong, they were just early. Eventually, we ended up using all of those railroads and all of those telecom fibers, and much more. This has led a lot of people to speculate that big investment bubbles might actually be beneficial to the economy, since manias leave behind a surplus of cheap infrastructure that can be used to power future technological advances and new business models.

But for anyone who gets caught up in the crash, the future benefits to society are of cold comfort. So a lot of people are worrying that there’s going to be a crash in the AI data center industry, and thus in Big Tech in general, if AI industry revenue doesn’t grow fast enough to keep up with the capex boom over the next few years.

A data center bust would mean that Big Tech shareholders would lose a lot of money, like dot-com shareholders in 2000. It would also slow the economy directly, because Big Tech companies would stop investing. But the scariest possibility is that it would cause a financial crisis.

Financial crises tend to involve bank debt. When a financial bubble and crash is mostly a fall in the value of stocks and bonds, everyone takes losses and then just sort of walks away, a bit poorer — like in 2000. Jorda, Schularick, and Taylor (2015) survey the history of bubbles and crashes, and they find that debt (also called “credit” and “leverage”) is a key predictor of whether a bubble ends up hurting the real economy.

So when I look at this entire landscape, it seems to me that some of the basic conditions of a financial crisis are at least starting to fall into place:

  1. We have a big story about why “this time is different” — the idea that AI will change everything, and that data centers will thus earn huge returns.
  2. We have a large and increasing amount of debt being used to fund one single sector of the economy (data centers), meaning that the loans’ default probability is probably highly correlated.
  3. We have an opaque corner of the financial system (private credit) that has recently grown from a tiny piece of the system to a very significant piece.
  4. We have systemically important lenders (banks, and possibly insurance companies) enmeshed in the new sector in a multitude of ways.

So far, the danger doesn’t scream “2008”. But if you wait until 2008 to start worrying, you’re going to get 2008.

Actual “Abundance” co-author Derek Thompson isn’t quite willing to call it a bubble, but he has written a piece titled “How AI Conquered the US Economy: A Visual FAQ.

Here’s a taste:

The American economy has split in two. There’s a rip-roaring AI economy. And there’s a lackluster consumer economy.

You see it in the economic statistics. Last quarter, spending on artificial intelligence outpaced the growth in consumer spending. Without AI, US economic growth would be meager.

You see it in stocks. In the last two years, about 60 percent of the stock market’s growth has come from AI-related companies, such as Microsoft, Nvidia, and Meta. Without the AI boom, stock market returns would be putrid.

You see it in the business data. According to Stripe, firms that self-describe as “AI companies” are dominating revenue growth on the platform, and they’re far surpassing the growth rate of any other group.

Nobody can say for sure whether the AI boom is evidence of the next Industrial Revolution or the next big bubble. All we know is that it’s happening. We can all stop talking about “what will happen if AI dominates the economy at such-and-such future date?” No, the AI economy is here and now. We’re living in it, for better or worse.

In a follow up piece, Thompson warns of “The Looming Social Crisis of AI Friends and Chatbot Therapists.

AI engineers set out to build god. But god is many things. Long before we build a deity of knowledge, an all-knowing entity that can solve every physical problem through its technical omnipotence, it seems we have built a different kind of god: a singular entity with the power to talk to the whole planet at once.

No matter what AI becomes, this is what AI already is: a globally scaled virtual interlocutor that can offer morsels of life advice wrapped in a mode of flattery that we have good reason to believe may increase narcissism and delusions among young and vulnerable users, respectively. I think this is something worth worrying about, whether you believe AI to be humankind’s greatest achievement or the mother of all pointless infrastructure bubbles. Long before artificial intelligence fulfills its purported promise to become our most important economic technology, we will have to reckon with it as a social technology.

Thompson’s “Abundance” co-author Ezra Klein struggles with similar questions for The New York Times but in a more “aw shucks” manner and throws in the obligatory pop culture reference to show he’s got that common touch:

I don’t know whether A.I. will look, in the economic statistics of the next 10 years, more like the invention of the internet, the invention of electricity or something else entirely. I hope to see A.I. systems driving forward drug discovery and scientific research, but I am not yet certain they will. But I’m taken aback at how quickly we have begun to treat its presence in our lives as normal. I would not have believed in 2020 what GPT-5 would be able to do in 2025. I would not have believed how many people would be using it, nor how attached millions of them would be to it.

But we’re already treating it as borderline banal — and so GPT-5 is just another update to a chatbot that has gone, in a few years, from barely speaking English to being able to intelligibly converse in virtually any imaginable voice about virtually anything a human being might want to talk about at a level that already exceeds that of most human beings.

I find myself thinking a lot about the end of the movie “Her,” in which the A.I.s decide they’re bored of talking to human beings and ascend into a purely digital realm, leaving their onetime masters bereft. It was a neat resolution to the plot, but it dodged the central questions raised by the film — and now in our lives.

What if we come to love and depend on the A.I.s — if we prefer them, in many cases, to our fellow humans — and then they don’t leave?

Thompson and Klein, in their patented speak-slow-enough-for-the-midwits-to-keep-up method, might actually be circling around what I fear is the actual use case for Large Language Models: personalized persuasion at scale.

I have a feeling Peter Thiel and Alex Karp of Palantir, for example, have read this article in Nature:

The potential of generative AI for personalized persuasion at scale

Matching the language or content of a message to the psychological profile of its recipient (known as “personalized persuasion”) is widely considered to be one of the most effective messaging strategies. We demonstrate that the rapid advances in large language models (LLMs), like ChatGPT, could accelerate this influence by making personalized persuasion scalable. Across four studies (consisting of seven sub-studies; total N = 1788), we show that personalized messages crafted by ChatGPT exhibit significantly more influence than non-personalized messages. This was true across different domains of persuasion (e.g., marketing of consumer products, political appeals for climate action), psychological profiles (e.g., personality traits, political ideology, moral foundations), and when only providing the LLM with a single, short prompt naming or describing the targeted psychological dimension. Thus, our findings are among the first to demonstrate the potential for LLMs to automate, and thereby scale, the use of personalized persuasion in ways that enhance its effectiveness and efficiency. We discuss the implications for researchers, practitioners, and the general public.

As a plodding midwit myself, I am tempted to write off the LLM boom as a financial disaster in the making and a technological dead end that will never attain its stated goal of “inventing god”, but reading the study above and watching the video below from Benn Jordan make me fear there is a different game being played by Peter Thiel, Sam Altman, Elon Musk et al.

I’m sure many of you in the Naked Capitalism community are way ahead of me on understanding this (or are smart enough to poke holes in his argument), but for me, Jordan’s insights were revelatory.

Rather than trying to make money, the aspiring AI barons are hoarding power in the form of information.

Some key points from Jordan’s video:

In early 2020 I started becoming increasingly interested in finance. I know, yuck.

This was mostly trying to make sense of the logic behind what the Federal Reserve did in response to Covid…the more and more I learned the less and less it made sense.

So I started deeply analyzing stocks from companies that I was personally and professionally familiar with such as Spotify which led to me starting a private fund that strategically shorted those companies. I even ended up getting hired as a freelance consultant for Venture Capital firms: Me, a musician with no higher education whatsoever.

The more time that I spent looking at and studying the big picture the more that I doubted myself as if the only two options were option A: me being unable to understand what every 25-year-old day trader understands; or option B: I’m right and every hedge fund manager and billionaire is wrong.

But what if there’s a third option? What if capitalism’s death wasn’t a mistake or something that the ultra wealthy were trying to avoid?

What if the pesky burden of labor laws and taxation could be avoided by deprioritizing the goal of financial profit or maybe even money itself?

This video is going to explain or explore what I and an increasing number of economists think is happening or has already happened to capitalism.

More importantly I’m going to try to encapsulate the unsettling socioeconomic reality that we’re living in.

To help properly explain post capitalism I need to remind you of the utter absurdity of just having a billion dollars.

Not only is it impossible to spend on even the most lavish things that you would desire, but it’s also very much not like having a garage full of cash.

For example if Elon Musk with his value of $450 billion (at the time of me recording this) negotiated a deal to buy South Africa with its total gross domestic product valued at $381 billion, he wouldn’t be able to afford it due to the value of his holdings dropping significantly as he liquidated the equity.

If 13% of Tesla’s shares immediately went for public sale it would create an enormous Surplus and the company’s value would crash all net worth does is obfuscate things.

Cash is only powerful when you’re broke and it’s only valuable when you need or want something that you otherwise couldn’t get without cash.

What I instead want my viewers to be concerned with is your personal levels of power or control over the things that you earn consume and trade.

Jordan then explains the private equity take over of the U.S. economy before explaining the implications of this:

The point of all this maneuvering is decreasingly about turning a profit and increasingly about transitioning to a rent-based economy that hedges inflation.

Profit and even money itself is only really valuable when it buys power.

If a societal system can be engineered to gain power without profit or money, capitalism begins to exist as something to exploit within that system under its umbrella.

This sounds hyperbolic but it needs to be said a lot more often. Most of us do not realize that we are in a constant battle with billionaires who don’t want us to own things but rent them.

When we don’t own things we lose control over our own budgets lives decisions and ultimately our own destinies.

Democratic elections don’t work when news and information are no longer democratized. For example, in most dictatorships the government will have an office that permits or denies which stories can run in the newspaper.

In 2025 every piece of news that the average voter sees is manipulated by a third party aggregator. Even if you go to something like Google News and select the generalized headline areas your feed will be different than mine.

This disequilibrating of information was concerning enough when it was being done for ad revenue and profit, but now it’s about a lot more than that whether capitalism is healthy or unhealthy. At any given point point in time (capitalism) has always provided the framework for negotiation between profit and labor.

In the last 2 years AI was being shoved in my face virtually everywhere. As I’m sure you have experienced as well.

My Google searches now open a sidebar telling me about my previous search that I don’t have patience to sort out; or, if I want to make sure that a replacement vacuum hose works with my model of shop vac on Amazon, instead of searching through reviews and answers, I now have to talk to a bot named Rufus about it.

This is happening with emails, with messaging apps, image editors, everything.

I was initially really really confused by this because virtually nobody likes it. It changes familiar workflows of users which risk somebody leaving your ecosystem and it’s really expensive: just one round of training an LLM or large language model can cost over $200 million and that’s not to mention the hugely increased processing power that needs to be done every time that you query that model.

All of this is just happening automatically, just to beg us to use it. Why?

The reason that so many cloud-based companies are dumping so much money and resources into pushing this technology on everyone is to capture the part of our routine that comparison shops or researches or asks for advice or arrives at a logical conclusion about something.

If you become friends with chat GPT or Gemini or Rufus or Siri and you consult with their vast informational resources on a more personal and casual level you’re not only supplying them with the purest form of sentiment analysis but you’re helping them build a system that will further control not only your decisions on what to buy or subscribe to but what hobby you’ll take up next summer.

Not only is it influencing your decision on where to buy an engagement ring or where and when to go on vacation to propose to your partner, there’s nothing stopping it from influencing your decision on whether you should get married at all.

It would be very naive to think of digital assistance as your assistant. They are very much not there to help you but to help their owners.

The conclusion that most of us has is that they’re there to help their owners make more money but in the case of things like the Cambridge Analytica scandal or Twitter quite literally and intentionally becoming an interactive right-wing propaganda tool, there’s a lot more to be worried about than money.

What we ultimately have is money being extracted to systems that do not grow or scale the way a capitalist system does by expanding the assembly line or opening more chain restaurants or researching and patenting a new type of refrigerant.

I am personally convinced that the market will crash again and more money will be borrowed and printed that will ultimately be invested into the further transformation into a rent-based economy.

During World War I America’s debt to GDP ratio was 35%. In World War II, as we strained every single resource that we had, it skyrocketed to 121% meaning that we had 21% more debt than value as a nation.

Our debt to GDP ratio once again skyrocketed after the 2008 crisis and it just kept rising. Right now we are at 123% higher than it has ever been in the history of our economy. Our national debt as of February 2025 is $36.5 trillion and our total debt is $102 trillion.

I would say that we’re well past the point of no return, but I don’t even know what that is because this is so unprecedented.

I assume that it would be easier to change the value and concept of money than it would be to repay that level of debt.

It’s worth pointing out that the primary owner of that debt isn’t China or Japan or even our own Federal Reserve it’s us investors.

As this cycle continues to snowball money will be less and less valuable not only because of inflation but in principle when the government is so closely entangled with the people who have benefited the most from this mass extraction of wealth from a capitalist economy it stops making sense to think of wealth is money.

Unfortunately for oligarchs money can’t buy everything. Money alone can’t change laws. It can’t force people to dress the way you want them to dress or to identify in a way that aligns with how you see them.

It can’t make ideologies against the law. It can’t allow you to change the race culture or ethnicity of your neighbors. You can’t buy someone else’s choice to be or not be a parent.

To graduate, or transcend from those unfathomable limits of what money’s wealth can get you, you need to control the power of information so that you can manipulate the world that you live in.

I believe that that is exactly what’s happening and unfortunately, I think it’s going to get a lot worse before it gets better.

Jordan’s thesis has explanatory value beyond “whoa, these billionaires are a bunch of dumbasses driving their industry and our economy off a cliff.”

Maybe AI is more of a power grab than an economic play, and maybe it’s even a deliberate attempt to crash the economy ala 2008 and accelerate our transition further away from a capitalist economy as we understand it and toward the ultimate rentier’s paradise.

Or, for the rest of us, the perfect dystopia.

Print Friendly, PDF & Email

5 comments

  1. Judith

    The “personalized persuasion” could well be used to create quite an Orwellian society. That goes well beyond the financial environment into attempts to control people’s thoughts and behavior.

    Reply
    1. Nat Wilson Turner Post author

      That I’m afraid is the real play and why they’re so desperate to push this despite all the warning bells.

      Reply
  2. lyman alpha blob

    RE: the Noah Smith excerpt –

    “… the big capex spenders weren’t wrong, they were just early. Eventually, we ended up using all of those railroads…”

    I’d recommend Smith read a book I’ve reference a few times over the years – Railroaded: The Transcontinentals and the Making of Modern America. The author wrote the book from SF and Seattle and he explicitly mentions in the introduction that comparisons to the modern tech industry are intentional. He points out that railroad barons grossly overbuilt tracks, sometimes with one company’s right next to another’s, due to the need to compete and put each other out of business rather than cooperating with each other. These tracks were not all eventually used over the long term, and during the construction phase there was a ton of corruption, lots of wasted resources, buffalo were deliberately driven to near extinction, and one could argue that the entire concept of racism was constructed during this period as a deliberate attempt by railroad barons to pit one group against another to keep wages down.

    But, railroads did allow people and freight to get from one place to another. “AI” on the other hand seems much more like a solution in search of a problem. I’m perfectly capable of summarizing my own emails, thank you very much. I could be wrong here and freely admit that I failed to see the long term utility of email back in the day, but the AI bubble seems a lot more like Tulipmania.

    As for Jordan’s concerns about debt to GDP ratio, I’d say they are a little overwrought. As the NC commentariat is aware, debts in a sovereign currency can always be repaid by issuing more money. Also, I’m not convinced that “AI” will ever be used for anything more than a few legitimate niche uses, and I suspect the rest of the population will give it up once the novelty wears off. with a few diehards clinging to it like they do their google glass or VR headset.

    That being said, these techbro clowns clearly have many delusions of grandeur and many of them would love to create a panopticon under their control. Here’s hoping they tear each other to pieces trying to be The One.

    Reply
    1. Nat Wilson Turner Post author

      There’s a lot more to Jordan’s argument than debt to GDP ratio, that was just an easy to excerpt point.

      The idea that the techbros are knowingly playing a different game than profit/loss AND that LLMs aren’t really about what we’re being told they’re about rings true to me.

      His explanation includes the bubble as a deliberate accelerationist play. Makes sense to me.

      Reply

Leave a Reply

Your email address will not be published. Required fields are marked *