Latest Posts (17 found)

The AI Bubble's Impossible Promises

Readers: I’ve done a very generous “free” portion of this newsletter, but I do recommend paying for premium to get the in-depth analysis underpinning the intro. That being said, I want as many people as possible to get the general feel for this piece. Things are insane, and it’s time to be realistic about what the future actually looks like. We’re in a bubble. Everybody says we’re in a bubble. You can’t say we’re not in a bubble anymore without sounding insane, because everybody is now talking about how OpenAI has promised everybody $1 trillion — something you could have read about two weeks ago on my premium newsletter . Yet we live in a chaotic, insane world, where we can watch the news and hear hand-wringing over the fact that we’re in a bubble , read article after CEO after article after CEO after analyst after investor saying we’re in a bubble, yet the market continues to rip ever-upward on increasingly more-insane ideas , in part thanks to analysts that continue to ignore the very signs that they’re relied upon to read . AMD and OpenAI signed a very strange deal where AMD will give OpenAI the chance to buy 160 million shares at a cent a piece, in tranches of indeterminate size, for every gigawatt of data centers OpenAI builds using AMD’s chips, adding that OpenAI has agreed to buy “six gigawatts of GPUs.” This is a peculiar way to measure GPUs, which are traditionally measured in the price of each GPU , but nevertheless, these chips are going to be a mixture of AMD’s mi450 instinct GPUs — which we don’t know the specs of! — and its current generation mi350 GPUs , making the actual scale of these purchases a little difficult to grasp, though the Wall Street Journal says it would “result in tens of billions of dollars in new revenue” for AMD . This AMD deal is weird , but one that’s rigged in favour of Lisa Su and AMD. OpenAI doesn’t get a dollar at any point - it has work out how to buy those GPUs and figure out how to build six further gigawatts of data centers on top of the 10GW of data centers it promised to build for NVIDIA and the seven-to-ten gigawatts that are allegedly being built for Stargate , bringing it to a total of somewhere between 23 and 26 gigawatts of data center capacity. Hell, while we’re on the subject, has anyone thought about how difficult and expensive it is to build a data center?  Everybody is very casual with how they talk about Sam Altman’s theoretical promises of trillions of dollars of data center infrastructure , and I'm not sure anybody realizes how difficult even the very basics of this plan will be. Nevertheless, everybody is happily publishing stories about how Stargate Abilene Texas — OpenAI’s massive data center with Oracle — is “open, ” by which they mean two buildings, and I’m not even confident both of them are providing compute to OpenAI yet. There are six more of them that need to get built for this thing to start rocking at 1.2GW — even though it’s only 1.1GW according to my sources in Abilene. But, hey, sorry — one minute — while we’re on that subject, did anybody visiting Abilene in the last week or so ever ask whether they’ll have enough power there?  Don’t worry, you don’t need to look. I’m sure you were just about to, but I did the hard work for you and read up on it, and it turns out that Stargate Abilene only has 200MW of power — a 200MW substation that, according to my sources, has only been built within the last couple of months, with 350MWs of gas turbine generators that connect to a natural gas power plant that might get built by the end of the year . Said turbine is extremely expensive, featuring volatile pricing ( for context, natural gas price volatility fell in Q2 2025…to 69% annualized ) and even more volatile environmental consequences , and is, while permitted for it ( this will download the PDF of the permit ), impractical and expensive to use long-term.  Analyst James van Geelen, founder of Citrini Research recently said on Bloomberg’s Odd Lots podcast that these are “not the really good natural gas turbines” because the really good ones would take seven years to deliver due to a natural gas turbine shortage . But they’re going to have to do. According to sources in Abilene, developer Lancium has only recently broken ground on the 1GW substation and five transformers OpenAI’s going to need to build out there , and based on my conversations with numerous analysts and researchers, it does not appear that Stargate Abilene will have sufficient power before the year 2027.  Then there’s the question of whether 1GW of power actually gets you 1GW of compute. This is something you never see addressed in the coverage of OpenAI’s various construction commitments, but it’s an important point to make. Analyst Daniel Bizo, Research Director at the Uptime Institute, explained that 1 gigawatt of power is only sufficient to power (roughly) 700 megawatts of data center capacity . We’ll get into the finer details of that later in this newsletter, but if we assume that ratio is accurate, we’re left with a troubling problem. That figure represents a 1.43 PUE — Power Usage Effectiveness — and if we apply that to Stargate Abilene, we see that it needs at least 1.7GW of power, and currently only has 200MW. Stargate Abilene does not have sufficient power to run at even half of its supposed IT load of 1.2GW, and at its present capacity — assuming that the gas turbines function at full power — can only hope to run 370MW to 460MW of IT load. I’ve seen article after article about the gas turbines and their use of fracked gas — a disgusting and wasteful act typical of OpenAI — but nobody appears to have asked “how much power does a 1.2GW data center require?” and then chased it with “how much power does Stargate Abilene have?” The answer is not enough, and the significance of said “not enough” is remarkable. Today, I’m going to tell you, at length, how impossible the future of generative AI is.  Gigawatt data centers are a ridiculous pipe dream, one that runs face-first into the walls of reality.   The world’s governments and media have been far too cavalier with the term “gigawatt,” casually breezing by the fact that Altman’s plans require 17 or more nuclear reactors’ worth of power , as if building power is quick and easy and cheap and just happens. I believe that many of you think that this is an issue of permitting — of simply throwing enough money at the problem — when we are in the midst of a shortage in the electrical grade steel and transformers required to expand America’s (and the world’s) power grid. I realize it’s easy to get blinded by the constant drumbeat of “ gargoyle-like tycoon cabal builds 1GW  data center ” and feel that they will simply overwhelm the problem with money, but no, I’m afraid that isn’t the case at all, and all of this is so silly, so ridiculous, so cartoonishly bad that it threatens even the seemingly-infinite wealth of Elon Musk, with xAI burning over a billion dollars a month and planning to spend tens of billions of dollars building the Colossus 2 data center , dragging two billion dollars from SpaceX in his desperate quest to burn as much money as possible for no reason.  This is the age of hubris — a time in which we are going to watch stupid, powerful and rich men fuck up their legacies by finding a technology so vulgar in its costs and mythical outcomes that it drives the avaricious insane and makes fools of them.  Or perhaps this is what happens when somebody believes they’ve found the ultimate con — the ability to become both the customer and the business, which is exactly what NVIDIA is doing to fund the chips behind Colossus 2. According to Bloomberg , NVIDIA is creating a company — a “special purpose vehicle” — that it will invest $2 billion in, along with several other backers. Once that’s done, the special purpose vehicle will then use that equity to raise debt from banks, buy GPUs from NVIDIA, and then rent those GPUs to Elon Musk for five years. Hell, why make it so complex? NVIDIA invested money in a company specifically built to buy chips from it, which then promptly handed the money back to NVIDIA along with a bunch of other money, and then whatever happened next is somebody else’s problem. Actually, wait — how long do GPUs last, exactly? Four years for training ? Three years ? The A100 GPU started shipping in May 2020 , and the H100 (and the Hopper GPU generation) entered full production in September 2022 , meaning that we’re hurtling at speed toward the time in which we’re going to start seeing a remarkable amount of chips start wearing down, which should be a concern for companies like Microsoft, which bought 150,000 Hopper GPUs in 2023 and 485,000 of them in 2024 . Alright, let me just be blunt: the entire economy of debt around GPUs is insane. Assuming these things don’t die within five years ( their warranties generally end in three ), their value absolutely will, as NVIDIA has committed to releasing a new AI chip every single year , likely with significant increases to power and power efficiency. At the end of the five year period, the Special Purpose Vehicle will be the proud owner of five-year-old chips that nobody is going to want to rent at the price that Elon Musk has been paying for the last five years. Don’t believe me? Take a look at the rental prices for H100 GPUs that went from $8-an-hour in 2023 to $2-an-hour in 2024 , or the Silicon Data Indexes (aggregated realtime indexes of hourly prices) that show H100 rentals at around $2.14-an-hour and A100 rentals at a dollar-an-hour , with Vast.AI offering them at as little as $0.67 an hour . This is, by the way, a problem that faces literally every data center being built in the world , and I feel insane talking about it. It feels like nobody is talking about how impossible and ridiculous all of this is. It’s one thing that OpenAI has promised one trillion dollars to people — it’s another that large swaths of that will be spent on hardware that will, by the end of these agreements, be half-obsolete and generating less revenue than ever. Think about it. Let’s assume we live in a fantasy land where OpenAI is somehow able to pay Oracle $300 billion over 5 years — which, although the costs will almost certainly grow over time, and some of the payments are front-loaded, averages out to $5bn each month, which is a truly insane number that’s in excess of what Netflix makes in revenue.  Said money is paying for access to Blackwell GPUs, which will, by then, be at least two generations behind, with NVIDIA’s Vera Rubin GPUs due next year . What happens to that GPU infrastructure? Why would OpenAI continue to pay the same rental rate for five-year-old Blackwell GPUs?   All of these ludicrous investments are going into building data centers full of what will, at that point, be old tech.  Let me put it in simple terms: imagine you, for some reason, rented an M1 Mac when it was released in 2020 , and your rental was done in 2025, when we’re onto the M4 series . Would you expect somebody to rent it at the same price? Or would they say “hey, wait a minute, for that price I could rent one of the newer generation ones.” And you’d be bloody right!  Now, I realize that $70,000 data center GPUs are a little different to laptops, but that only makes their decline in value more profound, especially considering the billions of dollars of infrastructure built around them.  And that’s the problem. Private equity firms are sinking $50 billion or more a quarter into theoretical data center projects full of what will be years-old GPU technology, despite the fact that there’s no real demand for generative AI compute , and that’s before you get to the grimmest fact of all: that even if you can build these data centers, it will take years and billions of dollars to deliver the power, if it’s even possible to do so. Harvard economist Jason Furman estimates that data centers and software accounted for 92% of GDP growth in the first half of this year , in line with my conversation with economist Paul Kedrosky from a few months ago .  All of this money is being sunk into infrastructure for an “AI revolution” that doesn’t exist, as every single AI company is unprofitable, with pathetic revenues ( $61 billion or so if you include CoreWeave and Lambda, both of which are being handed money by NVIDIA ), impossible-to-control costs that have only ever increased , and no ability to replace labor at scale ( and especially not software engineers ).   OpenAI needs more than a trillion dollars to pay its massive cloud compute bills and build 27 gigawatts of data centers, and to get there, it needs to start making incredible amounts of money, a job that’s been mostly handed to Fidji Simo, OpenAI’s new CEO of Applications , who is solely responsible for turning a company that loses billions of dollars into one that makes $200 billion in 2030 with $38 billion in profit . She’s been set up to fail, and I’m going to explain why. In fact, today I’m going to explain to you how impossible all of this is — not just expensive , not just silly , but actively impossible within any of the timelines set .  Stargate will not have the power it needs before the middle of 2026 — the beginning of Oracle’s fiscal year 2027, when OpenAI has to pay it $30 billion for compute — or, according to The Information , choose to walk away if the capacity isn’t complete. And based on my research, analysis and discussions with power and data center analysts, gigawatt data centers are, by and large, a pipedream, with their associated power infrastructure taking two to four years, and that’s if everything goes smoothly. OpenAI cannot build a gigawatt of data centers for AMD by the “second half of 2026.”   It haven’t even announced the financing, let alone where the data center might be, and until it does that it’s impossible to plan the power, which in and of itself takes months before you even start building.   Every promise you’re reading in the news is impossible. Nobody has even built a gigawatt data center, and more than likely nobody ever will . Stargate Abilene isn’t going to be ready in 2026, won’t have sufficient power until at best 2027, and based on the conversations I’ve had it’s very unlikely it will build that gigawatt substation before the year 2028.   In fact, let me put it a little simpler: all of those data center deals you’ve seen announced are basically bullshit. Even if they get the permits and the money, there are massive physical challenges that cannot be resolved by simply throwing money at them.  Today I’m going to tell you a story of chaos, hubris and fantastical thinking. I want you to come away from this with a full picture of how ridiculous the promises are, and that’s before you get to the cold hard reality that AI fucking sucks.

2 views

OpenAI Is Just Another Boring, Desperate AI Startup

What is OpenAI? I realize you might say "a foundation model lab" or "the company that runs ChatGPT," but that doesn't really give the full picture of everything it’s promised, or claimed, or leaked that it was or would be. No, really, if you believe its leaks to the press... To be clear, many of these are ideas that OpenAI has leaked specifically so the media can continue to pump up its valuation and continue to raise the money it needs — at least $1 Trillion over the next four or five years, and I don't believe the theoretical (or actual) costs of many of the things I've listed are included. OpenAI wants you to believe it is everything , because in reality it’s a company bereft of strategy, focus or vision. The GPT-5 upgrade for ChatGPT was a dud — an industry-wide embarrassment for arguably the most-hyped product in AI history, one that ( as I revealed a few months ago ) costs more to operate than its predecessor, not because of any inherent capability upgrade, but how it actually processes the prompts its user provides — and now it's unclear what it is that this company does.   Does it make hardware? Software? Ads? Is it going to lease you GPUs to use for your own AI projects? Is it going to certify you as an AI expert ? Notice how I've listed a whole bunch of stuff that isn't ChatGPT, which will, if you look at The Information's reporting of its projections, remain the vast majority of its revenue until 2027, at which point "agents" and "new products including free user monetization" will magically kick in. In reality, OpenAI is an extremely boring (and bad!) software business. It makes the majority of its revenue selling subscriptions to ChatGPT, and apparently had 20 million paid subscribers (as of April) and 5 million business subscribers (as of August, though 500,000 of them are Cal State University seats paid at $2.50 a month ). It also loses incredibly large amounts of money. Yes, I realize that OpenAI also sells access to its API, but as you can see from the chart above, it is making a teeny tiny sliver of revenue from it in 2025, though I will also add that this chart has a little bit of green for "agent" revenue, which means it's very likely bullshit. Operator, OpenAI's so-called agent, is barely functional , and I have no idea how anyone would even begin to charge money for it outside of "please try my broken product." In any case, API sales appear to be a very, very small part of OpenAI's revenue stream, and that heavily suggests a lack of interest in integrating its models at scale. Worse still, this effectively turns OpenAI into an AI startup. Think about it: if OpenAI can't make the majority of its money through "innovating" in the development of large language models (LLMs), then it’s just another company plugging LLMs into its software. While ChatGPT may be a very popular product, it is, by definition (and in its name!) a GPT wrapper, with the few differences being that OpenAI pays its own immediate costs, has the people necessary to continue improving its own models, and also continually makes promises to convince people it’s anything other than just another AI startup. In fact, the only real difference is the amount of money backing it. Otherwise, OpenAI could be literally any foundation model company, and with a lack of real innovation within those models, it’s just another startup trying to find ways to monetize generative AI, an industry that only ever seems to lose money . As a result, we should start evaluating OpenAI as just another AI startup, as its promises do not appear to mesh with any coherent strategy, other than " we need $1 trillion dollars ." There does not seem to be much of a plan on a day-to-day basis, nor does there seem to be one about what OpenAI should be, other than that OpenAI will be a consumer hardware, consumer software, enterprise SaaS and data center operator, as well as running a social network. As I've discussed many times , LLMs are inherently flawed due to their probabilistic nature."Hallucinations" — when a model authoritatively states something is true when it isn't (or takes an action that seems the most likely course of action, even if it isn't the right one) — are a " mathematically inevitable " according to OpenAI's own research feature of the technology, meaning that there is no fixing their most glaring, obvious problem , even with "perfect data." I'd wager the reason OpenAI is so eager to build out so much capacity while leaking so many diverse business lines is an attempt to get away from a dark truth: that when you peel away the hype, ChatGPT is a wrapper, every product it makes is a wrapper, and OpenAI is pretty fucking terrible at making products. Today I'm going to walk you through a fairly unique position: that OpenAI is just another boring AI startup lacking any meaningful product roadmap or strategy, using the press as a tool to pump its bags while very rarely delivering on what it’s promised. It is a company with massive amounts of cash, industrial backing, and brand recognition, and otherwise is, much like its customers, desperately trying to work out how to make money selling products built on top of Large Language Models. OpenAI lives and dies on its mythology as the center of innovation in the world of AI, yet reality is so much more mediocre. Its revenue growth is slowing, its products are commoditized, its models are hardly state-of-the-art, the overall generative AI industry has lost its sheen , and its killer app is a mythology that has converted a handful of very rich people and very few others. OpenAI spent, according to The Information , 150% ($6.7 billion in costs) of its H1 2025 revenue ($4.3 billion) on research and development, producing the deeply-underwhelming GPT-5 and Sora 2, an app that I estimate costs it upwards of $5 for each video generation, based on Azure's published rates for the first Sora model , though it's my belief that these rates are unprofitable, all so that it can gain a few more users. To be clear, R&D is good, and useful, and in my experience, the companies that spend deeply on this tend to be the ones that do well. The reason why Huawei has managed to outpace its American rivals in several key areas — like automotive technology and telecommunications — is because it spends around a quarter of its revenue on developing new technologies and entering new markets, rather than stock buybacks and dividends. The difference is that said R&D spending is both sustainable and useful, and has led to Huawei becoming much a stronger business, even as it languishes on a Treasury Department entity list that effectively cuts it off from US-made or US-origin parts or IP . Considering that OpenAI’s R&D spending was 38.28% of its cash-on-hand by the end of the period (totalling $17.5bn, which we’ll get to later), and what we’ve seen as a result, it’s hard to describe it as either sustainable or useful.     OpenAI isn't innovative, it’s exploitative, a giant multi-billion dollar grift attempting to hide how deeply unexciting it is, and how nonsensical it is to continue backing it . Sam Altman is an excellent operator, capable of spreading his mediocre, half-baked mantras about how 2025 was the year AI got smarter than us , or how we'll be building 1GW data centers each week (something that, by my estimations, takes 2.5 years), taking advantage of how many people in the media, markets and global governments don't know a fucking thing about anything. OpenAI is  also getting desperate. Beneath the surface of the media hype and trillion-dollar promises is a company struggling to maintain relevance, its entire existence built on top of hype and mythology. And at this rate, I believe it’s going to miss its 2025 revenue projections, all while burning billions more than anyone has anticipated. OpenAI is a social media company, this week launching Sora 2, a social feed entirely made up of generative video . OpenAI is a workplace productivity company, allegedly working on its own productivity suite to compete with Microsoft . OpenAI is a jobs portal, announcing in September it was "developing an AI-powered hiring platform ," which it will launch 'by mid-2026. OpenAI is an ads company, and is apparently trying to hire an an ads chief , with the (alleged) intent to start showing ads in ChatGPT "by 2026." OpenAI is a company that would sell AI compute like Microsoft Azure or Amazon Web Services, or at least is considering being one, with CFO Sarah Friar telling Bloomberg in August that it is not "actively looking" at such an effort today but will "think about it as a business down the line, for sure." OpenAI is a fabless semiconductor design company, launching its own AI chips in, again, 2026 with Broadcom , but only for internal use. OpenAI is a consumer hardware company, preparing to launch a device by the end of 2026 or early 2027 and hiring a bunch of Apple people to work on it , as well as considering — again, it’s just leaking random stuff at this point to pump up its value — a smart speaker, a voice recorder and AR glasses. OpenAI is also working on its own browser , I guess.

0 views

The Case Against Generative AI

Soundtrack: Queens of the Stone Age - First It Giveth Before we go any further: This is, for the third time this year, the longest newsletter I've ever written, weighing in somewhere around 18,500 words. I've written it specifically to be read at your leisure — dip in and out where you'd like — but also in one go.  This is my comprehensive case that yes, we’re in a bubble, one that will inevitably (and violently) collapse in the near future. I'll also be cutting this into a four-part episode starting tomorrow on my podcast Better Offline . I deeply appreciate your time. If you like this newsletter, please think about subscribing to the premium, which I write weekly. Thanks for reading. Alright, let’s do this one last time . In 2022, a (kind-of) company called OpenAI surprised the world with a website called ChatGPT that could generate text that sort-of sounded like a person using a technology called Large Language Models (LLMs), which can also be used to generate images, video and computer code.  Large Language Models require entire clusters of servers connected with high-speed networking, all containing this thing called a GPU — graphics processing units. These are different to the GPUs in your Xbox, or laptop, or gaming PC. They cost much, much more, and they’re good at doing the processes of inference (the creation of the output of any LLM) and training (feeding masses of training data to models, or feeding them information about what a good output might look like, so they can later identify a thing or replicate it). These models showed some immediate promise in their ability to articulate concepts or generate video, visuals, audio, text and code. They also immediately had one glaring, obvious problem: because they’re probabilistic, these models can’t actually be relied upon to do the same thing every single time. So, if you generated a picture of a person that you wanted to, for example, use in a story book, every time you created a new page, using the same prompt to describe the protagonist, that person would look different — and that difference could be minor (something that a reader should shrug off), or it could make that character look like a completely different person. Moreover, the probabilistic nature of generative AI meant that whenever you asked it a question, it would guess as to the answer, not because it knew the answer, but rather because it was guessing on the right word to add in a sentence based on previous training data. As a result, these models would frequently make mistakes — something which we later referred to as “hallucinations.”  And that’s not even mentioning the cost of training these models, the cost of running them, the vast amounts of computational power they required, the fact that the legality of using material scraped from books and the web without the owner’s permission was (and remains) legally dubious, or the fact that nobody seemed to know how to use these models to actually create profitable businesses.  These problems were overshadowed by something flashy, and new, and something that investors — and the tech media — believed would eventually automate the single thing that’s proven most resistant to automation: namely, knowledge work and the creative economy.  This newness and hype and these expectations sent the market into a frenzy, with every hyperscaler immediately creating the most aggressive market for one supplier I’ve ever seen. NVIDIA has sold over $200 billion of GPUs since the beginning of 2023 , becoming the largest company on the stock market and trading at over $170 as of writing this sentence only a few years after being worth $19.52 a share . While I’ve talked about some of the propelling factors behind the AI wave — automation and novelty — that’s not a complete picture. A huge reason why everybody decided to “do AI” was because the software industry’s growth was slowing , with SaaS (Software As A Service) company valuations stalling or dropping , resulting in  the terrifying prospect of companies having to “ under promise and over deliver ” and “be efficient.” Things that normal companies — those whose valuations aren’t contingent on ever-increasing, ever-constant growth — don’t have to worry about, because they’re normal companies.  Suddenly, there was the promise of a new technology — Large Language Models — that were getting exponentially more powerful, which was mostly a lie but hard to disprove because “powerful” can mean basically anything, and the definition of “powerful” depended entirely on whoever you asked at any given time, and what that person’s motivations were.  The media also immediately started tripping on its own feet, mistakenly claiming OpenAI’s GPT-4 model tricked a Taskrabbit into solving a CAPTCHA ( it didn’t — this never happened), or saying that “ people who don’t know how to code already [used] bots to produce full-fledged games, ” and if you’re wondering what “full-fledged” means, it means “pong” and a cobbled-together rolling demo of SkyRoads, a game from 1993 . The media (and investors) helped peddle the narrative that AI was always getting better, could do basically anything, and that any problems you saw today would be inevitably solved in a few short months, or years, or, well, at some point I guess.  LLMs were touted as a digital panacea, and the companies building them offered traditional software companies the chance to plug these models into their software using an API, thus allowing them to ride the same generative AI wave that every other company was riding.  The model companies similarly started going after individual and business customers, offering software and subscriptions that promised the world, though this mostly boiled down to chatbots that could generate stuff, and then doubled down with the promise of “agents” — a marketing term that’s meant to make you think “autonomous digital worker” but really means “ broken digital product .” Throughout this era, investors and the media spoke with a sense of inevitability that they never really backed up with data. It was an era based on confidently-asserted “vibes.” Everything was always getting better and more powerful, even though there was never much proof that this was truly disruptive technology, other than in its ability to disrupt apps you were using with AI — making them worse by, for example, suggesting questions on every Facebook post that you could ask Meta AI, but which Meta AI couldn’t answer. “AI” was omnipresent, and it eventually grew to mean everything and nothing. OpenAI would see its every move lorded over like a gifted child, its CEO Sam Altman called the “ Oppenheimer of Our Age ,” even if it wasn’t really obvious why everyone was impressed. GPT-4 felt like something a bit different, but was it actually meaningful?  The thing is, Artificial Intelligence is built and sold on not just faith, but a series of myths that the AI boosters expect us to believe with the same certainty that we treat things like gravity, or the boiling point of water.  Can large language models actually replace coders? Not really, no, and I’ll get into why later in the piece. Can Sora — OpenAI’s video creation tool — replace actors or animators? No, not at all, but it still fills the air full of tension because you can immediately see who is pre-registered to replace everyone that works for them.  AI is apparently replacing workers, but nobody appears to be able to prove it! But every few weeks a story runs where everybody tries to pretend that AI is replacing workers with some poorly-sourced and incomprehensible study , never actually saying “someone’s job got replaced by AI” because it isn’t happening at scale, and because if you provide real-world examples, people can actually check. To be clear, some people have lost jobs to AI, just not the white collar workers, software engineers, or really any of the career paths that the mainstream media and AI investors would have you believe.  Brian Merchant has done excellent work covering how LLMs have devoured the work of translators , using cheap, “almost good” automation to lower already-stagnant wages in a field that was already hurting before the advent of generative AI, with some having to abandon the field, and others pushed into bankruptcy. I’ve heard the same for art directors, SEO experts, and copy editors, and Christopher Mims of the Wall Street Journal covered these last year .  These are all fields with something in common: shitty bosses with little regard for their customers who have been eagerly waiting for the opportunity to slash contract labor. To quote Merchant, “the drumbeat, marketing, and pop culture of ‘powerful AI’ encourages and permits management to replace or degrade jobs they might not otherwise have.”  Across the board, the people being “replaced” by AI are the victims of lazy, incompetent cost-cutters who don’t care if they ship poorly-translated text. To quote Merchant again, “[AI hype] has created the cover necessary to justify slashing rates and accepting “good enough” automation output for video games and media products.” Yet the jobs crisis facing translators speaks to the larger flaws of the Large Language Model era, and why other careers aren’t seeing this kind of disruption. Generative AI creates outputs , and by extension defines all labor as some kind of output created from a request. In the case of translation, it’s possible for a company to get by with a shitty version, because many customers see translation as “what do these words say,” even though ( as one worker told Merchant ) translation is about conveying meaning. Nevertheless, “translation” work had already started to condense to a world where humans would at times clean up machine-generated text, and the same worker warned that the same might come for other industries. Yet the problem is that translation is a heavily output-driven industry, one where (idiot) bosses can say “oh yeah that’s fine” because they ran an output back through Google Translate and it seemed fine in their native tongue. The problems of a poor translation are obvious, but the customers of translation are, it seems, often capable of getting by with a shitty product. The problem is that most jobs are not output-driven at all, and what we’re buying from a human being is a person’s ability to think.   Every CEO talking about AI replacing workers is an example of the real problem: that most companies are run by people who don’t understand or experience the problems they’re solving, don’t do any real work, don’t face any real problems, and thus can never be trusted to solve them. The Era of the Business Idiot is the result of letting management consultants and neoliberal “free market” sociopaths take over everything, leaving us with companies run by people who don’t know how the companies make money, just that they must always make more. When you’re a big, stupid asshole, every job that you see is condensed to its outputs, and not the stuff that leads up to the output, or the small nuances and conscious decisions that make an output good as opposed to simply acceptable, or even bad.  What does a software engineer do? They write code! What does a writer do? They write words! What does a hairdresser do? They cut hair!  Yet that’s not actually the case.  As I’ll get into later, a software engineer does far more than just code, and when they write code they’re not just saying “what would solve this problem?” with a big smile on their face — they’re taking into account their years of experience, what code does, what code could do , all the things that might break as a result, and all of the things that you can’t really tell from just looking at code , like whether there’s a reason things are made in a particular way. A good coder doesn’t just hammer at the keyboard with the aim of doing a particular task. They factor in questions like: How does this functionality fit into the code that’s already here? Or, if someone has to update this code in the future, how do I make it easy for them to understand what I’ve written and to make changes without breaking a bunch of other stuff? A writer doesn’t just “write words.” They jostle ideas and ideals and emotions and thoughts and facts and feelings into a condensed piece of text, explaining both what’s happening and why it’s happening from their perspective, finding nuanced ways to convey large topics, none of which is the result of a single (or many) prompts but the ever-shifting sand of a writer’s brain.  Good writing is a fight between a bunch of different factors: structure, style, intent, audience, and prioritizing the things that you (or your client) care about in the text. It’s often emotive — or at the very least, driven or inspired by a given emotion — which is something that an AI simply can’t replicate in a way that’s authentic and believable.  And a hairdresser doesn’t just cut hair, but cuts your hair, which may be wiry, dry, oily, long, short, healthy, unhealthy, on a scalp with particular issues, at a time of year when perhaps you want to change length, at a time that fits you, in “the way you like” which may be impossible to actually write down but they get it just right. And they make conversation, making you feel at ease while they snip and clip away at your tresses, with you having to trust that they’ll get it right.  This is the true nature of labor that executives fail to comprehend at scale: that the things we do are not units of work, but extrapolations of experience, emotion, and context that cannot be condensed in written meaning. Business Idiots see our labor as the result of a smart manager saying “do this,” rather than human ingenuity interpreting both a request and the shit the manager didn’t say. What does a CEO do? Uhhh, um, well, a Harvard study says they spend 25% of their time on “people and relationships,” 25% on “functional and business unit reviews,” 16% on “organization and culture,” and 21% on “strategy,” with a few percent here and there for things like “professional development.”  That’s who runs the vast majority of companies: people that describe their work predominantly as “looking at stuff,” “talking to people” and “thinking about what we do next.” The most highly-paid jobs in the world are impossible to describe, their labor described in a mish-mash of LinkedInspiraton, yet everybody else’s labor is an output that can be automated. As a result, Large Language Models seem like magic. When you see everything as an outcome — an outcome you may or may not understand, and definitely don’t understand the process behind, let alone care about — you kind of already see your workers as LLMs.   You create a stratification of the workforce that goes beyond the normal organizational chart, with senior executives — those closer to the class level of CEO — acting as those who have risen above the doldrums of doing things to the level of “decisionmaking,” a fuzzy term that can mean everything from “making nuanced decisions with input from multiple different subject-matter experts” to, as ServiceNow Bill McDermott did in 2022 , “[make] it clear to everybody [in a boardroom of other executives], everything you do: AI, AI, AI, AI, AI.”  The same extends to some members of the business and tech media that have, for the most part, gotten by without having to think too hard about the actual things the companies are saying.  I realize this sounds a little mean, and I must be clear it doesn’t mean that these people know nothing , just that it’s been possible to scoot through the world without thinking too hard about whether or not something is true. When Salesforce said back in 2024 that its “Einstein Trust Layer” and AI would be “transformational for jobs,” the media dutifully wrote it down and published it without a second thought. It fully trusted Marc Benioff when he said that Agentforce agents would replace human workers , and then again when he said that AI agents are doing “ 30% to 50% of all the work in Salesforce itself ,” even though that’s an unproven and nakedly ridiculous statement.  Salesforce’s CFO said earlier this year that AI wouldn’t boost sales growth in 2025 . One would think this would change how they’re covered, or how seriously one takes Marc Benioff.  It hasn’t, because nobody is paying attention. In fact, nobody seems to be doing their job. This is how the core myths of generative AI were built: by executives saying stuff and the media publishing it without thinking too hard.  AI is replacing workers! AI is writing entire computer programs! AI is getting exponentially more-powerful! What does “powerful” mean? That the models are getting better on benchmarks that are rigged in their favor, but because nobody fucking explains it , regular people are regularly told that AI is “powerful.”  The only thing “powerful” about generative AI is its mythology. The world’s executives, entirely disconnected from labor and actual production, are doing the only thing they know how to — spend a bunch of money and say vague stuff about “AI being the future.” There are people — journalists, investors, and analysts — that have built entire careers on filling in the gaps for the powerful as they splurge billions of dollars and repeat with increasing desperation that “the future is here” as absolutely nothing happens. You’ve likely seen a few ridiculous headlines recently. One of the most recent, and most absurd, is that that OpenAI will pay Oracle $300 billion over four years , closely followed with the claim that NVIDIA will “invest” “$100 billion” in OpenAI to build 10GW of AI data centers , though the deal is structured in a way that means that OpenAI is paid “progressively as each gigawatt is deployed,” and OpenAI will be leasing the chips (rather than buying them outright) . I must be clear that these deals are intentionally made to continue the myth of generative AI, to pump NVIDIA, and to make sure OpenAI insiders can sell $10.3 billion of shares .   OpenAI cannot afford the $300 billion, NVIDIA hasn’t sent OpenAI a cent and won’t do so if it can’t build the data centers, which OpenAI most assuredly can’t afford to do.  NVIDIA needs this myth to continue, because in truth, all of these data centers are being built for demand that doesn’t exist, or that — if it exists — doesn’t necessarily translate into business customers paying huge amounts for access to OpenAI’s generative AI services.  NVIDIA, OpenAI, CoreWeave and other AI-related companies hope that by announcing theoretical billions of dollars (or hundreds of billions of dollars) of these strange, vague and impossible-seeming deals, they can keep pretending that demand is there, because why else would they build all of these data centers, right?   That, and the entire stock market rests on NVIDIA’s back . It accounts for 7% to 8% of the value of the S&P 500, and Jensen Huang needs to keep selling GPUs. I intend to explain later on how all of this works, and how brittle it really is. The intention of these deals is simple: to make you think “this much money can’t be wrong.” It can. These people need you to believe this is inevitable, but they are being proven wrong, again and again, and today I’m going to continue doing so.  Underpinning these stories about huge amounts of money and endless opportunity lies a dark secret — that none of this is working, and all of this money has been invested in a technology that doesn’t make much revenue and loves to burn millions or billions or hundreds of billions of dollars. Over half a trillion dollars has gone into an entire industry without a single profitable company developing models or products built on top of models. By my estimates, there is around $44 billion of revenue in generative AI this year (when you add in Anthropic and OpenAI’s revenues to the pot, along with the other stragglers) and most of that number has been gathered through reporting from outlets like The Information, because none of these companies share their revenues, all of them lose shit tons of money , and their actual revenues are really, really small. Only one member of the Magnificent Seven (outside of NVIDIA) has ever disclosed its AI revenue — Microsoft, which stopped reporting in January 2025, when it reported “$13 billion in annualized revenue,” so around $1.083 billion a month.   Microsoft is a sales MACHINE. It is built specifically to create or exploit software markets, suffocating competitors by using its scale to drive down prices, and to leverage the ecosystem that it’s created over the past few decades. $1 billion a month in revenue is chump change for an organization that makes over $27 billion a quarter in PROFITS .  Don’t worry Satya, I’ll come back to you later. “But Ed, the early days!” Worry not — I’ve got that covered .  This is nothing like any other era of tech. There has never been this kind of cash-rush, even in the fiber boom . Over a decade, Amazon spent about one-tenth of the capex that the Magnificent Seven spent in two years on generative AI building AWS — something that now powers a vast chunk of the web, and has long been Amazon’s most profitable business unit .  Generative AI is nothing like Uber , with OpenAI and Anthropic’s true costs coming in at about $159 billion in the past two years, approaching five times Uber’s $30 billion all-time burn. And that’s before the bullshit with NVIDIA and Oracle. Microsoft last reported AI revenue in January . It’s October this week. Why did it stop reporting this number, you think? Is it because the numbers are so good it couldn’t possibly let people know? As a general rule, publicly traded companies — especially those where the leadership are compensated primarily in equity — tend to brag about their successes, in part because said bragging boosts the value of the thing that the leadership gets paid in. There’s no benefit to being shy. Oracle literally made a regulatory filing to boast it had a $30 billion customer , which turned out to be OpenAI, who eventually agreed (publicly) to spend $300 billion in compute over five years .  Which is to say that Microsoft clearly doesn’t have any good news to share, and as I’ll reveal later, they can’t even get 3% of their 440 million Microsoft 365 subscribers to pay for Microsoft 365 Copilot.  If Microsoft can’t sell this shit, nobody can.  Anyway, I’m nearly done, sorry, you see, I’m writing this whole thing as if you’re brand new and walking up to this relatively unprepared, so I need to introduce another company.  In 2020, a splinter group jumped off of OpenAI, funded by Amazon and Google to do much the same thing as OpenAI but pretend to be nicer about it until they have to raise from the Middle East . Anthropic has always been better at coding for some reason, and people really like its Claude models.  Both OpenAI and Anthropic have become the only two companies in generative AI to make any real progress, either in terms of recognition or in sheer commercial terms, accounting for the majority of the revenue in the AI industry.  In a very real sense, the AI industry’s revenue is OpenAI and Anthropic. In the year where Microsoft recorded $13bn in AI revenues, $10 billion came from OpenAI’s  spending on Microsoft Azure. Anthropic burned $5.3 billion last year — with the vast majority of that going towards compute . Outside of these two companies, there’s barely enough revenue to justify a single data center. Where we sit today is a time of immense tension. Mark Zuckerberg says we’re in a bubble , Sam Altman says we’re in a bubble , Alibaba Chairman and billionaire Joe Tsai says we’re in a bubble , Apollo says we’re in a bubble , nobody is making money and nobody knows why they’re actually doing this anymore, just that they must do it immediately.  And they have yet to make the case that generative AI warranted any of these expenditures.  That was undoubtedly the longest introduction to a newsletter I’ve ever written, and the reason why I took my time was because this post demands a level of foreshadowing and exposition, and because I want to make it make sense to anyone who reads it — whether they’ve read my newsletter for years, or whether they’re only just now investigating their suspicions that generative AI may not be all it’s cracked up to be.  Today I will make the case that generative AI’s fundamental growth story is flawed, and explain why we’re in the midst of an egregious bubble. This industry is sold by keeping things vague, and knowing that most people don’t dig much deeper than a headline, a problem I simply do not have.  This industry is effectively in service of two companies — OpenAI and NVIDIA — who pump headlines out through endless contracts between them and subsidiaries or investments to give the illusion of activity.  OpenAI is now, at this point, on the hook for over a trillion dollars , an egregious sum for a company that already forecast billions in losses, with no clear explanation as to how it’ll afford any of this beyond “we need more money” and the vague hope that there’s another Softbank or Microsoft waiting in the wings to swoop in and save the day.  I’m going to walk you through where I see this industry today, and why I see no future for it beyond a fiery apocalypse.  While everybody (reasonably!) harps on about hallucinations — which, to remind you, is when a model authoritatively states something that isn’t true — but the truth is far more complex, and far worse than it seems.  You cannot rely on a large language model to do what you want. Even the most highly-tuned models on the most expensive and intricate platform can’t actually be relied upon to do exactly what you want.  A “hallucination” isn’t just when these models say something that isn’t true. It’s when they decide to do something wrong because it seems the most likely thing to do, or when a coding model decides to go on a wild goose chase, failing the user and burning a ton of money in the process.  The advent of “reasoning” models — those engineered to ‘think’ through problems in a way reminiscent of a human — and the expansion of what people are (trying) to use LLMs for demands that the definition of an AI hallucination be widened, not merely referring to factual errors, but fundamental errors in understanding the user’s request or intent, or what constitutes a task, in part because these models cannot think and do not know anything .  However successful a model might be in generating something good *once*, it will also often generate something bad, or it’ll generate the right thing but in an inefficient and over-verbose fashion. You do not know what you’re going to get each time, and hallucinations multiply with the complexity of the thing you’re asking for, or whether a task contains multiple steps (which is a fatal blow to the idea of “agents.”  You can add as many levels of intrigue and “reasoning” as you want, but Large Language Models cannot be trusted to do something correctly, or even consistently, every time. Model companies have successfully convinced everybody that the issue is that users are prompting the models wrong, and that people need to be “trained to use AI,” but what they’re doing is training people to explain away the inconsistencies of Large Language Models, and to assume individual responsibility for what is an innate flaw in how large language models work.  Large Language Models are also uniquely expensive. Many mistakenly try and claim this is like the dot com boom or Uber, but the basic unit economics of generative AI are insane. Providers must purchase tens or hundreds of thousands of GPUs each costing $50,000 a piece, and hundreds of millions or billions of dollars of infrastructure for large clusters. And that’s without mentioning things like staffing, construction, power, or water.   Then you turn them on and start losing money. Despite hundreds of billions of GPUs sold, nobody seems to make any money, other than NVIDIA, the company that makes them, and resellers like Dell and Supermicro who buy the GPUs, put them in servers, and sell them to other people.  This arrangement works out great for Jensen Huang, and terribly for everybody else.  I am going to explain the insanity of the situation we find ourselves in, and why I continue to do this work undeterred. The bubble has entered its most pornographic, aggressive and destructive stage, where the more obvious it becomes that they’re cooked, the more ridiculous the generative AI industry will act — a dark juxtaposition against every new study that says “generative AI does not work” or new story about ChatGPT’s uncanny ability to activate mental illness in people.  So, let’s start simple: NVIDIA is a hardware company that sells GPUs, including the consumer GPUs that you’d see in a modern gaming PC, but when you read someone say “GPU” within the context of AI, they mean enterprise-focused GPUs like the A100, H100, H200, and more modern GPUs like the Blackwell-series B200 and GB200 (which combines two GPUs with an NVIDIA CPU).  These GPUs cost anywhere from $30,000 to $50,000 (or as high as $70,000 for the newer Blackwell GPUs), and require tens of thousands of dollars more of infrastructure — networking to “cluster” server racks of GPUs together to provide compute, and massive cooling systems to deal with the massive amounts of heat they produce, as well as the servers themselves that they run on, which typically use top-of-the-line data center CPUs, and contain vast quantities of high-speed memory and storage. While the GPU itself is likely the most expensive single item within an AI server, the other costs — and I’m not even factoring in the actual physical building that the server lives in, or the water or electricity that it uses — add up.  I’ve mentioned NVIDIA because it has a virtual monopoly in this space. Generative AI effectively requires NVIDIA GPUs, in part because it’s the only company really making the kinds of high-powered cards that generative AI demands, and  because NVIDIA created something called CUDA — a collection of software tools that lets programmers write software that  runs on GPUs, which were traditionally used primarily for rendering graphics in games.  While there are open-source alternatives , as well as alternatives from Intel (with its ARC GPUs) and AMD (Nvidia’s main rival in the consumer space), these aren’t nearly as mature or feature-rich.  Due to the complexities of AI models, one cannot just stand up a few of these things either — you need clusters of thousands, tens of thousands, or hundreds of thousands of them for it to be worthwhile, making any investment in GPUs in the hundreds of millions or billions of dollars, especially considering they require completely different data center architecture to make them run. A common request — like asking a generative AI model to parse through thousands of lines of code and make a change or an addition — may use multiple of these $50,000 GPUs at the same time, and so if you aspire to serve thousands, or millions of concurrent users, you need to spend big. Really big.  It’s these factors — the vendor lock-in, the ecosystem, and the fact that generative AI only works when you’re buying GPUs at scale — that underpin the rise of Nvidia. But beyond the economic and technical factors, there are human ones, too.   To understand the AI bubble is to understand why CEOs do the things they do. Because an executive’s job is so vague , they can telegraph the value of their “labor” by spending money on initiatives and making partnerships. AI gave hyperscalers the excuse to spend hundreds of billions of dollars on data centers and buy a bunch of GPUs to go in them, because that, to the markets, looks like they’re doing something. By virtue of spending a lot of money in a frighteningly short amount of time, Satya Nadella received multiple glossy profiles , all without having to prove that AI can really do anything, be it a job or make Microsoft money.  Nevertheless, AI allowed CEOs to look busy, and once the markets and journalists had agreed on the consensus opinion that “AI would be big,” all that these executives had to do was buy GPUs and “do AI.”   We are in the midst of one of the darkest forms of software in history, described by many as an unwanted guest invading their products, their social media feeds, their bosses’ empty minds, and resting in the hands of monsters. Every story of its success feels bereft of any real triumph, with every literal description of its abilities involving multiple caveats about the mistakes it makes or the incredible costs of running it.  Generative AI exists for two reasons: to cost money, and to make executives look busy. It was meant to be the new enterprise software and the new iPhone and the new Netflix all at once, a panacea where software guys pay one hardware guy for GPUs to unlock the incredible value creation of the future.  Generative AI was always set up to fail, because it was meant to be everything, was talked about like it was everything, is still sold like it’s everything, yet for all the fucking hype, it all comes down to two companies: OpenAI, and, of course, NVIDIA. NVIDIA was, for a while, living high on the hog. All CEO Jensen Huang had to do every three months was saying “check out these numbers” and the markets and business journalists would squeal with glee, even as he said stuff like “ the more you buy the more you save ,” in part tipping his head to the (very real and sensible) idea of accelerated computing, but framed within the context of the cash inferno that’s generative AI, seems ludicrous.  Huang’s showmanship  worked really well for NVIDIA for a while, because for a while the growth was easy. Everybody was buying GPUs. Meta, Microsoft, Amazon, Google (and to a lesser extent Apple and Tesla) make up 42% of NVIDIA’s revenue , creating, at least for the first four, a degree of shared mania where everybody justified buying tens of billions of dollars of GPUs a year by saying “the other guy is doing it!” This is one of the major reasons the AI bubble is happening, because people conflated NVIDIA’s incredible sales with “interest in AI,” rather than everybody buying GPUs. Don’t worry, I’ll explain the revenue side a little bit later. We’re here for the long haul. Anyway, NVIDIA is facing a problem — that the only thing that grows forever is cancer .  On September 9 2025, the Wall Street Journal said that NVIDIA’s “wow” factor was fading , going from beating analyst estimates in by nearly 21% in its Fiscal Year Q2 2024 earnings to scraping by with a mere 1.52% beat in its most-recent earnings — something that for any other company, would be a good thing, but framed against the delusional expectations that generative AI has inspired, is a figure that looks nothing short of ominous. Per the Wall Street Journal: In any other scenario, 56% year-over-year growth would lead to an abundance of Dom Perignon and Huang signing hundreds of boobs , but this is NVIDIA , and that’s just not good enough. Back in February 2024, NVIDIA was booking 265% year-over-year growth , but in its February 2025 earnings, NVIDIA only grew by a measly 78% year-over-year .  It isn’t so much that NVIDIA isn’t growing , but that to grow year-over-year at the rates that people expect is insane. Life was a lot easier when NVIDIA went from $6.05 billion in revenue in Q4 FY2023 to $22 billion in revenue in Q4 FY2024 , but for it to grow even 55% year-over-year from Q2 FY2026 ( $46.7 billion ) to Q2 2027 would require it to make $72.385 billion in revenue in the space of three months, mostly from selling GPUs (which make up around 88% of its revenue).   This would put Nvidia in the ballpark of Microsoft ( $76 billion in the last quarter ) and within the neighborhood of Apple ( $94 billion in the last quarter ), predominantly making money in an industry that a year-and-a-half ago barely made the company $6 billion in a quarter.  And the market needs NVIDIA to perform, as the company makes up 8% of the value of the S&P 500 . It’s not enough for it to be wildly profitable, or have a monopsony on selling GPUs, or for it to have effectively 10x’d their stock in a few years. It must continue to grow at the fastest rate of anything ever, making more and more money selling these GPUs to a small group of companies that immediately start losing money once they plug them in.  While a few members of the Magnificent Seven could be depended on to funnel tens of billions of dollars into a furnace each quarter, there were limits , even for companies like Microsoft, which had bought over 485,000 GPUs in 2024 alone . To take a step back, companies like Microsoft, Google and Amazon make their money by either selling access to Large Language Models that people incorporate into their products, or by renting out servers full of GPUs to run inference (as said previously, the process to generate an output by a model or series of models) or train AI models for companies that develop and market models themselves, namely Anthropic and OpenAI.  The latter revenue stream of which is where Jensen Huang found a solution to that eternal growth problem: the neocloud, namely CoreWeave, Lambda and Nebius.  These businesses are fairly straightforward. They own (or lease) data centers that they then fill full of servers that are full of NVIDIA GPUs, which they then rent on an hourly basis to customers, either on a per-GPU basis or in large batches for larger customers, who guarantee they'll use a certain amount of compute and sign up to long-term (i.e. more than an hour at a time) commitments. A neocloud is a specialist cloud compute company that exists only to provide access to GPUs for AI, unlike Amazon Web Services, Microsoft Azure and Google Cloud, all of which have healthy businesses selling other kinds of compute, with AI (as I’ll get into later) failing to provide much of a return on investment.  It’s not just the fact that these companies are more specialized than, say, Amazon’s AWS or Microsoft Azure. As you’ve gathered from the name, these are new, young, and in almost all cases, incredibly precarious businesses — each with financial circumstances that would make a Greek finance minister blush.  That’s because setting up a neocloud is expensive . Even if the company in question already has data centers — as CoreWeave did with its cryptocurrency mining operation — AI requires completely new data center infrastructure to house and cool the GPUs , and those GPUs also need paying for, and then there’s the other stuff I mentioned earlier, like power, water, and the other bits of the computer (the CPU, the motherboard, the memory and storage, and the housing).  As a result, these neoclouds are forced to raise billions of dollars in debt, which they collateralize using the GPUs they already have , along with contracts from customers, which they use to buy more GPUs. CoreWeave, for example, has $25 billion in debt on estimated revenues of $5.35 billion , losing hundreds of millions of dollars a quarter. You know who also invests in these neoclouds? NVIDIA! NVIDIA is also one of CoreWeave’s largest customers (accounting for 15% of its revenue in 2024), and just signed a deal to buy $6.3 billion of any capacity that CoreWeave can’t otherwise sell to someone else through 2032 , an extension of a $1.3 billion 2023 deal reported by the Information . It was the anchor investor ($250 million) in CoreWeave’s IPO , too. NVIDIA is currently doing the same thing with Lambda, another neocloud that NVIDIA invested in, which also  plans to go public next year. NVIDIA is also one of Lambda’s largest customers, signing a deal with it this summer to rent 10,000 GPUs for $1.3 billion over four years . In the UK, NVIDIA has just invested $700 million in Nscale , a former crypto miner that has never built an AI data center , and that has, despite having no experience, committed $1 billion (and/or 100,000 GPUs) to an OpenAI data center in Norway . On Thursday, September 25, Nscale announced it had closed another funding round, with NVIDIA listed as a main backer — although it’s unclear how much money it put in . It would be safe to assume it’s another few hundred million.  NVIDIA also invested in Nebius , an outgrowth of Russian conglomerate Yandex, and Nebius provides, through a partnership with NVIDIA, tens of thousands of dollars’ worth of compute credits to companies in NVIDIA’s Inception startup program. NVIDIA’s plan is simple: fund these neoclouds, let these neoclouds load themselves up with debt, at which point they buy GPUs from NVIDIA, which can then be used as collateral for loans, along with contracts from customers, allowing the neoclouds to buy even more GPUs. It’s like that Robinhood infinite money glitch… …except, that is, for one small problem. There don’t appear to be that many customers. As I went into recently on my premium newsletter , NVIDIA funds and sustains Neoclouds as a way of funnelling revenue to itself, as well as partners like Supermicro and Dell , resellers that take NVIDIA GPUs and put them in servers to sell pre-built to customers. These two companies made up 39% of NVIDIA’s revenues last quarter .  Yet when you remove hyperscaler revenue — Microsoft, Amazon, Google, OpenAI and NVIDIA — from the revenues of these neoclouds, there’s barely $1 billion in revenue combined, across CoreWeave, Nebius and Lambda . CoreWeave’s $5.35 billion revenue is predominantly made up of its contracts with NVIDIA, Microsoft (offering compute for OpenAI), Google ( hiring CoreWeave to offer compute for OpenAI ), and OpenAI itself, which has promised CoreWeave $22.4 billion in business over the next few years. This is all a lot of stuff , so I’ll make it really simple: there is no real money in offering AI compute, but that isn’t Jensen Huang’s problem, so he will simply force NVIDIA to hand money to these companies so that they have contracts to point to when they raise debt to buy more NVIDIA GPUs.  Neoclouds are effectively giant private equity vehicles that exist to raise money to buy GPUs from NVIDIA, or for hyperscalers to move money around so that they don’t increase their capital expenditures and can, as Microsoft did earlier in the year , simply walk away from deals they don’t like. Nebius’ “$17.4 billion deal” with Microsoft even included a clause in its 6-K filing that Microsoft can terminate the deal in the event the capacity isn’t built by the delivery dates, and Nebius has already used the contract to raise $3 billion to… build the data center to provide compute for the contract. Here, let me break down the numbers: From my analysis, it appears that CoreWeave, despite expectations to make that $5.35 billion this year, has only around $500 million of non-Magnificent Seven or OpenAI AI revenue in 2025 , with Lambda estimated to have around $100 million in AI revenue , and Nebius around $250 million without Microsoft’s share , and that’s being generous. In simpler terms, the Magnificent Seven is the AI bubble, and the AI bubble exists to buy more GPUs, because (as I’ll show) there’s no real money or growth coming out of this, other than in the amount that private credit is investing — “ $50 billion a quarter, for the low end, for the past three quarters .” I dunno man, let’s start simple: $50 billion a quarter of data center funding is going into an industry that has less revenue than Genshin Impact . That feels pretty bad. Who’s gonna use these data centers? How are they going to even make money on them? Private equity firms don’t typically hold onto assets, they sell them or take them public. Doesn’t seem great to me! Anyway, if AI was truly the next big growth vehicle, neoclouds would be swimming in diverse global revenue streams. Instead, they’re heavily-centralized around the same few names, one of which (NVIDIA) directly benefits from their existence not as a company doing business, but as an entity that can accrue debt and spend money on GPUs. These Neoclouds are entirely dependent on a continual flow of private credit from firms like Goldman Sachs ( Nebius , CoreWeave , Lambda for its IPO ), JPMorgan ( Lambda , Crusoe , CoreWeave ), and Blackstone ( Lambda , CoreWeave ), who have in a very real sense created an entire debt-based infrastructure to feed billions of dollars directly to NVIDIA, all in the name of an AI revolution that's yet to arrive. The fact that the rest of the neocloud revenue stream is effectively either a hyperscaler or OpenAI is also concerning. Hyperscalers are, at this point, the majority of data center capital expenditures , and have yet to prove any kind of success from building out this capacity, outside, of course, Microsoft’s investment in OpenAI, which has succeeded in generating revenue while burning billions of dollars .  Hyperscaler revenue is also capricious, but even if it isn’t, why are there no other major customers? Why, across all of these companies, does there not seem to be one major customer who isn’t OpenAI?  The answer is obvious: nobody that wants it can afford it, and those who can afford it don’t need it.  It’s also unclear what exactly hyperscalers are doing with this compute, because it sure isn’t “making money.” While Microsoft makes $10 billion in revenue from renting compute to OpenAI via Microsoft Azure, it does so at-cost, and was charging OpenAI $1.30-per-hour for each A100GPU it rents, a loss of $2.2 an hour per GPU , meaning that it is  likely losing money on this compute, especially as SemiAnalysis has the total cost per hour per GPU at around $1.46 with the cost of capital and debt associated for a hyperscaler, though it’s unclear if that’s for an H100 or A100 GPU. In any case, how do these neoclouds pay for their debt if the hyperscalers give up, or NVIDIA doesn’t send them money, or, more likely, private credit begins to notice that there’s no real revenue growth outside of circular compute deals with neoclouds’ largest supplier, investor and customer? They don’t! In fact, I have serious concerns that they can’t even build the capacity necessary to fulfil these deals, but nobody seems to worry about that. No, really! It appears to be taking Oracle and Crusoe around 2.5 years per gigawatt of compute capacity . How exactly are any of these neoclouds (or Oracle itself) able to expand to actually capture this revenue? Who knows! But I assume somebody is going to say “OpenAI!” Here’s an insane statistic for you: OpenAI will account for — in both its own revenue (projected $13 billion) and in its own compute costs ($16 billion, according to The Information, although that figure is likely out of date, and seemingly only includes the compute it’ll use, and not that it has committed to build, and thus has spent money on) — about 50% of all AI revenues in 2025. That figure takes into account the $400m ARR for ServiceNow, Adobe, and Salesforce ; the $35bn in revenue for the Magnificent Seven from AI (not profit, and based on figures from the previous year); revenue from neoclouds like CoreWeave, Nebius, and Lambda; and the estimated revenue from the entire generative AI industry (including Anthropic and other smaller players, like Perplexity and Anysphere) for a total of $55bn.OpenAI is the generative AI industry — and it’s a dog of a company. As a reminder, OpenAI has leaked that it’ll burn $115 billion in the next four years , and based on my estimates, it needs to raise more than $290 billion in the next four years based on its $300 billion deal with Oracle alone. That alone is a very, very bad sign, especially as we’re three years and $500 billion or more into this hype cycle with few signs of life outside of, well, OpenAI promising people money. Credit to Anthony Restaino for this horrifying graphic : This is not what a healthy, stable industry looks like. Alright, well, things can’t be that bad on the software side. As I covered on my premium newsletter a few weeks ago , everybody is losing money on generative AI, in part because the cost of running AI models is increasing , and in part because the software itself doesn’t do enough to warrant the costs associated with running them, which are already subsidized and unprofitable for the model providers .  Outside of OpenAI (and to a lesser extent Anthropic), nobody seems to be making much revenue, with the most “successful” company being Anysphere, makers of AI coding tool Cursor, which hit $500 million ‘annualized” ( so $41.6 million in one month ) a few months ago, just before Anthropic and OpenAI jacked up the prices for “priority processing” on enterprise queries , raising its operating costs as a result. In any case, that’s some piss-poor revenue for an industry that’s meant to be the future of software. Smartwatches are projected to make $32 billion this year , and as mentioned, the Magnificent Seven expects to make $35 billion or so in revenue from AI this year . Even Anthropic and OpenAI seem a little lethargic, both burning billions of dollars while making, by my estimates, no more than $2 billion and $6.26 billion in 2025 so far , despite projections of $5 billion and $13 billion respectively.  Outside of these two, AI startups are floundering, struggling to stay alive and raising money in several-hundred million dollar bursts as their negative-gross-margin businesses flounder.  As I dug into a few months ago , I could find only 12 AI-powered companies making more than $8.3 million a month, with two of them slightly improving their revenues, specifically AI search company Perplexity ( which has now hit $150 million ARR, or $12.5 million in a month ) and AI coding startup Replit ( which also hit $150 million ARR in September ).  Both of these companies burn ridiculous amounts of money. Perplexity burned 164% of its revenue on Amazon Web Services, OpenAI and Anthropic last year , and while Replit hasn’t leaked its costs, The Information reports its gross margins in July were 23% , which doesn’t include the costs of its free users, which you simply have to do with LLMs as free users are capable of costing you a hell of a lot of money. Problematically, your paid users can also cost you more than they bring in as well. In fact, every user loses you money in generative AI, because it’s impossible to do cost control in a consistent manner. A few months ago, I did a piece about Anthropic losing money on every single Claude Code subscriber , and I’ll walk you through it in a very simplified fashion: Anthropic is, to be clear, the second-largest model developer, and has some of the best AI talent in the industry. It has a better handle on its infrastructure than anyone outside of big tech and OpenAI. It still cannot seem to fix this problem, even with weekly rate limits . While one could assume that Anthropic is simply letting people run wild, my theory is far simpler: even the model developers have no real way of limiting user activity, likely due to the architecture of generative AI. I know it sounds insane, but at the most advanced level, model providers are still prompting their models, and whatever rate limits may be in place appear to, at times, get completely ignored, and there doesn’t seem to be anything they can do to stop it. No, really. Anthropic counts amongst its capitalist apex predators one lone Chinese man who spent $50,000 of their compute in the space of a month fucking around with Claude Code. Even if Anthropic was profitable — it isn’t, and will burn billions this year — a customer paying $200-a-month running up $50,000 in costs immediately devours the margin of any user running the service that day , if not that week or month. Even if Anthropic’s costs are half the published rates, one guy amounted to 125 users’ monthly revenue.  That’s not a real business! That’s a bad business with out-of-control costs, and it doesn’t appear anybody has these costs under control. A few weeks ago, Replit — an unprofitable AI coding company — released a product called “ Agent 3 .” which promised to be “10x more autonomous” and offer “infinitely more possibilities,” “[testing] and [fixing] its code, constantly improving your application behind the scenes in a reflection loop.” In reality, this means you’d go and tell the model to build something and it would “go do it,” and you’ll be shocked to hear that these models can’t be relied upon to “go and do” anything. Please note that this was launched a few months after Replit raised its prices, shifting to obfuscated “ effort-based ” pricing that would charge “the full scope of the agent’s work.” Agent 3 has been a disaster. Users found tasks that previously cost a few dollars were spiralling into the hundreds of dollars, with The Register reporting one customer found themselves with a $1000 bill after a week: Another user complained that “costs skyrocketed, without any concrete results”: As I previously reported, in late May/early June, both OpenAI and Anthropic cranked up the pricing on their enterprise customers , leading to Replit and Cursor both shifting their prices. This abuse has now trickled down to their customers. Replit has now released an update that lets you choose how autonomous you want Agent 3 to be , which is a tacit admission that you can’t trust coding LLMs to build software. Replit’s users are still pissed off, complaining that Replit is charging them for activity when the agent doesn’t do anything , a consistent problem across its Reddit. While Reddit is not the full summation of all users across every company, it’s a fairly good barometer of user sentiment, and man, are users pissy. Traditionally, Silicon Valley startups have relied upon the same model of “grow really fast and burn a bunch of money, then “turn the profit lever.” AI does not have a “profit lever,” because the raw costs of providing access to AI models are so high ( and they’re only increasing ) that the basic economics of how the tech industry sells software don’t make sense. I’ll reiterate something I wrote a few weeks ago : In simpler terms, it is very, very difficult to imagine what one user — free or otherwise — might cost, and thus it’s hard to charge them on a monthly basis, or tell them what a service might cost them on average. This is a huge problem with AI coding environments.  According to The Information , Claude Code was driving “nearly $400 million in annualized revenue, roughly doubling from a few weeks ago” on July 31 2025.  That annualized revenue works out to about $33 million a month in revenue for a company that predicts it will make at least $416 million a month by the end of the year, and for a product that has become the most-popular coding environment in the world, from the second-largest and best-funded AI company in the world. …is that it? Is that all that’s happening here?  $33 million dollars, all of it unprofitable, after it felt, at least based on social media chatter and discussing with multiple different software engineers, that Claude Code had become ubiquitous with anything to do with LLMs. To be clear, Anthropic’s Sonnet and Opus models are consistently some of the most popular for programming on Openrouter , an aggregator of LLM usage, and Anthropic has been consistently-named as “ the best at coding .”  Some bright spark out there is going to say that Microsoft’s Github Copilot has 1.8 million paying subscribers , and guess what, that’s true, and in fact, I reported it! Here’s another fun fact: the Wall Street Journal reported that Microsoft loses “on average more than $20-a-month-per-user,” with “...some users [costing] the company as much as $80.” And that’s for the most-popular product! If you believe the New York Times or other outlets that simply copy and paste whatever Dario Amodei says , you’d think that the reason that software engineers are having trouble finding work is because their jobs are being replaced by AI. This grotesque , abusive , manipulative and offensive lie has been propagated throughout the entire business and tech media without anybody sitting down and asking whether it’s true, or even getting a good understanding of what it is that LLMs can actually do with code. Members of the media, I am begging you, stop doing this. I get it, every asshole is willing to give a quote saying that “ coding is dead ,” and that every executive is willing to burp out some nonsense about replacing all of their engineers , but I am fucking begging you to either use these things yourself, or speak to people that do. I am not a coder. I cannot write or read code. Nevertheless, I am capable of learning , and have spoken to numerous software engineers in the last few months, and basically reached a consensus of “this is kind of useful, sometimes.” However, one very silly man once said that I don’t speak to people who use these tools , so I went and spoke to three notable, experienced software engineers, and asked them to give me the straight truth about what coding LLMs can do.  In simple terms, LLMs are capable of writing code , but can’t do software engineering, because software engineering is the process of understanding, maintaining and executing code to produce functional software, and LLMs do not “learn,” cannot “adapt,” and (to paraphrase Brown), break down the more of your code and variables you ask them to look at at once. It’s very easy to believe that software engineering is just writing code, but the reality is that software engineers maintain software , which includes writing and analyzing code among a vast array of different personalities and programs and problems . Good software engineering harkens back to Brian Merchant’s interviews with translators — while some may believe that translators simply tell you what words mean, true translation is communicating the meaning of a sentence , which is cultural, contextual, regional, and personal, and often requires the exercise of creativity and novel thinking.  My editor, Matthew Hughes, gave an example of this in his newsletter :  Similarly, coding is not just “a series of text that programs a computer,” but a series of interconnected characters that refers to other software in other places that must also function now and explain, on some level, to someone who has never, ever seen the code before, why it was done this way.  This is, by the way, why we are still yet to get any tangible proof that AI is replacing software engineers…because it can’t.  Of all the fields supposedly at risk from “AI disruption,” coding feels (or felt) the most tangible, if only because the answer to “can you write code with LLMs” wasn’t an immediate, unilateral no.  The media has also been quick to say that AI “writes software,” which is true in the same way that ChatGPT “writes novels”. In reality, LLMs can generate code, and do some software engineering-adjacent tasks, but, like all Large Language Models, break down and go totally insane, hallucinating more as the tasks get more complex . And, as I pointed out earlier, software engineering is not just coding. It involves thinking about problems, finding solutions to novel challenges, designing stuff in a way that can be read and maintained by others, and that’s (ideally) scalable and secure. The whole fucking point of an “AI” is that you hand shit off to it! That’s what they’ve been selling it as! That’s why Jensen Huang told kids to stop learning to code, as with AI, there’s no point .  And it was all a lie. Generative AI can’t do the job of a software engineer, and it fails while  also costing abominable amounts of money.  Coding LLMs seem like magic at first, because they (to quote a conversation with Carl Brown) make the easy things easier, but they also make the harder things harder. They don’t even speed up engineers — they actually make them slower ! Yet coding is basically the only obvious use case for LLMs.  I’m sure you’re gonna say “but I bet the enterprise is doing well!” and you are so very, very wrong. Before I go any further, let’s establish some facts: All of this is to say that Microsoft has one of the largest commercial software empires in history, thousands (if not tens of thousands) of salespeople, and thousands of companies that literally sell Microsoft services for a living . And it can’t sell AI. A source that has seen materials related to sales has confirmed that, as of August 2025, Microsoft has around eight million active licensed users of Microsoft 365 Copilot, amounting to a 1.81% conversion rate across the 440 million Microsoft 365 subscribers. This would amount to, if each of these users paid annually at the full rate of $30-a-month, to about $2.88 billion in annual revenue for a product category that makes $33 billion a fucking quarter. And I must be clear, I am 100% sure these users aren’t all paying $30 a month. The Information reported a few weeks ago that Microsoft has been “reducing the software’s price with more generous discounts on the AI features, according to customers and salespeople,” heavily suggesting discounts had already been happening. Enterprise software is traditionally sold at a discount anyway — or, put a different way, with bulk pricing for those who sign up a bunch of users at once.  In fact, I’ve found evidence that it’s been doing this a while, with a 15% discount on annual Microsoft 365 Copilot subscriptions for orders of 10-to-300 seats mentioned by an IT consultant back in late 2024 , and another that’s currently running through September 30, 2025 through Microsoft’s Cloud Solution Provider program , with up to 2400 licenses discounted if you pay upfront for the year. Microsoft seems to do this a lot, as I found another example of an offer that ran from January 1 2025 through March 31 2025 . An “active” user is someone who has taken one action on Copilot in any Microsoft 365 app in the space of 28 days. Now, I know. That word, active. Maybe you’re thinking “Ed, this is like the gym model! There are unpaid licenses that Microsoft is getting paid for!”  Fine! Let’s assume that Microsoft also has, based on research that suggests this is the case for all software companies, another 50% — four million — of paid Copilot licenses that aren’t being used. That still makes this 12 million users, which is still a putrid 2.72% conversion rate. So, why aren’t people paying for Copilot? Let’s hear from someone who talked to The Information : Microsoft 365 Copilot has been such a disaster that Microsoft will now integrate Anthropic’s models in an attempt and make them better.  Oh, one other thing: sources also confirm GPU utilization for Microsoft 365’s enterprise Copilot is barely scratching 60%.  I’m also hearing that less than SharePoint — another popular enterprise app from Microsoft with 250 million users — had less than 300,000 weekly active users of its AI copilot features in August. So, The Information reported a few months ago that Microsoft’s projected AI revenues would be $13 billion, with $10 billion of that from OpenAI, leaving about $3 billion of total revenue across Microsoft 365 Copilot and any other foreseeable feature that Microsoft sells with “AI” on it. This heavily suggests that Microsoft is making somewhere between $1.5 billion and $2 billion on Azure or Microsoft 365 Copilot, though I suppose there are other places it could be making AI revenue too. Right? I guess. In any case, Microsoft’s net income (read: profit) in its last quarterly earnings was $27.2 billion. One of the comfortable lies that people tell themselves is that the AI bubble is similar to the fiber boom, or the dot com bubble, or Uber, or that we’re in the “growth stage,” or that “this is what software companies do, they spend a bunch of money then “ pull the profit lever .”  This is nothing like anything you’ve seen before, because this is the dumbest shit that the tech industry has ever done.  AI data centers are nothing like fiber, because there are very few actual use cases for these GPUs outside of AI, and none of them are remotely hyperscale revenue drivers. As I discussed a month or so ago , data center development accounted for more of America’s GDP growth than all consumer spending combined, and there really isn’t any demand for AI in general, let alone at the scale that these hundreds of billions of dollars are being sunk into.  The conservative estimate of capital expenditures related to data centers is around $400 billion, but given the $50 billion a quarter in private credit, I’m going to guess it breaks $500 billion, all to build capacity for an industry yet to prove itself. And this NVIDIA-OpenAI “$100 billion funding” news should only fill you full of dread, but also it isn’t fucking finalized, stop reporting it as if it’s done, I swear to god- Anyway, according to CNBC , “the initial $10 billion tranche is locked in at a $500 billion valuation and expected to close within a month or so once the transaction has been finalized,” with “successive $10 billion rounds are planned, each to be priced at the company’s then-current valuation as new capacity comes online.”  At no point is anyone asking how, exactly, OpenAI builds data centers to fill full of these GPUs. In fact, I am genuinely shocked (and a little disgusted!) by how poorly this story has been told. Let’s go point by point: To be clear, when I say OpenAI needs at least $300 billion over the next four years, that’s if you believe its projections, which you shouldn’t .  Let’s walk through its (alleged) numbers, while plagiarizing myself :  According to The Information , here's the breakdown (these are projections): OpenAI's current reported burn is $116 billion through 2030, which means there is no way that these projections include $300 billion in compute costs, even when you factor in revenue. There is simply no space in these projections to absorb that $300 billion, and from what I can tell, by 2029, OpenAI will have actually burned more than $290 billion, assuming that it survives that long, which I do not believe it will. Don’t worry, though. OpenAI is about to make some crazy money . Here are the projections that CFO Sarah Friar signed off on : Just so we are clear, OpenAI intends to 10x its revenue in the space of four years, selling software and access to models in an industry with about $60 billion of revenue in 2025. How will it do this? It doesn’t say. I don’t know OpenAI CFO Sarah Friar, but I do know that signing off on these numbers is, at the very least, ethically questionable.  Putting aside the ridiculousness of OpenAI’s deals, or its funding requirements, Friar has willfully allowed Sam Altman and OpenAI to state goals that defy reality or good sense, all to take advantage of investors and public markets that have completely lost the plot.  I need to be blunter: OpenAI has signed multiple different deals and contracts for amounts it cannot afford to pay, that it cannot hope to raise the money to pay for, that defy the amounts of venture capital and private credit available, all to sustain a company that will burn $300 billion and has no path to profitability of any kind. So, as I said above, CNBC reported on September 23, 2025 that the NVIDIA deal will be delivered in $10 billion tranches, the first of which is “expected to close within a month,” and the rest delivered “as new capacity comes online.” This is, apparently, all part of a plan to build 10GW of data center capacity with NVIDIA. A few key points: So, let’s start simple: data centers take forever to build. As I said previously, based on current reports, it’s taking Oracle and Crusoe around 2.5 years per gigawatt of data center capacity, and nowhere in these reports does one reporter take a second to say “hey, what data centers are you talking about?” or “hey, didn’t Sam Altman say back in July that he was building 10GW of data center capacity with Oracle? ” But wait, now Oracle and OpenAI have done another announcement that says they’re only doing 7GW, but they’re “ahead of schedule” on 10GW?  Wait, is NVIDIA’s 10GW the same 10GW as Oracle and OpenAI are working on? Is it different? Nobody seems to know or care! Anyway, I cannot be clear enough how unlikely it is that (as NVIDIA has said) “ the first gigawatt of NVIDIA systems will be deployed in the second half of 2026 ,” and that’s if it has bought the land and got the permits and ordered the construction, none of which has happened yet. But let’s get really specific on costs!  Crusoe’s 1.2GW of compute for OpenAI is a $15 billion joint venture , which means a gigawatt of compute runs about $12.5 billion. Abilene’s eight buildings are meant to hold 50,000 NVIDIA GB200 GPUs and their associated networking infrastructure, so let’s say a gigawatt is around 333,333 Blackwell GPUs. Though this math is a little funky due to NVIDIA promising to install its new Rubin GPUs in these theoretical data centers, that means these data centers will require a little under $200 billion worth of GPUs. By my maths that’s $325 billion.  I’m so tired of this. A number of you have sent me the following image with some sort of comment about how “this is how it’ll work,” and you are wrong, because this is neither how it works nor how it will work nor accurate on any level. In the current relationship, NVIDIA Is Not Sending OpenAI $100 Billion, nor will it send it that much money, because 90% of OpenAI’s funding is gated behind building 9 or 10 gigawatts of data center capacity. In the current relationship, OpenAI does not have the money to pay Oracle. Also, can Oracle even afford to give that much money to NVIDIA? It had negative free cash flow last quarter , already has $104 billion in debt , and its biggest new customer cannot afford a single fucking thing it’s promised. The only company in this diagram that actually can afford to do any of this shit is NVIDIA, and even then it only has $56 billion cash on hand . In any case, as I went over on Friday, OpenAI has promised about a trillion dollars between compute contracts across Oracle, Microsoft, Google and CoreWeave, 17 Gigawatts of promised data centers in America between NVIDIA and “Stargate,” several more gigawatts of international data centers, custom chips from Broadcom, and their own company operations. How exactly does this get paid for?  Nobody seems to ask these questions! Why am I the asshole doing this? Don’t we have tech analysts that are meant to analyse shit? AHhhhh- Every time I sit down to write about this subject the newsletters seem to get longer, because people are so painfully attached to the norms and tropes of the past. This post is, already, 17,500 words — a record for this newsletter — and I’ve still not finished editing and expanding it.  What we’re witnessing is one of the most egregious wastes of capital in history, sold by career charlatans with their reputations laundered by a tech and business media afraid to criticize the powerful and analysts that don’t seem to want to tell their investors the truth. There are no historic comparisons here — even Britain’s abominable 1800s railway bubble, which absorbed half of the country’s national income , created valuable infrastructure for trains, a vehicle that can move people to and from places. GPUs are not trains, nor are they cars, or even CPUs. They are not adaptable to many other kinds of work, nor are they “the infrastructure of the future of tech,” because they’re already quite old and with everybody focused on buying them, you’d absolutely see one other use case by now that actually mattered. GPUs are expensive, power-hungry, environmentally destructive and require their own kinds of cooling and server infrastructure, making every GPU data center and environmental and fiscal bubble unto themselves. And, whereas the Victorian train infrastructure still exists in the UK — though it has been upgraded over the years — a GPU has a limited useful lifespan. These are cards that can — and will — break after a period of extended usage, whether that period is five years or later, and they’ll inevitably be superseded by something better and more powerful, meaning that the resale value of that GPU will only go down, with a price depreciation that’s akin to a new car.  I am telling you, as I have been telling you for years, again and again and again , that the demand is not there for generative AI, and the demand is never, ever arriving. The only reason anyone humours any of this crap is the endless hoarding of GPUs to build capacity for a revolution that will never arrive. Well, that and OpenAI, a company built and sold on lies about ChatGPT’s capabilities . ChatGPT’s popularity — and OpenAI’s hunger for endless amounts of compute — have created the illusion of demand due to the sheer amount of capacity required to keep their services operational, all so they can burn $8 billion or more in 2025 and, if my estimates are right, nearly a trillion dollars by 2030 . This NVIDIA deal is a farce — an obvious attempt by the largest company on the American stock market to prop up the one significant revenue-generator in the entire industry, knowing that time is running out for it to create new avenues for eternal growth. I’d argue that NVIDIA’s deal also shows the complete contempt that these companies have for the media. There are no details about how this deal works beyond the initial $10 billion, there’s no land purchased, no data center construction started, and yet the media slurps it down without a second thought. I am but one man, and I am fucking peculiar. I did not learn financial analysis in school, but I appear to be one of the few people doing even the most basic analysis of these deals, and while I’m having a great time doing so, I am also exceedingly frustrated at how little effort is being put into prying apart these deals. I realize how ridiculous all of this sounds. I get it. There’s so much money being promised to so many people, market rallies built off the back of massive deals , and I get that the assumption is that this much money can’t be wrong, that this many people wouldn’t just say stuff without intending to follow through, or without considering whether their company could afford it.  I know it’s hard to conceive that hundreds of billions of dollars could be invested in something for no apparent reason, but it’s happening, right god damn now, in front of your eyes, and I am going to be merciless on anyone who attempts to write a “how could we see this coming?”  Generative AI has never been reliable, has always been unprofitable, and has always been unsustainable, and I’ve been saying so since February 2024 . The economics have never made sense, something I’ve said repeatedly since April 2024 , and when I wrote “How Does OpenAI Survive?” in July 2024 , I had multiple people suggest I was being alarmist. Here’s some alarmism for you: the longer it takes for OpenAI to die, the more damage it will cause to the tech industry.  On Friday, when I put out my piece on OpenAI needing a trillion dollars , I asked analyst Gil Luria if the capital was there to build the 17 Gigawatts that OpenAI had allegedly planned to build. He said the following: That doesn’t sound good! Anyway, as I discussed earlier, venture capital could run out in six quarters, with investor and researcher Jon Sakoda estimating that there will only be around $164 billion of dry powder (available capital) in US VC firms by the end of 2025. In July, The French Tech Journal reported (using Pitchbook data) that global venture capital deal activity reached its lowest first-half total since 2018, with $139.4 billion in deal value in the first half of 2025, down from $183.4 billion in the first half of 2024, meaning that any further expansion or demands for venture capital from OpenAI will likely sap the dwindling funds available from other startups. Things get worse when you narrow things to US venture capital. In a piece from April , EY reported that VC-backed investment in US companies hit $80 billion in Q1 2025, but “one $40 billion deal” accounted for half of the investment — OpenAI’s $40 billion deal of which only $10 billion has actually closed, and that didn’t happen until fucking June . Without the imaginary money from OpenAI, US venture would have declined by 36%. The longer that OpenAI survives, the longer it will sap the remaining billions from the tech ecosystem, and I expect it to extend its tendrils to private credit too. The $325 billion it needs just to fulfil its NVIDIA contract, albeit over 4 years, is an egregious sum that I believe exceeds the available private capital in the world. Let’s get specific, and check out the top 10 private equity firms’ available capital!  Assuming that all of this capital is currently available, the top 10 private equity firms in the world have around $477 billion of available capital. We can, of course, include investment banks — Goldman Sachs had around $520 billion cash in hand available at the end of its last quarter , and JPMorgan over $1.7 trillion , but JP Morgan has only dedicated $50 billion in direct lending commitments as of February 2025 , and while Goldman Sachs expanded its direct private credit lending by $15 billion back in June , that appears to be an extension of its “more than $20 billion” direct lending close from mid-2024 . Include both of those, and that brings us up to — if we assume that all of these funds are available — $562 billion in capital and about $164 billion in US venture available to spend, and that’s meant to go to more places than just OpenAI. Sure, sure, there’s more than just the top 10 private equity firms and there’s venture money outside of the US, but what could it be? Like, another $150 billion? You see, OpenAI needs to buy those GPUs, and it needs to build those data centers, and it needs to pay its thousands of staff and marketing and sales costs too. While OpenAI likely wouldn’t be the ones raising the money for the data centers — and honestly, I’m not sure who would do it at this point? — somebody is going to need to build TWENTY GIGAWATTS OF DATA CENTERS if we’re to believe both Oracle and NVIDIA You may argue that venture funds and private credit can raise more, and you’re right! But at this point, there have been few meaningful acquisitions of AI companies, and zero exits from the billions of dollars put into data centers.  Even OpenAI admits in its own announcement about new Stargate sites that this will be a “$400 billion investment over 3 years.” Where the fuck is that money coming from? Is OpenAI really going to absorb massive chunks of all available private credit and venture capital for the next few years?  And no, god, stop saying the US government will bail this out. It will have to bail out hundreds of billions of dollars, there is no scenario where it’s anything less than that, and I’ve already been over this. While the US government has spent equivalent sums in the past to support private business (the total $440 billion dispersed during the Great Recession’s TARP program, where the Treasury bought toxic assets from investment banks to stop them from imploding a la Lehman, springs to mind), it’s hard to imagine any case where OpenAI is seen as vital to the global financial system — and the economic health of the US — as the banking system.  Sure, we spent around $1tn — if we’re being specific, $953bn — on the Paycheck Protection Program during the Covid era, but that was to keep people employed at a time when the economy outside of Zoom and Walmart had, for all intents and purposes, ceased to exist. There was an urgency that doesn’t apply here. If OpenAI goes tits up, Softbank loses some money — nothing new there — and Satya Nadella has to explain why he spent tens of billions of dollars on a bunch of data centers filled with $50,000 GPUs that are, at this point, ornamental.  And while there will be — and have been — disastrous economic consequences, they won’t be as systemically catastrophic as that of the pandemic, or the global financial crisis. To be clear, it’ll be bad, but not as bad .   And there’s also the problem of moral hazard — if the government steps in, what’s to stop big tech chasing its next fruitless rainbow? — and optics. If people resented bailing out the banks after they acted like profligate gamblers and lost, how will they feel bailing out f ucking Sam Altman and Jensen Huang ?  I do apologize for the length of this piece, but the significance of this bubble requires depth. There is little demand, little real money, and little reason to continue, and the sheer lack of responsibility and willingness to kneel before the powerful fills me full of angry bile. I understand many journalists are not in a position where they can just write “this shit sounds stupid,” but we have entered a deeply stupid era, and by continuing to perpetuate the myth of AI, the media guarantees that retail investors and regular people’s 401Ks will suffer. It is now inevitable that this bubble bursts. Deutsche Bank has said the AI boom is unsustainable outside of tech spending “remaining parabolic,” which it says “is highly unlikely,” and Bain Capital has said that $2 trillion in new revenue is needed to fund AI’s scaling , and even that math is completely fucked as it talks about “AI-related savings”: Even when stared in the face by a ridiculous idea — $2 trillion of new revenue in a global software market that’s expected to be around $817 billion in 2025 — Bain still oinks out some nonsense about the “savings from applying AI in sales, marketing, customer support and R&D,” yet another myth perpetuated I assume to placate the fucking morons sinking billions into this. Every single “vibe coding is the future,” “the power of AI,” and “AI job loss” story written perpetuates a myth that will only lead to more regular people getting hurt when the bubble bursts. Every article written about OpenAI or NVIDIA or Oracle that doesn’t explicitly state that the money doesn’t exist, that the revenues are impossible, that one of the companies involved burns billions of dollars and has no path to profitability, is an act of irresponsible make believe and mythos. I am nobody. I am not a financier. I am not anybody special. I just write a lot, and read a lot, and can do the most basic maths in the world. I am not trying to be anything other than myself, nor do I have an agenda, other than the fact that I like doing this and I hate how this story is being told. I never planned for this newsletter to get this big, and now that it has, I’m going to keep doing the same thing every week. I also believe that the way to stop this happening again is to have a thorough and well-sourced explanation of everything as it happens, ripping down the narratives as they’re spun and making it clear who benefits from them and how and why they’re choosing to do so. When things collapse, we need to be clear about how many times people chose to look the other way, or to find good faith ways to interpret bad faith announcements and leak. So, how could we have seen this coming? I don’t know. Did anybody try to fucking look?

0 views

OpenAI Needs A Trillion Dollars In The Next Four Years

Shortly before publishing this newsletter, I spoke with Gil Luria, Managing Director and Analyst at D.A. Davidson , and asked him whether the capital was there to build the 17 Gigawatts of capacity that OpenAI has promised. He said the following: There is quite literally not enough money to build what OpenAI has promised. A few days ago, NVIDIA and OpenAI announced a partnership that would involve NVIDIA “investing $100 billion” into OpenAI, and the reason I put that in quotation marks is the deal is really fucking weird. Based on the text of its own announcement, NVIDIA “intends to invest up to $100 billion in OpenAI progressively as each gigawatt is deployed,” except CNBC reported a day later that “[the] initial $10 billion tranche is locked in at a $500 billion valuation and expected to close within a month or so once the transaction has been finalized,” which also adds the important detail that this deal isn’t even god damn finalized. In any case , OpenAI has now committed to building 10 Gigawatts of data center capacity at a non-specific location with a non-specific partner, so that it can unlock $10 billion of funding per gigawatt installed. I also want to be clear that it has not explained where these data centers are, or who will build them, or, crucially, who will actually fund them. The very next day, OpenAI announced five more data centers planned “as part of the Stargate initiative,” “bringing Stargate’s current planned capacity to nearly 7 gigawatts,” which is when things get a little confusing.  Altman said back in July that Oracle and OpenAI were “committed to delivering 10GW of new compute capacity for Stargate,” and they were adding an additional 4.5GW of capacity in the US on top of the 1.2GW that was already planned in Abilene Texas. In fact, the Shackelford data center that is allegedly “new” is the 1.4GW facility I talked about a few weeks ago , though I see no mention of the Wisconsin one tied to the $38 billion loan raised by Vantage Data Centers . This announcement involves a site in Doña Ana County, New Mexico and an “undisclosed location in the Midwest,” which I assume is Wisconsin, and eager members of the media could, I dunno, look it up, look up any of this stuff, use the internet to look up the news, all the news is so easily found. But here’s my favourite part of the story: To be clear, the Lordstown, Ohio site is not a data center, at least according to SoftBank , who said it will be a “data center equipment manufacturing facility.” The Milam County data center is likely this one , and it doesn’t appear that SB Energy has even broken ground. Anyway, I want to get really specific about this, because the rest of the media is reporting these stories as if these data centers will pop up overnight, and the money will magically appear, and that there will, indeed, be enough of it to go around. Based on current reports, it’s taking Oracle and Crusoe around 2.5 years per gigawatt of data center capacity. Crusoe’s 1.2GW of compute for OpenAI is a $15 billion joint venture , which means a gigawatt of compute runs about $12.5 billion. Abilene’s 8 buildings are meant to hold 50,000 NVIDIA GB200 GPUs and their associated networking infrastructure, so let’s say a gigawatt is around 333,333 Blackwell GPUs at $60,000 a piece, so about $20 billion a gigawatt.  So, each gigawatt is about $32.5 billion. For OpenAI to actually receive its $100 billion in funding from NVIDIA will require them to spend roughly $325 billion — consisting of $125 billion in data center infrastructure costs and $200 billion in GPUs.  If you’re reporting this story without at least attempting to report these numbers, you are failing to give the general public the full extent of what these companies are promising. According to the New York Times , OpenAI has “agreements in place to build more than $400 billion in data center infrastructure” but also has now promised to spend $400 billion with Oracle over the next five years. What the fuck is going on? Are we just reporting any old shit that somebody says? Oracle hasn’t even got the money to pay for those data centers! Oracle is currently raising $15 billion in bonds to get a start on…something, even though $15 billion is a drop in the bucket for the sheer scale and cost of these data centers. Thankfully, Vantage Data Centers is raising $25 billion to handle the Shackelford ( ready, at best, in mid-to-late 2027 ) and Port Washington Wisconsin ( we have no idea, it doesn’t even appear Vantage has broken ground ) data center plans, allowing Oracle to share the burden of data centers that will likely not be built until fucking 2027 at the earliest. Anyway, putting all of that aside, OpenAI has now made multiple egregious, ridiculous, fantastical and impossible promises to many different parties, in amounts ranging from $50 million to $400 billion, all of which are due within the next five years. It will require hundreds of billions of dollars — either through direct funding, loans, or having partners like Oracle or NVIDIA take the burden, though at this point I believe both companies are genuinely failing their investors by not protecting them from Clammy Sam Altman, a career liar who somehow believes he can mobilize nearly a trillion dollars and have the media print anything he says, mostly because they will print anything he says, even when he says he wants to build 1 Gigawatt of AI infrastructure a week . Today, I’m going to go into detail about every single promise made by Sam Altman and his cadre of charlatans, and give you as close to a hard dollar amount as I can as what it would cost to meet these promises. To be clear, I am aware that in some of these cases another party will take on the burden of capital — but these dollars must be raised, and OpenAI must make sure they are raised. I’ll also get into the raw costs of running OpenAI, and how dire things look when you add everything up. In fact, based on my calculations, OpenAI needs at least $500 billion just to fund its own operations, and at least $432 billion or more through partners or associated entities raising debt just to make it through the next few years. And that's if OpenAI hits the insane revenue targets it's set!

0 views

Is There Any Real Money In Renting Out AI GPUs?

NVIDIA has become a giant, unhealthy rock on which the US markets — and to some extent the US economy — sits, representing 7-to-8% of the value of the market and a large percentage of the $400 billion in expected AI data center capex expected to be spent this year, which in turn made up for more GDP growth than all consumer spending combined. I originally started writing this piece about something else entirely — the ridiculous Oracle deal, what the consequences of me being right might be, and a lot of ideas that I'll get to later, but I couldn't stop looking at what NVIDIA is doing. To be clear, NVIDIA is insane, making 88% of its massive revenues from selling the distinct GPUs and associated server hardware to underpin the inference and training of Large Language Models, a market it effectively created by acquiring Mellanox for $6.9 billion in 2019 , and its underlying hardware that allowed for the high-speed networking to connect massive banks of servers and GPUs together, a deal now under investigation by China's antitrust authorities . Since 2023, NVIDIA has made an astonishing amount of money from its data center vertical , going from making $47 billion in the entirety of their Fiscal Year 2023 to making $41.1 billion in its last quarterly earnings alone. What's even more remarkable is how little money anyone is making as a result, with the combined revenues of the entire generative AI industry unlikely to cross $40 billion this year , even when you include companies like AI compute company CoreWeave, which expects to make a little over $5 billion or so this year , though most of that revenue comes from Microsoft, OpenAI (funded by Microsoft and Google, who are paying CoreWeave to provide compute to OpenAI, despite OpenAI already being a client of CoreWeave, both under Microsoft and in their own name)...and now NVIDIA itself, which has now agreed to buy $6.3 billion of any unsold cloud compute through, I believe, the next four years. Hearing about this deal made me curious. Why is NVIDIA acting as a backstop to CoreWeave? And why are they paying to rent back thousands of its GPUs for $1.5 billion over four years from Lambda , another AI compute company it invested in? The answer is simple: NVIDIA is effectively incubating its own customers, creating the contracts necessary for them to raise debt to buy GPUs — from NVIDIA, of course — which can, in turn, be used as collateral for further loans to buy even more GPUs . These compute contracts are used by AI compute companies as a form of collateral — proof of revenue to reassure creditors that they're good for the money so that they can continue to raise mountains of debt to build more data centers to fill with more GPUs from NVIDIA. This has also created demand for companies like Dell and Supermicro, companies that accounted for a combined 39% of NVIDIA's most recent quarterly revenues . Dell and Supermicro buy GPUs sold by NVIDIA and build the server architecture around them necessary to provide AI compute, reselling them to companies like CoreWeave and Lambda, who also buy GPUs of their own and have preferential access from NVIDIA. You'll be shocked to hear that NVIDIA also invested in both CoreWeave and Lambda, that Supermicro also invested in Lambda, and that Lambda also gets its server hardware from Supermicro. While this is the kind of merciless, unstoppable capitalism that has made Jensen Huang such a success, there's an underlying problem — that these companies become burdened with massive debt, used to send money to NVIDIA, Supermicro (an AI server/architecture reseller), and Dell (another reseller that works directly with CoreWeave ), and there doesn't actually appear to be mass market demand for AI compute, other than the voracious hunger to build more of it. In a thorough review of just about everything ever written about them, I found a worrying pattern within the three major neoclouds (CoreWeave, Lambda, and Nebius): a lack of any real revenue outside of Microsoft, OpenAI, Meta, Amazon, and of course NVIDIA itself, and a growing pile of debt raised in expectation of demand that I don't believe will ever arrive. To make matters worse, I've also found compelling evidence that all three of these companies lack the capacity to actually serve massive contracts like OpenAI's $11.9 billion deal with CoreWeave ( and an additional $4 billion added a few months later ), or Nebius' $17.4 billion deal with Microsoft , both of which were used to raise debt for each company. On some level, NVIDIA's Neocloud play was genius, creating massive demand for its own GPUs, both directly and through resellers, and creating competition with big tech firms like Microsoft's Azure Cloud and Amazon Web Services, suppressing prices in cloud compute and forcing them to buy more GPUs to compete with CoreWeave's imaginary scale. The problem is that there is no real demand outside of big tech's own alleged need for compute. Across the board, CoreWeave, Nebius and Lambda have similar clients, with the majority of CoreWeave's revenue coming from companies offering compute to OpenAI or NVIDIA's own "research" compute. Neoclouds exist as an outgrowth of NVIDIA, taking on debt using GPUs as collateral , which they use to buy more GPUs, which they then use as collateral along with the compute contracts they sign with either OpenAI, Microsoft, Amazon or Google. Beneath the surface of the AI "revolution" lies a dirty secret: that most of the money is one of four companies feeding money to a company incubated by NVIDIA specifically to buy GPUs and their associated hardware. These Neoclouds are entirely dependent on a continual flow of private credit from firms like Goldman Sachs ( Nebius , CoreWeave , Lambda for its IPO ), JPMorgan ( Lambda , Crusoe , CoreWeave ), and Blackstone ( Lambda , CoreWeave ), who have in a very real sense created an entire debt-based infrastructure to feed billions of dollars directly to NVIDIA, all in the name of an AI revolution that's yet to arrive. Those billions — an estimated $50 billion a quarter for the last three quarters at least — will eventually have the expectation of some sort of return, yet every Neocloud is a gigantic money loser, with CoreWeave burning $300 million in the last quarter with expectations to spend more than $20 billion in capital expenditures in 2025 alone . At some point the lack of real money in these companies will make them unable to pay their ruinous debt, and with NVIDIA's growth already slowing, I think we're watching a private credit bubble grow with no way for any of the money to escape. I'm not sure where it'll end, but it's not going to be pretty. Let's begin.

0 views

Oracle and OpenAI Are Full Of Crap

This week, something strange happened. Oracle, a company that had just missed on its earnings and revenue estimates, saw a more-than-39% single day bump in its stock , leading a massive market rally. Why? Because it said its remaining performance obligations — contracts signed that its customers yet to pay — had increased by $317 billion from the previous quarter, with CNBC reporting at the time that this was likely part of Oracle and OpenAI's planned additional 4.5 gigawatts of data center capacity being built in the US . Analysts fawned over Oracle — again, as it missed estimates — with TD Cowen's Derrick Wood saying it was a "momentous quarter" (again, it missed ) and that these numbers were "really amazing to see," and Guggenheim Securities' John DiFucci said he was "blown away." Deutsche Bank's Brad Zelnick added that "[analysts] were all kind of in shock, in a very good way." RPOs, while standard (and required) accounting practice and based on actual signed contracts, are being used by Oracle as a form of marketing. Plans change, contracts can be canceled (usually with a kill fee, but nevertheless), and, especially in this case, clients can either not have the money to pay or die for the very same reason they can't pay. In Oracle's case, it isn’t simply promising ridiculous growth, it is effectively saying it’ll become the dominant player in all cloud compute. A day after Oracle's earnings and a pornographic day of market swings, the Wall Street Journal reported that OpenAI and Oracle had signed a $300 billion deal , starting in "2027," though the Journal neglected to say whether that was the year or Oracle’s FY2027 (which starts June 1 2026). Oracle claims that it will make $18 billion in cloud infrastructure revenue in FY2026, $32 billion in FY2027, $73 billion in FY2028, $114 billion in FY2029, and $144 billion in FY2030. While all of this isn't necessarily OpenAI (as it adds up to $381 billion), it's fair to assume that the majority of it is. This means — as the $300 billion of the $317 billion of new contracts added by Oracle, and assuming OpenAI makes up 78% of its cloud infrastructure revenue ($300 billion out of $381 billion) — that OpenAI intends to spend over $88 billion fucking dollars in compute by FY2029, and $110 billion dollars in compute, AKA nearly as much as Amazon Web Services makes in a year , in FY2030. A sidenote on percentages, and how I'm going to talk about this going forward. If I'm honest, there's also a compelling argument that more of it is OpenAI. Who else is using this much compute? Who has agreed, and why?  In any case, if you trust Oracle and OpenAI, this is what you are believing: I want to write something smart here, but I can't get away from saying that this is all phenomenally, astronomically, ridiculously stupid. OpenAI, at present, has made about $6.26 billion in revenue this year , and it leaked a few days ago that it will burn $115 billion " through 2029 ," a statement that is obviously, patently false. Let's take a look at this chart from The Information : A note on "free cash flow." Now, these numbers may look a little different because OpenAI is now leaking free cash flow instead of losses, likely because it lost $5 billion in 2024 , which included $1 billion in losses from "research compute amortization," likely referring to spreading the cost of R&D out across several years, which means it already paid it. OpenAI also lost $700 million from its revenue share with Microsoft. In any case, this is how OpenAI is likely getting its "negative $2 billion" number." Personally, I don't like this as a means of judging this company's financial health, because it's very clear it’s using it to make its losses seem smaller than they are. The Information also reports that OpenAI will, in totality, spend $350 billion in compute from here until 2030, but claims it’ll only spend $100 billion on compute in that year. If I'm honest, I believe it'll be more based on how much Oracle is projecting. OpenAI represents $300 billion of the $317 billion of new cloud infrastructure revenue it’ll from 2027 through 2030, which heavily suggests that OpenAI will be spending more like $140 billion in that year. As I'll reveal in this piece, I believe OpenAI's actual burn is over $290 billion through 2029, and these leaks were intentional to muddy the waters around how much their actual costs would be. There is no way a $116 billion burnrate from 2025 to 2029 includes these costs, and I am shocked that more people aren't doing the basic maths necessary to evaluate this company. The timing of the leak — which took place on September 5, 2025, five days before the Oracle deal was announced — always felt deeply suspicious, as it's unquestionably bad news... unless, of course, you are trying to undersell how bad your burnrate is. I believe that OpenAI's leaked free cash flow projections intentionally leave out the Oracle contract as a means of avoiding scrutiny. I refuse to let that happen. So, even if OpenAI somehow had the money to pay for its compute — it won't, but it projects, according to The Information, that it’ll make one hundred billion dollars in 2028 — I'm not confident that Oracle will actually be able to build the capacity to deliver it. Vantage Data Centers, the partner building the sites, will be taking on $38 billion of debt to build two sites in Texas and Wisconsin , only one of which has actually broken ground from what I can tell, and unless it has found a miracle formula that can grow data centers from nothing, I see no way that it can provide OpenAI with $70 billion or more of compute in FY2027. Oracle and OpenAI are working together to artificially boost Oracle's stock based on a contract that is, from everything I can see, impossible for either party to fulfill. The fact that this has led to such an egregious pump of Oracle's stock is an utter disgrace, and a sign that the markets and analysts are no longer representative of any rational understanding of a company's value. Let me be abundantly clear: Oracle and OpenAI's deal says nothing about demand for GPU compute. OpenAI is the largest user of compute in the entirety of the generative AI industry. Anthropic expects to burn $3 billion this year (so we can assume that its compute costs are $3 billion to $5 billion, Amazon is estimated to make $5 billion in AI revenue this year, so I think this is a fair assumption ), and xAI burns through a billion dollars a month . CoreWeave expects about $5.3 billion of revenue in 2025 , and per The Information Lambda, another AI compute company, made more than $250 million in the first half of 2025. If we assume that all of these companies were active revenue participants (we shouldn't, as xAI mostly handles its own infrastructure), I estimate the global compute market is about $40 billion in totality, at a time when AI adoption is trending downward in large companies according to Apollo's Torsten Sløk . And yes, Nebius signed a $17.4 billion, four-year-long deal with Microsoft , but Nebius now has to raise $3 billion to build the capacity to acquire "additional compute power and hardware, [secure] land plots with reliable providers, and [expand] its data center footprint," because Nebius, much like CoreWeave, and, much like Oracle, doesn't have the compute to service these contracts. All three have seen a 30% bump in their stock in the last week. In any case, today I'm going to sit down and walk you through the many ways in which the Oracle and OpenAI deal is impossible to fulfill for either party. OpenAI is projecting fantastical growth in an industry that's already begun to contract, and Oracle has yet to even start building the data centers necessary to provide the compute that OpenAI allegedly needs.

0 views

Why Everybody Is Losing Money On AI

Hello and welcome to another premium newsletter. Thanks as ever for subscribing, and please email me at [email protected] to say hello. As I've written again and again , the costs of running generative AI do not make sense. Every single company offering any kind of generative AI service — outside of those offering training data and services like Turing and Surge — is, from every report I can find, losing money, and doing so in a way that heavily suggests that there's no way to improve their margins. In fact, let me explain an example of how ridiculous everything has got, using points I'll be repeating behind the premium break. Anysphere is a company that sells a subscription to their AI coding app Cursor, and said app predominantly uses compute from Anthropic via their models Claude Sonnet 4.1 and Opus 4.1. Per Tom Dotan at Newcomer , Cursor sends 100% of their revenue to Anthropic, who then takes that money and puts it into building out Claude Code, a competitor to Cursor. Cursor is Anthropic's largest customer. Cursor is deeply unprofitable, and was that way even before Anthropic chose to add "Service Tiers," jacking up the prices for enterprise apps like Cursor . My gut instinct is that this is an industry-wide problem. Perplexity spent 164% of its revenue in 2024 between AWS, Anthropic and OpenAI . And one abstraction higher (as I'll get into), OpenAI spent 50% of its revenue on inference compute costs alone , and 75% of its revenue on training compute too (and ended up spending $9 billion to lose $5 billion). Yes, those numbers add up to more than 100%, that's my god damn point. Large Language Models are too expensive, to the point that anybody funding an "AI startup" is effectively sending that money to Anthropic or OpenAI, who then immediately send that money to Amazon, Google or Microsoft, who are yet to show that they make any profit on selling it. Please don't waste your breath saying "costs will come down." They haven't been, and they're not going to. Despite categorically wrong boosters claiming otherwise , the cost of inference — everything that happens from when you put a prompt in to generate an output from a model — is increasing, in part thanks to the token-heavy generations necessary for "reasoning" models to generate their outputs, and with reasoning being the only way to get "better" outputs, they're here to stay (and continue burning shit tons of tokens). This has a very, very real consequence. Christopher Mims of the Wall Street Journal reported last week that software company Notion — which offers AI that boils down to "generate stuff, search, meeting notes and research" — had AI costs eat 10% of its profit margins to provide literally the same crap that everybody else does. As I discussed a month or two ago, the increasing cost of AI has begun a kind of subprime AI crisis , where Anthropic and OpenAI are having to charge more for their models and increasing the price on their enterprise customers to boot. As discussed previously, OpenAI lost $5 billion and Anthropic $5.3 billion in 2024, with OpenAI expecting to lose upwards of $8 billion and Anthropic — somehow — only losing $3 billion in 2025. I have severe doubts that these numbers are realistic, with OpenAI burning at least $3 billion in cash on salaries this year alone , and Anthropic somehow burning two billion dollars less on revenue that has, if you believe its leaks, increased 500% since the beginning of the year . Though I can't say for sure, I expect OpenAI to burn at least $15 billion in compute costs this year alone , and wouldn't be surprised if its burn was $20 billion or more. At this point, it's becoming obvious that it is not profitable to provide model inference, despite Sam Altman recently saying that OpenAI was. He no doubt is trying to play silly buggers with the concept of gross profit margins — suggesting that inference is "profitable" as long as you don't include training, staff, R&D, sales and marketing, and any other indirect costs. I will also add that OpenAI pays a discounted rate on its compute . In any case, we don't even have one — literally one — profitable model developer, one company that was providing these services that wasn't posting a multi-million or billion-dollar loss. In fact, even if you remove the cost of training models from OpenAI's 2024 revenues ( provided by The Information ), OpenAI would still have lost $2.2 billion fucking dollars. One of you will say "oh, actually, this is standard accounting." If that's the case, OpenAI had a 10% gross profit margin in 2024, and while OpenAI has leaked that it has a 48% gross profit margin in 2025 , Altman also claimed that GPT-5 scared him, comparing it to the Manhattan Project . I do not trust him. Generative AI has a massive problem that the majority of the tech and business media has been desperately avoiding discussing: that every single company is unprofitable, even those providing the models themselves. Reporters have spent years hand-waving around this issue, insisting that "these companies will just work it out," yet never really explaining how they'd do so other than "the cost of inference will come down" or "new silicon will bring down the cost of compute." Neither of these things have happened, and it's time to take a harsh look at the rotten economics of the Large Language Model era. Generative AI companies — OpenAI and Anthropic included — lose millions or billions of dollars, and so do the companies building on top of them, in part because the costs associated with delivering models continue to increase. Integrating Large Language Models into your product already loses you money, at a price where the Large Language Model provider (EG: OpenAI and Anthropic) is losing money. I believe that generative AI is, at its core, unprofitable, and that no company building their core services on top of models from Anthropic or OpenAI has a path to profitability outside of massive, unrealistic price increases. The only realistic path forward for generative AI firms is to start charging their users the direct costs for running their services, and I do not believe users will be enthusiastic to do so, because the amount of compute that the average user costs vastly exceeds the amount of money that the company generates from a user each month. As I'll discuss, I don't believe it's possible for these companies to make a profit even with usage-based pricing, because the outcomes that are required to make things like coding LLMs useful require a lot more compute than is feasible for an individual or business to pay for. I will also go into how ludicrous the economics behind generative AI have become, with companies sending 100% or more of their revenue directly to cloud compute or model providers. And I'll explain why, at its core, generative AI is antithetical to the way that software is sold, and why I believe this doomed it from the very beginning.

0 views

AI Bubble 2027

Soundtrack: The Dillinger Escape Plan - One Of Us Is The Killer An MIT study found that 95% of organizations are getting "zero return" from generative AI, seemingly every major outlet is now writing an "are we in a bubble?" story, and now Meta has frozen AI hiring . Things are looking bleak for the AI bubble, and people are getting excited that this loathsome, wasteful era might be coming to an end. As a result, I'm being asked multiple times a day when the bubble might burst. I admit I'm hesitant to make any direct timelines — we are in a time (to quote former Federal Reserve Board chairman Alan Greenspan) of irrational exuberance , where the markets have oriented themselves around something very silly and very, very expensive. Regardless, Anthropic is apparently raising as much as $10 billion in its next funding round , and OpenAI allegedly hit $1 billion in revenue in July , which brings it in line with my estimate that it’s made about $5.26 billion in revenue in 2025 so far , The bubble "bursting" is not Sam Altman declaring that we're in a bubble — it's a series of events that lead to big tech pulling away from this movement, investment money drying up, and the current slate of AI companies withering and dying because they can't raise more money, can't go public, and can't sell to anybody . Anyway, earlier in the year, a bunch of credulous oafs wrote an extremely long piece of fan fiction called " AI 2027 ," beguiling people like Kevin Roose with its "gloominess." Written with a deep seriousness and a lot of charts, AI 2027 makes massive leaps of logic, with its fans rationalizing taking it seriously by saying that the five authors "have the right credentials." In reality, AI 2027 is written to fool people that want to be fooled and scare people that are already scared, its tone consistently authoritative as it suggests that a self-learning agent is on the verge of waking up, a thing that is so remarkably stupid that anyone who took this seriously should be pantsed again and again. These men are also cowards. They choose to use fake company names like "OpenBrain" to mask their predictions instead of standing behind them with confidence. I get that extrapolating years into the future is scary — but these grifting losers can't even commit to a prediction! Nevertheless, I wanted to take a run at something similar myself, though not in the same narrative format. In this piece, I'm going to write out what conditions I believe will burst the bubble. Some of this will be extrapolations based on my own knowledge, sources, and writing hundreds of thousands of words on this subject. I am not going to write a strict timeline, but I am going to write how some things could go, and how (and why) they'll lead to the bubble bursting. For the bubble to burst, I see a few necessary conditions, though reality is often far more boring and annoying. Regardless, it's important to know that this is a bubble driven by vibes not returns, and thus the bubble bursting will be an emotional reaction. This post is, in many respects, a follow-on to my previous “pale horse” article (called Burst Damage) . Many of my original pale horses have already come true — Anthropic and OpenAI have pushed price increases and rate limits, there is already discord in AI investment, Meta is already considering downsizing its AI team , and OpenAI has done multiple different "big, stupid magic tricks," the chief of them being the embarrassing launch of GPT-5, a "smart, efficient router" that I reported last week was quite the opposite . This time, I'm going to write out the linchpin events that will shock the system, and how they might bring about the bubble bursting. I should also be clear, and I will get to after the premium break, that this will be a series of events rather than one big one, though there are big ones to look out for. I also think it might "take a minute," because "the bubble bursting" will be a succession of events that could take upwards of a year to fully happen. That’s been true for every bubble. Although people associate the implosion of the housing bubble with the “one big event” of Lehman Brothers collapsing in 2008, the reality is that it was preceded and followed by a bunch of other events, equally significant though not quite as dramatic.  One VC (who you'll read about shortly) predicted it will take 6 quarters to run out of funding entirely based on the current rate of investment, putting us around February 2027 for things to have truly collapsed. Here's a list of some of the things that I believe will have to happen for this era to be truly done, and the ones that I believe are truly essential. The common thread through all of these points is that they are predominantly impossible to ignore. So far this bubble has inflated because the problems with AI — such as "it doesn't make any money" and "burns billions of dollars" — have been dismissed until very recently as the necessary costs of the beautiful AI revolution. Now that things have begun to unravel, the intensity of criticism will increase gradually, rather than in one big movement that makes everyone say "we hate AI." And it isn't just because of the money. CEOs like Tobias Lutke of Shopify have oriented their companies’ entire culture around AI , demanding in his case that "employees must demonstrate why AI cannot be used before requesting additional resources." Generative AI is, on some level, a kind of dunce detector — its flimsy and vague use cases having enough juice to impress the clueless Business Idiots who don't really engage with the production that makes their companies money. The specious, empty hype of Large Language Models — driven by a tech and business media that has given up on trying to understand them — symbolizes a kind of magic to these empty-headed goobers, and unwinding their "AI-first" cultures will be difficult...right up until the first guy does it, at which point everybody will follow. AI has taken such a hold on our markets because it's symbolic of a few things: In any case, I am going to try and write the things that I think will happen, in detail. I'll go into more conditions in this piece, and as discussed, I'm going to make some informed guesses, extrapolations, and give my thoughts about how things collapse. I predict that the impact of Large Language Models over the next decade will be enormous, not in its actual innovation or returns, but in its ability to expose how little our leaders truly know about the world or labor, how willing many people are to accept whatever the last thing a smart-adjacent person said, and how our markets and economy are driven by people with the most tenuous grasp on reality. This will be an attempt to write down what I believe could happen in the next 18 months, the conditions that might accelerate the collapse, and how the answers to some of my open questions — such as how these companies book revenue and burn compute — could influence outcomes. This...is AI Bubble 2027.

0 views

How To Argue With An AI Booster

Editor's Note: For those of you reading via email, I recommend opening this in a browser so you can use the Table of Contents. This is my longest newsletter - a 16,000-word-long opus - and if you like it, please subscribe to my premium newsletter. Thanks for reading! In the last two years I've written no less than 500,000 words, with many of them dedicated to breaking both existent and previous myths about the state of technology and the tech industry itself. While I feel no resentment — I really enjoy writing, and feel privileged to be able to write about this and make money doing so — I do feel that there is a massive double standard between those perceived as "skeptics" and "optimists." To be skeptical of AI is to commit yourself to near-constant demands to prove yourself, and endless nags of " but what about ?" with each one — no matter how small — presented as a fact that defeats any points you may have. Conversely, being an "optimist" allows you to take things like AI 2027 — which I will fucking get to — seriously to the point that you can write an entire feature about fan fiction in the New York Times and nobody will bat an eyelid. In any case, things are beginning to fall apart. Two of the actual reporters at the New York Times (rather than a "columnist") reported out last week that Meta is yet again "restructuring" its AI department for the fourth time, and that it’s considering "downsizing the A.I. division overall," which sure doesn't seem like something you'd do if you thought AI was the future. Meanwhile, the markets are also thoroughly spooked by an MIT study covered by Fortune that found that 95% of generative AI pilots at companies are failing , and though MIT NANDA has now replaced the link to the study with a Google Form to request access, you can find the full PDF here , in the kind of move that screams "PR firm wants to try and set up interviews." Not for me, thanks! In any case, the report is actually grimmer than Fortune made it sound, saying that "95% of organizations are getting zero return [on generative AI]." The report says that "adoption is high, but transformation is low," adding that "...few industries show the deep structural shifts associated with past general-purpose technologies such as new market leaders, disrupted business models, or measurable changes in customer behavior." Yet the most damning part was the "Five Myths About GenAI in the Enterprise," which is probably the most wilting takedown of this movement I've ever seen: These are brutal, dispassionate points that directly deal with the most common boosterisms. Generative AI isn't transforming anything, AI isn't replacing anyone, enterprises are trying to adopt generative AI but it doesn't fucking work , and the thing holding back AI is the fact it doesn't fucking work. This isn't a case where "the enterprise" is suddenly going to save these companies, because the enterprise already tried, and it isn't working. An incorrect read of the study has been that the "learning gap" that makes these things less useful, when the study actually says that "...the fundamental gap that defines the GenAI divide [is that users resist tools that don't adapt, model quality fails without context, and UX suffers when systems can't remember." This isn't something you learn your way out of. The products don't do what they're meant to do, and people are realizing it. Nevertheless, boosters will still find a way to twist this study to mean something else. They'll claim that AI is still early, that the opportunity is still there, that we " didn't confirm that the internet or smartphones were productivity boosting ," or that we're in "the early days" of AI, somehow, three years and hundreds of billions and thousands of articles in. I'm tired of having the same arguments with these people, and I'm sure you are too. No matter how much blindly obvious evidence there is to the contrary they will find ways to ignore it. They continually make smug comments about people "wishing things would be bad" or suggesting you are stupid — and yes, that is their belief! — for not believing generative AI is disruptive.  Today, I’m going to give you the tools to fight back against the AI boosters in your life. I’m going to go into the generalities of the booster movement — the way they argue, the tropes they cling to, and the ways in which they use your own self-doubt against you.  They’re your buddy, your boss, a man in a gingham shirt at Epic Steakhouse who won't leave you the fuck alone, a Redditor, a writer, a founder or a simple con artist — whoever the booster in your life is, I want you to have the words to fight them with.  So, this is my longest newsletter ever, and I built it for quick reference - and, for the first time, gave you a Table of Contents. So, an AI booster is not, in many cases, an actual fan of artificial intelligence. People like Simon Willison or Max Woolf who actually work with LLMs on a daily basis don’t see the need to repeatedly harass everybody, or talk down to them about their unwillingness to pledge allegiance to the graveyard smash of generative AI. In fact, the closer I’ve found somebody to actually building things with LLMs , the less likely they are to emphatically argue that I’m missing out by not doing so myself. No, the AI booster is symbolically aligned with generative AI. They are fans in the same way that somebody is a fan of a sports team, their houses emblazoned with every possible piece of tat they can find, their Sundays living and dying by the success of the team, except even fans of the Dallas Cowboys have a tighter grasp on reality. Kevin Roose and Casey Newton are two of the most notable boosters, and — as I’ll get into later in this piece — neither of them have a consistent or comprehensive knowledge of AI. Nevertheless, they will insist that “ everybody is using AI for everything ” — a statement that even a booster should realize is incorrect based on the actual abilities of the models.  But that’s because it isn’t about what’s actually happening , it’s about allegiance. AI symbolizes something to the AI booster — a way that they’re better than other people, that makes them superior because they (unlike “cynics” and “skeptics”) are able to see the incredible potential in the future of AI, but also how great it is today , though they never seem to be able to explain why outside of “it replaced search for me!” and “I use it to draw connections between articles I write,” which is something I do without AI using my fucking brain. Boosterism is a kind of religion, interested in finding symbolic “proof” that things are getting “better” in some indeterminate way, and that anyone that chooses to believe otherwise is ignorant.  I’ll give you an example. Thomas Ptacek’s “ My AI Skeptic Friends Are All Nuts ” was catnip for boosters — a software engineer using technical terms like “interact with Git” and “MCP,” vague charts, and, of course, an extremely vague statement that says hallucinations aren’t a problem: Is it?  Anyway, my favourite part of the blog is this: Nobody projects more than an AI booster. They thrive on the sense they’re oppressed and villainized after years of seemingly every outlet claiming they’re right regardless of whether there’s any proof. They sneer and jeer and cry constantly that people are not showing adequate amounts of awe when an AI lab says “ we did something in private, we can’t share it with you, but it’s so cool ,” and constantly act as if they’re victims as they spread outright misinformation, either through getting things wrong or never really caring enough to check.  Also, none of the booster arguments actually survive a thorough response, as Nik Suresh proved with his hilarious and brutal takedown of Ptacek’s piece . There are, I believe, some people who truly do love using LLMs, yet they are not the ones defending them. Ptacek’s piece drips with condescension, to the point that it feels like he’s trying to convince himself how good LLMs are, and because boosters are eternal victims, he wrote them a piece that they could send around to skeptics and say “heh, see?” without being able to explain why it was such a brutal takedown, mostly because they can’t express why other than “ well this guy gets it!”   One cannot be the big, smart genius that understands the glory and power of AI while also acting like a scared little puppy every time somebody tells them it sucks. In fact, that’s a great place to start. When you speak to an AI booster, you may get the instinct to shake them vigorously, or respond to their post by saying to do something with your something , or that they’re “stupid.” I understand the temptation, but you want to keep a head on a swivel — they thrive on victimisation.   I’m sorry if you are an AI booster and this makes you feel bad. Please reflect on your work and how many times you’ve referred to somebody who didn’t understand AI in a manner that suggested they were ignorant, or tried to gaslight them by saying “AI was powerful” while providing no actionable ways in which it is. You cannot — and should not! — allow these people to act as if they are being victimized or “othered.”  First and foremost: there are boosters at pretty much every major think tank, government agency and media outlet. It’s extremely lucrative being a booster. You’re showered with panel invites, access to executives, and are able to get headlines by saying how scared you are of the computer with ease. Being a booster is the easy path! Being a critic requires you to constantly have to explain yourself in a way that boosters never have to.  If a booster says this to you, ask them to explain: There is no answer here, because this is not a coherent point of view. Boosters are more successful, get more perks and are in general better-treated than any critic. Fundamentally, these people exist in the land of the vague. They will drag you toward what's just on the horizon, but never quite define what the thing that dazzles you will be, or when it will arrive.  Really, their argument comes down to one thought: you must get on board now, because at some point it'll be so good you'll feel stupid for not believing something that kind of sucks wouldn't be really good.  If this line sounds familiar, it’s because you’ve heard it a million times before, most notably with crypto.  They will make you define what would impress you, which isn't your job, in the same way finding a use case for them isn't your job. In fact, you are the customer! Here’s a great place to start: say “that’s a really weird thing to say!” It is peculiar to suggest that somebody doesn’t get how to use a product, and that we, as the customer, must justify ourselves to our own purchases. Make them justify their attitude.  Just like any product, we buy software to serve a need.   This is meant to be artificial *intelligence* — why is it so fucking stupid that I have to work out why it's useful? The answer, of course, is that it has no intellect, is not intelligent, and Large Language Models are being pushed up a mountain by a cadre of people who are either easily impressed or invested — either emotionally or financially — in its success due to the company they keep or their intentions for the world.  If a booster suggests you “just don’t get it,” ask them to explain: Their use cases will likely be that AI has replaced search for them, that they use it for brainstorming or journaling, proof-reading an article, or looking through a big pile of their notes (or some other corpus of information) and summarizing it or pulling out insights.  If a booster refers to AI “being powerful” and getting “more powerful,” ask them: The core of the AI booster’s argument is to make you feel bad. They will suggest you are intentionally not liking A.I. because you're a hater, or a cynic, or a Luddite. They will suggest that you are ignorant for not being amazed by ChatGPT. To be clear, anyone with a compelling argument doesn’t have to make you feel bad to convince you. The iPhone - and to be clear, I am referring to the concept of the smartphone and its utility, I am aware that there was marketing for the iPhone - didn’t need a fucking marketing campaign to explain why one device that can do a bunch of things you already find useful was good.  You don't have to be impressed by ANYTHING by default, and any product — especially software — designed to make you feel stupid for "not getting it" is poorly designed. ChatGPT is the ultimate form of Silicon Valley Sociopathy — you must do the work to find the use cases, and thank them for being given the chance to do so.  A.I. is not even good, reliable software! It resembles the death of the art of technology — inconsistent and unreliable by definition, inefficient by design, financially ruinous, and ADDS to the cognitive load of the user by requiring them to be ever-vigilant.  So, here’s a really easy way to deal with this : if a booster ever suggests you are stupid or ignorant, ask them why it’s necessary to demean you to get their point across! Even if you are unable to argue on a technical level, make them explain why the software itself can’t convince you. Boosters will do everything they can to pull you off course. If you say that none of these companies make money, they’ll say it’s the early days. If you say AI companies burn billions, they’ll say the cost of inference is coming down. If you say the industry is massively overbuilding, they’ll say that this is actually just like the dot com boom and that the infrastructure will be picked up and used in the future. If you say there are no real use cases, they’ll say that ChatGPT has 700 million weekly users.  Every time there’s the same god damn arguments, so I’ve sat down and written as many of them as I can think of. Print this and feed it to your local booster today. Anytime a booster says “AI will,” tell them to stop and explain what AI can do, and if they insist, ask them both when to expect the things they’re talking about, and if they say “very soon,” ask them to be more specific. Get them to agree to a date, then call them on that date. There’s that “will” bullshit again. Agents don’t work! They don’t work at all. The term “agent” means, to quote Max Woolf , “a workflow where the LLM can make its own decisions, [such as in the case of] web search [where] the LLM is told “you can search the web if you need to” then can output “I should search the web” and do so.” Yet “agent” has now become a mythical creature that means “totally autonomous AI that can do an entire job.” if anyone tells you “agents are…” you should ask them to point to one. If they say “coding,” please demand that they explain how autonomous these things are, and if they say that they can “refactor entire codebases,” ask them what that means, and also laugh at them.  Here’s a comprehensive rundown , but here’s a particularly important part: Long story short, agents are not autonomous, they do not replace jobs, they cannot “replace coders,” they are not going to do so because probabilistic models are a horrible means of taking precise actions, and almost anyone who brings up agents as a booster is either misinformed or in the business of misinformation. Let's start with a really simple question: what does this actually mean? In many cases, I think they're referring to AI as being "like the early days of the internet." "The early days of the internet" can refer to just about anything. Are we talking about dial-up? DSL? Are we talking about the pre-platform days when people accessed it via Compuserve or AOL? Yes, yes, I remember that article from Newsweek, I already explained it here : In any case, one guy saying that the internet won't be big doesn't mean a fucking thing about generative AI and you are a simpleton if you think it does. One guy being wrong in some way is not a response to my work. I will crush you like a bug. If your argument is that the early internet required expensive Sun Microsystems servers to run, Jim Covello of Goldman Sachs addressed that by saying that the costs "pale in comparison ," adding that we also didn't need to expand our power grid to build the early Web. This is a straight-up lie. Sorry! Also, as Jim Covello noted , there were hundreds of presentations in the early 2000s that included roadmaps that accurately fit how smartphones rolled out, and that no such roadmap exists for generative AI. The iPhone was also an immediate success as a thing that people paid for , with Apple selling four million units in the space of six months . Hell, in 2006 (the year before the iPhone launch), there was an estimated 17.7 million worldwide smartphone shipments (mostly from BlackBerry and other companies building on Windows Mobile, with Palm vacuuming up the crumbs), though to be generous to the generative AI boosters, I’ll disregard those.  The original Attention Is All You Need paper — the one that kicked off the transformer-based Large Language Model era — was published in June 2017. ChatGPT launched in November 2022. Nevertheless, if we're saying "early days" here, we should actually define what that means. As I mentioned above , people paid for the iPhone immediately, despite it being a device that was completely and utterly new. While there was a small group of consumers that might have used similar devices ( like the iPAQ ), this was a completely new kind of computing, sold at a premium, requiring you to have a contract with a specific carrier (Cingular, now known as AT&T). Conversely, ChatGPT's "annualized" revenue in December 2023 was $1.6 billion (or $133 million a month), for a product that had, by that time, raised over $10 billion , and while we don't know what OpenAI lost in 2023, reports suggest it burned over $5 billion in 2024 . Big tech has spent over $500 billion in capital expenditures in the last 18 months , and all told — between investments of cloud credits and infrastructure — will likely sink over $600 billion by year's-end. The "early days" of the internet were defined not by its lack of investment or attention, but by its obscurity. Even in 2000 — around the time of the dot-com bubble — only 52% of US adults used the internet , and it would take another 19 years for 90% of US adults to do so. These early days were also defined by its early functionality. The internet would become so much more because of the things that hyper-connectivity allowed us to do, and both faster internet connections and the ability to host software in the cloud would change, well, everything. We could define what “better” would mean, and make reasonable predictions about what people could do on a “better” internet.  Yet even in those early days, it was obvious why you were using the internet, and how it might grow from there. One did not have to struggle to explain why buying a book online might be useful, or why a website might be a quicker reference than having to go to a library, or why downloading a game or a song might be a good idea. While habits might have needed adjusting, it was blatantly obvious what the value of the early internet was. It's also unclear when the early days of the internet ended. Only 44% of US adults had access to broadband internet by 2006 . Were those the early days of the internet?  The answer is "no," and that this point is brought up by people with a poor grasp of history and a flimsy attachment to reality. The early days of the internet were very, very different to any associated tech boom since, and we need to stop making the comparison. The internet also grew in a vastly different information ecosystem. Generative AI has had the benefit of mass media — driven by the internet! — along with social media (and social pressure) to "adopt AI" for multiple years. According to Pew, as of mid-June 2025, 34% of US adults have used ChatGPT , with 79% saying they had "heard at least a little about it." Furthermore, ChatGPT has always had a free version. On top of that, a study from May 2023 found that over 10,900 news headlines mentioned ChatGPT between November, 2022 and March, 2023 , and a BrandWatch report found that in the first five months of its release, ChatGPT received over 9.24 million mentions on social media . Nearly 80% of people have heard of ChatGPT, and over a quarter of Americans have used it. If we're defining "the early days" based on consumer exposure, that ship has sailed. If we're defining "the early days" by the passage of time , it's been 8 years since Attention Is All You Need, and three since ChatGPT came out. While three years might not seem like a lot of time, the whole foundation of an "early days" argument is that in the early days, things do not receive the venture funding, research, attention, infrastructural support or business interest necessary to make them "big." In 2024, nearly 33% of all global venture funding went to artificial intelligence , and according to The Information, AI startups have raised over $40 billion in 2025 alone , with Statista adding that AI absorbed 71% of VC funding in Q1 2025 . These numbers also fail to account for the massive infrastructure that companies like OpenAI and Anthropic don't have to pay for. The limitations of the early internet were two-fold: In generative AI's case, Microsoft, Google, and Amazon have built out the "fiber optic cables" for Large Language Models. OpenAI and Anthropic have everything they need. They have (even if they say otherwise) plenty of compute, access to the literal greatest minds in the field, the constant attention of the media and global governments, and effectively no regulations or restrictions stopping them from training their models on the works of millions of people , or destroying our environment . They have already had this support. OpenAI was allowed to burn half a billion dollars on a training run for GPT-4.5 and 5 . If anything, the massive amounts of capital have allowed us to massively condense the time in which a bubble goes from "possible" to "bursting and washing out a bunch of people," because the tech industry has such a powerful follower culture that only one or two unique ideas can exist at one time. The "early days" argument hinges on obscurity and limited resources, something that generative AI does not get to whine about. Companies that make effectively no revenue can raise $500 million to do the same AI coding bullshit that everybody else does. In simpler terms, these companies are flush with cash, have all the attention and investment they could possibly need, and are still unable to create a product with a defined, meaningful, mass-market use case. In fact, I believe that thanks to effectively infinite resources, we've speed-run the entire Large Language Model era, and we're nearing the end. These companies got what they wanted. Bonus trick: ask them to tell you what “the fiber boom” was. So, a little history. The "fiber boom" began after the telecommunications act of 1996 deregulated large parts of America's communications infrastructure, creating a massive boom — a $500 billion one to be precise, primarily funded with debt : In one sense, explaining what happened to the telecom sector is very simple: the growth in capacity has vastly outstripped the growth in demand. In the five years since the 1996 bill became law, telecommunications companies poured more than $500 billion into laying fiber optic cable, adding new switches, and building wireless networks. So much long-distance capacity was added in North America, for example, that no more than two percent is currently being used. With the fixed costs of these new networks so high and the marginal costs of sending signals over them so low, it is not a surprise that competition has forced prices down to the point where many firms have lost the ability to service their debts. No wonder we have seen so many bankruptcies and layoffs. This piece, written in 2002 , is often cited as a defense against the horrifying capex associated with generative AI, as that fiber optic cable has been useful for delivering high-speed internet. Useful, right? This period was also defined by a gluttony of over-investment, ridiculous valuations and outright fraud . In any case, this is not remotely the same thing and anyone making this point needs to learn the very fucking basics of technology. GPUs are built to shove massive amounts of compute into one specific function again and again, like generating the output of a model (which, remember, mostly boils down to complex maths). Unlike CPUs, a GPU can't easily change tasks, or handle many little distinct operations, meaning that these things aren't going to be adopted for another mass-scale use case. In simpler terms, this was not an infrastructure buildout. The GPU boom is a heavily-centralized, capital expenditure-funded asset bubble where a bunch of chips will sit in warehouses waiting for somebody to make up a use case for them, and if an endearing one existed, we'd already have it because we already have all the fucking GPUs. You are describing fan fiction. AI 2027 is fan fiction. Anyone who believes in it is a mark! It doesn’t matter if all of the people writing the fan fiction are scientists, or that they all have “the right credentials.” They themselves say that AI 2027 is a “guess,” an “extrapolation” (guess)  with “expert feedback” (someone editing your fan fiction), and involves “experience at OpenAI” (there are people that worked on the shows they write fan fiction about).  I am not going to go line-by-line to cut this apart anymore than I am going to write a lengthy takedown of someone’s erotic Banjo Kazooie story, because both are fictional. The entire premise of this nonsense is that at one point someone invents a self-learning “agent” that teaches itself stuff, and it does a bunch of other stuff as a result, with different agents with different numbers after them. There is no proof this is possible, nobody has done it, nobody will do it. AI 2027 was written specifically to fool people that wanted to be fooled, with big charts and the right technical terms used to lull the credulous into a wet dream and New York Times column where one of the writers folds their hands and looks worried. It was also written to scare people that are already scared. It makes big, scary proclamations, with tons of links to stuff that looks really legitimate but, when you piece it all together, is still fan fiction.  My personal favourite part is “Mid 2026: China Wakes Up,” which involves China’s intelligence agencies trying to steal OpenBrain’s agent (no idea who this company could be referring to, I’m stumped!), before the headline of “AI Takes Some Jobs” after OpenBrain released a model oh god I am so bored even writing up this tripe!  Sarah Lyons put it well , arguing that AI 2027 (and AI in general) is no different from the spurious “ spectral evidence ” used to accuse someone of being a witch during the Salem Witch Trials: Anyway, AI 2027 is fan fiction, nothing more, and just because it’s full of fancy words and has five different grifters on its byline doesn’t mean anything. Bonus trick: Ask them to explain whether things have actually got cheaper, and if they say they have, ask them why there are no profitable AI companies. If they say “they’re currently in growth stage,” ask them why there are no profitable AI companies. At this point they should try and kill you. In an interview on a podcast from earlier in the year , journalist Casey Newton said the following about my work: Newton then says — several octaves higher, showing how mad he isn't — that "[he] thought what [he] said was very civil" and that there are "things that are true and there are things that are false, like you can choose which ones you wanna believe." I am not going to be so civil. Other than the fact that Casey refers to "micro-innovations" (?) and "DeepSeek being on a curve that was expected," he makes — as many do — two very big mistakes, ones that I personally would not have said in a sentence that begun with suggesting that I knew how the technology works.  Inference — and I've gotten this one wrong in the past too! — is everything that happens from when you put a prompt in to generate an output. It's when an AI, based on your prompt, "infers" meaning. To be more specific, and quoting Google , "...machine learning inference is the process of running data points into a machine learning model to calculate an output such as a single numerical score." Casey will try and weasel out of this one and say this is what he meant. It wasn't.  Casey, like many people who talk about stuff without learning about it first, is likely referring to the fact that the price of tokens for some models has gone down in some cases. The problem, however, is that these are raw token costs, not actual expressions or evaluations of token burn in a practical setting. Worse still… Well, the cost of inference actually went up . In an excellent blog for Kilocode , Ewa Szyszka explained: Token consumption per application grew a lot because models allowed for longer context windows and bigger suggestions from the models. The combination of a steady price per token and more token consumption caused app inference costs to grow about 10x over the last two years. To explain in really simple terms, while the costs of old models may have decreased, new models cost about the same, and the "reasoning" that these models do actually burn way, way more tokens. When these new models "reason," they break a user's input and break into component parts, then run inference on each one of those parts. When you plug an LLM into an AI coding environment, it will naturally burn an absolute ton of tokens, in part because of the large amount of information you have to load into the prompt (and the "context window," or the information you load in with your prompt, with token burn increasing with the size of that information), and in part because generating code is inference-intensive. In fact, the inference costs are so severe that Szyszka says that "... combination of a steady price per token and more token consumption caused app inference costs to grow about 10x over the last two years ." I refuse to let this point go, because people love to say "the cost of inference is going down" when the cost of inference has increased, and they do so to a national audience, all while suggesting I am wrong somehow. I am not wrong. In fact, software development influencer Theo Browne recently put out a video called " I was wrong about AI costs (they keep going up) ," which he breaks down as follows: The price drops have, for the most part, stopped. See the below chart from The Information : You cannot, at this point, fairly evaluate whether a model is "cheaper" just based on its cost-per-tokens, because reasoning models are inherently built to use more tokens to create an output. Reasoning models are also the only way that model developers have been able to improve the efficacy of new models, using something called "test-time compute" to burn extra tokens to complete a task. And in basically anything you're using today, there's gonna be some sort of reasoning model, especially if you're coding. The cost of inference has gone up. Statements otherwise are purely false, and are the opinion of somebody who does not know what he's talking about. ...maybe? It sure isn't trending that way, nor has it gone down yet. I also predict that there's going to be a sudden realization in the media that it's going up, which has kind of already started. The Information had a piece recently about it , where they note that Intuit paid $20 million to Azure last year (primarily for access to OpenAI's models), and is on track to spend $30 million this year, which "outpaces the company's revenue growth in the same period, raising questions about how sustainable the spending is and how much of the cost it can pass along to customers." The problem here is that the architecture underlying Large Language Models is inherently unreliable. I imagine OpenAI's introduction of the router to ChatGPT-5 is an attempt to moderate both the costs of the model chosen and reduce the amount of exposure to reasoning models for simple queries — though Altman was boasting on August 10th about the "significant increase" in both free and paid users' exposure to reasoning models . Worse still, a study written up by VentureBeat found that open-weight models burn between 1.5 to 4 times more tokens, in part due to a lack of token efficiency, and in particular thanks to — you guessed it! — reasoning models: The findings challenge a prevailing assumption in the AI industry that open-source models offer clear economic advantages over proprietary alternatives. While open-source models typically cost less per token to run, the study suggests this advantage can be “easily offset if they require more tokens to reason about a given problem.” And models keep getting bigger and more expensive, too.  Because model developers hit a wall of diminishing returns, and the only way to make their models do more was to make them burn more tokens to generate a more accurate response (this is a very simple way of describing reasoning, a thing that OpenAI launched in September 2024 and others followed). As a result, all the "gains" from "powerful new models" come from burning more and more tokens. The cost-per-million-token number is no longer an accurate measure of the actual costs of generative AI, because it's much, much harder to tell how many tokens a reasoning model may burn, and it varies (as Theo Browne noted) from model to model. In any case, there really is no changing this path. They are out of ideas. So, I've heard this argument maybe 50 times in the last year, to the point that I had to talk about it in my July 2024 piece " How Does OpenAI Survive ." Nevertheless, people make a few points about Uber and AI that I think are fundamentally incorrect, and I'll break them down for you. I've seen this argument a lot, and it's one that's both ahistorical and alarmingly ignorant of the very basics of society. So, OpenAI got a $200 million defense contract with an "estimated completion date of July 2026," and is selling ChatGPT Enterprise to the US government for a dollar a year ( along with Anthropic, which sells access to Claude for the same price, Even Google is undercutting them, selling Gemini access at 47 cents for a year ). You're probably reading that and saying "oh no, that means the government has paid them now, they're never going away, " and I cannot be clear enough that you believing this is the intention of these deals . These are built specifically to make you feel like these things are never going away. This is also an attempt to get "in" with the government at a rate that makes "trying" these models a no-brainer. ... and??????? "The government is going to have cheap access to AI software" does not mean that "the government relies on AI software." Every member of the government having access to ChatGPT — something that is not even necessarily the case! — does not make this software useful, let alone essential , and if OpenAI burns a bunch of money "making it work for them," it still won't be essential, because Large Language Models are not actually that useful for doing stuff! Uber used lobbyist Bradley Tusk to steamroll local governments into allowing Uber to operate in their cities, but Tusk did not have to convince local governments that Uber was useful or have to train people how to use it. Uber's "too big to fail" moment was that local cabs kind of fucking sucked just about everywhere. Did you ever try and take a yellow cab from Downtown Manhattan to Hoboken New Jersey? Or Brooklyn? Or Queens? Did you ever try to pay with a credit card? How about trying to get a cab outside of a major metropolitan area? Do you remember how bad that was? I am not glorifying Uber the company, but the experience that Uber replaced was very, very bad. As a result, Uber did become too big to fail, because people now rely upon it because the old system sucked. Uber used its masses of venture capital to keep prices low to get people used to it too, but the fundamental experience was better than calling a cab company and hoping that they showed up. I also want to be clear this is not me condoning Uber, take public transport if you can! To be clear, Uber has created a new kind of horrifying, extractive labor practice which deprives people of benefits and dignity, paying off academics to help the media gloss over the horrors of its platform . It is also now having to increase prices . What, exactly, is the "essential" experience of generative AI? What essential experience are we going to miss if ChatGPT disappears tomorrow? And on an enterprise or governmental level: what exactly are these tools doing for governments that would make removing them so painful? What use cases? What outcomes? Uber's "essential" nature is that millions of people use it in place of regular taxis, and it effectively replaced decrepit, exploitative systems like the yellow cab medallions in New York with its own tech-enabled exploitation system that, nevertheless, worked far better for the user. Sidenote: although I acknowledge that the disruption that Uber brought to the medallion system had horrendous consequences for the owners of said medallions — some of whom had paid more than a million dollars for the privilege to drive a New York taxi cab , and were burdened under mountains of debt.  There is no such use case with ChatGPT, or any other generative AI system. You cannot point to one use case that is anywhere near as necessary as cabs in cities, and indeed the biggest use cases — things like brainstorming and search — are either easily replaced by any other commoditized LLM or literally already exist with Google Search . Sorry, this is a really simple one. These data centers are not, in and of themselves, driving much economic growth other than in the costs of building them . As I've discussed again and again , there's maybe $40 billion in revenue and no profit coming out of these companies. There isn't any economic growth! They're not holding up anything! These data centers, once built, also create very little economic activity. They don't create jobs , they take up massive amounts of land and utilities, and they piss off and poison their neighbors . If anything, letting these things die would be a political win. There is no "great loss" associated with the death of the Large Language Model era. Taking away Uber would genuinely affect people's ability to get places. So, the classic (and wrong!) argument about OpenAI and companies like OpenAI is that "Uber burned a bunch of money and is now "cash-flow positive" or "profitable ."  Let's talk about raw losses, and where people are making this assumption. Uber lost $24.9 billion in the space of four years (2019 to 2022), in part because of the billions it was spending on sales and marketing and R&D — $4.6 billion and $4.8 billion respectively in 2019 alone . It also massively subsidized the cost of rides — which is why prices had to increase — and spent heavily on driver recruitment, burning cash to get scale, the classic Silicon Valley way. This is absolutely nothing like how Large Language Models are growing, and I am tired of defending this point. OpenAI and Anthropic burn money primarily through compute costs and specialized talent. These costs are increasing, especially with the rush to hire every single AI scientist at the most expensive price possible . There are also essential, immovable costs that neither OpenAI nor Anthropic have to shoulder — the construction of the data centers necessary to train and run inference for their models, which I will get to in a little bit. Yes, Uber raised $33.5 billion ( through multiple rounds of post-IPO debt , though it raised about $25 billion in actual funding). Yes, Uber burned an absolute ass-ton of money. Yes, Uber has scale. But Uber was not burning money as a means of making its product functional or useful. Furthermore, the costs associated with Uber — and its capital expenditures from 2019 through 2024 were around $2.2 billion! — are miniscule compared to the actual costs of OpenAI and Anthropic. Both OpenAI and Anthropic lost around $5 billion in 2024, but their infrastructure was entirely paid for by either Microsoft, Google or Amazon. While we don't know how much of this infrastructure is specifically for OpenAI or Anthropic, as the largest model developers it's fair to assume that a large chunk — at least 30% — of Amazon and Microsoft's capital expenditures have been to support these loads (I leave out Google as it's unclear whether it’s expanded its infrastructure for Anthropic, but we know Amazon has done so ). As a result, the true "cost" of OpenAI and Anthropic is at least ten times what Uber burned. Amazon spent $83 billion in capital expenditures in 2024 and expects to spend $105 billion in 2025 . Microsoft spent $55.6 billion in 2024 and expects to spend $80 billion this year . Based on my (conservative) calculations, the true "cost" of OpenAI is around $82 billion, and that only includes capex from 2024 onward , based on 30% of Microsoft's capex (as not everything has been invested yet in 2025, and OpenAI is not necessarily all of the capex) and the $41.4 billion of funding it’s received so far . The true cost of Anthropic is around $77.1 billion, including all its funding and 30% of Amazon's capex from the beginning of 2024. These are inexact comparisons, but the classic argument is that Uber "burned lots of money and worked out okay," when in fact the combined capital expenditures from 2024 onwards that are necessary to make Anthropic and OpenAI work are each — on their own — four times the amount Uber burned in over a decade. I also believe that these numbers are conservative. There's a good chance that Anthropic and OpenAI dominate the capex of Amazon and Microsoft, in part because what the fuck else are they buying all these GPUs for, as their own AI services don't seem to be making that much money at all. Anyway, to put it real simple, AI has burned more in the last two years than Uber burned in ten, Uber didn't burn money in the same way, didn't burn much by way of capital expenditures, didn't require massive amounts of infrastructure, and isn't remotely the same in any way, shape or form, other than it burned a lot of money — and that burning wasn’t because it was trying to build the core product, but rather trying to scale. I covered this in depth in the Hater's Guide To The AI Bubble , but the long and short of it is that AWS is a platform, a necessity with an obvious choice, and has burned about ten percent of what Amazon et. al has burned chasing generative AI, and had proven demand before building it. Also, AWS was break-even in three years. OpenAI was founded in fucking 2015, and even if you start from November 2022, by AWS standards it should be break-even! Amazon Web Services was created out of necessity — Amazon's infrastructure needs were so great that it effectively had to build both the software and hardware necessary to deliver a store that sold theoretically everything to theoretically anywhere, handling both the traffic from customers, delivering the software that runs Amazon.com quickly and reliably, and, well, making sure things ran in a stable way. It didn't need to come up with a reason for people to run web applications — they were already doing so themselves, but in ways that cost a lot, were inflexible, and required specialist skills. AWS took something that people already did, and what there was a proven demand for, and made it better. Eventually, Google and Microsoft would join the fray. As I’ve discussed in the past, this metric is basically “monthx12,” and while it’s a fine measure for high-gross-margin businesses like SaaS companies, it isn’t for AI. It doesn’t account for churn (when people leave). It also is a number intentionally used to make a company sound more successful — so you can say “$200 million annualized revenue” instead of “$16.6 million a month.” Also, if they’re saying this number, it’s likely that number isn’t consistent! Simple answer: why have literally none of them done this yet? Why not one? There’s that “will” bullshit, once again, always about the “will.” We do not know how thinking works in humans and thus cannot extrapolate it to a machine, and at the very least human beings have the ability to re-evaluate things and learn, a thing that LLMs cannot do and will never do.  We do not know how to get to AGI. Sam Altman said in June that OpenAI was “now confident [they knew] how to build AGI as we have traditionally understood it.” In August, Altman said that AGI was “not a super useful term,” and that “the point of all this is it doesn’t really matter and it’s just this continuing exponential of model capability that we’ll rely on for more and more things.” So, yeah, total bullshit.  Even Meta’s Chief AI Scientist says it isn’t possible with transformer-based models . We don’t know if AGI is possible, anyone claiming they do is lying. This, too, is hogwash, nothing different than your buddy’s friend’s uncle who works at Nintendo that says Mario is coming to PlayStation. Ilya Sutskever and Mira Murati raised billions for companies with no product, let alone a product road map, and they did so because they saw a good opportunity for a grift and to throw a bunch of money at compute. Also: if someone from “deep within the AI industry” has told somebody “big things are coming,” they are doing so to con them or make them think they have privileged information. Ask for specifics.  This argument is poised as a comeback to my suggestion that AI isn't particularly useful, a proof point that this movement is not inherently wasteful, or that there are, in fact, use cases for ChatGPT that are lasting, meaningful or important. I disagree. In fact, I believe ChatGPT — and LLMs in general — have been marketed based on lies of inference. Ironic, I know. I also have grander concerns and suspicions about what OpenAI considers a “user” and how it counts revenue, I’ll get into that later in the week on my premium newsletter, which you should subscribe to. Here’s a hint though: 500,000 of OpenAI’s “ 5 million business customers ” are from its $15 million deal with Cal State University , which works out to around $2.50-a-user-a-month. It’s also started doing $1-a-month trials of its $30-a-month “Teams” subscription , and one has to wonder how many of those are counted in that total, and for how long. I do not know the scale of these offers, nor how long OpenAI has been offering them. A Redditor posted about the deal a few months ago , saying that OpenAI was offering up to 5 seats at once. In fact, I've found a few people talking about these deals, and even one adding that they were offered an annual $10-a-month ChatGPT Plus subscription , with one person saying a few weeks ago saying they'd seen people offered this deal for canceling their subscription . Suspicious. But there’s a greater problem at play. So, ChatGPT has 700 million weekly active users . OpenAI has yet to provide a definition — and yes, I've asked! — which means that an "active" user could be defined as somebody who has gone to ChatGPT once in the space of a week. This term is extremely flimsy, and doesn't really tell us much. Similarweb says that in July 2025 ChatGPT.com had 1.287 billion total visits , making it a very popular website. What do these facts actually mean , though? As I said previously, ChatGPT has had probably the most sustained PR campaign for anything outside of a presidency or a pop star. Every single article about AI mentions OpenAI or ChatGPT, every single feature launch — no matter how small — gets a slew of coverage. Every single time you hear "AI" you’re made to think of "ChatGPT” by a tech media that has never stopped to think about their role in hype, or their responsibility to their readers. And as this hype has grown, the publicity compounds, because the natural thing for a journalist to do when everybody is talking about something is to talk about it more . ChatGPT's immediate popularity may have been viral, but the media took the ball and ran with it, and then proceeded to tell people it did stuff it did not. People were pressured to try this service then under false pretenses, something that continues to this day.  I'll give you an example. On March 15 2023, Kevin Roose of the New York Times would say that OpenAI's GPT-4 was " exciting and scary ," exacerbating (his words!) "...the dizzy and vertiginous feeling I’ve been getting whenever I think about A.I. lately," wondering if he was experiencing "future shock," then described how it was an indeterminate level of "better" and something that immediately sounded ridiculous: In one test, conducted by an A.I. safety research group that hooked GPT-4 up to a number of other systems, GPT-4 was able to hire a human TaskRabbit worker to do a simple online task for it — solving a Captcha test — without alerting the person to the fact that it was a robot. The A.I. even lied to the worker about why it needed the Captcha done, concocting a story about a vision impairment. That doesn't sound remotely real! I went and looked up the paper , and here is the entire extent of what OpenAI shared: This safety card led to the perpetration of one of the earliest falsehoods — and most eagerly-parotted lies — that ChatGPT and generative AI is capable of "agentic" actions. Outlet after outlet — led by Kevin Roose — eagerly interpreted an entire series of events that took place that doesn't remotely make sense, starting with the fact that this is not something you can hire a Taskrabbit to do . Or, at the very least, without a contrived situation where you create an empty task and ask them to complete it. Why not use Mechanical Turk? Or Fiverr? They’ve tons of people offering this service !  But I'm a curious little critter, so I went further and followed their citation to a link on METR's research page . It turns out that what actually happened was METR had a researcher copy paste the generated responses from the model and otherwise handle the entire interaction with Taskrabbit , and based on the plurality of "Taskrabbit contractors," it appears to have taken multiple tries. On top of that, it appears that OpenAI/METR were prompting the model on what to say , which kind of defeats the point. Emphases mine, and comments in [brackets]: It took me five whole minutes to find this piece — which is cited on the GPT-4 system card — read it, then write this piece. It did not require any technical knowledge other than the ability to read stuff.  It is transparently, blatantly obvious that GPT-4 did not "hire" a Taskrabbit or, indeed, make any of these actions — it was prompted to, and they do not show the prompts they used, likely because they had to use so many of them. Anyone falling for this is a mark, and OpenAI should have gone out of its way to correct people. Instead, they sat back and let people publish outright misinformation. Roose, along with his co-host Casey Newton, would go on to describe this example at length on a podcast that week , describing an entire narrative where “the human actually gets suspicious” and “GPT 4 reasons out loud that it should not reveal that [it is] a robot,” at which point “the TaskRabbit solves the CAPTCHA.” During this conversation, Newton gasps and says “oh my god” twice, and when he asks Roose “how does the model understand that in order to succeed at this task, it has to deceive the human?” Roose responds “we don’t know, that is the unsatisfying answer,” and Newton laughs and states “we need to pull the plug. I mean, again, what?” Credulousness aside, the GPT-4 marketing campaign was incredibly effective, creating an aura that allowed OpenAI to take advantage of the vagueness of its offering as people — including members of the media — willfully filled in the blanks for them. Altman has never had to work to sell this product. Think about it — have you ever heard OpenAI tell you what ChatGPT can do, or gone to great lengths to describe its actual abilities? Even on OpenAI's own page for ChatGPT , the text is extremely vague: Scrolling down, you're told ChatGPT can "write, brainstorm, edit and explore ideas with you." It can "generate and debug code, automate repetitive tasks, and [help you] learn new APIs." With ChatGPT you can "learn something new...dive into a hobby...answer complex questions" and "analyze data and create charts." What repetitive tasks? Who knows. How am I learning? Unclear. It's got thinking built in! What that means is unclear, unexplained, and thus allows a user to incorrectly believe that ChatGPT has a brain. To be clear, I know what reasoning means , but this website does not attempt to explain what "thinking" means. You can also "offload complex tasks from start to finish with an agent," which can, according to OpenAI , "think and act, proactively choosing from a toolbox of agentic skills to complete tasks for you using its own computer." This is an egregious lie, employing the kind of weasel-wording that would be used to torture "I.R. Baboon" for an eternity.  Precise in its vagueness, OpenAI's copy is honed to make reporters willing to simply write down whatever they see and interpret it in the most positive light. And thus the lie of inference began. What "ChatGPT" meant was muddied from the very beginning, and thus ChatGPT's actual outcomes have never been fully defined. What ChatGPT "could do" became a form of folklore — a non-specific form of "automation" that could "write code" and "generate copy and images," that can "analyze data," all things that are true but one can infer much greater meaning from. One can infer that "automation" means the automation of anything related to text, or that "write code" means "write the entirety of a computer program." OpenAI's ChatGPT agent is not, by any extension of the word, " already a powerful tool for handling complex tasks ," but it has not, in any meaningful sense, committed to any actual outcomes. As a result, potential users — subject to a 24/7 marketing campaign — have been pushed toward a website that can theoretically do anything or nothing, and have otherwise been left to their own devices. The endless gaslighting, societal pressure, media pressure, and pressure from their bosses has pushed hundreds of millions of people to try a product that even its creators can't really describe. As I've said in the past, OpenAI is deliberately using Weekly Active Users so that it doesn't have to publish its monthly active users, which I believe would be higher. Why wouldn't it do this? Well, OpenAI has 20 million paying ChatGPT subscribers and five million "business customers," with no explanation of what the difference might be. This is already a mediocre (3.5%) conversion rate, yet its monthly active users (which are likely either 800 million or 900 million, but these are guesses!) would make that rate lower than 3%, which is pretty terrible considering everybody says this shit is the future. I also am tired of having people claim that "search" or "brainstorm" or "companions" are a lasting, meaningful business models. So, OpenAI announced that it has hit its first $1 billion month on August 20, 2025 on CNBC, which brings it exactly in line with my estimated $5.26 billion in revenue that I believe it has made as of the end of July . However, remember what the MIT study said: enterprise adoption is high but transformation is low .  There are tons of companies throwing money at AI, but they are not seeing actual returns. OpenAI's growth as the single-most-prominent company in AI (and if we're honest, one of the most prominent in software writ large) makes sense, but at some point will slow, because the actual returns for the businesses aren't there. If there were, we'd have one article where we could point at a ChatGPT integration that at scale helped a company make or save a bunch of money, written in plain English and not a gobbledygook of " profit improvement ." Also… OpenAI is projected to make $12.7 billion in 2025. How exactly will it do that? Is it really making $1.5 billion a month by the end of the year? Even if it does, is the idea that it keeps burning $10 billion or more every year into eternity? What actual revenue potential does OpenAI have long-term? Its products are about as good as everyone else's, cost about the same, and do the same things. ChatGPT is basically the same product as Claude or Grok or any number of different LLMs. The only real advantages that OpenAI has are infrastructure and brand recognition. These models have clearly hit a wall where training is hitting diminishing returns, meaning that its infrastructural advantage is that they can continue providing its service at scale, nothing more. It isn't making its business cheaper, other than the fact that it mostly hasn’t had to pay for it...other than the site in Abilene Texas where it’s promised Oracle $30 billion a year by 2028 . I'm sorry, I don't buy it! I don't buy that this company will continue growing forever, and its stinky conversion rate isn't going to change anytime soon. How? Literally…how!  How? How! HOW???  Nobody ever answers this question! “Efficiencies”? If you’re going to say GPT-5 — here’s a scoop I have about how it’s less efficient !  It's very, very, very common for people to conflate "AI" with "generative AI." Make sure that whatever you're claiming or being told is actually about Large Language Models, as there are all sorts of other kinds of machine learning that people love to bring up. LLMs have nothing to do with Folding@Home, autonomous cars, or most disease research. A lot of people think that they're going to tell me "I use this all the time!" and that'll change my mind. I cannot express enough how irrelevant it is that you have a use case, as every use case I hear is one of the following: This would all be fine and dandy if people weren't talking about this stuff as if it was changing society. None of these use cases come close to explaining why I should be impressed by generative AI. It also doesn't matter if you yourself have kind of a useful thing that AI did for you once. We are so past the point when any of that matters. AI is being sold as a transformational technology, and I am yet to see it transform anything. I am yet to hear one use case that truly impresses me, or even one thing that feels possible now that wasn't possible before. This isn't even me being a cynic — I'm ready to be impressed! I just haven't been in three fucking years and it's getting boring. Also, tell me with a straight face any of this shit is worth the infrastructure. One of the most braindead takes about AI and coding is that "vibe coding" is "allowing anyone to build software." While technically true, in that one can just type "build me a website" into one of many AI coding environments, this does not mean it is functional or useful software. Let's make this really clear: AI cannot "just handle coding ." Read this excellent piece by Colton Voege , then read this piece by Nik Suresh . If you contact me about AI and coding without reading these I will send them to you and nothing else, or crush you like a car in a garbage dump, one or the other. Also, show me a vibe coded company. Not a company where someone who can code has quickly spun up some features, a fully-functional, secure, and useful app made entirely by someone who cannot code. You won't be able to find this as it isn't possible. Vibe Coding is a marketing term based on lies, peddled by people who have either a lack of knowledge or morals. Are AI coding environments making people faster? I don't think so! In fact, a recent study suggested they actually make software engineers slower . The reason that nobody is vibe coding an entire company is because software development is not just "put a bunch of code in a pile and hit "go," and oftentimes when you add something it breaks something else. This is all well and good if you actually understand code — it's another thing entirely when you are using Cursor or Claude Code like a kid at an arcade machine turning the wheel repeatedly and pretending they're playing the demo. Vibe coders are also awful for the already negative margins of most AI coding environments, as every single thing they ask the model to do is imprecise, burning tokens in pursuit of a goal they themselves don't understand. "Vibe coding" does not work, it will not work, and pretending otherwise is at best ignorance and at worst supporting a campaign built on lies. If you are an AI booster, please come up with better arguments. And if you truly believe in this stuff, you should have a firmer grasp on why you do so. It's been three years, and the best some of you have is "it's real popular!" or "Uber burned a lot of money!" Your arguments are based on what you wish were true rather than what's actually true, and it's deeply embarrassing. Then again, there are many well-intentioned people who aren't necessarily AI boosters who repeat these arguments, regardless of how thinly-framed they are, in part because we live in a high-information, low-processing society where people tend to put great faith in people who are confident in what they say and sound smart. I also think the media is failing on a very basic level to realize that their fear of missing out or seeming stupid is being used against them. If you don't understand something, it's likely because the person you're reading or hearing it from doesn't either. If a company makes a promise and you don't understand how they'd deliver on it, it's their job to explain how, and your job to suggest it isn't plausible in clear and defined language . This has gone beyond simple "objectivity" into the realm of an outright failure of journalism. I have never seen more misinformation about the capabilities of a product in my entire career, and it's largely peddled by reporters who either don't know or have no interest in knowing what's actually possible, in part because all of their peers are saying the same nonsense. As things begin to collapse — and they sure look like they're collapsing, but I am not making any wild claims about "the bubble bursting" quite yet — it will look increasingly more deranged to bluntly publish everything that these companies say. Never have I seen an act of outright contempt more egregious than Sam Altman saying that GPT-5 was actually bad, and that GPT-6 will be even better . Members of the media: Sam Altman does not respect you. He is not your friend. He is not secretly confiding in you. He thinks you are stupid and easily-manipulated, and will print anything he says, largely in part because many members of the media will print exactly what he says whenever he says it. To be clear, if you wrote about it and actively mocked it , that's fine. But let's close by discussing the very nature of AI skepticism, and the so-called "void" between those who "hate" AI and those who "love" AI, from the perspective of one of the more prominent people in the "skeptic" side. Critics and skeptics are not given the benefit of grace, patience, or, in many cases, hospitality when it comes to their position. While they may receive interviews and opportunities to "give their side," it is always framed as the work of a firebrand, an outlier, somebody with dangerous ideas that they must eternally justify. They are demonized, their points under constant scrutiny, their allegiances and intentions constantly interrogated for some sort of moral or intellectual weakness. "Skeptic" and "critic" are words said with a sneer or trepidation — that the listener should be suspicious that this person isn't agreeing that AI is the most powerful, special thing ever. To not immediately fall in love with something that everybody is talking about is to be framed as a "hater," to have oneself introduced with the words "not everybody agrees..." on 40% of appearances. By comparison, AI boosters are the first to get TV appearances and offers to be on panels, their coverage featured prominently on Techmeme, selling slop-like books called shit like The Future Of Intelligence: Masters Of The Brain featuring 18 interviews with different CEOs that all say the same thing. They do not have to justify their love — they simply have to remember all the right terms, chirping out "test-time compute" and "the cost of inference is going down" enough times to summon Wario Amodei to give them an hour-long interview where he says "the models, they are, in years, going to be the most powerful school teacher ever built." And yeah, I did sell a book, because my shit fucking rocks. I have consistent, deeply-sourced arguments that I've built on over the course of years. I didn't "become a hater" because I'm a "contrarian," I became a hater because the shit that these fucking oafs have done to the computer pisses me off. I wrote The Man Who Killed Google Search because I wanted to know why Google Search sucked. I wrote Sam Altman, Freed because at the time I didn't understand why everybody was so fucking enamoured with this damp sociopath. Everything I do comes from genuine curiosity and an overwhelming frustration with the state of technology. I started writing this newsletter with 300 subscribers and 60 views, and have written it as an exploration of subjects that grows as I write. I do not have it in me to pretend to be anything other than what I am, and if that is strange to you, well, I'm a strange man, but at least I'm an honest one. I do have a chip on my shoulder, in that I really do not like it when people try to make other people feel stupid, especially when they do so as a means of making money for themselves or somebody else. I write this stuff out because I have an intellectual interest, I like writing, and by writing, I am able to learn about and process my complex feelings about technology. I happen to do so in a manner that hundreds of thousands of people enjoy every month, and if you think that I've grown this by "being a hater," you are doing yourself the disservice of underestimating me, which I will use to my advantage by writing deeper, more meaningful, more insightful things than you. I have watched these pigs ruin the computer again and again, and make billions doing so, all while the media celebrates the destruction of things like Google, Facebook, and the fucking environment in pursuit of eternal growth. I cannot manufacture my disgust, nor can I manufacture whatever it is inside me that makes it impossible to keep quiet about the things I see. I don't know if I take this too seriously or not seriously enough, but I am honoured that I am able to do it, and have 72,000 of you subscribed to find out when I do so.

0 views

How Does GPT-5 Work?

Welcome to another premium edition of Where's Your Ed At! Please subscribe to it so I can continue to drink 80 Diet Cokes a day. Email me at [email protected] with the subject "premium" if you ever want to chat. I realize this is before the paywall, so if you email me without paying, no promises I don't respond with the lyrics to Cheeseburger In Paradise . Also: this is an open call — if you've tried prompt caching with GPT-5 on OpenAI's API, please reach out! You've probably heard a lot about GPT-5 this week, with takes ranging from " it's just good at stuff " to SemiAnalysis' wild statement that " GPT-5 [is setting] the stage for Ad Monetization and the SuperApp ," a piece that makes several assertions about how the "router" that underpins GPT-5 is somehow the secret way that OpenAI will inject ads. Here's a quote: This...did not make a ton of sense to me. Why would this be the case? The article also makes a lot of claims about the "value" of a question and how ChatGPT could — I am serious — "agentically reach out to lawyers" based on a query. In fact, I'm not sure this piece reflects how GPT-5 works at all. To be fair on SemiAnalysis, it's not as if OpenAI gave them much help. Here's what it says : There is a really, really important distinction to make here: that GPT-5, as described above, is referring to GPT-5 as part of ChatGPT. OpenAI's API-based access to GPT-5 models does not route them, nor does OpenAI offer access to its router, or any other associated models. How do I know this? Because I went and found out how ChatGPT-5 actually works. In discussions with a source at an infrastructure provider familiar with the architecture, it appears that ChatGPT-5 is, in fact, potentially more expensive to run than previous models, and due to the complex and chaotic nature of its architecture, can at times burn upwards of double the tokens per query. ChatGPT-5 is also significantly more convoluted, plagued by latency issues, and is more compute-intensive thanks to OpenAI's new "smarter, more efficient" model. In simple terms, every user prompt on ChatGPT — whether it's on the auto, "Fast," "Thinking Fast" or "Thinking" tab — starts by putting the user's prompt before the "static prompt," which is a hidden prompt where instructions like "You are ChatGPT, you are a Large Language Model, You Are A Helpful Chatbot" and so on goes. These static prompts are different with each model you use - a reasoning model will have a different instruction set than a more chat-focused one, such as “think hard about a particular problem before giving an answer.” This becomes an issue when you use multiple different models in the same conversation, because the router — the thing that selects the right model for the request — has to look at the user prompt. It can’t consider the static instructions first. The order has to be flipped for the whole thing to work. Put simpler: Previous versions of ChatGPT would take the static prompt, and then (invisibly) append the user prompt onto it. ChatGPT-5 can’t do that.  Every time you use ChatGPT-5, every single thing you say or do can cause it to do something different. Attach a file? Might need a different model. Ask it to "look into something and be detailed?" Might trigger a reasoning model. Ask a question in a weird way? Sorry, the router's gonna need to send you to a different model.  Every single thing that can happen when you ask ChatGPT to do something may trigger the "router" to change model, or request a new tool, and each time it does so requires a completely fresh static prompt, regardless of whether you select Auto, Thinking, Fast or any other option. This, in turn, requires it to expend more compute, with queries consuming more tokens compared to previous versions.  As a result, ChatGPT-5 may be "smart," but it sure doesn't seem "efficient." To play Devil's Advocate, OpenAI likely added the routing model as a means of creating more sophisticated outputs for users, and, I imagine, with the intention of cost-saving. Then again, this may just be the thing that it had ready to ship — after all, GPT-5 was meant to be " the next great leap in AI ," and the pressure was on to get it out the door. By creating a system that depends on an external routing model — likely another LLM — OpenAI has removed the ability to cache the hidden instructions that dictate how the models generate answers in ChatGPT, creating massive infrastructural overhead. Worse still, this happens with every single "turn" (IE: message) on ChatGPT-5, regardless of the model you choose, creating endless infrastructural baggage with no real way out that only compounds based on how complex a user's queries get.  Could OpenAI make a better router? Sure! Does it have a good router today? I don't think so! Every time you message ChatGPT it has the potential to change model or tooling based on its own whims, each time requiring a fresh static prompt. It doesn't even need to be a case where a user asks ChatGPT-5 to "think," and based on my tests with GPT-5, sometimes just asking it a four-word question can trigger it to "think longer" for no apparent reason. OpenAI has created a product with latency issues and an overwhelmingly convoluted routing system that's already straining capacity, to the point that this announcement feels like OpenAI is walking away from its API entirely. Unlike the GPT-4o announcement , which mentions the API in the first paragraph, the GPT-5 announcement has no reference to it, and a single reference to developers when talking about coding. Sam Altman has already hinted that he intends to deprecate any "new API demand " — though I imagine he'll let anyone in who will pay for priority processing . ChatGPT-5 feels like the ultimate comeuppance for a company that was never forced to build a product, choosing instead to bolt increasingly-complex "tools" onto the sides of models in the hopes that one would magically appear. Now each and every "feature" of ChatGPT burns even more money than it did before.  ChatGPT-5 feels like a product that was rushed to market by a desperate company that had to get something out the door. In simpler terms, OpenAI gave ChatGPT a middle manager.

0 views

The Enshittification of Generative AI

Thanks for subscribing to Where’s Your Ed At Premium, please shoot me an email at [email protected] if you ever have any questions. Yesterday, OpenAI launched GPT-5 , a new “flagship” model of some sort that’s allegedly better at coding and writing, but upon closer inspection it feels like the same old shit it’s been shoveling for the last year or two.  Sure, I’m being dismissive, but three years and multiple half-billion-dollar training runs later , OpenAI has delivered us a model that is some indeterminate level of “better” that “scared” Sam Altman , and immediately began doing what some Twitter users called “ chart crimes ” with its supposed coding benchmark charts.  This also begs the question: what is GPT-5? WIRED calls it a “flagship language model,” but OpenAI itself calls it a “unified system with a smart, efficient model that answers most questions, a deeper reasoning model, and a real-time router that quickly decides which[model]  to use based on conversation type, complexity, tool needs, and your explicit intent.” That sure sounds like two models to me, and not necessarily new ones! Altman, back in February , said that GPT-5 was “a system that integrates a lot of our technology, including o3.” It is a little unclear what GPT-5 — or at least the one accessed through ChatGPT — is. According to Simon Willison , there’s three sub-models — a regular, mini and a nano model, “which can each be run at one of four reasoning levels” if you configure them using the API. When it comes to what you access on ChatGPT, however, you’ve got two options — GPT-5 and GPT-5-Thinking, with the entire previous generation of GPT models no longer available for most users to access. I believe GPT-5 is part of a larger process happening in generative AI — enshittification , Cory Doctorow’s term for when platforms start out burning money offering an unlimited, unguarded experience to attract their users, then degrade and move features to higher tiers as a means of draining the blood from users.  With the launch of GPT-5, OpenAI has fully committed to enshittifying its consumer and business subscription products, arbitrarily moving free users to a cheaper model and limiting their ability to generate images, and removing the ability to choose which model you use in its $20, $35 and “enterprise” subscriptions, moving any and all choice to its “team” and $200-a-month “pro” subscriptions.  OpenAI’s justification is an exercise in faux-altruism , framing “taking away all choice” as a “real-time router that quickly decides which [model] to use.” ChatGPT Plus and Team members now mostly have access to two models — GPT-5 and GPT-5-Thinking — down from the six they had before.  This distinction is quite significant. Where users once could get hundreds of messages a day on OpenAI’s o4-mini-high and o4-mini reasoning models, GPT-5 for ChatGPT Plus subscribers offers 200 reasoning (GPT-5-thinking) messages a week , with 80 GPT-5 messages every 3 hours which allow you to ask it to “think” about its answer, shoving you over to an undisclosed reasoning model. This may seem like a good deal, OpenAI is likely putting you on the cheapest model whenever it can in the name of “the best choice.” While Team accounts have “unlimited” access to GPT-5, they still face the same 200-reasoning-messages-a-week limit , and while yes, you could ask it to “think” more, do you think that OpenAI is going to give you their best reasoning models? Or will they, as they said, “bring together the best of their previous models” and “choose the right one for the job”? Furthermore, OpenAI is permanently sunsetting ChatGPT access to every model that doesn’t start with GPT-5 on August 14th except for customers of its most expensive subscription tier. OpenAI will (and it appears this applies to the $200-a-month "Pro" plan too, I'm told by reporter Joanna Stern )- reduce your model options to two or three choices (Chat, Thinking and Pro), and will choose whatever sub-model it sees fit in the most opaque way possible. GPT-5 is, by definition, a “trust me bro” product.  OpenAI is trying to reduce the burden of any particular user on the system under the guise of providing the “smartest, fastest model,” with “smartest” defined internally in a way that benefits the company, marketed as “choosing the best model for the job.”  Let's see how users feel! An intrepid Better Offline listener pulled together some snippets from r/ChatGPT , where users are mourning the loss of GPT-4o , furious at the loss of other models and calling GPT-5, in one case, " the biggest peice (sic) of garbage even as a paid user ," who says that "projects are absolutely brain-dead now." One user said that GPT-5 is " the biggest bait-and-switch in AI history ," another said that OpenAI " deleted a workfow of 8 models overnight, with no prior warning ," and another said that " ChatGPT 5 is the worst model ever ." In fact, there are so many of these posts that I could find posts to link to for every word of this paragraph in under five minutes. Yet OpenAI isn’t just screwing over consumers. Developers that want to integrate OpenAI’s model now have access to “priority processing” — previously an enterprise-only feature ( see this archive from July 21st 2025 ) to guarantee low latency and uptime. While this sounds like something altruistic, or a new beneficial feature, I’m not convinced. I believe there’s only one reason to do this: that OpenAI intends to, or will be forced to due to capacity constraints, start degrading access to its API.  As with every model developer, we have no real understanding of what may or may not lead to needing “reliable, high-speed performance” from API access, but the suggestion here is that failing to pay OpenAI’s troll toll will put your API access in the hole. That toll is harsh, too, nearly doubling the API price on each model , and while the Priority Processing Page has pricing for all manner of models, its pricing page reduces the options down to two models — GPT-5 and GPT-5-mini, suggesting it may not intend to provide priority access in perpetuity.  OpenAI is far from alone in turning the screws on its customers. As I’ll explain, effectively every consumer generative AI company has started some sort of $200-a-month “pro” plan — Perplexity Max, Gemini ($249.99 a month before discounts), Cursor Ultra, Grok Heavy (which is $300 a month!), and, of course, Anthropic, whose $100-a-month and $200-a-month plans allowed Claude Code users to spend anywhere from 100% to 10,000% of their monthly subscription in API calls . This led to rate limits starting August 28 2025 — a conveniently-placed date to allow Anthropic to close as much as $5 billion in funding before its users churn.  Worse still, Anthropic burned all of that cash to get Claude Code to $400 million in annualized revenue according to The Information — around $33 million in monthly revenue that will almost certainly evaporate as its customers hit week-long rate limits on a product that’s billed monthly.  These are not plans created for “power users.” They are the actual price points at which these things need to be to be remotely sustainable, though Sam Altman said earlier in the year that ChatGPT Pro’s $200-a-month subscription was losing OpenAI money . And with GPT-5, meaningful functionality — the ability to choose the specific model you want for a task — is being completely removed for ChatGPT Plus and Team subscribers . This is part of an industry-wide enshittification of generative AI, where the abominable burn rates behind these products are forcing these companies to take measures ranging from minor to drastic.  The problem, however, is that these businesses have yet to establish truly essential products, and even when they create something popular — like Claude Code — they can’t make it popular without burning horrendous amounts of cash. The same goes for Cursor, and I believe just about every other major product built on top of Large Language Models. And I believe that when they try to adjust pricing to reflect their actual costs, that popularity will begin to wane. I believe we’re already seeing that with Claude Code, based on the sentiment I’ve seen on the tool’s Reddit page, although I’m also wary of making any sweeping statements right now, as it’s just too early to say.  The great enshittification of AI has begun.

0 views

AI Is A Money Trap

In the last week, we’ve had no less than three different pieces asking whether the massive proliferation of data centers is a massive bubble, and though they, at times, seem to take the default position of AI’s inevitable value, they’ve begun to sour on the idea that it’s going to happen soon. Meanwhile, quirked-up threehundricorn OpenAI has either raised or is about to raise another $8.3 billion in cash , less than two months since it raised $10 billion from SoftBank and a selection of venture capital firms.  I hate to be too crude, but where the fuck is this money going? Is OpenAI just incinerating capital? Is it compute? Is it salaries? Is it compute? Is it to build data centers, because SoftBank isn’t actually building anything for Stargate ?  The Information suggested OpenAI is using the money to build data centers — possibly the only worse investment it can make other than generative AI, and it’s one that it can’t avoid because OpenAI also is somehow running out of compute. And now they're in " early-stage discussions " about an employee share sale that would value the company at $500 billion, a ludicrous number that shows we're leaving the realm of reality. To give you some context, Shopify's market cap is $197 billion, Salesforce's is $248 billion, and Netflix's is $499 billion. Do you really think that OpenAI is worth more than these companies? Do you think they're worth more than AMD at a $264 billion market cap? Do you? Amongst this already-ridiculous situation sits the issue of OpenAI and Anthropic’s actual revenues, which I wrote about last week , and have roughly estimated to be $5.26 billion and $1.5 billion respectively (as of July). In any case, these estimates were made based on both companies’ predilection for leaking their “annualized revenues,” or monthx12.  This extremely annoying term is one that I keep bringing up because it’s become the de-facto way for generative AI companies to express their revenue, and both OpenAI and Anthropic are leaking them intentionally, and doing so in a way that suggests they’re not using even the traditional ways of calculating them. OpenAI leaked on July 30 2025  that it was at $12 billion annualized revenue — so around $833 million in a 30-day period — yet two days later on August 1 2025 the New York Times reported they were at $13 billion annualized revenue , or $1.08 billion of monthly revenue. It’s very clear OpenAI is not talking in actual calendar months, at which point we can assume something like a trailing 30 day window (as in the “month” is just 30 days rather than a calendar month). We can, however, declaratively say that it’s not doing “the month of June” or “the month of July” because if it was, OpenAI wouldn’t have given two vastly different god damn numbers in the same two day period. That doesn’t make any sense. There are standard ways to handle annualized revenue, and it's clear they're not following them. And to be even clearer, while I can’t say for certain, I believe these leaks are deliberate. OpenAI’s timing matches exactly with fundraising.  On Anthropic’s side, these revenues are beginning to get really weird. Anthropic went from making $72 million ( $875 million annualized ) in January to $433 million in July — or at least, it leaked on July 1, 2025 that it was at $4 billion annualized to The Information ($333 million a month) and claimed it had reached $5 billion annualized revenue ($416 million) to Bloomberg on July 29 2025 .  How’d it get there? I’m guessing it was from cranking up prices on Cursor, and we’ve had the confirmation that’s the case thanks to The Information reporting that $1.4 billion of its annualized revenue is from its top two customers (so around $116 million a month), the biggest of which is Cursor. Confusingly, The Information also says that Anthropic’s Claude Code is “generating nearly $400 million in annualized revenue, roughly doubling from just a few weeks ago,” meaning about $33 million of monthly revenue.  In any case, I think Cursor is a huge indicator of the current fragility of the bubble — and the fact that for most AI startups, there’s simply no way out, because being acquired or going public does not appear to be a viable route.  I know it sounds a little insane, but I believe that Cursor is the weak point of the entire AI bubble, and I’ll explain why, and how this could go. This is, by no means, inevitable, but I cannot work out what Cursor does other than this. Cursor, at this point, faces two options: die, or get acquired. This is not an attack on anyone who works at the company, nor anything personal. The unit economics of this business do not make sense and yet, on some level, its existence is deeply important to the valley’s future.  OpenAI? OpenAI couldn’t acquire Windsurf because it was too worried Microsoft would get the somehow-essential IP of one of what feels like a hundred different AI-powered coding environments . It also already tried and failed to buy Cursor , and if I’m honest, I bet Cursor would sell now. Honestly, Cursor fucked up bad not selling then. It could have got $10 billion and Sam Altman would’ve had to accelerate the funding clause. It would’ve been so god-damn sick, but now the only “sick” thing here is Cursor’s fragile, plagued business model.  How about Anthropic? Eh! It already has their own extremely-expensive coding environment, Claude Code, which I estimated loses the company 100% to 10,000% of a subscription per-customer a few weeks ago , and now Anthropic is adding weekly limits on accounts , which will, I believe, create some of the most gnarly churn in SaaS history. Also, does Anthropic really want to acquire its largest customer? Also, with what money? It’s not raising $5 billion to bail out Cursor. Anthropic needs that to feed directly into Andy Jassy’s pocket to keep offering increasingly-more-complex models that never quite seem to be good enough.  Google ? It just sort-of-bought Windsurf ! It can’t do that again. It’s already given out the participation trophy multiple billions of dollars to investors and founders so nobody has to get embarrassed about this, and then allowed Cognition to pick up the scraps of a business that made $6.83 million a month after burning $143 million of investor capital (TechCrunch reports Windsurf was left with $100 million in cash post-acquisition). TechCrunch also reports that Cognition paid $250 million for what remained , and that this deal didn’t actually pay out the majority of Windsurf’s employees, Meta ? If I’m Cursor’s CEO, I am calling Mark Zuckerberg and pretending that I think the only person in the world who can usher in the era of Superintelligence is the guy who burned more than $45 billion on the metaverse and believes that not wearing AI glasses in the future will be a disadvantage . I would be saying all manner of shit about the future, and that the only way to do this was to buy my AI-powered coding startup that literally can’t afford to exist. And that really is the problem. These companies are all going through the same motions that every company before them did — raise as much money as possible, get as big as possible, and eventually scale to the point you’re fat with enterprise cash.  Except the real problem is that, just like big tech’s new gluttony of physical real estate it's taken on, generative AI companies are burdened with a constant and aggressive form of cloud debt — the endless punishment of the costs of accessing the API for generative AI models that always seem to get a little better, but never in such a way that anything really changes other than how much Anthropic and OpenAI are going to need at the end of the month or they break your startup’s legs. I’m not even trying to be funny! Anthropic raised its prices on Cursor so severely it broke its already-unprofitable business model. These products — while also, for the most part, not producing that much revenue — need to be sold with users being aware of (and sensitive to) the cost of providing them, and Cursor’s original product was $20-a-month for 500 “fast requests” of different models, in the same way that accessing Claude Code on any subscription is either $20, $100, or $200 a month rather than paying per API call, because these companies all sell products that shield the customer from the actual costs of running the services. The irony is that, despite being willing to kill these companies by fundamentally changing the terms upon which they access these models, Anthropic is also, in some way, dependent on Cursor, Replit, and other similar firms continuing to buy tokens at the same rate as before, as that consumption is baked into its ARR figures, as well as the forward-looking revenue projections.  It is, in some sense, a Kobayashi Maru . Anthropic has an existential need to screw over its customers by hiking rates and imposing long-term commitments, but its existence is also, in some way, predicated on these companies continuing to exist. If Cursor and Replit both die, that’s a significant chunk of Anthropic's API business gone in a flash — and, may I remind you, that significantly overshadows its subscription business (making it almost like an inverse of OpenAI, where subscriptions drive the bulk of revenue).  Anthropic’s future is wedded to Cursor, and I just don’t see how Cursor survives, let alone exits, or gets subsumed by another company in a way that mirrors how acquisitions have worked since…ever.   If Cursor does not sell for a healthy amount — I’m talking $10 billion plus, and I mean actually sell, not “the founders are hired in a strange contractual agreement that pays out investors and its assets are sold to Rick from Pawn Stars” — it will prove that no generative AI company, to this date, has actually been successful. In reality, I expect a Chumlee-esque deal that helps CEO Michael Truell buy a porsche while his staff makes nothing. Is Cursor worth $10 billion? Nope! No matter how good its product may or may not be, it is not good enough to be sold at a price that doesn’t require Cursor to incinerate hundreds of millions of dollars with no end in sight. And this ultimately gives us the real conundrum — why aren’t generative AI startups selling?   Before we go any further, there have been some acquisitions, but they are sparse, and seem almost entirely centered around bizarre acqui-hires and confusing fire sales. AMD bought Silo AI, “the largest private AI lab in Europe,” in August 2024 for $665 million , which appears to be the only real acquisition in generative AI history, and appears to be partially based on Silo’s use of AMD’s GPUs .  Elsewhere, NVIDIA bought OctoAI for an estimated $250 million in September 2024 , after buying Brev.dev in July 2024 for an undisclosed sum , and then Gretel in March 2025 . Yet in all three cases these are products to deploy generative AI, and not products built on top of generative AI or AI models. Canva bought “generative AI content and research company” Leonardo.AI in July 2024 for an undisclosed sum. Really, the only significant one I’ve seen was on July 29 2025 — publicly-traded customer service platform NICE buying AI-powered customer service company Cognigy in a $955 million deal . According to Cxtoday, Cognigy expects about $85 million in revenue this year , though nobody appears to be talking about costs. However, Cognigy, according to some sources , charges tens or hundreds of thousands per contract for its “AI voice agents” that can “understand and respond to user input in a natural way.” Great! We’ve got one real-deal “company built on models” acquisition, and it’s a company that most people haven’t heard of making around $7 million a month.  Let’s take a look at the others.  Outside of one very industry-specific acquisition, there just doesn’t seem to be the investor hunger to buy a company valued at $9.9 billion . And you have to ask why. If AI is, as promised, the thing that’ll radically change our economy, and these companies are building the tools that’ll bring about that change, why does nobody want to buy them?  And, in the broader term, what does it mean when these companies — those with $10bn, or in the case of OpenAI, $300bn valuations — can’t be bought, and can’t go public? Where does this go? What happens next? What’s the gameplan here? How will the venture firms that ploughed billions of capital into these businesses bring a return for their LPs if there are no IPOs or buyouts?  The economic implications of these questions are, quite frankly, terrifying — especially when you consider the importance that VC has historically held in building the US tech ecosystem, and they raise further questions about the impact of an AI bubble on companies that are promising, and do have a viable business model, and a product with actual fit, but won’t be able to actually raise any cash.  Great! I would believe it was possible if it had ever, ever happened, which it has not.  I’m not even being sarcastic or rude. It has just not happened. No company that actually stakes their entire product on generative AI appears to be able to make money. Glean, a company that makes at best $8.3 million a month ($100 million annualized revenue) said it had $550 million in cash December of last year , and then had to raise $150 million in June of this year . Where did that money go? Why does a generative search engine product with revenues that are less than a third of the Cincinnati Reds baseball team need half a billion dollars to make $8.3 million a month?  I’m not saying these companies are unnecessary, so much as they may very well be impossible to run as real businesses. This isn’t even a qualitative judgment of any one generative AI company. I’m just saying, if any of these were good businesses, they would be either profitable or being acquired in actual deals, and there would be good businesses by now.  The amount of cash they are burning does not suggest they’re rapidly approaching any kind of sane burn rate, or we would have heard. Putting aside any kind of skepticism I have, anything you may hold against me for what I say or the way I say it, where are the profitable companies? Why isn’t there one, outside of the companies creating data to train the AI models, or Nvidia? We’re three years in, and we haven’t had one. We also have had no exits and no IPOs. There has been no cause for celebration, no validation of a business model through another company deciding that it was necessary to continue its dominance by raising funds on the public market, or allowing actual investors — flawed though they may be — act as the determiner of their value.  It is unclear what the addition of Windsurf’s intellectual property adds to Cognition, much like it’s a little unclear what differentiates Cognition’s so-called AI-powered software engineer “Devin” from anything else on the market. I hear Goldman is paying for it , and said the stupidest shit I’ve ever heard to CNBC that nevertheless shows how little it’s actually paying for: “We’re going to start augmenting our workforce with Devin, which is going to be like our new employee who’s going to start doing stuff on the behalf of our developers,” Argenti told CNBC. “Initially, we will have hundreds of Devins [and] that might go into the thousands, depending on the use cases.”  Hundreds of Devins = hundreds of seats. At a very optimistic 500 users at the highest-end pricing of $500-a-month (if it’s $20-a-month, Cognition is making a whole, at most, less than $20,000 a month) — and let’s assume that it does a discount at enterprise scale, because that always happens — that’s $250,000 a month! Wow! $3 million in revenue? On a trial basis? Amazing! In fact, I can’t find a shred of evidence that Cognition otherwise makes much money. Despite currently raising $300 million at a $10 billion valuation , I can find no information about Cognition’s revenues beyond one comment from The Information from July 2024, when Cognition raised at a $2 billion valuation : Cognition’s fundraise is the latest example of AI startups raising capital at sky-high valuations despite having little or no revenue.”  In a further move per The Information that is both a pale horse and a deeply scummy thing to do, Cognition has now laid off 30 people from the Windsurf team, and is now offering the remaining 200 buyouts equal to 9 months of salary and, I assume, the end of any chance to accrue further stock in Cognition. CEO Scott Wu said the following in the email telling Windsurf employees about the layoffs and buyouts: “We don’t believe in work-life balance—building the future of software engineering is a mission we all care so deeply about that we couldn’t possibly separate the two,” he said. “We know that not everyone who joined Windsurf had signed up to join Cognition where we spend 6 days at the office and clock 80+ hour weeks.” All that piss, vinegar, and burning of the midnight oil does not appear to have created a product that actually matters. I realize this is a little cold, but if you’re braying and smacking your chest about your hard-charging, 6-days-a-week office culture, you should be able to do better than “we have one publicly-known customer and nobody knows our revenue.” Maybe it’s a little simpler: Cognition paid $250 million to acquire Windsurf so that it could, after the transaction, say they have $82 million in annualized revenue. If that’s the case, this is one of the dodgiest, weirdest acquisitions I’ve seen in my life — two founders getting a few hundred million dollars between them and their investors, and a few of their colleagues moving with them to Google, leaving the rest of the staff effectively jobless or in Hell with little payoff for their time working at Windsurf. I can only imagine how it must have felt to go from being supposedly acquired by OpenAI to this farcical “rich get richer” bullshit. It also suggests that the actual underlying value of Windsurf’s IP was $250 million . So, I ask, why, exactly, is Cognition worth $10 billion? And why did it have to raise $300 million after raising “hundreds of millions” according to Bloomberg in March? Where is the money going? It doesn’t seem to have great revenue, Carl Brown of the Internet of Bugs revealed it faked the demo of “Devin the AI powered software developer” last year , and Devin doesn’t even rank on SWE-benchmark , the industry standard for model efficacy at coding tasks.  At best, it’s now acquired their own unprofitable coding environment and the smidgen of revenue associated. How would Cognition go public? What is the actual exit path for Cognition, or any other generative AI startup?  And that, right there, is Silicon Valley’s own housing crisis, except instead of condos houses they can’t afford with sub-prime adjustable rate mortgages, venture capitalists have invested in unprofitable, low-revenue startups with valuations that they can never sell at. And, like homeowners in the dismal years of 2008 and 2009, they’re almost certainly underwater — they just haven’t realized it yet. Where consumers were unable to refinance their mortgages to bring their monthly payments down, generative AI startups face pressure to continually raise at higher and higher valuations to keep up with their costs, with each one making it less likely their company will survive.  The other difference is that, in the case of the housing crisis, those who were able to hold onto their properties eventually saw their equity recover to their pre-crash levels, in part because housing is essential and because its price is influenced just as much by supply and demand, as it is the ability for people to finance the purchase of properties, and when the population increases, so too does the demand for housing. None of that is true with AI. There’s a finite number of investors, a finite number of companies, and a finite amount of capital — and those companies are only as valuable as the expectations that investors have for them, and as the broader sentiment towards AI.  Who is going to buy Cognition? Because the only other opportunity for the investors who put the money into this company to make money here — let alone to recoup their initial investment —  is for Cognition to go public. Do you think Cognition will go public? How about Cursor? It’s worth $9.9 billion, and there was a rumour that it was raising at a valuation of $18 billion to $20 billion back in June . Do you see Perplexity, at a valuation of $18 billion , selling to another company? The alternative, as discussed, is that Perplexity, a company with 15 million users and, at $150 million annualized revenue , is still making less than half of the revenue of the Cincinnati Reds baseball team ( $325 million in annual revenue, and that’s real money, not “annualized revenue”), must go public. Perplexity has, at this point, raised over a billion dollars to lose $68 million in 2024 on $34 million of revenue.  By comparison, the Cincinnati Reds is a great business, with a net monthly income of $29 million, all to provide a service that upsets and humiliates millions of people from Ohio every year for the pleasure of America.  Putting aside the Reds, what exactly is it that Perplexity could offer to the public markets as a stock, or to an acquirer? Apple considered acquiring it in June , but Apple tends to acquire the companies it wants to integrate into the core business ( as was the case with Siri ), which makes me think that Perplexity leaked information about a deal that was never really serious. Hell, Meta talked about acquiring it too . Isn’t it weird that two different companies talked about buying Perplexity but neither of them did it? CEO Aravind Srivinas said in July that he wanted to “ remain independent ,” which is a weird thing to say after talking to two giant multi-trillion-dollar market cap tech firms about selling to them.  It’s almost as if nobody actually wants to buy Perplexity, or any of these sham companies, which I know sounds mean, but if you are worth billions or tens of billions of dollars and you can’t make more than a bottom-tier baseball team in fucking Ohio , you are neither innovative nor deserving of said valuation.  But really, my pissiness and baseball comparisons aside, what exactly is the plan for these companies? They don’t make enough money to survive without a continuous flow of venture capital, and they don’t seem to make impressive sums of money even when allowed to burn as much as they’d like. These companies are not being forced to live frugally, or at least have yet to be made to, perhaps because they’re all actively engaged at spending as much money as possible in pursuit of finding an idea that makes more money than it loses. This is not a rational or reasonable way to proceed. Yes, there are startups that can justify burning capital. Yes, there are companies that have burned hundreds of millions of dollars to find their business models, or billions in the case of Uber, but none of these companies are like those companies in the generative AI space. GenAI businesses don’t have the same economics, nor do they have the same total addressable markets. If you’re going to say “Amazon Web Services,” I already explained why you’re wrong a few weeks ago . These startups are their VC firms’ subprime mortgages, overstuffed valuations with no exit route, and no clear example of how to sell them or who to sell them to. The closest they’ve got is using generative AI startups as beauty pageants for guys wearing Patagonia, finding ways to pretend that the guy who runs an AI startup — sorry, AI lab — is some sort of mysterious genius versus just another founder in just another bubble with just another overstuffed valuation.  The literal only liquidity mechanism (outside of Cognigy) that generative AI has had so far is “selling AI talent to big tech at a premium.” Nobody has gone or is going public, and if they are not going public, the only route for these companies is to either become profitable — which they haven’t — or sell to somebody, which they do not. But I’ve been dancing around the real reason they won’t sell: because, fundamentally, generative AI does not let companies build something new . Anyone that builds a generative AI product is ultimately just prompting the model, albeit in increasingly more-complex ways at the scale of something like Claude Code — though Anthropic has the advantage of being one of the main veins of infrastructure. This means that a generative AI company owns very few unique things beyond their talent, and will forever be at the mercy of any and all decisions that their model provider makes, such as increasing prices or creating competing products.  I know it sounds ludicrous, but this is the reality of these companies. While there are some companies that have some unique training and models, none of them seem to be building interesting or unique products as a result.  If your argument is that these things take some time — how long do they have?  No, really! So many of you have said that “this is what happens, they burn a bunch of money, they grow, and then…” and then you stop short because the next thing you say is “turn profitable by getting enterprise customers.” Nobody can do the first part and few can do the second part in anything approaching a consistent fashion.  But really, how long should we give them? Three years?  Perplexity’s had three years and a billion dollars, it doesn’t seem to be close to profitable. How long does Perplexity deserve, exactly? An eternity?  Every single example of a company that has “burned a lot of money and then not done so in the end” has been a company with a physical thing or connections to the real world, with the exception of Facebook, which was never the kind of cash-burning monstrosity that generative AI is.  There has never been a software company that has just chewed through hundreds of millions — or billions — of dollars and then suddenly became profitable, mostly because the magical valuations of software have been in their ability to transcend infrastructure. One’s unit economics in the sales of software like Microsoft Office or providing access to Instagram do not require the most powerful graphics processing units run at full tilt at all times, and those are products that people like and want to use every day.  I get people saying “they’re in the growth stage!” about a few companies, but when all of them are unprofitable, and even the unprofitable ones outside of OpenAI and Anthropic aren’t really making impressive amounts of money anyway? C’mon! This isn’t anything like any boom that leads to something, and it’s because the economics do not make sense. And that’s before we get to OpenAI and Anthropic! So, as a reminder, OpenAI appears to have burned at least ten billion dollars in the last two months. It is has just raised another $8.3 billion dollars ( after raising $10 billion in June according to the New York Times ), and intends to receive around $22.5 billion by the end of year from SoftBank, and that is assuming it becomes a for-profit entity by the end of the year, and if that doesn’t happen, the round gets cut to $20 billion total, meaning that SoftBank would only be on the hook for a further $1.7 billion. I am repeating myself, but I need you to really get this: OpenAI just got $10 billion in June 2025, and had to raise another $8.3 billion in August 2025. That is an unbelievable cash burn, one dwarfing any startup in history, rivalled only by xAI, makers of “Grok, the racist LLM,” losing it over $1 billion a month . I should be clear that if OpenAI does not convert to a for-profit, there is no path forward. To continue raising capital, OpenAI must have the promise of an IPO. It must go public , because at a valuation of $300 billion, OpenAI can no longer be acquired, because nobody has that much money and, if let’s be real, nobody actually believes OpenAI is worth that much. The only way to prove that anybody does is to take OpenAI public, and that will be impossible if it cannot convert. And, ironically, Softbank’s large and late-stage participation makes any exit harder, as early investors will see their holdings diluted as a percentage of total equity — or whatever the hell we’re calling it. While a normal company could just issue equity, and deal with the dilution that way, OpenAI’s structure necessitates a negotiation where companies can obstruct the entire process if they see fit.  Speaking of companies that might obstruct that transition, let’s talk about Microsoft.  As I asked in my premium newsletter a few weeks ago , what if Microsoft doesn’t want OpenAI to convert? It owns all the IP, it owns access to all OpenAI’s research, and already runs most of its infrastructure. While — assuming a best-case scenario — that it would end up owning a massive chunk of the biggest tech startup of all time (I’m talking about equity, not OpenAI’s current profit-sharing units), Microsoft might also believe that it stands more to gain by letting AI die and assuming its role in the AI ecosystem.  Embrace. Extend. Extinguish.  But let’s assume it converts, and OpenAI now…has to continue raising money at a rate that will require it, allegedly, to only need to raise $17 billion in 2027.  That number doesn’t make sense, considering it already had to bring forward its $8.3 billion fundraise by at least three months, but let’s stick with that idea. OpenAI believes it will be profitable, somehow, by 2030, and even if we assume that, that means it intends to burn over a hundred billion dollars to get there. Is the plan to take OpenAI public, dumping a toxic asset onto the public markets, only to let it flounder and convulse and die for all to see? Can you imagine OpenAI’s S-1? How well do you think this company would handle a true financial audit from a major accounting firm?  If you want to know what that looks like, google “WeWork,” which went from tech industry darling to joke in a matter of days, in part because it was forced to disclose how bad things actually were on its S-1. No, really, read this article . With that in mind, I feel similarly about Anthropic. Nobody is buying this company at $ 170 billion , and thus the only way to access liquidity would be to take it public, and show the world how a company that made $72 million in January 2025 and then more than $400 million in July 2025 also loses $3 billion or more after revenue, and then let the market decide on its fair price.  The arguments against my work always come down to “costs will go down” and “these products will become essential.” Outside of ChatGPT, there’s really no proof that these products are anything remotely essential, and I argue there’s very little about ChatGPT that Microsoft couldn’t provide with rate limits via Copilot.  I’d also argue that “essential” is a very subjective term. Essential — in the sense that some people use it as search — doesn’t mean that it’s useful for enterprises, or the majority of people.  And, I guess, ChatGPT somehow makes $1 billion a month in revenue selling access to premium versions of ChatGPT — though I’m not 100% sure how. Assuming it has 20 million customers paying $20 a month, that’s $400 million a month, then 5 million business customers paid an average of $100 each, that’s $900 million…and is that average really that good? Are that many people paying $35 a month, or $50, or $200? OpenAI doesn’t break out the actual revenues behind these numbers for a reason, and I believe that reason is “they don’t look as good.” What’s OpenAI’s churn like? And does it really, as I wrote last week , end the year making more than Spotify at $1.5 billion a month?  We don’t know, and OpenAI (much like Anthropic) has never shared actual revenues, choosing instead to leak to the media and hope to obfuscate the actual amounts of money being spent on its services.  Anyway, long story short, these companies are unprofitable with no end in sight, don’t even make that much money in most cases, are valued more than anybody would ever buy them for, do not have much in the way of valuable intellectual property, and the two biggest players burn billions of dollars more than they make. Even if this were going to happen — it will not! — who would they give the money to and for how long? Would they give it to all the startups? Is every startup going to get a Paycheck Protection Program but for generative AI? How would that play out in rural red districts (where big tech has never been popular), which are being hit with both massive cuts to welfare, as well as the shockwaves of a trade war that has made American agricultural exports (like feedstocks, which previously went to China by the shipload) less appealing worldwide?  So they bail out OpenAI, then stuff it full of government contracts to the tune of $15 billion a year, right? Sorry, just to be clear, that’s the low end of what this would take to do, and they’ll have to keep doing it forever, until Sam Altman can build enough data centers to…keep burning billions, because there’s no actual plan to make this profitable.  Say this happens. Now what? America has a bullshit generative AI company attached to the state that doesn’t really innovate and doesn’t really matter in any meaningful way, except that it owns a bunch of data centers?  I don’t think this happens! I think this is a silly idea, and the most likely situation would be that Microsoft would unhinge its jaw and swallow OpenAI and its customers whole. Hey, did you know that Microsoft’s data center construction is down year-over-year , and it’s basically signed no new data center leases? I wonder why it isn’t building these new data centers for OpenAI? Who knows.  Stargate isn’t saving it, either. As I wrote previously, Stargate doesn’t actually exist beyond the media hype it generated . And yes, OpenAI is offering ChatGPT at $1 for a year to US government workers - and I cannot express how little this means other than that they are horribly desperate. This product doesn't do enough to make it essential, and this fire sale doesn't change anything. Anyway, does the government do this for everybody? Because everyone else is gonna need it as none of these companies can go public as they all suffer from the burden of generative AI. And, if the government does it, will it also subsidize the compute of for-profit companies like Cursor? To what end? Where is the limit? I think this is a question that we have to seriously consider at this point, because its ramifications are significant. If I’m honest, I think the future of LLMs will be client-side on egregiously-expensive personal setups for enthusiasts, and in a handful of niche enterprise roles. Large Language Models do not scale profitably, and their functionality is not significant enough to justify the costs of running them. By immediately applying old economics — the idea that you would pay a monthly fee to have relatively-unlimited access — companies like OpenAI and Anthropic immediately trained users to use their products in a way that was antithetical to their costs.  Then again, had these models been served in a way that was mindful of their costs, there would likely have been no way to even get this far. If OpenAI is making a billion dollars a month, it is possibly losing that much (or more) after revenue, and that’s the money it can get selling the product in a form that can never turn profitable. If OpenAI charged in line with its actual costs, would it even be able to justify a freely-available version of ChatGPT, outside of a few free requests?  The revenue you see today is what people are willing to pay for a product that loses money, and I cannot imagine they would pay as much if the companies in question charged their costs. If I’m wrong, Cursor will be just fine, and that’s assuming that Cursor’s current hobbled form is even profitable, which it has not said it is. So, you’ve got an entire industry of companies that struggle to do anything other than lose a lot of money. Great.  And now we have a massive expansive data centre buildout, the likes of which we’ve never seen, all to capture demand for a product that nobody makes much money selling. This, naturally, leads to an important question: how do these people building data centers actually make money? Last week, the Wall Street Journal published one of the more worrying facts I’ve seen in the last two years: The massive buildout of data centers — and the associated physical gear like chips, servers, and raw materials for building them — has become a massive, dominant economic force…building capacity for an industry that is yet to prove it can make real revenues.  And no, Microsoft talking about its Azure revenue in its last quarterly earnings for the first time is not the same thing , as it stopped explicitly stating their AI revenue in January (when it was $13 billion annualized). Anyway, AI capex allegedly — though I have some questions about this figure! — accounts for 1.2% of the US GDP in the first half of the year, and accounted for more than half of the (to quote the Wall Street Journal) “already-sluggish” 1.2% growth rate of the US economy.  Another Wall Street Journal piece published a few days later discussed how data center development is souring the free cash flow for big tech, turning them from the kind of “asset-light” businesses that the markets love into entities burdened by physical real estate and their associated costs: These numbers are all very scary, and I mean that sincerely, but they also fail to express why. How much was actually spent on AI capex in the US? One would think two different articles on this subject would include that number versus a single quarter’s worth, but from my estimates, I expect capital expenditures from the Magnificent Seven alone to crest $200 billion in the first half of 2025, with Axios estimating they’d spend around $400 billion this year .  Most articles are drafting off of a blog from Paul Kedrosky , who estimates total AI capex would be somewhere in the region of $520 billion in total for the year, which felt conservative to me, so I did the smart thing and asked him. Kedrosky noted that these numbers focus entirely on the four big spenders — Microsoft, Google, Meta and Amazon, and his own estimated $312 billion capex, and the 1.2% number came from the assumption that the US GDP in 2025 will be around $28 trillion (which, I add, is significantly lower than other forecasts, which puts it closer to $30 trillion ).  Kedrosky, in his own words, was trying to be conservative, using public data and then building his analysis from there. I, personally, believe his estimate is too conservative — because it doesn’t factor in the capital expenditures from Oracle, which (along with Crusoe) is building the vast Abilene Texas data center for OpenAI , or any private data center developers sinking cash into AI capex.  When I asked him to elaborate, he estimated that “...AI spend, all-in, was around half of 3.0%  Q2 real GDP growth, so 2-3x the lower bound, given multipliers, debt, etc. it could be half of US GDP full-year GDP growth.” That’s so cool! Half of the US economy’s growth came from building data centers for generative AI, which has the combined revenue of a little more than the fucking smart watch industry in 2024 .   Another troubling point is that big tech doesn’t just buy data centers and then use them, but in many cases pays a construction company to build them, fills them with GPUs and then leases them from a company that runs them , meaning that they don’t have to personally staff up and maintain them. This creates an economic boom for construction companies in the short term, as well as lucrative contracts for ongoing support… as long as the company in question still wants them. While Microsoft or Amazon might use a data center and, indeed, act as if it owns it, ultimately somebody else is holding the bag and the ultimate responsibility for the data centers. One such company is QTS, a data center developer that leases to both Amazon and Meta according to the New York Times , which was acquired by Blackstone in 2021 for $10 billion. Since then, Blackstone has used commercial mortgage-backed securities — I know! — to raise over $8.7 billion since then to sink into QTS’ expansion , and as of mid-July said it’d be investing $25 billion in AI data centers and energy . Blackstone, according to the New York Times, sees “strong demand from tech companies,” who are apparently “willing to sign what they describe as airtight leases for 15 to 20 years to rent out data center space.”  Yet the Times also names another problem — the “unanswered question” of how these private equity firms actually exit these situations. Blackstone, KKR and other asset management firms do not buy companies with the intention of syphoning off revenue, but to pump them up and sell them to another company. Much like AI startups, it isn’t obvious who would buy QTS at what I imagine would be a $25 billion or $30 billion valuation, meaning that Blackstone would have to take them public. Similarly, KKR’s supposed $50 billion partnership with investment firm Energy Capital partners to build data centers and their associated utilities does not appear to have much of an exit plan either. And let’s not forget Oracle, OpenAI, and Crusoe’s abominable mess in Abilene Texas, where Oracle is paying for the $40 billion of GPUs and Crusoe is spending $15 billion raised from Blue Owl Capital and Primary Digital infrastructure to build data centers for OpenAI, a company that loses billions of dollars a year. Why? So that OpenAI can, allegedly starting in 2028, pay Oracle $30 billion a year for compute , and yes, I am being fully serious.  To be clear, OpenAI, by my estimates, has only made around $5.26 billion this year (and will have trouble hitting its $12.7 billion revenue projection for 2025), and will likely lose more than $10 billion to do so. Oracle will, according to The Information, owe Crusoe $1 billion in payments across the 15 year span of its lease . How does Crusoe afford to pay back its $15 billion in loans? Beats me! The Information says it’s raising $1 billion to “take on cloud giants” by “earning construction management fees and rent, and it can sell its stake in the project upon reaching certain completion milestones,” while also building its own AI compute, making the assumption that the demand is there outside of hyperscalers. Then there’s CoreWeave, my least-favourite company in the world. As I discussed a few months ago, CoreWeave is burdened by obscene debt and a horrifying cash burn, and has seen its stock spike to a high of $183 on June 20, 2025 to around $111 as of writing this sentence, which has led to its all-stock attempt to acquire developer Core Scientific for $9 billion to start to fall apart as shareholders balk at the worrisome drop in CoreWeave’s stock price. CoreWeave has, since going public, had to borrow billions of dollars to fund its obscene capital expenditures to handle the upcoming October 2025 start date for OpenAI’s $11.9 billion, 5-year-long deal for compute , which is also when CoreWeave must start paying off its largest loan. CoreWeave lost $314 million in its last earnings, and I see no path to profitability or, honestly, its ability to keep doing business if the market sours. Coreweave, I add, is pretty much reliant on Microsoft as its primary customer. While this relationship has been fairly smooth (so far, and as far as we know), this dependence also presents an existential threat to Coreweave, and is part of the reason why I’m so pessimistic about its survival. Microsoft has its own infrastructure, and has every incentive to cut out middlemen when it's able to meet supply with the demand it itself owns (or leases, rather than subcontracts out), simply because middlemen add costs and shrink margins . If Microsoft walks, what’s left? How does it service its ongoing obligations, and its mountain of debt?  In all of these cases, data center developers seem to have very few options as to making actual money. We have companies spending billions of dollars to vastly expand their data center footprint, but very little evidence that doing so results in revenue let alone some sort of payoff, and similarly, the actual capital expenditures they’re making are…much smaller than those of big tech.  Digital Realty Trust — one of the largest developers with over 300 data centers worldwide and $5.55 billion in revenue in 2024 — only spent $3.5 billion in capex last quarter , and Equinix ( $8.7 billion revenue in 2024 ), which has 270 of them, put capex at $3.5 billion too . NTT Global Data Centers, which has over 160 data centers, has dedicated $10 billion in capital expenditures “through 2027” to build out data centers .  Yet in many of these cases, it’s because these companies are — to quote a source of mine — “functionally obsolete for this cycle,” because legacy data centers are not plug-and-play ready for GPUs to slot into. Any investment in capex by these companies would have to be for both GPUs and either building or retrofitting (basically ripping the insides out of old) data centers. This means that the money flowing into AI data centers is predominantly going to neoclouds like CoreWeave and Crusoe, and all seems to flow back to private equity firms that never thought about where the cashout might be. Blackstone led CoreWeave’s $7.5 billion loan with Magnetar Capital , and Crusoe signed a deal a week ago with infrastructure firm Blackstone-owned Tallgrass to build a data center in Wyoming , all of which seems very good for Blackstone unless you think “how does it actually make money here,” as private equity firms do not generally like to hold assets longer than five years.  Even if it did, its capital expenditures are a drop in the bucket in the grand scheme of things. Assuming Crusoe burns, as The Information suggests it will , as much as $4 billion in 2025, CoreWeave spends as much as $20 billion, Digital Realty Trust spends $14 billion, Global Data Centers spends $3.33 billion (that’s $10bn over 3 years), and Equinix spends $14 billion. That’s $55.33 billion in AI capex spent in 2025 from the largest developers of data centers in the world. For some context, as discussed above, $102 billion was spent by Meta, Alphabet, Microsoft and Amazon in the last quarter.  Private equity may ultimately face the same problem as many AI startups: there is no clear exit strategy for these investments. In the absence of real liquidity, firms will likely resort to all manner of financial engineering (read: bullshit) —marking up portfolio companies using internally generated valuations, charging fees on those inflated marks, and using those marks to entice new commitments from limited partners. Compounding this is their ability to lend increasing amounts of capital to their own portfolio companies via affiliated private credit vehicles—effectively recycling capital and pushing valuation risk further down the line. This kind of self-reinforcing leverage loop is particularly opaque in private credit, which now underpins much of the AI infrastructure buildout. The complexity of these arrangements makes it hard to anticipate the full economic fallout if the cycle breaks down, but the systemic risk is building. In any case, the supposed “AI capex boom” that is driving the US economy is not, as reported, driven by the massive interest in building out AI infrastructure for a variety of customers.  The reality is simple: the majority of all AI capex is from big tech, which is a massive systemic weakness in our economy. While some might say that “AI capex” has swallowed the US economy, I think it’s more appropriate to say that Big Tech Capex Has Swallowed The US Economy. I also want to be clear that the economy — which is the overall state of the country’s production and consumption of stuff, and the flow of money between participants in said economy — and the markets (as in the stock market) are very different things, but the calculations from Kedrosky and others have now allowed us to see where one might hit the other. You see, the markets do not actually represent reality . While Microsoft, Amazon, Google, and Meta might want you to think there’s a ton of money in AI, their growth is mostly from selling further iterations and contracts for their existing stuff, or in the case of Meta further increasing its ad revenue. The economy is where things are actually bought and sold, representing the economic effects of both the things happening to build out AI and selling access to services and the AI models themselves. I recognize this is simplistic, but I am laying it out for a reason. As I discussed at length in the Hater’s Guide to the AI Bubble , NVIDIA is the weak point in the stock market, representing roughly 19% of the value of the Magnificent 7, which in turn makes up about 35% of the value of the US stock market. The associated Magnificent Seven stocks have seen a huge boom through their own growth, which has been mistakenly and incorrectly attributed to revenue from AI, which as I laid out previously, is about $35 to $40 billion in the last two years. Nevertheless, the markets can continue to be irrational because all they care about is “number going up,” as the “value” of a stock is oftentimes disconnected from the value of the company itself, instead associated with its propensity for growth. GDP and other measurements of the economy aren’t really something you can fudge quite as easily ( at least, in transparent, democratic societies ), nor can you say a bunch of fancy words to make people feel better in the event that growth stalls or declines.  This leads me to my principle worry: that “AI capex” is actually a term for the expenditures of four companies, namely Microsoft, Amazon, Google and Meta, with NVIDIA’s GPU sales being part of that capex too. While we can include others like Oracle, Musk’s xAI, and various Neoclouds like CoreWeave and Crusoe — who, according to D.A. Davidson’s Gil Luria, will account for about 10% of NVIDIA’s GPU sales in 2025 — the reality is that whatever economic force is being driven by “AI investment” is really just four companies building and leasing data centers to burn on generative AI, a product that makes a relatively small amount of money before losing a great deal more.  42% of NVIDIA’s revenue comes from the Magnificent Seven ( per Laura Bratton at Yahoo Finance ), which naturally means that big tech is the lynchpin of investment in data centers.  I’ll put it far more simply: if AI capex represents such a large part of our GDP and economic growth, our economy does, on some level, rest on the back of Microsoft, Google, Meta and Amazon and their continued investment in AI. What should worry everybody is that Microsoft — which makes up 18.9% of NVIDIA’s revenue — has signed basically no leases in the last 12 months, and its committed datacenter construction and land purchases are down year-over-year .  While its capex may not have dipped yet (in part because the chip-heavy nature of generative AI means that capex isn’t exclusively dominated by property), it’s now obvious that if it does there will be direct effects on both the US economy and stock market, as Microsoft is part of what amounts to a stimulus package propping up America’s economic growth.   And not to repeat the point too much, but big tech has yet to actually turn anything resembling a profit on these data centers, and isn’t making much revenue at all out of generative AI. How, exactly, does this end? What is the plan here? Is big tech going to spend hundreds of billions a year in capital expenditures on generative AI in perpetuity? Will they continue to buy more and more NVIDIA chips as they do so?  At some point, surely these companies have built enough data centers? Surely, at some point, they’ll run out of space to put these GPUs in? Is the plan to, by then, make so much money from AI that it won’t matter? What does NVIDIA do at that point? And how does the US economy rebound from the loss of activity that follows? As I’ve said again and again, the generative AI bubble is, and always has been, fundamentally irrational, and inherently gothic, playing in the ruins, patterns and pathways of previous tech booms despite this one having little or no resemblance to them. Though the tech industry loves to talk about building a glorious future, its present is one steeped in rituals of decay and death, where the virtues of value creation and productivity take a backseat to burning billions and lying to the public again and again and again . The way in which the media has participated in these lies is disgusting. Venture capital, still drunk off the fumes of 2021, keeps running the old playbook: shove as much money into a company as possible in the hopes you can dump it onto an acquirer or the public markets, only to get high on their own supply, pushing valuations to the point that there is no possible liquidity event for the majority of big private AI companies as a result of their overstuffed valuations, burdensome business models and lack of any real intellectual property. And, like the rest of the AI bubble, Silicon Valley’s only liquidity path out of the bubble is big tech itself. Without Google, Character.ai and Windsurf’s founders would likely have been left for dead, and the same goes for Inflection, and I’d even argue Scale AI, whose $14.3 billion “investment” from Meta effectively decapitated the company, removing its CEO Alexandr Wang, leaving the rest of the company to die, laying off 14% of its staff and 500 contractors mere weeks after its CEO and investors cashed in.  In fact, generative AI is turning out to be a fever dream entirely made up by big tech. OpenAI would be dead if it wasn’t for the massive infrastructure provided by Microsoft at-cost in return for rights to its IP, research, and the ability to sell its models on top of the tens of billions of dollars of venture capital thrown into its billion-dollar cash incinerator. Anthropic would be dead if both Google and Amazon — the latter of which provides much of its infrastructure — hadn’t invested billions in keeping it alive so that it can burn $3 billion or more in 2025 while fucking over its enterprise customers and rate limiting the rest .  The generative AI industry is, at its core, unnatural. It does not make significant revenue compared to its unbelievable costs, nor does it have much revenue potential. It requires, unlike just about every software revolution, an unbelievable amount of physical infrastructure to run, and because nobody but big tech can afford to build the infrastructure necessary, creates very little opportunity for competition or efficiency. As the markets are in the throes of the growth-at-all-costs Rot Economy, they have failed to keep big tech in line, conflating big tech’s ability to grow with growth driven as a result of their capital expenditures. Sensible, reasonable markets would notice the decay of free cash flow or the ridiculousness of big tech’s capex bonanza, but instead they clap and squeal every time Satya Nadella jingles his keys. What is missing is any real value generation. Again, I tell you, put aside any feelings you may have about generative AI itself , and focus on the actual economic results of this bubble. How much revenue is there? Why is there no profit? Why are there no exits? Why does big tech, which has sunk hundreds of billions of dollars into generative AI, not talk about the revenues they’re making? Why, for three years straight, have we been asked to “just wait and see,” and for how long are we going to have to wait to see it? What’s incredible is that the inherently compute-intensive nature of generative AI basically requires the construction of these facilities, without actually representing whether they are contributing to the revenues of the companies that operate the models (like Anthropic or OpenAI, or any other business that builds upon them). As the models get more complex and hungry, more data centers get built — which hyperscalers book as long-term revenue, even though it’s either subsidised by said hyperscalers, or funded by VC money. This, in turn, stimulates even more capex spending. And without having to answer any basic questions about longevity or market fit.  Yet the worst part of this financial farce is that we’ve now got a built-in economic breaking point in the capex from AI. At some point capex has to slow — if not because of the lack of revenues or massive costs associated, but because we live in a world with finite space , and when said capex slow happens, so will purchases of NVIDIA GPUs, which will in turn, as proven by Kedrosky and others, slow America’s economic growth. And that growth is pretty much based on the whims of four companies, which is an incredibly risky and scary proposition. I haven’t even dug into the wealth of private credit deals that underpin buildouts for private AI “neoclouds” like CoreWeave, Crusoe, Nebius, and Lambda, in part because their economic significance is so much smaller than big tech’s ugly, meaningless sprawl.  To quote Kedrosky :  You can’t bail this out, because there is nothing to bail out. Microsoft, Meta, Amazon and Google have plenty of money and have proven they can spend it. NVIDIA is already doing everything it can to justify people spending more on its GPUs. There’s little more it can do here other than soak up the growth before the party ends.  That capex reduction will bring with it a reduction in expenditures on NVIDIA GPUs, which will take a chunk out of the US stock market. Although the stock market isn’t the economy, the two things are inherently linked, and the popping of the AI bubble will have downstream ramifications, just like the dot com bubble did on the wider economy. Expect to see an acceleration in layoffs and offshoring, in part driven by a need for tech companies to show — for the first time in living memory — fiscal restraint. For cities where tech is a major sector of the economy — think Seattle and San Francisco — there’ll be knock-on effects to those companies and individuals that support the tech sector (like restaurants, construction companies building apartments, Uber drivers, and so on). We’ll see a drying-up of VC funding. Pension funds will take a hit — which will affect how much people have to spend in retirement. It’ll be grim.  Worse than that is the fact that these data centers will be, by definition, non-performing assets, and one that inflicted an opportunity cost that’ll be almost impossible to calculate. While a house, once built and sold, technically falls into that category (it doesn’t add to any economic productivity), people at least need somewhere to live. Shelter is an essential component of life. You can live without a data center the size of Manhattan.  What would have happened if companies like Microsoft and Meta instead spent the money on things that actually drove productivity, or created a valuable competitive business that drove economic activity? Hell, even if they just gave everyone a 10% raise, it would have likely been better for the economy than this, if we’re factoring in things like consumer spending.  It’s just waste. Profligate, pointless waste.  In summary, we’re already facing the prospect of a recession, and though I am not an economist, I can imagine it being a particularly nasty one given that the Magnificent Seven accounted for 47.87% of the Russell 1000 Index’s returns in 2024 . Even if big tech somehow makes this crap profitable, it’s hard to imagine that they’ll counterbalance any capex reduction with revenue, because there doesn’t seem to be that much revenue in generative AI to begin with. This is what happens when you allow the Rot Economy to run wild, building the stock market and tech industry on growth over everything else. This is what happens when the tech media repeatedly fails to hold the powerful to account, catering to their narratives and making excuses for their abominable, billion-dollar losses and mediocre, questionably-useful products.  Waffle on all you want about the so-called “agentic era” or “annualized revenues” that make you hot under the collar — I see no reason for celebration about an industry with no exit plans and needless capital expenditures that appear to be one of the few things keeping the American economy growing.  I have been writing about the tech industry’s obsession with generative AI for two years, and never have I felt more grim. Before this was an economic uncertainty — a way that our markets might contract, that big tech might take a big haircut, that a bunch of money might be wasted but otherwise the world would keep turning.  It feels as if everything is aligning for disaster, and I fear there’s nothing that can be done to avert it.

0 views

How Much Money Do OpenAI And Anthropic Actually Make?

Hello and welcome to the latest premium edition of Where's Your Ed At, I appreciate any and all of your subscriptions. I work very hard on these, and they help pay for the costs of running Ghost and, well, my time investigating different things. If you're on the fence, subscribe! I promise it's worth it. I also want to give a HUGE thank you to Westin Lee , a writer who has written about business and the use of AI, who was the originator of the whole "what if we used ARR to work out what these people make?" idea. He's been a tremendous help, and I recommend you check out his work. If you're an avid reader of the business and tech media, you'd be forgiven for thinking that OpenAI has made (or will make) in excess of $10 billion this year, and Anthropic in excess of $4 billion. Why? Because both companies have intentionally reported or leaked their "annualized recurring revenue" – a month's revenue multiplied by 12. OpenAI leaked yesterday to The Information that it hit $12 billion in "annual recurring revenue" – suggesting that its July 2025 revenues were around $1 billion. The Information reported on July 1 2025 that Anthropic's annual run rate was $4 billion – meaning that its revenue for the month of June 2025 was around $333 million. Then, yesterday, it reported that the run rate was up to $5 billion.  These do not, however, mean that their previous months were this high, nor do they mean that they've "made" anything close to these numbers. Annualized recurring revenue is one of the most regularly-abused statistics in the startup world, and can mean everything from "[actual month]x12" to "[30 day period of revenue]x12" and in most cases it's a number that doesn't factor in churn. Some companies even move around the start dates for contracts as a means of gaming this number.  ARR, also, doesn’t factor seasonality of revenue into the calculations. For example, you’d expect ChatGPT to have peaks and troughs that correspond with the academic year, with students cancelling their subscriptions during the summer break. If you use ARR, you’re essentially taking one month and treating it as representative of the entire calendar year, when it isn’t.  These companies are sharing (or leaking) their annualized revenues for a few reasons: In any case, I want to be clear this is a standard metric in non-public Software-as-a-Service (SaaS) businesses. Nothing is inherently wrong with the metric, save for its use and what's being interpreted from it. Nevertheless, there has been a lot of reporting on both OpenAI and Anthropic's revenues that has created incredible confusion in the market that benefits both companies, making them seem far more successful than they really are, and giving them credit for revenue they are yet to book. Before I dive into this — and yes, before the premium break — I want to establish some facts. The intention of either reporting or leaking their annualized revenue numbers was to make you think that OpenAI would hit its projected $12.7 billion revenue number, and Anthropic would hit its "optimistic" $4 billion number, because those "annualized revenue" figures sure seem to have the word "annual" in them. Yet through an historic analysis of reported annual recurring revenue numbers over the past three years, I've found things to be a little less certain. You see, when a company reports their "annual recurring revenue," what they're actually telling you is how much they made in a month, and I've sat down and found every single god damn bit of reporting about these numbers, calculating (based on the compound growth necessary between the months of reported monthly revenue) how much these companies are actually making in cash. My analysis, while imperfect (as we lack the data for certain months), aligns closely enough with projections that I am avoiding any egregious statements. OpenAI and Anthropic's previous projections were fairly accurate, though as I'll explain in this piece, I believe their new ones are egregious and ridiculous. More importantly, in all of these stories, there was only one time that these companies shared their revenues — when OpenAI shared its $10 billion runrate in May, though the July $12 billion ARR leak is likely intentional too. In fact, I believe both were an intentional attempt to mislead the general public into believing the company was more successful than it is. Based on my analysis, OpenAI made around $3.616 billion in revenue in 2024, and so far in 2025 has made, by my calculations, around $5.266 billion in revenue as of the end of July . This is also a slower growth rate than it’s experienced so far in the year. Going from $5.5 billion in annualized revenue in December 2024 to $10 billion annualized in May 2025 was a compound growth rate of around 12.7%. The "jump" from $10 billion ARR to $12 billion ARR is 9.54%. While I realize this may not seem like a big drop, every single penny counts, and percentage point shifts are worth hundreds of millions (if not billions) of dollars. OpenAI has been projected to make $12.7 billion in revenue in 2025. Making this number will be challenging, and require OpenAI to grow by 14%, every single month, without fail. For OpenAI to hit this number will require it to make nearly $2 billion a month in revenue by the end of the year to account for the disparity with the earlier months in the year when it made far, far less. I also have serious suspicions about how much OpenAI actually made in May, June and July 2025.  While The Information reported OpenAI hit $12 billion in annualized revenue, they did so in an obtuse way : Yet the New York Times, mere days later, reported $13 billion annualized revenue : First and foremost, it’s incredibly fucking suspicious that two very different numbers were reported here so closely, and even more so that the June 9 2025 announcement of OpenAI hitting $10 billion in annualized revenue was not, as I had originally believed, discussing the month of May 2025. This likely means that OpenAI is not using standard annualized revenue metrics - which would traditionally mean “the last month’s revenue multiplied by 12,” and instead choosing “if all the monthly subscribers and contracts that are currently paying us on this day, June 9 2025, were to be multiplied by 12, we’d have $X annualized revenue.” This is astronomically fucking dodgy. For the sake of this analysis, I am assuming any announcement of annualized revenue refers to the month previous. So, for example, when OpenAI announced they hit $10 billion in annualized revenue, I am going to assume this is for the month of May 2025. This analysis is going to favour the companies in question. If OpenAI “hit $10 billion annualized” in or around June 9 2025, it likely means that their May revenues were lower than that. Similarly, OpenAI “hitting” $12 billion in annualized revenue (announced end of July 2025) - which I have factored into my analysis - is considered the revenue they hit in July 2025.  In reality, this is likely to credit them more revenue than they deserve. If June’s annualized revenue was $10 billion, that means they made $833 million, rather than the $939 million I credit them with for the month.  One cannot hit $12 billion AND $13 billion annualized in one month unless you are playing extremely silly games with the numbers in question, such as moving around when you start a 30 day period to artificially inflate things. In any case, my analysis for OpenAI’s revenue for August is around $13.145 billion - so in line with a “$13 billion annualized” figure. In any case, I am sticking with my analysis as it stands. However, the timing of these annualized revenue leaks now makes me doubt the veracity of their previous leaks, in the sense that there’s every chance that they too are either inflated or used in a deceptive manner. Based on these numbers, OpenAI's current growth rate is around 9.54% — and at that current pace, it will finish the year at around $11.89 billion in revenue. This is an impressive number, meaning it’d be making over $1.5 billion a month in revenue by December 2025 — but such an impressive number will be difficult to reach, and mean it has something in the region of $18 billion annualized revenue by the end of the year. I also question whether it can make it, and even if it does, how it could possibly afford to serve that revenue long-term. In Anthropic's case, I am extremely confident, based on its well-reported annualized revenues, that Anthropic has, through July 2025, made around $ 1.5 billion in revenue. This is, of course, assuming that their annualized revenue leaks are for calendar months, and if they're not, this number could actually be lower. This is not a question of opinion. Other than April, we have ARR for every single month of the year.  Bloomberg is now reporting that Anthropic sees its revenue rate " maybe [going] to $9 billion annualized by year-end ," which, to use a technical term, is total bullshit, especially as this number was leaked as Anthropic is fundraising. In any case, I believe Anthropic can beat its base case estimates. It will almost certainly cross $2 billion in revenue, but I also believe that revenue growth is slowing for these companies, and the amount of cash we credit them as actually making is decidedly more average than "annualized revenue" would have you believe.

0 views

Is SoftBank Still Backing OpenAI?

Earlier in the week, the Wall Street Journal reported that SoftBank and OpenAI's "$500 billion" "AI Project" was now setting a "more modest goal of building a small data center by year-end." One might be forgiven for being a little confused here, as there is, apparently, a Stargate Data Center being built in Abilene Texas. Yet the Journal added another detail: Confusing, right? One might also be confused by the Bloomberg story called " Inside The First Stargate AI Data Center ," which had the subheadline "OpenAI, Oracle and SoftBank hope that the site in Texas is the first of many across the US." More-confusingly, the piece talked about Stargate LLC, which OpenAI, Oracle and SoftBank were (allegedly) shareholders of. Yet I have confirmed that SoftBank never, ever had any involvement with the site in Abilene Texas. It didn't fund it, it didn't build it, it didn't choose the site and, in fact, does not appear to have anything to do with any data center that OpenAI uses. The data center many, many reporters have referred to as "Stargate" has nothing to do with the "Stargate data center project."  Any reports suggesting otherwise are wrong, and I believe that this is a conscious attempt at misleading the public by OpenAI and SoftBank. I confirmed the following with a PR representative from Crusoe, one of the developers of the site in Abilene Texas: And, as a reminder, Stargate as an entity was never formed — my source being the CEO of Oracle Safra Catz on Oracle’s earnings call . This is an astonishing — and egregious — act of misinformation on the part of Sam Altman and OpenAI. By my count, at least 15 different stories attribute the Abilene Texas data center to the Stargate project, despite the fact that SoftBank was never and has never been involved. One would forgive anyone who got this wrong, because OpenAI itself engaged in the deliberate deception in its own announcement of the Stargate Project [emphasis mine]: You can weasel-word all you want about how nobody has directly reported that SoftBank was or was not part of Abilene. This is a deliberate, intentional deception, perpetrated by OpenAI and SoftBank, who deliberately misled both the public and the press as a means of keeping up the appearance that SoftBank was deeply involved in (and financially obligated to) the Abilene site. Based on reporting that existed at the time but was never drawn together, it appears that Abilene was earmarked by Microsoft for OpenAI's use as early as July 2024, and never involved SoftBank in any way, shape or form. The "Stargate" Project, as reported, was over six months old when it was announced in January 2025, and there have been no additional sites added other than Abilene. In simpler terms, Stargate does not exist other than as a name that Sam Altman gives things to make them feel more special than they are, and SoftBank was never involved. Stargate does not exist as reported. The use of the term "Stargate" is an intentional act of deceit, but beneath the surface lies, in my mind, a much bigger story. Furthermore, I believe this deceit means that we should review any and all promises made by OpenAI and SoftBank, and reconsider any and all statements they've made, or that have been made about them. Let's review. According to reporting: Yet based on my research, it appears that SoftBank may not be able to — or want to — proceed with any of these initiatives other than funding OpenAI's current round, and evidence suggests that even if it intends to, SoftBank may not be able to afford investing in OpenAI further. I believe that SoftBank and OpenAI's relationship is an elaborate ruse, one created to give SoftBank the appearance of innovation, and OpenAI the appearance of a long-term partnership with a major financial institution that, from my research, is incapable of meeting the commitments it has made. In simpler terms, OpenAI and SoftBank are bullshitting everyone. I can find no tangible proof that SoftBank ever intended to seriously invest money in Stargate, and have evidence from its earnings calls that suggests SoftBank has no idea — or real strategy — behind its supposed $3-billion-a-year deployment of OpenAI software. In fact, other than the $7.5 billion that SoftBank invested earlier in the year, I don't see a single dollar actually earmarked for anything to do with OpenAI at all. SoftBank is allegedly going to send upwards of $20 billion to OpenAI by December 31 2025, and doesn't appear to have started any of the processes necessary to do so, or shown any signs it will. This is not a good situation for anybody involved.

0 views

The Hater's Guide To The AI Bubble

Hey! Before we go any further — if you want to support my work, please sign up for the premium version of Where’s Your Ed At, it’s a $7-a-month (or $70-a-year) paid product where every week you get a premium newsletter, all while supporting my free work too.  Also, subscribe to my podcast Better Offline , which is free. Go and subscribe then download every single episode. Here's parts 1 , 2 and 3 of the audio version of the Hater's Guide. One last thing: This newsletter is nearly 14,500 words. It’s long. Perhaps consider making a pot of coffee before you start reading.  Good journalism is making sure that history is actively captured and appropriately described and assessed, and it's accurate to describe things as they currently are as alarming. And I am alarmed. Alarm is not a state of weakness, or belligerence, or myopia. My concern does not dull my vision, even though it's convenient to frame it as somehow alarmist, like I have some hidden agenda or bias toward doom. I profoundly dislike the financial waste, the environmental destruction, and, fundamentally, I dislike the attempt to gaslight people into swearing fealty to a sickly and frail psuedo-industry where everybody but NVIDIA and consultancies lose money. I also dislike the fact that I, and others like me, are held to a remarkably different standard to those who paint themselves as "optimists," which typically means "people that agree with what the market wishes were true." Critics are continually badgered, prodded, poked, mocked, and jeered at for not automatically aligning with the idea that generative AI will be this massive industry, constantly having to prove themselves, as if somehow there's something malevolent or craven about criticism, that critics "do this for clicks" or "to be a contrarian." I don't do anything for clicks. I don't have any stocks or short positions. My agenda is simple: I like writing, it comes to me naturally, I have a podcast, and it is, on some level, my job to try and understand what the tech industry is doing on a day-to-day basis. It is easy to try and dismiss what I say as going against the grain because "AI is big," but I've been railing against bullshit bubbles since 2021 — the anti-remote work push ( and the people behind it ), the Clubhouse and audio social networks bubble , the NFT bubble , the made-up quiet quitting panic , and I even, though not as clearly as I wished, called that something was up with FTX several months before it imploded .  This isn't "contrarianism."  It's the kind of skepticism of power and capital that's necessary to meet these moments, and if it's necessary to dismiss my work because it makes you feel icky inside, get a therapist or see a priest. Nevertheless, I am alarmed , and while I have said some of these things separately, based on recent developments, I think it's necessary to say why.  In short, I believe the AI bubble is deeply unstable, built on vibes and blind faith, and when I say "the AI bubble," I mean the entirety of the AI trade. And it's alarmingly simple, too. But this isn’t going to be saccharine, or whiny, or simply worrisome. I think at this point it’s become a little ridiculous to not see that we’re in a bubble. We’re in a god damn bubble, it is so obvious we’re in a bubble, it’s been so obvious we’re in a bubble, a bubble that seems strong but is actually very weak, with a central point of failure. I may not be a contrarian, but I am a hater . I hate the waste, the loss, the destruction, the theft, the damage to our planet and the sheer excitement that some executives and writers have that workers may be replaced by AI — and the bald-faced fucking lie that it’s happening, and that generative AI is capable of doing so. And so I present to you — the Hater’s Guide to the AI bubble, a comprehensive rundown of arguments I have against the current AI boom’s existence. Send it to your friends, your loved ones, or print it out and eat it.   No, this isn’t gonna be a traditional guide, but something you can look at and say “oh that’s why the AI bubble is so bad.” And at this point, I know I’m tired of being gaslit by guys in gingham shirts who desperately want to curry favour with other guys in gingham shirts but who also have PHDs. I’m tired of reading people talk about how we’re “in the era of agents” that don’t fucking work and will never fucking work. I’m tired of hearing about “powerful AI” that is actually crap, and I’m tired of being told the future is here while having the world’s least-useful most-expensive cloud software shoved down my throat. Look, the generative AI boom is a mirage, it hasn’t got the revenue or the returns or the product efficacy for it to matter, everything you’re seeing is ridiculous and wasteful, and when it all goes tits up I want you to remember that I wrote this and tried to say something. As I write this, NVIDIA is currently sitting at $170 a share — a dramatic reversal of fate after the pummelling it took from the DeepSeek situation in January, which sent it tumbling to a brief late-April trip below $100 before things turned around.  The Magnificent 7 stocks — NVIDIA, Microsoft, Alphabet (Google), Apple, Meta, Tesla and Amazon — make up around 35% of the value of the US stock market, and of that, NVIDIA's market value makes up about 19% of the Magnificent 7. This dominance is also why ordinary people ought to be deeply concerned about the AI bubble. The Magnificent 7 is almost certainly a big part of their retirement plans, even if they’re not directly invested. Back in May, Yahoo Finance's Laura Bratton reported that Microsoft (18.9%), Amazon (7.5%), Meta (9.3%), Alphabet (5.6%), and Tesla (0.9%) alone make up 42.4% of NVIDIA's revenue. The breakdown makes things worse. Meta spends 25% — and Microsoft an alarming 47% — of its capital expenditures on NVIDIA chips, and as Bratton notes, Microsoft also spends money renting servers from CoreWeave, which analyst Gil Luria of D.A.Davidson estimates accounted for $8 billion (more than 6%) of NVIDIA's revenue in 2024. Luria also estimates that neocloud companies like CoreWeave and Crusoe — that exist only to prove AI compute services — account for as much as 10% of NVIDIA's revenue. NVIDIA's climbing stock value comes from its continued revenue growth. In the last four quarters, NVIDIA has seen year-over-year growth of 101%, 94%, 78% and 69%, and, in the last quarter , a little statistic was carefully brushed under the rug: that NVIDIA missed , though narrowly, on data center revenue. This is exactly what it sounds like — GPUs that are used in servers, rather than gaming consoles and PCs (. Analysts estimated it would make $39.4 billion from this category, and NVIDIA only (lol) brought in $39.1 billion . Then again, it could be attributed to its problems in China, especially as the H20 ban has only just been lifted . In any case, it was a miss! NVIDIA's quarter-over-quarter growth has also become aggressively normal — from 69%, to 59%, to 12%, to 12% again each quarter, which, again, isn't bad (it's pretty great!), but when 88% of your revenue is based on one particular line in your earnings, it's a pretty big concern, at least for me. Look, I'm not a stock analyst, nor am I pretending to be one, so I am keeping this simple: In simpler terms, 35% of the US stock market is held up by five or six companies buying GPUs. If NVIDIA's growth story stumbles, it will reverberate through the rest of the Magnificent 7, making them rely on their own AI trade stories. And, as you will shortly find out, there is no AI trade, because generative AI is not making anybody any money. I'm so tired of people telling me that companies are "making tons of money on AI." Nobody is making a profit on generative AI other than NVIDIA. No, really, I’m serious.  If they keep their promises, by the end of 2025, Meta, Amazon, Microsoft, Google and Tesla will have spent over $560 billion in capital expenditures on AI in the last two years, all to make around $35 billion. This is egregiously fucking stupid. As of January 2025, Microsoft's "annualized" — meaning [best month]x12 — revenue from artificial intelligence was around $13 billion , a number that it chose not to update in its last earnings, likely because it's either flat or not growing, though it could in its upcoming late-July earnings. Yet the problem with this revenue is that $10 billion of that revenue, according to The Information , comes from OpenAI's spend on Microsoft's Azure cloud, and Microsoft offers preferential pricing — " a heavily discounted rental rate that essentially only covers Microsoft's costs for operating the servers " according to The Information. In simpler terms, 76.9% of Microsoft's AI revenue comes from OpenAI, and is sold at just above or at cost, making Microsoft's "real" AI revenue about $3 billion, or around 3.75% of this year's capital expenditures , or 16.25% if you count OpenAI's revenue, which costs Microsoft more money than it earns. The Information reports that Microsoft made $4.7 billion in "AI revenue" in 2024, of which OpenAI accounted for $2 billion, meaning that for the $135.7 billion that Microsoft has spent in the last two years on AI infrastructure , it has made $17.7 billion, of which OpenAI accounted for $12.7 billion. Things do not improve elsewhere. An analyst estimates that Amazon, which plans to spend $105 billion in capital expenditures this year, will make $5 billion on AI in 2025 , rising, and I quote, "as much as 80%," suggesting that Amazon may have made a measly $2.77 billion in 2024 on AI in a year when it spent $83 billion in capital expenditures . [editor's note: this piece originally said "$1 billion" instead of "$2.77 billion" due to a math error, sorry!] Last year, Amazon CEO Andy Jassy said that “ AI represents for sure the biggest opportunity since cloud and probably the biggest technology shift and opportunity in business since the internet. " I think he's full of shit. Bank of America analyst Justin Post estimated a few weeks ago that Google's AI revenue would be in the region of $7.7 billion , though his math is, if I'm honest, a little generous: Google's "One" subscription includes increased cloud storage across Google Drive, Gmail and Google Photos, and added a $20-a-month "premium" plan in February 2024 that included access to Google's various AI models. Google has claimed that the "premium AI tier accounts for millions" of the 150 million subscribers to the service , though how many millions is impossible to estimate — but that won't stop me trying!  Assuming that $3.1 billion in 2025 revenue would work out to $258 million a month, that would mean there were 12.9 million Google One subscribers also paying for the premium AI tier. This isn't out of the realm of possibility — after all, OpenAI has 15.5 million paying subscribers — but Post is making a generous assumption here. Nevertheless, we'll accept the numbers as they are. And the numbers fuckin' stink! Google's $1.1 billion in workspace service revenue came from a forced price-hike on those who use Google services to run their businesses , meaning that this is likely not a number that can significantly increase without punishing them further. $7.7 billion of revenue — not profit! — on $75 billion of capital expenditures. Nasty! Someone's gonna get mad at me for saying this, but I believe that Meta is simply burning cash on generative AI. There is no product that Meta sells that monetizes Large Language Models, but every Meta product now has them shoved into them, such as your Instagram DMs oinking at you to generate artwork based on your conversation. Nevertheless, we do have some sort of knowledge of what Meta is saying due to the copyright infringement case Kadrey v. Meta . Unsealed judgment briefs revealed in April that Meta is claiming that "GenAI-driven revenue will be more than $2 billion," with estimates as high as $3 billion.  The same document also claims that Meta expects to make $460 billion to $1.4 trillion in total revenue through 2035, the kind of thing that should get you fired in an iron ball into the sun. Meta makes 99% of its revenue from advertising, and the unsealed documents state that it "[generates] revenue from [its] Llama models and will continue earning revenue from each iteration," and "share a percentage of the revenue that [it generates] from users of the Llama models...hosted by those companies," with the companies in question redacted. Max Zeff of TechCrunch adds that Meta lists host partners like AWS, NVIDIA, Databricks, Groq, Dell, Microsoft Azure, Google Cloud, and Snowflake, so it's possible that Meta makes money from licensing to those companies. Sadly, the exhibits further discussing these numbers are filed under seal. Either way, we are now at $332 billion of capital expenditures in 2025 for $28.7 billion of revenue, of which $10 billion is OpenAI's "at-cost or just above cost" revenue. Not great. Despite its prominence in the magnificent 7, Tesla is one of the least-exposed of the magnificent 7 to the AI trade, as Elon Musk has turned it into a meme stock company. That doesn't mean, of course, that Musk isn't touching AI. xAI, the company that develops racist Large Language Model "Grok" and owns what remains of Twitter, apparently burns $1 billion a month , and The Information reports that it makes a whopping $100 million in annualized revenue — so, about $8.33 million a month. There is a shareholder vote for Tesla to potentially invest in xAI , which will probably happen, allowing Musk to continue to pull leverage from his Tesla stock until the company's decaying sales and brand eventually swallow him whole. But we're not talking about Elon Musk today. Apple Intelligence radicalized millions of people against AI, mostly because it fucking stank. Apple clearly got into AI reluctantly, and now faces stories about how they " fell behind in the AI race ," which mostly means that Apple aggressively introduced people to the features of generative AI by force, and it turns out that people don't really want to summarize documents, write emails, or make "custom emoji," and anyone who thinks they would is a fucking alien. In any case, Apple hasn't bet the farm on AI, insomuch as it hasn't spent two hundred billion dollars on infrastructure for a product with a limited market that only loses money. To be clear, I am not saying that any of the Magnificent 7 are going to die — just that five companies' spend on NVIDIA GPUs largely dictate how stable the US stock market will be. If any of these companies (but especially NVIDIA) sneeze, your 401k or your kid’s college fund will catch a cold.  I realize this sounds a little simplistic, but by my calculations, NVIDIA's value underpins around 8% of the value of the US stock market. At the time of writing, it accounts for roughly 7.5% of the S&P 500 — an index of the 500 largest US publicly-traded companies. A disturbing 88% of Nvidia’s revenue comes from enterprise-scale GPUs primarily used for generative AI, of which five companies' spend makes up 42% of its revenue. In the event that any one of these companies makes significant changes to their investments in NVIDIA chips, it will eventually have a direct and meaningful negative impact on the wider market. NVIDIA's earnings are, effectively, the US stock market's confidence, and everything rides on five companies — and if we're honest, really four companies — buying GPUs for generative AI services or to train generative AI models. Worse still, these services, while losing these companies massive amounts of money, don't produce much revenue, meaning that the AI trade is not driving any real, meaningful revenue growth. Any of these companies talking about "growth from AI" or "the jobs that AI will replace" or "how AI has changed their organization" are hand-waving to avoid telling you how much money these services are actually making them. If they were making good money and experiencing real growth as a result of these services, they wouldn't shut the fuck up about it! They'd be in your ear and up your ass hooting about how much cash they were rolling in! And they're not, because they aren't rolling in cash, and are in fact blowing nearly a hundred billion dollars each to build massive, power-hungry, costly data centers for no real reason. Don’t watch the mouth — watch the hands. These companies are going to say they’re seeing growth from AI, but unless they actually show you the growth and enumerate it, they are hand-waving.  So, one of the most annoying and consistent responses to my work is to say that either Amazon or Amazon Web Services “ran at a loss,” and that Amazon Web Services — the invention of modern cloud computing infrastructure — “lost money and then didn’t.”  The thing is, this statement is one of those things that people say because it sounds rational. Amazon did lose money, and Amazon Web Services was expensive, that’s obvious, right?  The thing is, I’ve never really had anyone explain this point to me, so I am finally going to sit down and deal with this criticism, because every single person who mentions it thinks they just pulled Excalibur from the stone and can now decapitate me. They claim that because people in the past doubted Amazon — because, or in addition to the burn rate of Amazon Web Services as the company built out its infrastructure — that I too am wrong, because they were wrong about that. This isn't Camelot, you rube! You are not King Arthur! I will address both the argument itself and the "they" part of it too — because if the argument is that the people that got AWS wrong should not be trusted, then we should no longer trust them, the people actively propagandizing our supposed generative AI future. So, I'm honestly not sure where this argument came from, because there is, to my knowledge, no story about Amazon Web Services where somebody suggested its burnrate would kill Amazon. But let’s go back in time to the May 31 1999 piece that some might be thinking of, called " Amazon.bomb ," and how writer Jacqueline Doherty was mocked soundly for "being wrong" about Amazon, which has now become quite profitable. I also want to be clear that Amazon Web Services didn't launch until 2006, and Amazon itself would become reliably profitable in 2003 . Technically Amazon had opened up Amazon.com's web services for developers to incorporate its content into their applications in 2002, but what we consider AWS today — cloud storage and compute — launched in 2006. But okay, what did she actually say? Her argument, for the most part, is that Amazon was burning cash, had a ton of competition from other people doing similar things, and that analysts backed her up: Fair arguments for the time, though perhaps a little narrow-minded. The assumption wasn't that what Amazon was building was a bad idea, but that Amazon wouldn't be the ones to build it, with one saying: In simpler terms: Amazon's business model wasn't in question. People were buying shit online. In fact, this was just before the dot com bubble burst, and when optimism about the web was at a high point. Yet the comparison stops there — people obviously liked buying shit online, it was the business models of many of these companies — like WebVan — that sucked. Amazon Web Services was an outgrowth of Amazon's own infrastructure, which had to expand rapidly to deal with the influx of web traffic for Amazon.com, which had become one of the world's most popular websites and was becoming increasingly more-complex as it sold things other than books. Other companies had their own infrastructure, but if a smaller company wanted to scale, they’d basically need to build their own thing. It's actually pretty cool what Amazon did! Remember, this was the early 2000s, before Facebook, Twitter, and a lot of the modern internet we know that runs on services like Amazon Web Services, Microsoft Azure and Google Cloud. It invented the modern concept of compute ! But we're here to talk about Amazon Web Services being dangerous for Amazon and people hating on it. A November 2006 story from Bloomberg talked about Jeff Bezos' Risky Bet to "run your business with the technology behind his web site," saying that "Wall Street [wanted] him to mind the store." Bezos, referred to as a "one-time internet poster boy" that became "a post-dot-com piñata." Nevertheless, this article has what my haters crave: Pretty bad, right? My goose is cooked? All those analysts seem pretty mad! Except it's not, my goose is raw! Yours, however, has been in the oven for over a year!  Emphasis mine: That's right — the ongoing costs aren't the problem. Hey wait a second, that's a name! I can look up a name! Scott W. Devitt now works at Wedbush as its managing director of equity research, and has said AI companies would enter a new stage in 2025...god, just read this : The analyst says the market will enter phase two in 2025, with software companies and cloud provider stocks expected to see gains. He adds that cybersecurity companies could also benefit as the technology evolves. Dewitt specifically calls out Palantir, Snowflake, and Salesforce as those who would "gain." In none of these cases am I able to see the actual revenue from AI, but Salesforce itself said that it will see no revenue growth from AI this year . Palantir also, as discovered by the Autonomy Institute’s recent study , recently added to the following to its public disclosures : What I'm saying is that analysts can be wrong! And they can be wrong at scale! There is no analyst consensus that agrees with me. In fact, most analysts appear to be bullish on AI, despite the significantly-worse costs and total lack of growth! Yet even in this Hater's Parade, the unnamed journalist makes a case for Amazon Web Services: Nice try, chuckles! In 2015, the year that Amazon Web Services became profitable , Morgan Stanley analyst Katy Huberty believed that it was running at a "material loss," suggesting that $5.5 billion of Amazon's "technology and content expenses" was actually AWS expenses, with a "negative contribution of $1.3 billion." Here is Katy Huberty, the analyst in question, declaring six months ago that " 2025 [will] be the year of Agentic AI, robust enterprise adoption, and broadening AI winners ." So, yes, analysts really got AWS wrong. But putting that aside, there might actually be a comparison here! Amazon Web Services absolutely created a capital expenditures drain on Amazon. From Forbes’s Chuck Jones : In today's money, this means that Amazon spent $6.76 billion in capital expenditures on AWS in 2014. Assuming it was this much every year — it wasn't, but I want to make an example of every person claiming that this is a gotcha — it took $67.6 billion and ten years (though one could argue it was nine) of pure capital expenditures to turn Amazon Web Services into a business that now makes billions of dollars a quarter in profit . That's $15.4 billion less than Amazon's capital expenditures for 2024, and less than one-fifteenth its projected capex spend for 2025. And to be clear, the actual capital expenditure numbers are likely much lower, but I want to make it clear that even when factoring in inflation, Amazon Web Services was A) a bargain and B) a fraction of the cost of what Amazon has spent in 2024 or 2025. Some people compare Large Language Models and their associated services to Amazon Web Services, or services like Microsoft Azure or Google Cloud, and they are wrong to do so. Amazon Web Services, when it launched, comprised of things like (and forgive how much I'm diluting this) Amazon's Elastic Compute Cloud (EC2), where you rent space on Amazon's servers to run applications in the cloud, or Amazon Simple Storage (S3), which is enterprise-level storage for applications. In simpler terms, if you were providing a cloud-based service, you used Amazon to both store the stuff that the service needed and the actual cloud-based processing (compute, so like your computer loads and runs applications but delivered to thousands or millions of people).  This is a huge industry. Amazon Web Services alone brought in revenues of over $100 billion in 2024 , and while Microsoft and Google don't break out their cloud revenues, they're similarly large parts of their revenue, and Microsoft has used Azure in the past to patch over shoddy growth . These services are also selling infrastructure. You aren't just paying for the compute, but the ability to access storage and deliver services with low latency — so users have a snappy experience — wherever they are in the world. The subtle magic of the internet is that it works at all, and a large part of that is the cloud compute infrastructure and oligopoly of the main providers having such vast data centers. This is much cheaper than doing it yourself, until a certain point. Dropbox moved away from Amazon Web Services as it scaled . It also allows someone else to take care of maintenance of the hardware and make sure it actually gets to your customers. You also don't have to worry about spikes in usage, because these things are usage-based, and you can always add more compute to meet demand. There is, of course, nuance — security-specific features, content-specific delivery services, database services — behind these clouds. You are buying into the infrastructure of the infrastructure provider, and the reason these products are so profitable is, in part, because you are handing off the problems and responsibility to somebody else. And based on that idea, there are multiple product categories you can build on top of it, because ultimately cloud services are about Amazon, Microsoft and Google running your infrastructure for you. Large Language Models and their associated services are completely different, despite these companies attempting to prove otherwise, and it starts with a very simple problem: why did any of these companies build these giant data centers and fill them full of GPUs? Amazon Web Services was created out of necessity — Amazon's infrastructure needs were so great that it effectively had to build both the software and hardware necessary to deliver a store that sold theoretically everything to theoretically anywhere, handling both the traffic from customers, delivering the software that runs Amazon.com quickly and reliably, and, well, making sure things ran in a stable way. It didn't need to come up with a reason for people to run web applications — they were already doing so themselves, but in ways that cost a lot, were inflexible, and required specialist skills. AWS took something that people already did, and what there was a proven demand for, and made it better. Eventually, Google and Microsoft would join the fray.  And that appears to be the only similarity with generative AI — that due to the ridiculous costs of both the data centers and GPUs necessary to provide these services, it's largely impossible for others to even enter the market. Yet after that, generative AI feels more like a feature of cloud infrastructure rather than infrastructure itself. AWS and similar megaclouds are versatile, flexible and multifaceted. Generative AI does what generative AI does, and that's about it. You can run lots of different things on AWS. What are the different things you can run using Large Language Models? What are the different use cases, and, indeed, user requirements that make this the supposed "next big thing"? Perhaps the argument is that generative AI is the next AWS or similar cloud service because you can build the next great companies on the infrastructure of others — the models of, say, OpenAI and Anthropic, and the servers of Microsoft.  So, okay, let's humour this point too. You can build the next great AI startup, and you have to build it on one of the megclouds because they're the only ones that can afford to build the infrastructure. One small problem. Let's start by establishing a few facts: None of this is to say that one hundred million dollars isn't a lot of money to you and me, but in the world of Software-as-Service or enterprise software, this is chump change. Hubspot had revenues of $2.63 billion in its 2024 financial year. We're three years in, and generative AI's highest-grossing companies — outside OpenAI ($10 billion annualized as of early June ) and Anthropic ( $4 billion annualized as of July ), and both lose billions a year after revenue — have three major problems: But let's start with Anysphere and Cursor, its AI-powered coding app, and its $500 million of annualized revenue. Pretty great, right? It hit $200 million in annualized revenue in March , then hit $500 million annualized revenue in June after raising $900 million . That's amazing! Sadly, it's a mirage. Cursor's growth was a result of an unsustainable business model that it’s now had to replace with opaque terms of service, dramatically restricted access to models, and rate limits that effectively stop its users using the product at the price point they were used to. It’s also horribly unprofitable, and a sign of things to come for generative AI. A couple of weeks ago, I wrote up the dramatic changes that Cursor made to its service in the middle of June on my premium newsletter , and discovered that they timed precisely with Anthropic (and OpenAI to a lesser extent) adding "service tiers" and "priority processing," which is tech language for "pay us extra if you have a lot of customers or face rate limits or service delays." These price shifts have also led to companies like Replit having to make significant changes to its pricing model that disfavor users .  I will now plagiarise myself: In simpler terms, Cursor raised $900 million and very likely had to hand large amounts of that money over to OpenAI and Anthropic to keep doing business with them, and then immediately changed its terms of service to make them worse. As I said at the time: Cursor is the largest and most-successful generative AI company, and these aggressive and desperate changes to its product suggest A) that its product is deeply unprofitable and B) that its current growth was a result of offering a product that was not the one it would sell in the long term. Cursor misled its customers, and its current revenue is, as a result, highly unlikely to stay at this level. Worse still, the two Anthropic engineers who left to join Cursor two weeks ago just returned to Anthropic . This heavily suggests that whatever they saw at Cursor wasn’t compelling enough to make them stay. As I also said: I realize it's likely a little boring hearing about software as a service, but this is the only place where generative AI can really make money. Companies buying hundreds or thousands of seats are how industries that rely upon compute grow, and without that growth, they're going nowhere. To give you some context, Netflix makes about $39 billion a year in subscription revenue, and Spotify about $18 billion. These are the single-most-popular consumer software subscriptions in the world — and OpenAI's 15.5 million subscribers suggest that it can't rely on them for the kind of growth that would actually make the company worth $300 billion (or more). Cursor is, as it stands, the one example of a company thriving using generative AI, and it appears its rapid growth was a result of selling a product at a massive loss. As it stands today, Cursor's product is significantly worse, and its Reddit is full of people furious at the company for the changes. In simpler terms, Cursor was the company that people mentioned to prove that startups could make money by building products on top of OpenAI and Anthropic's models, yet the truth is that the only way to do so and grow is to burn tons of money. While the tempting argument is to say that Cursor’s "customers are addicted," this is clearly not the case, nor is it a real business model. This story also showed that Anthropic and OpenAI are the biggest threats to their customers, and will actively rent-seek and punish their success stories, looking to loot as much as they can from them. To put it bluntly: Cursor's growth story was a lie. It reached $500 million in annualized revenue selling a product it can no longer afford to sell, suggesting material weakness in its own business and any and all coding startups. It is also remarkable — and a shocking failure of journalism — that this isn’t in every single article about Anysphere. I'm serious! Perplexity? Perplexity only has $150 million in annualized revenue ! It spent 167% of its revenue in 2024 ($57m, its revenue was $34m) on compute services from Anthropic, OpenAI, and Amazon! It lost $68 million! And worse still, it has no path to profitability, and it’s not even anything new! It’s a search engine! Professional gasbag Alex Heath just did a flummoxing interview with Perplexity CEO Aravind Srivinas, who, when asked how it’d become profitable, appeared to experience a stroke: Aravind, do you smell toast? And don’t talk to me about “AI browsers,” I’m sorry, it’s not a business model. How are people going to make revenue on this, hm? What do these products actually do? Oh they can poorly automate accepting LinkedIn invites ? It’s like God himself has personally blessed my computer. Big deal!  In any case, it doesn't seem like you can really build a consumer AI startup that makes anything approaching a real company. Other than ChatGPT, I guess? Arguably the biggest sign that things are troubling in the generative AI space is that we use "annualized revenue" at all, which, as I've mentioned repeatedly, means multiplying a month by 12 and saying "that's our annualized!" The problem with this number is that, well, people cancel things. While your June might be great, if 10% of your subscribers churn in a bad month (due to a change in your terms of service), that's a chunk of your annualized revenue gone. But the worst sign is that nobody is saying the monthly figures, mostly because the monthly figures kinda suck! $100 million of annualized revenue is $8.33 million a month. To give you some scale, Amazon Web Services hit $189 million ($15.75 million a month) in revenue in 2008 , two years after founding, and while it took until 2015 to hit profitability, it actually hit break-even in 2009, though it invested cash in growth for a few years after. Right now, not a single generative AI software company is profitable, and none of them are showing the signs of the kind of hypergrowth that previous "big" software companies had. While Cursor is technically " the fastest growing SaaS of all time ," it did so using what amounts to fake pricing. You can dress this up as "growth stage" or "enshittification (it isn't by the way, generally price changes make things profitable, which this did not)," but Cursor lied. It lied to the public about what its product would do long-term. It isn't even obvious whether its current pricing is sustainable. Outside of Cursor, what other software startups are there? Everyone loves to talk about enterprise search company Glean — a company that uses AI to search and generate answers from your company's files and documents. In December 2024, Glean raised $260 million, proudly stating that it had over $550 million of cash in hand with "best-in-class ARR growth." A few months later in February 2025, Glean announced it’d "achieved $100 million in annual recurring revenue in fourth quarter FY25, cementing its position as one of the fastest-growing SaaS startups and reflecting a surging demand for AI-powered workplace intelligence." In this case, ARR could literally mean anything, as it appears to be based on quarters — meaning it could be an average of the last three months of the year, I guess? Anywho, in June 2025, Glean announced it had raised another funding round , this time raising $150 million , and, troublingly, added that since its last round, it had "...surpassed $100M in ARR." Five months into the fucking year and your monthly revenue is the same? That isn't good! That isn't good at all! Also, what happened to that $550 million in cash? Why did Glean need more? Hey wait a second, Glean announced its raise on June 18 2025, two days after Cursor's pricing increase and the same day that Replit announced a similar hike ! It's almost as if its pricing dramatically increased due to the introduction of Anthropic's Service Tiers and OpenAI's Priority Processing . I'm guessing, but isn't it kind of weird that all of these companies raised money about the same time? Hey, that reminds me. If you look at what generative AI companies do (note that the following is not a quality barometer), it's probably doing one of the following things: Every single generative AI company that isn't OpenAI or Anthropic does one or a few of these things, and I mean every one of them, and it's because every single generative AI company uses Large Language Models, which have inherent limits on what they can do. LLMs can generate, they can search, they can edit (kind of!), they can transcribe (sometimes accurately!) and they can translate (often less accurately). As a result, it's very, very difficult for a company to build something unique. Though Cursor is successful, it is ultimately a series of system prompts, a custom model that its users hate, a user interface and connections to models by OpenAI and Anthropic, both of whom have competing products and make money from Cursor and its competitors. Within weeks of Cursor's changes to its services, Amazon and ByteDance released competitors that, for the most part, do the same thing. Sure there's a few differences in how they're designed, but design is not a moat, especially in a high-cost, negative-profit business, where your only way of growing is to offer a product you can't afford to sustain. The only other moat you can build..is the services you provide, which, when your services are dependent on a Large Language Model, are dependent on the model developer, who, in the case of OpenAI and Anthropic, could simply clone your startup, because the only valuable intellectual property is theirs. You may say "well, nobody else has any ideas either," to which I'll say that I fully agree. My Rot-Com Bubble thesis suggests we're out of hypergrowth ideas, and yeah, I think we're out of ideas related to Large Language Models too. At this point, I think it's fair to ask — are there any good companies you can build on top of Large Language Models? I don't mean add features related to , I mean an AI company that actually sells a product that people buy at scale that isn't called ChatGPT. In previous tech booms, companies would make their own “models” — their own infrastructure, or the things that make them distinct from other companies — but the generative AI boom effectively changes that by making everybody build stuff on top of somebody else’s models , because training your own models is both extremely expensive and requires vast amounts of infrastructure. As a result, much of this “boom” is about a few companies — really two, if we’re honest — getting other companies to try and build functional software for them.  I wanted to add one note — that, ultimately, OpenAI and Anthropic are bad for their customers. Their models are popular (by which I mean their customers' customers will expect access to them) meaning that OpenAI and Anthropic can (as they did with Cursor) arbitrarily change pricing, service availability or functionality based on how they feel that day. Don't believe me? Anthropic cut off access to AI coding platform Windsurf because it looked like it might get acquired by OpenAI . Even by big tech standards this fucking sucks. And these companies will do it again! Because all Large Language Models require more data than anyone has ever needed , they all basically have to use the same data, either taken from the internet or bought from one of a few companies (Scale, Surge, Turing, Together, etc.). While they can get customized data or do customized training/reinforcement learning, these models are all transformer-based, and they all function similarly, and the only way to make them different is by training them, which doesn't make them much different , just better at things they already do. I already mentioned OpenAI and Anthropic's costs, as well as Perplexity's $50 million+ bill to Anthropic, Amazon and OpenAI off of a measly $34 million in revenue . These companies cost too much to run, and their functionality doesn't make enough money to make them make sense. The problem isn't just the pricing, but how unpredictable it is. As Matt Ashare wrote for CIO Dive last year , generative AI makes a lot of companies’ lives difficult through the massive spikes in costs that come from power users, with few ways to mitigate their costs. One of the ways that a company manages their cloud bills is by having some degree of predictability — which is difficult to do with the constant slew of new models and demands for new products to go with them, especially when said models can (and do) cost more with subsequent iterations. As a result, it's hard for AI companies to actually budget . "But Ed!" you cry, "What about AGENTS?" Let me tell you about agents . The term "agent" is one of the most egregious acts of fraud I've seen in my entire career writing about this crap, and that includes the metaverse. When you hear the word "agent," you are meant to think of an autonomous AI that can go and do stuff without oversight , replacing somebody's job in the process, and companies have been pushing the boundaries of good taste and financial crimes in pursuit of them. Most egregious of them is Salesforce's " Agentforce ," which lets you "deploy AI agents at scale" and "brings digital labor to every employee, department and business process." This is a blatant fucking lie. Agentforce is a god damn chatbot platform, it's for launching chatbots, they can sometimes plug into APIs that allow them to access other information, but they are neither autonomous nor "agents" by any reasonable definition. Not only does Salesforce not actually sell "agents," its own research shows that agents only achieve around a 58% success rate on single-step tasks , meaning, to quote The Register , "tasks that can be completed in a single step without needing follow-up actions or more information." On multi-step tasks — so, you know, most tasks — they succeed a depressing 35% of the time. Last week, OpenAI announced its own "ChatGPT agent" that can allegedly go "do tasks" on a "virtual computer." In its own demo , the agent took 21 or so minutes to spit out a plan for a wedding with destinations, a vague calendar and some suit options, and then showed a pre-prepared demo of the "agent" preparing an itinerary of how to visit every major league ballpark. In this example's case, "agent" took 23 minutes, and produced arguably the most confusing-looking map I've seen in my life. It also missed out every single major league ballpark on the East Coast — including Yankee Stadium and Fenway Park — and added a random stadium in the middle of the Gulf of Mexico. What team is that, eh Sam? The Deepwater Horizon Devils? Is there a baseball team in North Dakota?  I should also be clear this was the pre-prepared example. As with every Large Language Model-based product — and yes, that's what this is, even if OpenAI won't talk about what model — results are extremely variable. Agents are difficult, because tasks are difficult, even if they can be completed by a human being that a CEO thinks is stupid. What OpenAI appears to be doing is using a virtual machine to run scripts that its models trigger. Regardless of how well it works (it works very very poorly and inconsistently), it's also likely very expensive. In any case, every single company you see using the word agent is trying to mislead you. Glean's " AI agents " are chatbots with if-this-then-that functions that trigger events using APIs (the connectors between different software services), not taking actual actions, because that is not what LLMs can do. ServiceNow's AI agents that allegedly "act autonomously and proactively on your behalf" are , despite claiming they "go beyond ‘better chatbots,’" still ultimately chatbots that use APIs to trigger different events using if-this-then-that functions. Sometimes these chatbots can also answer questions that people might have, or trigger an event somewhere. Oh, right, that's the same thing. The closest we have to an "agent" of any kind is a coding agent, which can make a list of things that you might do on a software project and then go and generate the code and push stuff to Github when you ask them to, and they can do so "autonomously," in the sense that you can let them just run whatever task seems right. When I say "ask them to" or "go and" I mean that these agents are not remotely intelligent, and when let run rampant fuck up everything and create a bunch of extra work . Also, a study found that AI coding tools made engineers 19% slower . Nevertheless, none of these products are autonomous agents, and anybody using the term agent likely means "chatbot." And it's working because the media keeps repeating everything these companies say. I realize we've taken kind of a scenic route here, but I needed to lay the groundwork here, because I am well and truly alarmed. According to a UBS report from the 26th of June, the public companies running AI services are making absolutely pathetic amounts of money from AI: ServiceNow's use of "$250 million ACV" — so annual contract value — may be one of the more honest explanations of revenue I've seen, putting them in the upper echelons of AI revenue unless, of course, you think for two seconds, whether these are AI-specific contracts. Or, perhaps, are they contracts including AI? Eh, who cares. It's also year-long agreements that could churn, and according to Gartner, over 40% of "agentic AI" projects will be canceled by end of 2027 . And really, ya gotta laugh at Adobe and Salesforce, both of whom have talked so god damn much about generative AI and yet have only made around $100 million in annualized revenue from it. Pathetic! These aren't futuristic numbers! They're barely product categories! And none of this seems to include costs. I haven't really spent time on my favourite subject — OpenAI being a systemic risk to the tech industry . Anthropic is in a similar, but slightly better position — it is set to lose $3 billion this year on $4 billion of revenue . It also has no path to profitability, recently jacked up prices on Cursor, its largest customer , and had to put restraints on Claude Code after allowing users to burn 100% to 10,000% of their revenue . These are the actions of a desperate company. Nevertheless, OpenAI and Anthropic's revenues amount to, by my estimates, more than half of the entire revenue of the generative AI industry, including the hyperscalers. To be abundantly clear: the two companies that amount to around half of all generative artificial intelligence revenue are ONLY LOSING MONEY. I've said a lot of this before, which is why I'm not harping on about it, but the most important company in the entire AI industry needs to convert by the end of the year or it's effectively dead, and even if it does, it burns billions and billions of dollars a year and will die without continual funding. It has no path to profitability, and anyone telling you otherwise is a liar or a fantasist. Worse still, outside of OpenAI...what is there, really? As I wrote earlier in the year , there is really no significant adoption of generative AI services or products. ChatGPT has 500 million weekly users, and otherwise, it seems that other services struggle to get 15 million of them. And while the 500 million weekly users sounds — and, in fairness, is — impressive, there’s a world of difference between someone using a product as part of their job, and someone dicking around with an image generator, or a college student trying to cheat on their homework. Sidebar: Google cheated by combining Google Gemini with Google Assistant to claim that it has 350 million users. Don't care, sorry. This is worrying on so many levels, chief of which is that everybody has been talking about AI for three god damn years, everybody has said "AI" in every earnings and media appearance and exhausting blog post, and we still can't scrape together the bits needed to make a functional industry. I know some of you will probably read this and point to ChatGPT's users, and I quote myself here : It has, allegedly, 500 million weekly active users — and, by the last count, only 15.5 million paying subscribers, an absolutely putrid conversion rate even before you realize that the actual conversion rate would be monthly active subscribers. That’s how any real software company actually defines its metrics, by the fucking way. Why is this impressive? Because it grew fast? It literally had more PR and more marketing and more attention and more opportunities to sell to more people than any company has ever had in the history of anything. Every single industry has been told to think about AI for three years, and they’ve been told to do so because of a company called OpenAI. There isn’t a single god damn product since Google or Facebook that has had this level of media pressure, and both of those companies launched without the massive amount of media (and social media) that we have today.  ChatGPT is a very successful growth product and an absolutely horrifying business. OpenAI is a banana republic that cannot function on its own, it does not resemble Uber, Amazon Web Services, or any other business in the past other than WeWork, the other company that SoftBank spent way too much money on. And outside of ChatGPT, there really isn't anything else. Before I wrap up — I'm tired, and I imagine you are too — I want to address something. Yes, generative AI has functionality. There are coding products and search products that people like and pay for. As I have discussed above, none of these companies are profitable, and until one of them is profitable, generative AI-based companies are not real businesses. In any case, the problem isn't so much that LLMs "don't do anything," but that people talk about them doing things they can't do. I believe that the generative AI market is a $50 billion revenue industry masquerading as a $1 trillion one, and the media is helping. As I've explained at length, the AI trade is not one based on revenue, user growth, the efficacy of tools or significance of any technological breakthrough. Stocks are not moving based on whether they are making money on AI, because if they were, they'd be moving downward. However, due to the vibes-based nature of the AI trade, companies are benefiting from the press inexplicably crediting growth to AI with no proof that that's the case . OpenAI is a terrible business , and the only businesses worse than OpenAI are the companies built on top of it. Large Language Models are too expensive to run, and have limited abilities beyond the ones I've named previously, and because everybody is running models that all, on some level, do the same thing, it's very hard for people to build really innovative products on top of them. And, ultimately, this entire trade hinges on GPUs. CoreWeave was initially funded by NVIDIA, its IPO funded partially by NVIDIA, NVIDIA is one of its customers, and CoreWeave raises debt on the GPUs it buys from NVIDIA to build more data centers, while also using the money to buy GPUs from NVIDIA. This isn’t me being polemic or hysterical — this is quite literally what is happening, and how CoreWeave operates. If you aren’t alarmed by that, I’m not sure what to tell you. Elsewhere, Oracle is buying $40 billion in GPUs for the still-unformed Stargate data center project , and Meta is building a Manhattan-sized data center to fill with NVIDIA GPUs . OpenAI is Microsoft's largest Azure client — an insanely risky proposition on multiple levels, not simply in the fact that it’s serving the revenue at-cost but that Microsoft executives believed OpenAI would fail in the long term when they invested in 2023 — and Microsoft is NVIDIA's largest client for GPUs, meaning that any changes to Microsoft's future interest in OpenAI, such as reducing its data center expansion , would eventually hit NVIDIA's revenue. Why do you think DeepSeek shocked the market? It wasn't because of any clunky story around training techniques. It was because it said to the market that NVIDIA might not sell more GPUs every single quarter in perpetuity. Microsoft, Meta, Google, Apple, Amazon and Tesla aren't making much money from AI — in fact, they're losing billions of dollars on whatever revenues they do make from it. Their stock growth is not coming from actual revenue, but the vibes around "being an AI company," which means absolutely jack shit when you don't have the users, finances, or products to back them up. So, really, everything comes down to NVIDIA's ability to sell GPUs, and this industry, if we're really honest, at this point only exists to do so. Generative AI products do not provide significant revenue growth, its products are not useful in the way that unlocks significant business value, and the products that have some adoption run at such a grotesque loss. I realize I've thrown a lot at you, and, for the second time this year, written the longest thing I've ever written. But I needed to write this, because I'm really worried. We're in a bubble. If you do not think we're in a bubble, you are not looking outside. Apollo Global Chief Economist Torsten Slok said it last week . Well, okay, what he said was much worse: We are in a bubble. Generative AI does not do the things that it's being sold as doing, and the things it can actually do aren't the kind of things that create business returns, automate labor, or really do much more than one extension of a cloud software platform. The money isn't there, the users aren't there, every company seems to lose money and some companies lose so much money that it's impossible to tell how they'll survive. Worse still, this bubble is entirely symbolic. The bailouts of the Great Financial Crisis were focused on banks and funds that had failed because they ran out of money, and the TARP initiative existed to plug the holes with low-interest loans. There are few holes to plug here, because even if OpenAI and Anthropic somehow became eternal money-burners, the AI trade exists based on the continued and continually-increasing sale and use of GPUs. There are limited amounts of capital, but also limited amounts of data centers to actually put GPUs, and on top of that, at some point growth will slow at one of the Magnificent 7, at which point costs will have to come down from things that lose them tons of money, such as generative AI. Before you ask - You do not have proof for this statement! The cost of tokens going down is not the same thing as the cost of inference goes down! Everyone saying this is saying it because a guy once said it to them! You don't have proof! I have more proof for what I am saying! While it theoretically might be, all evidence points to larger models costing more money, especially reasoning-heavy ones like Claude Opus 4. Inference is not the only thing happening, and if this is your one response, you are a big bozo and doofus and should go back to making squeaky noises when you see tech executives or hear my name. Okay, so one argument is that these companies will use ASICs — customized chips for specific operations — to reduce the amount they're spending. A few thoughts: I am worried because despite all of these obvious, brutal and near-unfixable problems, everybody is walking around acting like things are going great with AI. The New York Times claims everybody is using AI for everything — a blatant lie, one that exists to prop up an industry that has categorically failed to deliver the innovations or returns that it promised, yet still receives glowing press from a tech and business media that refuses to look outside and see that the sky is red and frogs are landing everywhere. Other than the frog thing, I'm not even being dramatic. Everywhere you look in the AI trade, things get worse — no revenue, billions being burned, no moat, no infrastructure play, no comparables in history other than the dot com bubble and WeWork, and a series of flagrant lies spouted by the powerful and members of the press that are afraid of moving against market consensus. Worse still, despite NVIDIA's strength, NVIDIA is the market's weakness , through no fault of its own, really. Jensen Huang sells GPUs, people want to buy GPUs, and now the rest of the market is leaning aggressively on one company, feeding it billions of dollars in the hopes that the things they're buying start making them a profit. And that really is the most ridiculous thing. At the center of the AI trade sits GPUs that, on installation, immediately start losing the company in question money. Large Language Models burn cash for negative returns to build products that all kind of work the same way. If you're going to say I'm wrong, sit and think carefully about why. Is it because you don't want me to be right? Is it because you think "these companies will work it out"? This isn't anything like Uber, AWS, or any other situation. It is its own monstrosity, a creature of hubris and ignorance caused by a tech industry that's run out of ideas , built on top of one company . You can plead with me all you want about how there are actual people using AI. You've probably read the " My AI Skeptic Friends Are All Nuts " blog, and if you're gonna send it to me, read the response from Nik Suresh first . If you're going to say that I "don't speak to people who actually use these products," you are categorically wrong and in denial. I am only writing with this aggressive tone because, for the best part of two years, I have been made to repeatedly explain myself in a way that no AI "optimist" is made, and I admit I resent it. I have written hundreds of thousands of words with hundreds of citations, and still, to this day , there are people who claim I am somehow flawed in my analysis, that I'm missing something, that I am somehow failing to make my case. The only people failing to make their case are the AI optimists still claiming that these companies are making "powerful AI." And once this bubble pops, I will be asking for an apology. I love ending pieces with personal thoughts about stuff because I am an emotional and overly honest person, and I enjoy writing a lot. I do not, however, enjoy telling you at length how brittle everything is. An ideal tech industry would be one built on innovation, revenue, real growth based on actual business returns that helped humans be better, not outright lie about replacing them. All that generative AI has done is show how much lust there is in both the markets and the media for replacing human labor — and yes, it is in the media too. I truly believe there are multiple reporters who feel genuine excitement when they write scary stories about how Dario Amodei says white collar workers will be fired in the next few years in favour of "agents" that will never exist. Everything I’m discussing is the result of the Rot Economy thesis I wrote back in 2023 — the growth-at-all-costs mindset that has driven every tech company to focus on increasing quarterly revenue numbers, even if the products suck, or are deeply unprofitable, or, in the case of generative AI, both. Nowhere has there been a more pungent version of the Rot Economy than in Large Language Models, or more specifically GPUs . By making everything about growth, you inevitably reach a point where the only thing you know how to do is spend money, and both LLMs and GPUs allowed big tech to do the thing that worked before — building a bunch of data centers and buying a bunch of chips — without making sure they’d done the crucial work of “making sure this would create products people like.” As a result, we’re now sitting on top of one of the most brittle situations in economic history — our markets held up by whether four or five companies will continue to buy chips that start losing them money the second they’re installed. I am disgusted by how many people are unwilling or unable to engage with the truth, favouring instead a scornful, contemptuous tone toward anybody who doesn't believe that generative AI is the future. If you are a writer that writes about AI smarmily insulting people who "don't understand AI," you are a shitty fucking writer, because either AI isn't that good or you're not good at explaining why it's good. Perhaps it's both. If you want to know my true agenda, it's that I see something in generative AI and its boosters something I truly dislike. Large Language Models authoritatively state things that are incorrect because they have no concept of right or wrong. I believe that the writers, managers and executives that find it exciting do so because it gives them the ability to pretend to be intelligent without actually learning anything, to do everything they can to avoid actual work or responsibility for themselves or others. There is an overwhelming condescension that comes from fans of generative AI — the sense that they know something you don't, something they double down on. We are being forced to use it by bosses, or services we like that now insist it's part of our documents or our search engines, not because it does something, but because those pushing it need us to use it to prove that they know what's going on. To quote my editor Matt Hughes : "...generative AI...is an expression of contempt towards people, one that considers them to be a commodity at best, and a rapidly-depreciating asset at worst." I haven't quite cracked why, but generative AI also brings out the worst in some people. By giving the illusion of labor, it excites those who are desperate to replace or commoditize it. By giving the illusion of education, it excites those who are too idle to actually learn things by convincing them that in a few minutes they can learn quantum physics . By giving the illusion of activity, it allows the gluttony of Business Idiots that control everything to pretend that they do something. By giving the illusion of futurity , it gives reporters that have long-since disconnected from actual software and hardware the ability to pretend that they know what's happening in the tech industry. And, fundamentally, its biggest illusion is economic activity, because despite being questionably-useful and burning billions of dollars, its need to do so creates a justification for spending billions of dollars on GPUs and data center sprawl, which allows big tech to sink money into something and give the illusion of growth. I love writing, but I don't love writing this. I think I'm right, and it’s not something I’m necessarily happy about.  If I'm wrong, I'll explain how I'm wrong in great detail, and not shy away from taking accountability, but I really do not think I am, and that's why I'm so alarmed. What I am describing is a bubble, and one with an obvious weakness: one company's ability to sell hardware to four or five other companies, all to run services that lose billions of dollars. At some point the momentum behind NVIDIA slows. Maybe it won't even be sales slowing — maybe it'll just be the suggestion that one of its largest customers won't be buying as many GPUs. Perception matters just as much as actual numbers, and sometimes more, and a shift in sentiment could start a chain of events that knocks down the entire house of cards.  I don't know when, I don't know how, but I really, really don't know how I'm wrong. I hate that so many people will see their retirements wrecked, and that so many people intentionally or accidentally helped steer the market in this reckless, needless and wasteful direction, all because big tech didn’t have a new way to show quarterly growth. I hate that so many people have lost their jobs because companies are spending the equivalent of the entire GDP of some European countries on data centers and GPUs that won’t actually deliver any value.  But my purpose here is to explain to you, no matter your background or interests or creed or whatever way you found my work, why it happened. As you watch this collapse, I want you to tell your friends about why — the people responsible and the decisions they made — and make sure it’s clear that there are people responsible. Sam Altman, Dario Amodei, Satya Nadella, Sundar Pichai, Tim Cook, Elon Musk, Mark Zuckerberg and Andy Jassy have overseen a needless, wasteful and destructive economic force that will harm our markets (and by a larger extension our economy) and the tech industry writ large, and when this is over, they must be held accountable. And remember that you, as a regular person, can understand all of this. These people want you to believe this is black magic, that you are wrong to worry about the billions wasted or question the usefulness of these tools. You are smarter than they reckon and stronger than they know, and a better future is one where you recognize this, and realize that power and money doesn’t make a man righteous, right, or smart. I started writing this newsletter with 300 subscribers, and I now have 67,000 and a growing premium subscriber base. I am grateful for the time you’ve given me, and really hope that I continue to help you see the tech industry for what it currently is — captured almost entirely by people that have no interest in building the future.

0 views

The Remarkable Incompetence At The Heart Of Tech

Hello premium subscribers! Today I have the first guest post I've ever commissioned (read: paid) on Where's Your Ed At - Nik Suresh , one of the greatest living business and tech writers , best-known for his piece I Will Fucking Piledrive You If You Mention AI Again , probably my favourite piece of the AI era. I want to be clear that I take any guest writing on here very seriously, and do not intend to do this regularly. The quality bar is very high, which is why I started with Nik. I cannot express enough how much I love his work. Brainwash An Executive Today is amazing, as is his teardown of Contra Ptacek's All My AI Skeptic Friends Are Nuts . Nik is a software engineer, the executive director of an IT consultancy, and in general someone who actually understands software and the industries built around selling it. You can check out his work here and check out his team here . Ed asked me to write about why leaders around the world are constantly buying software they don’t need. He probably had a few high-profile companies in mind, like Snowflake. Put aside whether Snowflake is a good product – most people don’t know what a database is , so why on earth does a specialized and very expensive database have a market cap of $71B?  That’s a fair question – and being both a software engineer and the managing director at a tech consultancy, I can talk about what’s happening on the ground. And yes, people are buying software that they don’t need. I wish that was the extent of our problems. Pointless software purchases are a comparatively minor symptom of the seething rot and stunning incompetence at the core of most companies’ technical operations. Things are bad to a degree that sounds unbelievable to people that don’t have the background to witness or understand it firsthand. Here is my thesis: Most enterprise SaaS purchases are simply a distraction – total wishful thinking – for leaders that hope waving a credit card is going to absolve them of the need to understand and manage the true crisis in software engineering. Buying software has many desirable characteristics – everyone else is doing it, it can stall having to deliver results for years, and allows leaders to adopt a thin veneer of innovation. In reality, they’re  settling for totally conservative failure. The real crisis, the one they’re ignoring, is only resolved by deep systems thinking, emotional awareness, and an actual understanding of the domain they operate in. And that crisis, succinctly stated, is thus: our institutions are filled to burst with incompetents cosplaying as software engineers, forked-tongue vermin-consultants hawking lies to the desperate, and leaders who consider think reading Malcolm Gladwell makes you a profound intellectual (if you don’t understand why this is a problem, please report to my office for immediate disciplinary action).  I’m going to try and explain things as they’re actually like at normal companies. Welcome to my Hell, and hold your screams until the end. The typical team at a large organization – in a truly random office, of the sort that buys products like Salesforce but will otherwise never be in the news – might literally deliver nothing of value for years at a time. I know, I know, how can people be doing nothing for years? A day? Sure, everyone has an off-day. Weeks? Maybe. But years ? Someone’s going to notice eventually, right? Most industries have long-since been seized by a variety of tedious managerialism that’s utterly divorced from actually accomplishing any work, but the abstract nature of programming allows for software teams to do nothing to a degree that stretches credulity. Code can be reported as 90% done in meetings for years; there’s no physical artifact that non-programmers can use to verify it. There’s no wall with half the bricks laid, just lines of incomprehensible text which someone assures you constitutes Value.  This is a real, well-known phenomenon amongst software engineers, but no one believes us when we bring it up, because surely there’s no way profit-obsessed capitalists are spending millions of dollars on teams with no visible output. I know it sounds wild. I feel like I’ve been taking crazy pills for years. But enjoy some anecdotes: My first tech job was “data scientist,” a software engineering subspecialty focused on advanced statistical methods (or “AI” if you are inclined towards grifting). When a data scientist successfully applies statistical methods to solve a business problem, it’s called producing a “model.” My team produced no models in two years , but nonetheless received an innovation award from leadership, and they kept paying me six figures . I know small squads of data scientists with salaries totaling millions that haven’t deployed working models for twice that long. During my next job, at an entirely unrelated organization, I was tasked with finishing a website that had been “almost done” for a few years, whose main purpose was for a team to do some data entry. This is something that takes about a competent team two weeks – my current team regularly does more complicated things in that time. I finished in good time and handed it to the IT department to host, a task that should take a day if done very efficiently, or perhaps three months if you were dealing with extreme bureaucracy and a history of bad technical decisions. It’s been five years and the organization has to deploy the finished product. I later discovered that the company had spent four years trying before I joined . It’s just a website! When the internet was coming up, people famously hired teenagers to do this! I’m not even going to get into my third and fourth jobs, except to say they involved some truly spectacular displays of technical brilliance, such as discovering a team burning hundreds of thousands of dollars on Snowflake because they didn’t take thirty seconds to double-check any settings. I suspect that Snowflake’s annual revenue would drop by more than 20% if every team in the world spent five minutes ( actually five minutes, it was that easy) to make the change I did – editing a single number in the settings that has no negative side-effects for the typical business – but they’re also staffed by people that don’t read or study, so there’s no way to reach them. I warn every single friend who enters the software industry that unless they land a role with the top 1% of software engineering organizations, they are about to witness true madness. Without fail, they report back in six months with something along the lines of “I thought you were exaggerating.” In a private conversation about a year ago, an employee that left a well-known unicorn start-up confided: There are places where this doesn’t happen, but this madness is overwhelmingly the experience at companies that purchase huge enterprise products like Salesforce – the relationship between astonishing inefficiency and buying these products is so strong that it’s a core part of how my current team handles sales. We don’t waste time trying to sell to companies that use this stuff – it’s usually too late to save them – and I spend a lot of time tracking down companies in the process of being pitched this stuff by competing vendors. In 2023, software engineer Emmanuel Maggiore wrote : This will be totally unrecognizable to about half the software people in the world – those working at companies like Netflix which are famous for their software engineering cultures, or some of those working at startups where there isn’t enough slack to obfuscate. For everyone else, what I’ve described is Tuesday. Also as a note to my lovely fans who email me about the RSS feed "including the whole premium newsletter" - I give generous previews! The rest of the (premium) article follows.

0 views

Anthropic Is Bleeding Out

Hello premium customers! Feel free to get in touch at [email protected] if you're ever feeling chatty. And if you're not one yet, please subscribe and support my independent brain madness. Also, thank you to Kasey Kagawa for helping with the maths on this. Soundtrack: Killer Be Killed - Melting Of My Marrow Earlier in the week , I put out a piece about how Anthropic had begun cranking up prices on its enterprise customers, most notably Cursor, a $500 million Annualised Recurring Revenue (meaning month multiplied by 12) startup that is also Anthropic’s largest customer for API access to models like Claude Sonnet 4 and Opus 4. As a result, Cursor had to make massive changes to the business model that had let it grow so large in the first place, replacing (on June 17 2025, a few weeks after Anthropic’s May 22 launch of its  Claude Opus 4 and Sonnet 4 models) a relatively limitless $20-a-month offering with a much-more-limited $20-a-month package and a less-limited-but-still-worse-than-the-old-$20-tier $200-a-month subscription, pissing off customers and leading to most of the Cursor Subreddit turning into people complaining or discussing they’d cancel their subscription. Though I recommend you go and read the previous analysis, the long and short of it is that Anthropic increased the costs on its largest customer — a coding startup — about 8 days (on May 30 2025) after launching two models (Sonnet 4 and Claude Opus 4) specifically dedicated to coding. I concluded with the following: But the whole time I kept coming back to a question: why, exactly, would Anthropic do this? Was this rent seeking? A desperate attempt to boost revenue? An attempt to bring its largest customer’s compute demands under control as its regularly pushed Anthropic’s capacity to the limit ? Or, perhaps, it was a little simpler: was Anthropic having its own issues with capacity, and maybe even cash flow. Another announcement happened on May 22 2025 — Anthropic launched Claude Code , a version of Anthropic’s Claude that runs directly in your terminal (or integrates into your IDE) that uses Anthropic’s Claude models to write and manage code. This is, I realize, a bit of an oversimplification, but the actual efficacy or ability of Claude Code is largely irrelevant other than in the sheer amount of cloud compute it requires. As a reminder, Anthropic also launched its Claude Sonnet 4 and Opus 4 models on May 22 2025 , shortly followed by its Service Tiers, and then both Cursor and vibe-coding startup Replit’s price changes, which I covered last week. These are not the moves of a company brimming with confidence about its infrastructure or financial position, which made me want to work out why things might have got more expensive. And then I found out, and it was really, really fucking bad. Claude Code, as a product, is quite popular, along with its Sonnet 4 and Opus 4 models. It’s accessible via Anthropic’s $20-a-month “Pro” subscription (but only using the Claude Sonnet 4 model), or the $100 (5x the usage of Pro) and $200 (20x the usage of Pro) ”Max” subscriptions. While people hit rate limits, they seem to be getting a lot out of using it, to the point that you have people on Reddit boasting about running eight parallel instances of Claude Code . Something to know about software engineers is that they’re animals , and I mean that with respect. If something can be automated, a software engineer is at the very least going to take a look at automating it , and Claude Code, when poked in the right way, can automate a lot of things, though to what quality level or success rate I have no real idea, and while there are limits and restrictions , software engineers absolutely fucking love testing limits and getting around restrictions, and many I know see them as challenges to overcome. As a result, software engineers are running Claude Code at full tilt, to the limits, to the point that some set alarms to wake up during the night when their limits reset after five hours to maximize their usage, along with specialised dashboards to help them do so . One other note is that Claude Code’s functionality creates detailed logs about the amount of input or output tokens it’s using throughout the day to complete its tasks, including whether said tokens were written to or read from the cache. Software engineers also love numbers , and they also love deals , and thus somebody created CCusage, a tool that, using those logs, allows Claude Code users to see exactly how much compute you’re burning with your subscription, even though Anthropic is only charging you $20, $100 or $200-a-month. CCUsage compares these logs (which contain both how tokens were used and what models were run) to the up-to-date information from Anthropic’s API prices, and tells you exactly how much you’ve spent in compute ( here’s a more detailed run-down ). In simpler terms, CCusage is a relatively-accurate barometer of how much you are costing Anthropic at any given time, with the understanding that its costs may (we truly have no idea) be lower than the API prices they charge, though I add that based on how Anthropic is expected to lose $3 billion this year (that’s after revenue!) there’s a chance that it’s actually losing money on every API call. Nevertheless, there’s one much, much, much bigger problem: Anthropic is very likely losing money on every single Claude Code customer, and based on my analysis, appears to be losing hundreds or even thousands of dollars per customer. There is a gaping wound in the side of Anthropic, and it threatens financial doom for the company. Some caveats before we continue: The reason I’m leading with these caveats is because the numbers I’ve found about the sheer amount of money Claude Code’s users are burning are absolutely shocking.  In the event that they are representative of the greater picture of Anthropic’s customer base, this company is wilfully burning 200% to 3000% of each Pro or Max customer that interacts with Claude Code, and in each price point’s case I have found repeated evidence that customers are allowed to burn their entire monthly payment in compute within, at best, eight days, with some cases involving customers on a $200-a-month subscription burning as much as $10,000 worth of compute. Furthermore, it’s important to know that Anthropic’s models are synonymous with using generative AI to write or manage code. Anthropic’s models make up more than half of all tokens related to programming that flow through OpenRouter , a popular unified API for integrating models into services, and currently lead on major LLM coding benchmarks . Much of Cursor’s downfall has come from integrating both of these models, in particular the expensive (and I imagine compute intensive) Opus 4 model, which Anthropic only allows users of its $100-a-month or $200-a-month “Max” subscriptions to access. In my research, I have found more than thirty different reported instances of users that have spent in excess of the amount they pay Anthropic by a factor of no less than 100%. For sake of your sanity, I will split them up by their paid subscription, and how much more than it they spent.

0 views