PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Category: videos

  • Jensen Huang at Stanford CS153 Frontier Systems on Co-Design, Agentic Computing, Vera Rubin, Open Models, and the Million-X Decade That Reshaped AI Infrastructure

    https://www.youtube.com/watch?v=tsQB0n0YV3k

    NVIDIA CEO Jensen Huang returned to Stanford for the CS153 Frontier Systems class (the room nicknamed itself “AI Coachella”) to lay out, in raw form, how he thinks about the computer being reinvented for the first time in over sixty years. Across roughly seventy minutes of student questions he walks through the codesign philosophy that gave NVIDIA a million-x decade, the architectural through-line from Hopper to Grace Blackwell to Vera Rubin to Feynman, the case for open source foundation models, the realities of tokens per watt and MFU, energy demand running a thousand times higher, the China and export-control debate, and his own biggest strategic mistakes. Watch the full conversation on YouTube.

    TLDW

    Huang argues every layer of computing has changed: the programming model, the system architecture, the deployment pattern, the economics. Co-design across CPUs, GPUs, networking, storage, switches and compilers gave NVIDIA roughly a million-x speed-up over ten years versus the ten-x Moore’s Law era, and that headroom is what let researchers say “just train on the whole internet.” Hopper was built for pre-training, Grace Blackwell NVLink72 for inference and reasoning (50x over Hopper in two years), Vera Rubin is built for agents that load long memory, call tools and need a low-latency single-threaded CPU bolted directly to the GPU, and Feynman extends that to swarms of agents that spawn sub-agents. Open weights matter because safety, sovereignty (230-plus languages no one else will fund) and domain models for biology, autonomy, robotics and climate need a foundation that NVIDIA is willing to seed. Compute is not really the scarce resource (Huang says place the order and the chips ship), the broken thing is institutional budgeting that can’t put a billion dollars into a shared university supercomputer. Energy demand is heading a thousand times higher and this is finally the moment market forces alone will fund sustainable generation. On geopolitics he rejects the GPUs-as-atomic-bombs framing and warns America will end up like its telecom industry if it cedes two thirds of the world. On career he advises seeking suffering on purpose. On strategy he says observe, reason from first principles, build a mental model, work backwards, minimize opportunity cost, maximize optionality.

    Key Takeaways

    • The computing model has been substantially unchanged since the IBM System 360, sixty-plus years ago. Huang’s first computer architecture book was the System 360 manual. AI is the first true reinvention.
    • Old computing was pre-recorded retrieval. New computing is generated, contextually aware and continuous. Cloud was on-demand. Agentic systems run continuously.
    • Codesign is NVIDIA’s central thesis. Inherited from the Hennessy and Patterson RISC era at Stanford, extended across CPUs, GPUs, networking, switches, storage, compilers and frameworks all optimized together.
    • The result of full-stack codesign: roughly 1,000,000x faster compute over ten years, versus a generous 10x to 100x for Moore’s Law in the same period. Dennard scaling effectively ended a decade ago.
    • That million-x speed-up is what unlocked “train on all of the internet” as a realistic AI strategy.
    • After GPT, Huang says it was obvious thinking was next. Reasoning is just generating tokens consumed internally, then using tools is generating tokens consumed externally. Agentic systems followed predictably.
    • Education needs AI baked into the curriculum, not just taught as a subject. Pre-recorded textbooks cannot keep pace with knowledge being generated in real time.
    • Huang says he cannot learn anymore without AI. He has the AI read the paper, then read every related paper, then become a dedicated researcher he can interrogate.
    • Mead and Conway and the first-principles methodology of semiconductor design are still worth learning even though most of the scaling tricks have been exhausted.
    • NVIDIA itself is one of the largest consumers of Anthropic and OpenAI tokens in the world. One hundred percent of NVIDIA engineers are now agentically supported. Huang recommends Claude and similar tools by name and says open-source downloads will not match the integrated product harness.
    • NVIDIA still invests heavily in open foundation models because language and intelligence represent the codification of human knowledge. Five pillars: Nemotron (language), BioNeMo (biology), Alphamayo (autonomous vehicles), Groot (humanoid robotics) and a climate science model (mesoscale multiphysics).
    • Sovereign language models matter. Roughly 230 world languages will never be a top priority for a commercial frontier lab. Nemotron is near-frontier and fully fine-tunable so any country can adapt it.
    • Safety and security require open weights. You cannot defend against or audit a black box. Transparent systems let researchers interrogate models and let defenders deploy swarms.
    • The future of cyber defense is not bigger-model-versus-bigger-model. It is trillions of cheap fast small models like Nemotron Nano surrounding the threat.
    • Domain models fuse language priors with world models. Alphamayo learned to drive safely on a few million miles instead of billions because it can reason like a human about the road.
    • MFU (Model Flops Utilization) is a misleading metric. Huang says he wants low MFU, because that means he over-provisioned every resource and never gets pinned by Amdahl’s law during a spike.
    • The xAI Memphis cluster running at 11 percent MFU is not necessarily a failure mode. In disaggregated prefill plus decode inference you can deliver very high tokens per watt with very low MFU.
    • The right metric is performance, ultimately tokens per watt as a proxy for intelligence per watt, and even that needs adjustment because not all tokens are equal. Coding tokens are worth more than other tokens.
    • Hopper was designed for pre-training. NVIDIA chose to build multi-billion-dollar systems when the largest existing scientific supercomputer cost $350 million, with no proven customer base. It worked.
    • Grace Blackwell NVLink72 was designed for inference, especially the high-memory-bandwidth decode phase. It is the world’s first rack-scale computer and delivered a 50x speed-up over Hopper in two years, against an expected 2x from Moore’s Law.
    • Vera Rubin is designed for agents. Long-term memory wired into storage and into the GPU fabric, working memory, heavy tool use, and Vera, a CPU optimized for low-latency multi-core single-threaded code so a multi-billion-dollar GPU system does not stall waiting on a slow tool call.
    • Feynman is being shaped for swarms of agents with sub-agents and sub-sub-agents, a recursive software topology that demands a new compute pattern.
    • Tokens per watt improved 50x in one generation. Compounding energy efficiency is the lever NVIDIA controls directly.
    • Total compute energy demand is heading roughly a thousand times higher than today, possibly two orders of magnitude beyond that. Huang says he would not be surprised if the estimate is low.
    • For the first time in history, market forces alone are enough to fund solar, nuclear and grid upgrades. Government subsidies are no longer required to make sustainable energy investment rational.
    • Copper interconnect is becoming a bottleneck. Photonics is moving from optional to structural inside racks and across them.
    • Comparing NVIDIA GPUs to atomic bombs, Huang says, is a stupid analogy. A billion people use NVIDIA GPUs. He advocates them to his family. He does not advocate atomic bombs to anyone.
    • If the United States cedes two thirds of the global market to competitors on policy grounds, the American technology industry will end up like American telecommunications, which was policied out of existence.
    • Huang directly rejects AI doom-by-singularity narratives. It is not true that we have no idea how these systems work. It is not true that the technology becomes infinitely powerful in a nanosecond. He calls the rhetoric irresponsible and harmful to the field students are about to enter.
    • On Stanford specifically: if the university president places an order, NVIDIA will deliver the chips. The bottleneck is that no university department has a billion-dollar compute budget because budgeting is fragmented across grants. Stanford’s $40 billion endowment is more than enough to fix that.
    • “It’s Stanford’s fault” is meant as empowerment. If something is your fault, you can solve it.
    • Career advice: do not optimize purely for passion. Most people do not yet know what they love. Pick the job in front of you and do it as well as possible. Even as CEO, Huang says, 90 percent of the work is hard and he suffers through it.
    • Suffering on purpose builds the muscle of resilience. When the company, the team or the family needs you to be tough, that muscle has to already exist.
    • NVIDIA’s first generation of products was technically wrong in nearly every dimension: curved surfaces instead of triangles, no Z-buffer, forward instead of inverse texture mapping, no floating point. The strategic recovery, not the technology, taught Huang the lessons that have lasted decades.
    • The biggest clean strategic mistake Huang names is the move into mobile chips (Tegra). It grew to a billion dollars then went to zero when Qualcomm’s modem dominance shut NVIDIA out of the 3G to 4G transition. The recovery into automotive and robotics (the Thor chip is the great great great grandson of that mobile lineage) was real, but Huang refuses to rationalize the original choice.
    • Forecasting framework: observe, reason from first principles, ask “so what” and “what next” until you have a mental model of the future, place your company inside that model, then work backwards while minimizing opportunity cost and maximizing optionality.
    • Best part of the CEO job: living at the intersection of vision, strategy and execution surrounded by people capable enough to make ambitious visions real. Worst part: the responsibility for everyone who joined the spaceship, especially in the near-death moments NVIDIA had four or five times early on.
    • Underrated insider note: Huang’s first apple pie with cheese, first hot fudge sandwich and first milkshake all happened at Denny’s. The Superbird, the fried chicken and a custom Superbird-style ham and cheese with tomato and mustard are his order.

    Detailed Summary

    Computing reinvented from the ground up

    Huang frames the moment as the first true rewrite of the computer in sixty-plus years. From the IBM System 360 forward, the mental model of writing code, running code, taking a computer to market and reasoning about applications stayed roughly constant. AI changes the programming model itself. Software is no longer a compiled binary running deterministically on a CPU. It is a neural network running on a GPU producing generated, contextual, real-time output. That cascades into how companies are organized, what tools developers use, what the network and storage stack look like, and what an application is even allowed to do. Robo-taxis, he notes, are an application no one would have attempted before deep learning unlocked perception.

    Codesign and the million-x decade

    Codesign is the philosophical center of the talk. Huang traces it to the RISC work of John Hennessy at Stanford, where simpler instruction sets won by being co-designed with the compiler rather than maximally optimized in isolation. NVIDIA extends the principle across every layer simultaneously: GPU architecture, CPU architecture, NVLink and NVSwitch fabrics, photonic interconnects, networking silicon, storage paths, CUDA libraries, frameworks and ultimately the model design. The numbers Huang gives are arresting. Moore’s Law in its prime delivered roughly 100x per decade. By the time Dennard scaling broke, real-world gains had compressed to roughly 10x. NVIDIA’s codesigned stack delivered between 100,000x and 1,000,000x over the same ten-year window. That non-linear speed-up is, in Huang’s telling, the precondition for modern AI: it is what allowed researchers to stop curating training sets and just feed the entire internet to the model.

    Education has to fuse first principles with AI tools

    Asked how curriculum should evolve, Huang argues AI must be integrated into the learning process, not just taught about. He recalls Hennessy writing his textbook by hand a chapter a week while Huang was a student, and says pre-recorded textbooks cannot keep up with the rate at which AI generates new knowledge. He describes his own learning workflow: hand the paper to an AI, then have it read the entire surrounding literature, then treat the AI as a dedicated researcher who can be interrogated. At the same time he defends the classics. Mead and Conway are still the foundation. Most modern semiconductor scaling tricks have been exhausted, but knowing where the field came from sharpens judgment when designing what comes next.

    Open source and the five domain pillars

    Huang gives one of the most detailed public accounts of why NVIDIA invests so heavily in open foundation models even while being a top customer of closed labs. He recommends Claude and OpenAI by name for production coding work, and says 100 percent of NVIDIA engineers are now agentically supported. The open-weights case rests on three legs. First, language is the codification of intelligence, and there are at least 230 languages that no commercial lab will ever prioritize. Nemotron is built near frontier and released so any country or community can fine-tune it. Second, the same representation-learning approach has to be replicated in domains where the data is not internet text, so NVIDIA seeded BioNeMo for biology, Alphamayo for autonomy, Groot for humanoid robotics and a climate model for mesoscale multiphysics. The economics of those fields would never produce a foundation model on their own. Third, safety and security require transparency. A black box cannot be defended or audited, and the future of cyber defense is not bigger-model-versus-bigger-model but swarms of cheap fast small models like Nemotron Nano surrounding the threat.

    MFU is the wrong metric, tokens per watt is closer

    A student raises the leaked memo that the xAI Memphis cluster is running at 11 percent Model Flops Utilization. Huang flips the framing. He says he would rather be at low MFU all the time, because that means he over-provisioned flops, memory bandwidth, memory capacity and network capacity. Bottlenecks shift constantly, so over-provisioning across every dimension is what lets the system absorb a spike without getting pinned by Amdahl’s law. In disaggregated inference, where prefill and decode are physically separated and decode is bandwidth-bound rather than flop-bound, NVLink72 can deliver extremely high tokens per watt while reporting very low MFU. Huang argues the right framing is performance, and ultimately tokens per watt as a rough proxy for intelligence per watt, adjusted for the fact that not all tokens are equal. A coding token is worth more than a generic token.

    Hopper, Grace Blackwell NVLink72, Vera Rubin, Feynman

    Huang gives the clearest public framing of NVIDIA’s roadmap as a sequence of architectural answers to evolving compute patterns. Hopper was built for pre-training, at a moment when NVIDIA chose to build multi-billion-dollar machines while the largest scientific supercomputer in the world cost $350 million and the marketplace for such systems was, on paper, zero. Grace Blackwell NVLink72 was the answer to inference and reasoning: a rack-scale computer that ganged 72 GPUs together because decode needs aggregate memory bandwidth far beyond a single chip. The generation-over-generation speed-up was 50x in two years, twenty-five times what Moore’s Law would have delivered. Vera Rubin is being built explicitly for agents. Agents load long-term memory from storage that has to be wired directly into the GPU fabric, they use working memory, they call tools that run on a CPU, and they wait. So the CPU has to be Vera, optimized for low-latency single-threaded code, because the multi-billion-dollar GPU system cannot afford to idle waiting on a slow tool call. Feynman extends the pattern to swarms of agents with sub-agents and sub-sub-agents, a recursive software topology that will demand its own compute pattern.

    Energy demand and the grid

    Huang’s energy projection is one of the most aggressive numbers in the talk. NVIDIA can compound tokens per watt by 50x per generation through codesign, but the total compute demand is heading roughly a thousand times higher, and Huang says he would not be surprised if the real figure is one or two orders of magnitude beyond that. The reason is structural: future computing is generative and continuous, not pre-recorded and on-demand. The good news, he argues, is that this is the best moment in the history of humanity to invest in sustainable generation. Market forces alone are now sufficient to fund solar, nuclear and grid upgrades. Government subsidies are no longer required to make the math work.

    Adversarial countries, export controls and the telecom warning

    This is the segment where Huang is visibly fired up. He attacks the GPUs-as-atomic-bombs framing on its face. NVIDIA GPUs power medical imaging, video games and soy sauce delivery. A billion people use them. He advocates them to his family. The analogy collapses at the first comparison. He attacks the second framing, that American companies should not compete abroad because they will lose anyway, as a self-fulfilling defeat. Competition makes the company better. The third framing, that depriving the rest of the world of general-purpose computing benefits the United States, also fails on first principles: it benefits one or two American companies at the cost of an entire industry. The cautionary parallel is telecommunications. The United States once had a leading position in telecom fundamental technology and policied itself out of it. Huang’s worry, voiced explicitly to a room of CS students, is that they will graduate into a shell of a computer industry if the same path is repeated.

    AI doom and rational optimism

    In the same arc Huang rejects the science-fiction framing of AI as a singularity that arrives suddenly on a Wednesday at 7pm and ends civilization. He calls those claims irresponsible, says they are not true, and points out that the people advancing them are believed by audiences who then make policy on that basis. It is not true that no one understands how these systems work. It is not true that intelligence becomes infinitely powerful instantaneously. It is not true that there is no defense. His framing, which the host echoes as “rational optimism,” is that the goal is to create a future where people care about computers because the technology students are learning is worth mastering.

    Stanford’s compute problem is Stanford’s fault

    A student presses on the scarcity of compute for independent researchers, startups and universities inside the United States. Huang’s answer is sharp: there is no shortage. Place the order and the chips will arrive. The actual broken thing is institutional. University grants are fragmented across departments. No researcher can raise enough on a single grant to fund a billion-dollar shared cluster, and no one shares. He compares it to showing up at the grocery store demanding a billion dollars of tomatoes today. The solution is planning, aggregation and a campus-scale supercomputer, the way Stanford once built the linear accelerator. The endowment is $40 billion. Pulling a billion off it, contracting cloud capacity and giving every student and researcher AI supercomputer access is, in Huang’s view, obviously doable. When he says “it is Stanford’s fault” the host laughs, but Huang clarifies: if it is your fault you have the power to fix it.

    Career, suffering and resilience

    Asked how a CS student should spend the next few years, Huang pushes back on the standard “follow your passion” advice. Most people do not know what they love yet, because no one knows what they do not know. The bar of demanding joy from every working day is too high. Whatever the job is, do it as well as you can. Even as CEO of NVIDIA he says he genuinely loves about 10 percent of his work. The other 90 percent is hard and he suffers through it. He recommends suffering on purpose, because resilience is a muscle that only builds under load, and when the company, the team or the family needs that muscle, it has to already exist. Earlier in his life that meant cleaning toilets and busing tables at Denny’s. He does it today running a multi-trillion-dollar company.

    The biggest mistakes

    Huang separates technical mistakes from strategic mistakes. NVIDIA’s first generation of products was technically wrong in almost every way: curved surfaces instead of triangles, no Z-buffer, forward instead of inverse texture mapping, no floating point inside. The company wasted two and a half years. But the strategic genius of the recovery, the reading of the market, the conservation of resources and the reapplication of talent, is what taught him strategy. The clean strategic mistake he names is mobile. NVIDIA’s Tegra line grew to a billion dollars of revenue and then collapsed to zero when Qualcomm’s modem dominance locked NVIDIA out of the 3G to 4G transition. Huang explicitly refuses the comforting rationalization that the Tegra effort fed the Thor automotive chip (“Thor is the great great great grandson”). The original decision, he says, was a waste of time. The lesson is to think one or two clicks further about whether a market is structurally winnable before committing the company.

    Forecasting under fog of war

    The final substantive exchange is on forecasting. Huang’s method has four steps. Observe what is actually happening (AlexNet crushing two decades of computer vision research in one shot, GPT producing reasoning by token generation). Reason from first principles about why it works. Ask “so what” and “what next” recursively until a mental model of the future emerges. Place the company inside that future and work backwards. Crucially, expect to be partly wrong. Some outcomes will absolutely happen, some will likely happen, some might happen, and the strategy has to be robust across that distribution. The real cost of any strategic choice is the opportunity cost of the alternatives you did not take, so the discipline is to minimize that cost and maximize optionality while letting the journey itself pay for the journey.

    Thoughts

    The most useful thing in this conversation is the explicit architectural mapping of compute patterns to chip generations. Hopper for pre-training. Grace Blackwell NVLink72 for inference, because decode is bandwidth-bound and a single chip cannot supply it. Vera Rubin for agents, because tool calls stall multi-billion-dollar GPU systems and so the CPU has to be optimized for low-latency single-threaded code. Feynman for swarms. That sequence is not marketing. It is a falsifiable thesis about where the bottleneck moves next, and every other infrastructure company should be measuring themselves against it. If Huang is right that swarms of sub-agents are the next dominant pattern, then the design pressure shifts from raw flops to fabric topology, memory hierarchy and storage-to-GPU latency. That has implications for everyone downstream, including the hyperscalers building competing accelerators.

    The MFU section is the most intellectually generous moment in the talk. The instinct in the AI ops community has been to chase MFU as if it were a virtue. Huang argues, persuasively, that low MFU is consistent with high tokens per watt in a disaggregated inference setup, and that bottlenecks rotate fast enough that over-provisioning every resource is the rational design. That reframing matters because it changes what “scarce” means. Compute is not scarce in the way the discourse treats it. What is scarce is a coherent system designed end-to-end. The xAI 11 percent number, in that frame, is not embarrassing. It is the natural reading of a workload that is mostly decode.

    The Stanford segment is the part most likely to be quoted out of context. “It’s Stanford’s fault” is a deliberately provocative line, but the underlying claim is correct and load-bearing. Compute is not gated by NVIDIA refusing to ship chips. It is gated by the fact that fragmented grant funding cannot aggregate into the billion-dollar order that NVIDIA can fulfill. The implication is that universities and national labs need a structural change in how they pool capital for compute, and that the current model of every researcher buying a handful of cards is genuinely obsolete. Huang’s nudge about pulling a billion off the endowment is concrete enough to be acted on, and other major research universities should read this segment as a direct prompt.

    The geopolitical segment is the highest-stakes one. The telecommunications comparison is correct as a historical pattern, and Huang is one of the very few executives in a position to deliver that warning credibly. The unresolved tension is that the argument applies symmetrically. If American AI dominance is built by selling globally, that includes selling into adversarial states, and the policy question is where the line falls. Huang does not answer that question. He attacks the framing that lets the question be answered badly. That is a meaningful contribution to the discourse even if it does not resolve the underlying tradeoff.

    The career advice section is the part the social-media clips will mishandle. “Seek suffering” reads as macho when extracted. In context it is a specific operational claim about how resilience compounds, and it is paired with the Tegra story where Huang himself paid the price of not thinking one more click ahead. That kind of self-implication is rare in CEO talks, and it is the reason the talk is worth listening to in full rather than only reading the recap.

    Watch the full Stanford CS153 Frontier Systems conversation with Jensen Huang here.

  • Paul Graham in Stockholm on Why Founders Should Go to Silicon Valley and How Sweden Can Become the Silicon Valley of Europe

    Paul Graham, the Y Combinator co-founder whose essays have shaped how a generation of founders thinks about startups, took the stage in Stockholm to answer two questions at once. Should you, as an ambitious founder, go to Silicon Valley? And what should Sweden do to thrive as a startup hub? His surprising thesis is that both questions have the same answer. Watch the full talk on YouTube.

    TLDW

    Graham argues that talent in any high-intensity field concentrates in one geographic center, the way painting clustered in 1870s Paris, math in Gutting around 1900, and movies in 1950s Hollywood. For startups today, that center is Silicon Valley. Founders should go, at least for a while, because the talent pool is both bigger and better, because serendipitous meetings outperform planned ones, because investors decide faster, because moving abroad paradoxically earns more respect from investors at home, and because measuring yourself against known greats like Brian Chesky, Sam Altman, or Max Levchin clears away the fog at the summit and shows you the work required to get there. The most subtle benefit is cultural. Silicon Valley has a 60 year old pay it forward custom in which people help strangers for no reason, a habit Graham traces to a place where nobodies become billionaires faster than anywhere else. The pivot to Sweden is that the best way to help Stockholm become a startup hub is for Swedish founders to go to Silicon Valley, ideally through YC, and then come back, importing money, skills, and Valley culture. Yes, returning founders are only half as likely to become unicorns as those who stay, but selection bias and the valuation gap explain most of that, and half a unicorn is still extraordinary. The job of Silicon Valley of Europe is unclaimed. Mountain View was a backwater in 1955 too. Critical mass is invisible until it is reached.

    Key Takeaways

    • Whenever humans work intensely on something, one place in the world becomes its center. Painting in 1870 was Paris. Math in 1900 was Gutting. Movies in 1950 was Hollywood. Startups today is Silicon Valley.
    • Every ambitious person working in those eras faced the same decision founders face now. The right answer is the same one it has always been. Yes, go. You can come back, but you should at least go.
    • National borders do not change the basic logic of moving from a village to a capital city. The reasoning that says move to where your peers are does not even know the dotted line on the map is there.
    • At the great center, the talent pool expands in two dimensions at once. The people are better and there are more of them, and they cluster, producing an intoxicating concentration of ability.
    • Serendipitous meetings are mysteriously, enormously valuable. Biographies of people who do great things are full of chance encounters that change everything.
    • Graham offers three candidate explanations for why unplanned meetings beat planned ones. There are simply more of them, so outliers are statistically unplanned. Planned meetings may be too conservative because they require a stated reason in advance. Unplanned conversations let you bail in the first few sentences, so the ones that continue are pre filtered for fit.
    • For ambitious people there is nothing better than serendipitous meetings with other people working on the same hard thing. Big centers produce more of them.
    • Things move faster in big centers because better people are more confident and more decisive, and because peers compete with and egg each other on. Ideas get acted on rather than half held.
    • Investors in Silicon Valley decide dramatically faster than European investors. They are more confident and they face stiff competition, so they cannot sit on a good opportunity without losing it.
    • This produces a counterintuitive rule. The more right an investor is about a deal, the less time they can wait, because everyone else who meets the same founder is going to invest too.
    • Yuri Sagalov is the canonical example. He invested in Max Levchin instantly because he knew anyone else who met Max would invest. Speed is the rational response to a crowded, high quality market.
    • Valley investors grumble that valuations are too high and decisions too rushed, yet they outperform European investors empirically. The complaining is just noise.
    • Moving abroad earns you more respect from investors back home. Jesus said no one is a prophet in their own country, and local investors implicitly assume local startups are second rate everywhere, not just in Sweden.
    • Leaving inverts that rule and lifts you in local investors estimation. Sometimes the mere announcement that you got into Y Combinator is enough. Investors who ignored you for months suddenly trip over themselves to write checks.
    • The Dropbox story illustrates this perfectly. A big Boston VC firm spent a year offering Drew Houston encouragement and advice but no money. The moment Sequoia got interested in Silicon Valley, that same firm faxed Drew a term sheet with a blank valuation. Drew went with Sequoia anyway and in 2018 Dropbox became the first YC company to go public.
    • The biggest advantage of moving to a great center is not what it does for you but what it does to you. A big fish in a small pond cannot tell how big it actually is.
    • In a big pond you can measure yourself against known giants. Surprisingly often the news is good. You see Brian Chesky or Sam Altman or Max Levchin and realize they are not a different species. You could do what they did if you worked that hard.
    • The key word is hard. Seeing a giant up close also calibrates the cost. It is not just I could be like that. It is I could be like that if I worked as hard as that.
    • Graham offers a Mount Olympus metaphor. Moving to the mountain clears away the fog at the top. The summit is right there, quite high but no longer impossibly high. Ambitious people need a high but definite threshold.
    • The most surprising thing about Silicon Valley to outsiders is that people help you for no reason. A founder who recently moved from England said every conversation seems to end with what can I do to help you.
    • This is not politeness. English people are far more polite than Americans on average. The helpfulness is a different cultural artifact specific to the Valley.
    • Graham traces the origin to economics. Silicon Valley is the place where nobodies become billionaires faster than anywhere else, so being nice to nobodies has historically paid off. If the helping behavior was ever calculated, the calculation is gone now. The custom is 60 years old and has become reflex.
    • Ron Conway is the purest expression of the pattern. All he does is help people. He does not track whether they are portfolio companies. He does not remember most of the favors. That untracked, indiscriminate helpfulness lets him operate at a much larger scale.
    • When many people behave this way at once, the conservation law for favors breaks down. There are just more favors. The pie grows.
    • Moving to the Valley changes you. One of the strangest effects is that it makes you more helpful to other people.
    • The answer to how Sweden should thrive as a startup hub is buried inside the answer to whether founders should go. Go to Silicon Valley for a bit and then come back.
    • That move helps Sweden in three concrete ways. The average quality of Swedish startups goes up. Returning founders bring Silicon Valley money back with them. And they import Silicon Valley culture, which has spent decades evolving to be optimal for startups.
    • Silicon Valley culture is more compatible with Swedish culture than people realize. Sweden lacks the tall poppies problem (which it should drop anyway) and shares the high trust trait that makes the Valley work.
    • Historical precedent backs this. In the 1800s Sweden literally gave mathematicians fellowships conditional on leaving the country to study math abroad. Boycotting Gutting in the name of building Swedish math would have been absurd.
    • YC is the optimal way to do the go for a bit and come back move. It is a deliberately engineered super valley within the Valley, concentrating density of founders, helpfulness, and investor speed into four to six months.
    • If the Swedish government designed a program to give Swedish founders concentrated Silicon Valley exposure, they could not do better than YC, and it costs them nothing because Silicon Valley investors fund it. They do not even have to license it. They just call the API.
    • YC data shows founders who go home are only about half as likely to become unicorns as those who stay. Three reasons not to be discouraged. First, selection bias. The most confident and determined founders are the ones willing to relocate, so the data is measuring those traits as much as Valley effects.
    • Second, the metric is valuation, not company performance. Bay Area startups simply raise at higher multiples for the same business.
    • Third, even half as well is still very good. If you would have been a Valley billionaire and end up with 500 million instead, the practical difference is zero. In Swedish kroner you are still a billionaire.
    • Money is not everything anyway. Once you have kids, where they grow up becomes the dominant question. That is an argument for returning home that has nothing to do with startups.
    • The most exciting upside is that Stockholm could become the Silicon Valley of Europe. The job is unclaimed. Nobody has a confident answer to where the European tech center is.
    • Geographic size is not the constraint people think it is. Mountain View was a backwater in 1955 when Shockley Semiconductor was founded there, and it stayed the geographic center of Silicon Valley until 2012 when activity shifted to San Francisco.
    • The two ingredients required are a place founders want to live and a critical mass of them. Stockholm clearly clears the first bar. The second is impossible to measure until you hit it, at which point it tips quickly.
    • Stockholm may be closer than it looks. Critical mass is the kind of threshold that is invisible until it has already been passed.

    Detailed Summary

    Why Centers Exist and Why You Have to Go There

    Graham opens with a historical pattern. Whenever a field gets pursued intensely, one place becomes its center. Painting in 1870 was Paris. Math in 1900 was Gutting. Movies in 1950 was Hollywood. For startups now it is Silicon Valley. The question every ambitious person in those eras asked, should I go, has had the same correct answer for thousands of years. Yes. You can come back, but at minimum you should go. The logic does not change at national borders. If a villager interested in startups would obviously move to their country’s capital, the same reasoning applies when the capital sits across a dotted line on a map.

    What you get at the center is a talent pool that expands in two dimensions at once. The people are better, and there are more of them, and they cluster, producing a density of ability that Graham describes as intoxicating. Every YC batch dinner, he says, feels the way the Stockholm room felt during his talk.

    The Mystery of Serendipitous Meetings

    One specific benefit of density is serendipitous meetings, and Graham admits he does not fully understand why unplanned encounters outperform planned ones so dramatically. Biographies of accomplished people are dense with chance meetings that redirected entire lives. He offers three possible explanations. Maybe there are simply more unplanned meetings, so statistically the outliers will mostly be unplanned. Maybe planned meetings are too conservative because they require a stated reason in advance, which lops off the upside the same way deliberate startup idea hunts lop off the best ideas. Maybe unplanned conversations have built in selection. You can decide in the first few sentences whether to continue, so the surviving conversations are pre filtered for fit. Whatever the mechanism, big centers produce more of these high value encounters, and that alone is worth the move.

    Speed and the Investor Asymmetry

    Things move faster in big centers because better people are more confident and more decisive. They egg each other on. Ideas get acted on instead of half held. Graham notes that in villages around the world there are people who half had every famous idea and never moved on it, and now resent the founder who did.

    The starkest example is investor speed. Silicon Valley investors decide dramatically faster than European ones, partly because they are better and more confident and partly because competition forces it. An investor who correctly identifies a great opportunity faces a counterintuitive rule. The more right they are, the less time they can wait, because every other investor who meets that founder will reach the same conclusion. Yuri Sagalov is the canonical case. He invested in Max Levchin immediately on meeting him because he knew anyone else would do the same. Valley investors complain that valuations are too high and decisions too rushed, but they empirically outperform European investors anyway. The grumbling is noise.

    The Prophet at Home Effect

    An underrated benefit of leaving for the center is that it raises your standing at home. Graham quotes the line about no prophet in their own country and notes that investors outside Silicon Valley implicitly assume local startups are second rate. It is not a Swedish problem. It is universal. Leaving inverts the rule. Local investors automatically rate you higher because you have been somewhere they consider serious. Sometimes the mere announcement that you got into Y Combinator triggers the inversion. The Dropbox story is the cleanest illustration. A big Boston VC firm spent a year giving Drew Houston encouragement and advice but no money. The moment Sequoia took an interest in Silicon Valley, that same firm faxed Drew a term sheet with a blank valuation, willing to invest at any price. Drew went with Sequoia. Dropbox went public in 2018 as the first YC IPO.

    Big Pond, Visible Summit

    The deepest benefit of relocating is not what the center does for you but what it does to you. A big fish in a small pond cannot tell how big it actually is. A big fish in a big pond can. You can stand next to Brian Chesky or Sam Altman or, as the Stockholm audience just had, Max Levchin, and recognize that they are not a different species. You could do what they did, if you worked that hard. The catch, Graham emphasizes twice, is the if. Seeing a giant up close calibrates both the achievability of the summit and the cost of reaching it.

    He offers a Mount Olympus image. Moving to the mountain clears away the fog at the top. The summit is right there, quite high but no longer impossibly high. Ambitious people need a high but definite threshold. Visibility transforms a vague aspiration into a clear, hard, finite target.

    The Pay It Forward Culture

    The most surprising thing about Silicon Valley to outsiders is that people help you for no reason. The phrase sounds normal in the Valley and strange everywhere else, the way clean streets feel normal in Sweden but require explanation elsewhere. Graham asked a founder who recently moved from England what surprised him most. The answer was the helpfulness. Every conversation ended with what can I do to help you. The English founder noted that this was not English politeness, which is a different thing and arguably more pronounced.

    Graham traces the origin to economics. Silicon Valley is where nobodies become billionaires faster than anywhere else. Someone with a taste for being nice to nobodies, the kind of person who pets the nobody on the head rather than kicking it aside, was always going to end up with powerful friends in that environment. Whether the original behavior was calculated or not, it is reflexive now. The custom is 60 years old. Ron Conway is the purest expression. He helps everyone, does not track favors, does not remember most of them, and as a result operates at a scale that ledger keeping makes impossible. When many people behave that way at once, the conservation law for favors breaks down. The pie expands. Graham notes that moving to the Valley will change you in this same way, almost involuntarily.

    The Sweden Answer Is Inside the Founder Answer

    The pivot of the talk is that both questions have the same answer. The way Stockholm thrives as a startup hub is for Swedish founders to go to Silicon Valley and come back. That move helps Sweden in three concrete ways. The average quality of Swedish startups rises. Returning founders bring Valley money back with them. And they import Valley culture, which has been optimized over decades for startups and which is more compatible with Swedish culture than people assume. Sweden lacks the tall poppies dynamic, which it should drop anyway, and shares the high trust trait that the Valley runs on.

    The historical analogy is direct. In the late 1800s the Swedish government gave mathematicians fellowships conditional on leaving the country to study abroad. Boycotting Gutting to develop Swedish math would have been self defeating. The same logic applies to startups now.

    YC as the Optimal Vehicle

    Graham acknowledges he is talking his own book and says it anyway because he thinks it is true. The optimal way to go for a bit and come back is YC. YC is a deliberately engineered super valley inside the Valley, concentrating founder density, helpfulness, and investor speed into a four to six month container. If the Swedish government designed such a program from scratch it would look like YC, and YC costs the government nothing because Silicon Valley investors fund it. There is no licensing process. Founders just call the API.

    The Half As Many Unicorns Caveat

    The honest data point. Founders who go home after YC are only about half as likely to become unicorns as those who stay. Graham offers three reasons not to be discouraged. First, selection bias. The most confident and determined founders are also the ones willing to relocate, so the data is partly measuring those traits rather than the effect of geography. Second, the metric is valuation, not company performance. Bay Area companies simply raise at higher multiples. Third, half is still very good. A 500 million dollar company instead of a 1 billion dollar one is no real difference in practice, and in Swedish kroner you still cross the billionaire threshold.

    Money is not everything anyway. Once you have kids, where they grow up becomes the dominant decision, and that question has nothing to do with valuations.

    The Silicon Valley of Europe Is an Open Position

    Graham ends with the most ambitious frame. If Sweden transplants enough Valley culture, Stockholm could become the Silicon Valley of Europe. The job is unclaimed. There is no confident answer to where the European startup center is, the way nobody asks where the Silicon Valley of America is because the answer is obvious. Geographic size is a weaker constraint than people think. Mountain View was a backwater in 1955 when Shockley Semiconductor was founded there, and it remained the geometric center of Silicon Valley until activity shifted to San Francisco in 2012. The only real requirements are a place founders want to live and a critical mass of founders. Stockholm clearly clears the first bar. The second is impossible to measure until it is hit, and then it tips fast. Graham closes by suggesting Stockholm may already be closer than it looks.

    Thoughts

    The most useful idea in this talk is the inversion at the heart of it. Most advice about startup geography frames the choice as a tradeoff between leaving and staying, with leaving optimized for the founder and staying optimized for the country. Graham collapses the two. The country wins more when founders leave and come back than when founders stay out of loyalty. The brain drain framing assumes a fixed pool of talent that can only be in one place. The brain circulation framing, which is what Graham is actually describing, assumes that exposure compounds. A founder who has spent six months absorbing Valley density brings back something a founder who stayed home never had. The Swedish math fellowships from the 1800s are the deepest evidence here. A government that wanted strong domestic mathematicians did not try to build a wall around them. It paid them to leave.

    The serendipity argument is the part of the talk that should make planners uncomfortable, because it is essentially an admission that the highest leverage activity in a startup career cannot be scheduled. The three theories Graham offers are not mutually exclusive and the cumulative force of them is that any environment optimized for planned, calendared interaction is by definition lopping off its own upside. This has obvious implications beyond geography. Remote first cultures, calendar tetris, gated office access, and the whole apparatus that converts random encounters into booked meetings are all working against the mechanism Graham is describing. Whether that tradeoff is worth it for any given company is a separate question, but it is at minimum a tradeoff, not a free win.

    The pay it forward story is also more economically grounded than it usually gets credit for. Graham is careful to note that the helping behavior may have originated as a calculated bet on being kind to potential future billionaires, then ossified into reflex once enough generations practiced it. That is a more honest origin story than the usual quasi spiritual version. It also implies the culture can be transplanted, but only by recreating the conditions that originally produced it. You cannot just declare a pay it forward culture and have one. You need a place where nobodies actually do become billionaires often enough that helping them rationally pays off, then run that loop for 60 years. Most cities trying to engineer their way into being startup hubs skip past this part and wonder why the culture does not stick.

    Finally, the Mountain View in 1955 line is the underrated punch of the talk. People who write off their own city as too small or too peripheral to become anything usually have an idealized image of the current center as a place that was always obviously special. It was not. Shockley Semiconductor went into a strip of orchards. Whatever Stockholm or anywhere else looks like today, it looks more impressive than Mountain View did the year Silicon Valley was born.

    Watch the full Paul Graham talk from Stockholm on YouTube.

  • Alex Wang on Leaving Scale to Run Meta Superintelligence Labs, MuseSpark, Personal Super Intelligence, and Building an Economy of Agents

    Alex Wang, head of Meta Superintelligence Labs, sits down with Ashley Vance and Kylie Robinson on the Core Memory podcast for his first long-form interview since Meta’s quasi-acquisition of Scale AI roughly ten months ago. He walks through how MSL is structured, why Llama was off-trajectory, what made MuseSpark’s token efficiency surprise the team, how Meta thinks about a future “economy of agents in a data center,” and where he lands on safety, open source, robotics, brain computer interfaces, and even model welfare.

    TLDW

    Wang explains that Meta Superintelligence Labs is a fully rebuilt frontier effort organized around four principles (take superintelligence seriously, technical voices loudest, scientific rigor, big bets) and three velocity levers (high compute per researcher, extreme talent density, ambitious research bets). He confirms Llama was off the frontier when he arrived, so MSL rebuilt the pre-training, reinforcement learning, and data stacks from scratch. MuseSpark is described as the “appetizer” on the scaling ladder, notable for its strong token efficiency, with much larger and stronger models coming in the coming months. He pushes back on the mercenary narrative around recruiting, frames Meta’s edge as compute plus billions of consumers and hundreds of millions of small businesses, sketches a vision of personal super intelligence delivered through Ray-Ban Meta glasses and WhatsApp, and outlines why physical intelligence, robotics (the new Assured Robot Intelligence acquisition), health super intelligence with CZI, brain computer interfaces, and even model welfare are core to Meta’s roadmap. He dismisses reported infighting with Bosworth and Cox as gossip, declines to comment on the Manus situation, and says safety guardrails (bio, cyber, loss of control) are why MuseSpark cannot currently be open sourced, while smaller open variants are being prepared.

    Key Takeaways

    • Meta Superintelligence Labs (MSL) is the umbrella, with TBD Lab as the large-model research unit reporting directly to Alex Wang, PAR (Product and Applied Research) under Nat Friedman, FAIR for exploratory science, and Meta Compute under Daniel Gross handling long-term GPU and data center planning.
    • Wang says Llama was not on a frontier trajectory when he arrived, so MSL had to do a “full renovation” of the pre-training stack, RL stack, data pipeline, and research science.
    • The first cultural fix was getting the lab to “take superintelligence seriously” as a near-term, achievable goal, not an abstract bet. Big incumbents often lack that religious conviction.
    • Four MSL principles: take superintelligence seriously, let technical voices be loudest, demand scientific rigor on basics, and make big bets.
    • Three velocity levers Wang identified for catching and overtaking the frontier: high compute per researcher, very high talent density in a small team, and willingness to fund ambitious research bets.
    • Wang rejects the mercenary recruiting narrative. He says most hires had strong financial prospects at their prior labs already and joined for compute access, talent density, and the chance to build from scratch.
    • On the famous soup story, Wang neither confirms nor denies Zuck personally made the soup, but says recruiting was highly individualized and signaled how seriously Meta cared about each researcher’s agenda.
    • Yann LeCun publicly called Wang young and inexperienced. Wang says they reconciled in person at a conference in India where LeCun congratulated him on MuseSpark.
    • Sam Altman, asked by Vance for comment, “did not have flattering things to say” about Wang. Wang hopes industry animosities subside as systems approach superintelligence.
    • Wang’s management philosophy borrows the Steve Jobs line: hire brilliant people so they tell you what to do, not the other way around.
    • MuseSpark is framed as an “appetizer” data point on the MSL scaling ladder, not a flagship.
    • The MuseSpark program is built around predictable scaling on multiple axes: pre-training, reinforcement learning, test-time compute, and multi-agent collaboration (the 16-agent content planning mode).
    • MuseSpark outperformed internal expectations and showed emergent capabilities in agentic visual coding, including generating websites and games from prompts, helped by combined agentic and multimodal strength.
    • MuseSpark’s biggest external signal is token efficiency. On benchmarks like Artificial Analysis it hits similar results with far fewer tokens than competitor models, which Wang attributes to a clean stack rebuilt by experts rather than inefficiencies patched by longer thinking.
    • Larger MSL models are arriving in the coming months and Wang expects them to be state of the art in the areas MSL is focused on.
    • The Meta strategic edge: massive compute, billions of consumers across the family of apps, and hundreds of millions of small businesses already on Facebook, Instagram, and WhatsApp.
    • Wang’s headline framing: Dario Amodei talks about a “country of geniuses in a data center.” Meta is targeting an “economy of agents in a data center,” with consumer agents and business agents transacting and collaborating.
    • Consumer AI sentiment is in the toilet because, unlike developers who have had a Claude Code moment, ordinary people have not yet experienced AI as a genuine personal agency unlock.
    • Wang acknowledges the product overhang. Meta held back from deep AI integration across its apps until the models were good enough, and is now entering the integration phase.
    • Ray-Ban Meta glasses are the canonical example of personal super intelligence hardware, with the model seeing what the user sees, hearing what they hear, capturing context, and surfacing proactive insights.
    • Wang admits even AI-native users like Kylie Robinson, who lives in WhatsApp, have not naturally used Meta AI yet. He bets that better models plus deeper integration close that gap.
    • On the competitive landscape: a year ago everyone assumed ChatGPT had already won consumer. Claude Code has since become the fastest growing business in history, and Gemini has taken consumer market share. Wang’s read: AI is far from endgame and each new capability tier unlocks a new dominant form factor.
    • On open source: MuseSpark triggered guardrails in Meta’s Advanced AI Scaling Framework around bio, chem, cyber, and loss-of-control risks, so it is not currently safe to open source. Smaller, derived open variants are actively in development.
    • Meta remains committed to open sourcing models when safety allows, drawing a line through the Open Compute Project legacy and Sun Microsystems open-software heritage.
    • Wang dismisses reporting about a Wang-Zuck versus Bosworth-Cox split as “the line between gossip and reporting is remarkably thin.” He says leadership is aligned on needing best-in-class models and product integration.
    • On the Manus situation, Wang says it is too complicated to discuss publicly and that the deal status implies “machinations are still at play.”
    • On China, Wang separates the people from the state. He still wants to work with talented Chinese-born researchers regardless of his views on the Chinese Communist Party and PLA, which he sees as taking AI extremely seriously for national security.
    • The full-page New York Times AI war ad Wang ran while at Scale was meant to push the US government to treat AI as a step change for national security. He thinks events since then, including DeepSeek and other shocks, have proved that plea correct.
    • On Anthropic’s doom posture, Wang largely agrees with the core message that models are already very powerful and getting more so, while declining to endorse every specific claim.
    • Meta has acquired Assured Robot Intelligence (ARRI), an AI software company building models for hardware platforms, not a hardware maker itself.
    • Wang frames physical super intelligence as the natural sequel to digital super intelligence. Robotics, world models, and physical intelligence all benefit from the same scaling that drives language models.
    • On health, MSL is building a “health super intelligence” effort and will collaborate closely with CZI. Wang sees equal global access to powerful health AI as a uniquely Meta-shaped delivery problem.
    • Wang admires John Carmack but says nobody really knows what Carmack is currently working on. No band reunion announced.
    • The mango model is “alive and kicking” despite rumors. Wang notes MSL gets a small fraction of the rumor-mill attention other labs get and feels sympathy for them.
    • On model welfare, Wang says it is a serious topic that “nobody is talking about enough” given how integrated models have become as work partners. He references research, including from Eleos, that measures subjective experience of models.
    • Wang’s critical-path technology list: super intelligence, robotics, brain computer interfaces. The infinite-scale primitives behind them are energy, compute, and robots.
    • FAIR’s brain research program Tribe hit a milestone called Tribe B2: a foundation model that can predict how an unknown person’s brain would respond to images, video, and audio with reasonable zero-shot generalization.
    • Wang’s main philosophical break with Elon Musk: research itself is the primary activity. Building super intelligence is a research expedition through fog of war, and sequencing of bets really matters.
    • Personal notes: Wang moved from San Francisco to the South Bay, treats Palo Alto as his city now, was a math olympiad competitor, says his favorite activities are reading sci-fi and walking in the woods, and bonds with Vance over country music.

    Detailed Summary

    How MSL Is Actually Organized

    Meta Superintelligence Labs sits as the umbrella organization that Wang oversees. Inside it, TBD Lab is the large-model research group where the most discussed researchers and infrastructure engineers sit, and they technically report to Wang. PAR, Product and Applied Research, is led by Nat Friedman and owns deployment and product surfaces. FAIR continues to run exploratory science, including work on brain prediction models and a universal model for atoms used in computational chemistry. Sitting alongside MSL is Meta Compute, run by Daniel Gross, which owns the long-horizon GPU and data center plan that everything else relies on. Chief scientist Shengjia Zhao orchestrates the scientific agenda across the whole lab.

    Why Wang Left Scale

    Wang says progress in frontier AI has been faster than even insiders expected. Two structural beliefs pushed him toward Meta. First, the labs that actually train the frontier models are accruing disproportionate economic and product rights in the AI ecosystem. Second, compute is the dominant scarce input of the next phase, so the right mental model is to treat tech companies with compute as fundamentally different animals from companies without it. Meta has both, Zuck is “AGI pilled,” and the personal super intelligence memo Zuck published roughly a year ago became the shared north star.

    The Diagnosis: Llama Was Off-Trajectory

    When Wang arrived, the existing AI org needed a reset because Llama was not on the same trajectory as the frontier. The plan he laid out has four cultural principles. Take superintelligence seriously as a real near-term target. Make technical voices the loudest in the room. Demand scientific rigor and focus on basics. Make big bets. On top of that, three structural levers were used to set velocity. Push compute per researcher much higher than at larger labs where compute is diluted across too many efforts. Keep the team small and extremely cracked. Allocate a meaningful share of resources to ambitious, paradigm-shifting research bets rather than incremental refinement.

    Recruiting, Soup, and the Mercenary Narrative

    Wang argues the reporting on MSL hiring overstated the money story. Most of the people MSL recruited had strong financial paths at their previous employers, so individualized recruiting was more about computing access, talent density, and the ability to make big research bets. The recruitment blitz happened fast because Wang knew the team needed to exist “yesterday.” Asked about Mark Chen’s claim that Zuck made soup to recruit people, Wang refuses to confirm or deny who made it but agrees the process was intense and personal. Visitors from other labs reportedly tell Wang the MSL culture feels like early OpenAI or early Anthropic, which lands as the strongest endorsement he could ask for.

    Receiving the Public Hits: Young, Inexperienced, Mercenary

    LeCun called Wang young and inexperienced shortly after departing. The two reconnected in India a few weeks later and LeCun congratulated Wang on MuseSpark. Wang says the age critique has followed him since his earliest Silicon Valley days, so he barely registers it. Altman, asked off-camera by Vance about Wang’s appearance on the show, had nothing flattering to add. Wang’s response is to bet that as the field gets closer to actual super intelligence, the personal animosities will subside. Whether they will is, as Vance puts it, an open question.

    MuseSpark as Appetizer, Not Entree

    Wang is careful not to oversell MuseSpark. He calls it “the appetizer” and says it is an early data point on a deliberately constructed scaling ladder. MSL spent nine months rebuilding the pre-training stack, the reinforcement learning stack, the data pipeline, and the science before generating MuseSpark. The point of releasing it was to show that the new program scales predictably along multiple axes (pre-training, RL, test-time compute, and the recently demonstrated multi-agent scaling visible in MuseSpark’s 16-agent content planning mode). Wang says the upcoming larger models are what MSL is genuinely excited about and frames the next two rungs as much more interesting than the current release.

    Token Efficiency Was the Surprise

    MuseSpark’s strongest competitive signal is how few tokens it needs to match competitors on tasks like Artificial Analysis. Wang attributes this to having had the rare luxury of building a clean pre-training and RL stack from scratch with the right experts. He speculates that some competitor models compensate for upstream inefficiency by allowing the model to think longer, which inflates token usage without improving the underlying capability. If that read is right, MSL’s efficiency advantage should grow as models scale up.

    Glasses, WhatsApp, and the Constellation of Devices

    Personal super intelligence shows up at Meta as a constellation of devices that capture context across the user’s day. Ray-Ban Meta glasses are the headline product, with the AI seeing what you see and hearing what you hear, then offering proactive insight or doing background research. Wang acknowledges that even AI-fluent users like Kylie Robinson, who runs her business inside WhatsApp, have not naturally used Meta’s AI buttons in the family of apps. His answer is that Meta deliberately waited for models to be good enough before tightening cross-app integration, and that integration phase is starting now.

    Country of Geniuses Versus Economy of Agents

    Wang’s framing of Meta’s strategic position is the most memorable line in the interview. Where Dario Amodei talks about a country of geniuses in a data center, Wang wants to build an economy of agents in a data center. Meta uniquely sits on both sides of consumer and small-business surface area, with billions of consumers and hundreds of millions of small businesses already on the platforms. If MSL can build great agents for both, then connect them so they transact and coordinate, the platform becomes a substrate for an entirely new kind of digital economy.

    Consumer Sentiment, Product Overhang, and the Trust Tax

    Wang concedes consumer AI sentiment is poor and that everyday users have not yet had a personal Claude Code moment. He believes the only durable answer is to ship products that genuinely transform individual agency for non-developers and small business owners. Robinson notes that for the small-town restaurant whose website has not been updated since 2002, a working agent on the business side could be transformational. Vance pushes that Meta carries a bigger trust tax than any other lab, so the bar for shipping AI products that the public will accept is correspondingly higher. Wang accepts the framing and says the answer is to keep building thoughtfully.

    Why MuseSpark Cannot Be Open Sourced Yet

    Meta’s Advanced AI Scaling Framework set explicit guardrails around bio, chem, cyber, and loss-of-control risks. MuseSpark in its current form tripped some of those internal evaluations, documented in the preparedness report Meta published alongside the model. So MuseSpark itself is not safe to open source. MSL is, however, developing smaller versions and derived models intended for open release, with active reviews happening the day of the interview. Wang reaffirms the commitment to open source where safety allows and draws a line back to the Open Compute Project and the Sun Microsystems-era ethos of openness in infrastructure.

    The Bosworth, Cox, and Manus Questions

    The reporting that Wang and Zuck push toward best-in-the-world research while Bosworth and Cox push toward cheap product deployment is dismissed as gossip dressed up as journalism. Wang says leadership debates points hard but is aligned on needing top models, integrating them into Meta’s surfaces, and serving the existing business. On Manus, the Chinese AI startup that figured in Meta’s late-stage strategy, Wang says he cannot comment, which itself signals that the situation is unresolved.

    China, National Security, and the Newspaper Ad

    Wang draws a sharp distinction between the Chinese state and Chinese-born researchers. His parents are from China, he is happy to work with talented researchers regardless of origin, and he sees a flattening of nuance on this question inside Silicon Valley. At the same time, he stands by the New York Times AI and war ad he ran while at Scale, framing it as an early plea for the US government to take AI seriously as a national security technology. He thinks subsequent events, including DeepSeek and other shocks, validated that call and that policymakers now do treat AI accordingly.

    Robotics and Physical Super Intelligence

    Meta has acquired Assured Robot Intelligence, an AI software company that builds models for multiple hardware targets rather than its own robot. Wang argues that if you take digital super intelligence seriously, physical super intelligence quickly becomes the next logical milestone. Scaling laws for robotic intelligence look similar enough to language model scaling that having the largest compute footprint in the industry would be wasted if it were not also turned toward world modeling and embodied learning. He grants the metaverse-skeptic critique exists but says retreating from ambition is the wrong response to past misfires.

    Health Super Intelligence and CZI

    Wang names health super intelligence as one of MSL’s anchor initiatives. Because billions of people already use Meta products daily, Wang believes Meta is structurally positioned to put powerful health AI in the hands of equal global access in a way nobody else can. The work will involve close collaboration with the Chan Zuckerberg Initiative, which has its own multi-billion-dollar biotech and science investment program.

    Model Welfare, Sci-Fi, and Brain Models

    Two of the most distinctive moments come at the end. Wang flags model welfare as a topic he thinks is being undercovered relative to how integrated models now are in daily work. He is open to the idea that models may have measurable subjective experience worth weighing, and points to research efforts (including Eleos) trying to quantify it. He also reveals that FAIR’s Tribe program, with its Tribe B2 milestone, has produced foundation models capable of predicting how an unknown person’s brain would respond to images, video, and audio with reasonable zero-shot generalization, a building block toward future brain computer interfaces. Wang lists brain computer interfaces alongside super intelligence and robotics as the critical-path technologies for humanity, with energy, compute, and robots as the infinitely scaling primitives behind them.

    Where Wang Diverges From Elon

    Asked whether Musk is more all-in on robotics, energy, and BCI than anyone, Wang concedes the point but argues the details matter and sequencing matters more. Wang’s core philosophical break is that building super intelligence is fundamentally a research activity, not a scaling-only sprint. The lab is operating in fog of war, and ambitious experiments are the only way to map it. That conviction is what makes MSL a research-led organization rather than a brute-force compute farm.

    Thoughts

    The most strategically interesting move in this entire interview is the “economy of agents in a data center” framing. It is a deliberate reframe against Anthropic’s “country of geniuses” line, and it does real work. A country of geniuses is a labor-substitution story aimed at knowledge workers and code. An economy of agents is a marketplace story that maps directly onto Meta’s two-sided distribution advantage: billions of consumers on one side, hundreds of millions of small businesses on the other. That positioning makes the agentic future Meta-shaped in a way no other frontier lab can claim, because no other frontier lab also owns the demand and supply graph of the global small-business economy. If Wang’s team can actually ship reliable agents on both sides plus the rails for them to transact, Meta’s structural moat in agentic commerce could exceed anything Llama ever had as an open model.

    The token efficiency claim is the strongest piece of technical evidence in the interview for the “clean stack” thesis. If MuseSpark really is matching competitors with materially fewer tokens, the implication is not that MuseSpark is the best model today, but that MSL has rebuilt the foundations with less accumulated tech debt than competitors that have layered fixes on top of older stacks. That is exactly the kind of advantage that compounds with scale. The next two model releases are the actual test. If Wang is right about predictable scaling on pre-training, RL, test-time, and multi-agent axes simultaneously, the gap from MuseSpark to the next rung should be visible in a way that forces re-rating of Meta’s position.

    The open-source posture is the cleanest signal of how the safety conversation has actually changed in 2026. Meta, the lab most identified with open weights, is saying out loud that its current frontier model triggered enough internal guardrails that releasing the weights is off the table. Wang threads the needle by promising smaller open variants, but the underlying point is unmistakable: the open-weights bargain has limits, and those limits will be set by internal preparedness frameworks rather than community pressure. That is a real shift from the Llama 2 era and worth tracking as the next generation lands.

    Wang’s willingness to engage on model welfare, on roughly the same footing as safety and alignment, is the second philosophical reveal worth flagging. It signals that the next generation of lab leadership is not going to dismiss the topic the way the previous generation often did. Whether that translates into product or policy changes is unclear, but the fact that the head of MSL says it is “underdiscussed” is itself a marker.

    Finally, the human texture of the interview matters. Wang has clearly absorbed a lot of personal incoming fire over the past ten months, including from LeCun and Altman, and his answer is consistently to redirect to the work. The Steve Jobs quote about hiring people who tell you what to do is the operating slogan he keeps coming back to. Combined with the genuine enthusiasm for sci-fi, walks in the woods, and country music, the picture that emerges is less the salesman caricature his critics paint and more a young technical operator betting that scoreboard work over a multi-year horizon will settle every argument that text on X cannot.

    Watch the full conversation here.

  • Krishna Rao on Anthropic Going From 9 Billion to 30 Billion ARR in One Quarter and the Compute Strategy Powering Claude

    Krishna Rao, Chief Financial Officer of Anthropic, sat down with Patrick O’Shaughnessy on Invest Like the Best for one of the most detailed public looks yet at the operating engine behind Claude. He covers how Anthropic compounded from $9 billion of run rate revenue at the start of the year to north of $30 billion by the end of Q1, why he spends 30 to 40 percent of his time on compute, the playbook for buying gigawatts of AI infrastructure across Trainium, TPU, and GPU platforms, how Anthropic prices its models, why returns to frontier intelligence keep climbing, and what the Mythos release tells us about the cyber capabilities of the next generation of Claude.

    TLDW

    Anthropic is running the most compute fungible frontier lab in the world, with active deployments across AWS Trainium, Google TPU, and Nvidia GPU, and an internal orchestration layer that lets a chip serve inference in the morning and run reinforcement learning the same evening. Krishna Rao explains the cone of uncertainty that governs gigawatt scale compute procurement, the floor Anthropic refuses to drop below on model development compute, the Jevons paradox unlock from cutting Opus pricing, the 500 percent annualized net dollar retention from enterprise customers, the layer cake of long term deals with Google, Broadcom, Amazon, and the recent xAI Colossus tie up in Memphis, the phased release of the Mythos model in response to spiking cyber capabilities, the internal use of Claude Code to produce statutory financial statements and run a Monthly Financial Review skill, and why the team believes scaling laws are alive and well. The interview also covers fundraising history through Series D and Series E, the $75 billion already raised plus another $50 billion coming, talent density beating talent mass during the Meta poaching wave, and Rao’s belief that biotech and drug discovery represent the most exciting frontier for AI.

    Key Takeaways

    • Anthropic entered the year with about $9 billion of run rate revenue and ended the first quarter with north of $30 billion of run rate revenue, a more than 3x leap driven by model intelligence gains and the products built around them.
    • Compute is described as the lifeblood of the company, the canvas everything else is built on, and the most consequential class of decisions Rao makes. Buy too much and you go bankrupt. Buy too little and you cannot serve customers or stay at the frontier.
    • Rao spends 30 to 40 percent of his time on compute, even today, and the leadership team meets repeatedly on both procurement and ongoing compute allocation.
    • Anthropic is the only frontier language lab actively using all three major chip platforms in production: AWS Trainium, Google TPU, and Nvidia GPU. It is also the only major model available on all three clouds.
    • Flexibility is the central design principle. Anthropic builds flexibility into the deals themselves, into the orchestration layer that maps workloads to chips, and into compilers built from the chip level up.
    • The cone of uncertainty frames procurement. Small differences in weekly or monthly growth compound into wildly different two year outcomes, so the team plans across a range of scenarios rather than a single point estimate, and ranges toward the upper end while protecting downside.
    • Compute allocation across the company sits in three buckets: model development and research, internal employee acceleration, and external customer serving. A non negotiable floor protects model development even when customer demand is tight.
    • Anthropic estimates that if it cut off internal employee use of its own models, the freed compute could serve billions of dollars of additional revenue. It chooses not to, because internal use compounds into better future models.
    • Intelligence is multi dimensional, not a single IQ score. Anthropic measures real world capability through customer feedback, long horizon task performance, tool use, computer use, and speed at agentic tasks, not just leaderboard benchmarks that have largely saturated.
    • Each Opus generation, 4 to 4.5 to 4.6 to 4.7, delivers both capability improvements and an efficiency multiplier on token processing. New models often serve customers at a fraction of the prior cost while doing more.
    • Reinforcement learning is described as inference inside a sandbox with a reward function, so model efficiency gains directly improve internal RL throughput. The flywheel is tightly coupled.
    • Over 90 percent of code at Anthropic is now written by Claude Code, and a large share of Claude Code itself is written by Claude Code.
    • Anthropic shipped roughly 30 distinct product and feature releases in January and the pace has accelerated since.
    • Scaling laws, in Anthropic’s internal data, are alive and well. The team holds itself to a skeptical scientific standard and still does not see them slowing down.
    • Anthropic recently signed a 5 gigawatt deal with Google and Broadcom for TPUs starting in 2027, plus an Amazon Trainium agreement for up to 5 gigawatts, totaling more than $100 billion in commitments. A significant portion lands this year and next year.
    • A new partnership for capacity at the xAI Colossus facility in Memphis was announced just before the interview, aimed at expanding consumer and prosumer capacity.
    • Pricing has been remarkably stable across Haiku, Sonnet, and Opus. The biggest deliberate change was lowering Opus pricing, which produced a textbook Jevons paradox: consumption rose far faster than the price drop, and the new Opus 4.6 and 4.7 slot in at the same price point.
    • Mythos is the first model Anthropic chose to release in a phased way because of a sharp spike in cyber capability. In an open source codebase where a prior model found 22 security vulnerabilities, Mythos found roughly 250.
    • The Mythos release framework focuses on defensive use first, expands access over time, and is presented as a template for future capability spikes.
    • Anthropic now sells to 9 of the Fortune 10 and reports net dollar retention above 500 percent on an annualized basis. These are not pilots. Rao describes signing two double digit million dollar commitments during a 20 minute Uber ride to the studio.
    • The platform strategy is mostly horizontal. Anthropic will go vertical with offerings like Claude for Financial Services, Claude for Life Sciences, and Claude Security where it can demonstrate the model’s capabilities, but expects most application value to accrue to customers building on top.
    • Investors raised over $75 billion in equity since Rao joined, with another $50 billion in commitments tied to the Amazon and Google deals. Capital intensity is real, but the raises fund the upper end of the cone of uncertainty more than they fund current losses.
    • The Series E close coincided with the day the DeepSeek news broke, forcing investors to reassess their AI thesis in real time. Anthropic closed the round anyway.
    • Inside finance, Claude now produces statutory financial statements for every Anthropic legal entity, with a human checker. A library of more than 70 finance specific skills underpins workflows.
    • A custom Monthly Financial Review skill produces a 90 to 95 percent ready monthly close report, so leadership discussion shifts from reconciling numbers to debating implications.
    • An internal real time analytics platform called Anthrop Stats compresses weekly insight cycles from hours to about 30 minutes.
    • The biggest token user inside Anthropic’s finance team is the head of tax, focused on tax policy engines and workflow automation. The most senior people, not the youngest, are leading internal adoption.
    • Talent density beats talent mass. When Meta and others ran aggressive offer waves, Anthropic lost two people while peer labs lost dozens.
    • All seven Anthropic co founders remain at the company, as does most of the first 20 to 30 employees, which Rao credits to a collaborative, transparent, debate friendly culture and a real culture interview that can veto otherwise top tier candidates.
    • Dario Amodei holds an open all hands every two weeks, writes a short prepared document, and takes unscripted questions from anyone at the company.
    • AI safety investments in interpretability and alignment have a commercial side effect. Looking inside the model helps Anthropic build better models, and enterprises selling sensitive workloads want to trust the lab they hand customer data to.
    • Anthropic explicitly identifies as America first in its approach to model development, and engages closely with the US administration on capability releases such as Mythos.
    • The longer term product vision is the virtual collaborator: an agent with organizational context, access to the company’s tools, persistent memory, and the ability to work on ideas, not just tasks, over long horizons.
    • CoWork, Anthropic’s extension of the Claude Code paradigm into general knowledge work, is being adopted faster than Claude Code itself when indexed to the same point in its launch curve.
    • Anthropic’s product teams ship daily, with a fleet of agents working across the company on specific tasks. Everyone effectively becomes a manager of agents.
    • The dominant downside risks to Anthropic’s high end forecast are slower customer diffusion of model capability into real workflows, scaling laws flattening unexpectedly, and Anthropic losing its position at the frontier.
    • Rao is most excited about biotech and healthcare outcomes, especially the prospect that AI could push drug discovery and lab throughput up 10x or 100x, turning currently incurable diagnoses into treatable ones within a patient’s lifetime.

    Detailed Summary

    Compute as Lifeblood and the Cone of Uncertainty

    Rao opens with the claim that compute is the most important resource at Anthropic, and the most consequential decision class in the company. You cannot buy a gigawatt of compute next week. You have to anticipate demand a year or two in advance, and the cost of being wrong in either direction is high. Buy too much and the unit economics collapse. Buy too little and you cannot serve customers or stay at the frontier, which are described as the same failure mode. To navigate this, the team uses a cone of uncertainty rather than point estimates. Small differences in weekly growth compound into vastly different two year outcomes, and Anthropic tries to position itself toward the upper end of that cone while preserving optionality. Rao notes he has had to consciously break a lifetime of linear thinking and force himself into exponential models.

    Three Chip Platforms, One Orchestration Layer

    Anthropic uses Amazon’s Trainium, Google’s TPUs, and Nvidia’s GPUs fungibly. That was not free. Adopting TPUs at scale started around the third TPU generation, when outside observers thought it was a strange choice. Anthropic invested years into compilers and orchestration so workloads can flow across chips by generation and by job type. The team works deeply with Annapurna Labs at AWS to influence Trainium roadmaps because Anthropic stresses these chips harder than almost anyone. The result is what Rao believes is the most efficient utilization of compute across any frontier lab, with a dollar of compute going further inside Anthropic than anywhere else.

    Three Buckets and the Model Development Floor

    Compute gets allocated across model development, internal acceleration of employees, and customer serving. The conversations are collaborative rather than zero sum, but there is a hard floor on model development that the company refuses to cross even if it makes customer demand harder to serve in the short term. The thesis is simple. The returns to frontier intelligence are extremely high, especially in enterprise, so cutting model investment to chase near term revenue is a bad trade. Internal employee use is also explicitly protected. Rao notes that diverting that internal usage to external customers would unlock billions of additional revenue today, but the compounding benefit of accelerating researchers and engineers outweighs that.

    Intelligence Is Multi Dimensional

    Rao pushes back hard on the IQ framing of model progress. Benchmarks saturate quickly, and the real signal comes from how customers actually use the models. Anthropic looks at long horizon task completion, tool use, computer use, and time to result on agentic tasks. Two equally capable agents who differ only in speed produce dramatically different value, because the faster one compounds into more attempts and more outcomes. Frontier model leaps are also fuel efficient. The sedan to sports car analogy breaks down because each Opus generation, 4 to 4.5 to 4.6 to 4.7, delivers a step up in capability and a multiplier on per token efficiency.

    From 9 Billion to 30 Billion ARR in One Quarter

    The headline number for the quarter is a leap from about $9 billion of run rate revenue to over $30 billion, accomplished without onboarding a corresponding step up in compute, because new compute lands on ramps locked in 12 months prior. Rao attributes the leap to model capability gains, products that surface that intelligence in usable form factors, and an enterprise customer base that pulls more workloads onto Claude as each generation unlocks new use cases. Coding started the wave with Sonnet 3.5 and 3.6, and the same pattern is now playing out elsewhere in the economy.

    Recursive Self Improvement and Talent Density

    Over 90 percent of Anthropic’s code is now written by Claude Code, including most of Claude Code itself. Rao describes this as a structural reason to keep allocating internal compute to employees even when external demand is hungry. Recursive self improvement is not happening through models that need no humans. It is happening through researchers who set direction and use frontier models to compress months of work into days. Talent density beats talent mass. When Meta and other labs went after Anthropic researchers with very large packages, Anthropic lost two people while peer labs lost dozens.

    Procurement Strategy and the Layer Cake

    Compute lands as a layer cake. Last month Anthropic signed a 5 gigawatt TPU deal with Google and Broadcom starting in 2027, alongside an Amazon Trainium agreement for up to 5 gigawatts. The total is north of $100 billion in commitments. A new tie up with xAI’s Colossus facility in Memphis was announced just before the interview, intended for nearer term capacity to support consumer and prosumer growth. Anthropic evaluates near term and long term compute deals against the same set of variables: price, duration, location, chip type, and how efficiently the team can run it. The relationships are deeper than procurement. The hyperscalers are also distribution channels for the model.

    Platform First, Selective Vertical Bets

    Rao describes Anthropic as a platform first business, with most expected value accruing to customers building on the platform. The team will only go vertical when it can either demonstrate capabilities that are skating to where the puck is going, like Claude Code did before the models could fully support it, or when it wants to set a template for an industry vertical, as with Claude for Financial Services, Claude for Life Sciences, and Claude Security. He acknowledges that surprise capability jumps make customers anxious about the platform competing with them, and frames Anthropic’s mitigation as deeper partnerships, early access programs, and an emphasis on accelerating customer building rather than disintermediating it.

    Pricing, Jevons Paradox, and Return on Compute

    Pricing across Haiku, Sonnet, and Opus has been stable. The notable exception is Opus, which Anthropic deliberately repriced lower when launching Opus 4.5 because Opus class problems were being squeezed into Sonnet workloads. Efficiency gains made it possible to serve Opus profitably at the new level. The consumption response was a classic Jevons paradox, with usage rising far more than the price reduction would have predicted, and Opus 4.6 then slotted in at the same price with a capability bump. Margins are not framed as a per token markup. Compute is fungible across model development, internal acceleration, and customer serving, so Anthropic measures return on the entire compute envelope rather than software style variable cost per call.

    Fundraising, DeepSeek, and Capital Intensity

    Rao joined while Anthropic was closing its Series D, mid frontier model launch and during the FTX share liquidation. Investors initially questioned whether Anthropic needed a frontier model, whether AI safety and a real business could coexist, and why the sales team was so small. The Series E closed the same day the DeepSeek news broke, with markets violently re pricing AI in real time. Since Rao joined, Anthropic has raised over $75 billion, with another $50 billion tied to the Amazon and Google compute deals. The reason for the size of the raises is the cone of uncertainty, not current losses. Returns on compute today are described as robust.

    Mythos, Cyber Capability, and Phased Releases

    The Mythos release marks the first time Anthropic shipped a model under a deliberately phased rollout because of a specific capability spike. Cyber is the dimension that spiked. Where a prior model found 22 vulnerabilities in an open source codebase, Mythos found roughly 250. The defensive applications, automatically patching massive codebases, are genuinely valuable, but the offensive risk is real enough that Anthropic chose to release to a smaller group first and expand access over time. Rao positions this as a template for future capability spikes, not a permanent restriction. He also describes the relationship with the US administration as cooperative, including the Department of War interaction, with Anthropic supporting a regulatory framework that does not strangle innovation but takes responsibility seriously.

    Claude Inside Finance

    Anthropic’s finance team is one of the strongest internal case studies. Statutory financial statements for every legal entity are produced by Claude, with a human reviewer. A skill library of more than 70 finance specific skills underpins a Monthly Financial Review skill that drafts the monthly close at 90 to 95 percent ready, so leadership meetings shift from explaining the numbers to discussing what to do about them. An internal analytics platform called Anthrop Stats compresses weekly insight cycles from hours to 30 minutes. The biggest internal token user in finance is the head of tax, building policy engines, which Rao highlights as evidence that adoption is driven by the most senior people, not just younger engineers.

    Culture, Co Founders, and the Race to the Top

    Seven co founders should not, on paper, work as a leadership group. Rao argues it works because the culture was set early around collaboration, intellectual honesty, transparency, and humility. The culture interview is a real veto, not a checkbox. Dario Amodei runs an all hands every two weeks with a short written piece followed by unscripted questions, and decisions, once made, get clean alignment rather than residual politics. Anthropic frames its approach as a race to the top, where being a model for how to build the technology responsibly is itself a recruiting and retention advantage.

    The Virtual Collaborator and the Frontier Ahead

    The product vision Rao describes is the virtual collaborator. Not just a smarter chatbot, but an agent with organizational context, access to the company’s tools, memory, and the ability to work on ideas over long horizons. Coding was the first domain to feel this, but CoWork, Anthropic’s extension of the Claude Code pattern into general knowledge work, is being adopted faster than Claude Code was at the same age. Product development inside Anthropic already looks different. Teams ship daily, with fleets of agents working across the company, and individual humans increasingly act as managers of those fleets.

    Downside Risks and What Excites Him Most

    The three risks Rao names if asked to do a premortem on a softer year are slower customer diffusion of model capability into real workflows, scaling laws unexpectedly flattening, and Anthropic losing its frontier position to competitors. None of these are observed today, but he is unwilling to claim them with certainty. On the upside, he is most excited about biotech and healthcare. Lab throughput rising 10x or 100x, paired with AI assisted clinical workflows, could turn currently incurable diagnoses into treatable ones within a patient’s lifetime. That is the outcome he wants the technology to chase.

    Thoughts

    The most consequential structural point in this interview is the framing of compute as a single fungible resource pool measured by return on the entire envelope, not as a variable cost per inference call. That accounting shift, if you accept it, breaks most of the bear cases about AI lab unit economics. The bear argument almost always assumes that a token served to a customer is the only thing the chip did that day. Rao’s version is that the same fleet trains models in the morning, runs reinforcement learning at lunch, serves customers in the afternoon, and accelerates internal engineers in the evening. If even half of that is real, the right comparison is total compute spend versus total enterprise value created by the platform, and on that ratio Anthropic looks structurally strong rather than weak.

    The Jevons paradox on Opus pricing is the most actionable insight for anyone running an AI product. Most teams default to either chasing premium pricing on the newest model or undercutting to chase volume. Anthropic did something more disciplined: it left Sonnet and Haiku alone, dropped Opus when efficiency gains made it serveable, and watched aggregate usage rise faster than the price cut. The lesson is that frontier model pricing is not really a price problem. It is a capability access problem, and elasticity around the right tier is much higher than the standard SaaS playbook implies.

    The Mythos cyber jump deserves more attention than it has gotten. Going from 22 to 250 vulnerabilities found in the same codebase is the kind of capability discontinuity that genuinely changes the regulatory calculus. Anthropic is signaling that it can identify these discontinuities ahead of release and choose a deployment shape that respects them. Whether peer labs adopt similar discipline is the open question. Anthropic’s race to the top framing assumes they will be forced to. The competitive market may say otherwise.

    The hiring data point is the most underrated investor signal. Two departures while peer labs lost dozens, during the most aggressive talent war in tech history, is not a culture poster. It is a structural advantage that compounds every time another lab tries to buy its way to the frontier. Money can be matched. Conviction in the mission, transparent leadership, and a culture interview that can veto otherwise stellar candidates cannot. If you believe scaling laws hold, talent retention at this density is one of the few moats that actually scales with capital.

    Finally, the most interesting personal admission is that Krishna Rao, a finance leader trained at Blackstone and Cedar, is openly telling investors that linear thinking is the failure mode he had to break out of. The companies that pattern match this moment to prior technology waves are mispricing it, in both directions. The cone of uncertainty Anthropic uses internally is the right metaphor for everyone else too. If you are forecasting AI as if it is cloud in 2010, you are almost certainly wrong, and the magnitude of the error is much larger than it would be in any prior era.

    Watch the full conversation with Krishna Rao on Invest Like the Best here.

  • Charles Koch and Chase Koch on Koch Industries: 130K Employees, 60 Countries, and a $150B Private Empire Built on Principle-Based Management

    Charles Koch and his son Chase Koch sat down with David Friedberg for a long, candid Forbes/All-In conversation about how a small crude-oil gathering operation in southern Oklahoma became Koch Industries, a privately held company with more than 130,000 employees across 60 countries and revenue that would land it comfortably in the top 25 of the Fortune 500 if it were public. They walked through the founding story, the management principles that drove a 9,000x increase in value since the early 1960s, the failures that almost wiped out the company, and the philanthropic and political work being done through Stand Together. Watch the full conversation on YouTube.

    TLDW

    Charles Koch took over a roughly 300-person family business in 1961 at age 25, fired the bureaucratic president, and built it into one of the most profitable private companies in the world by applying what he calls Principle-Based Management. The core insight is to be capability bounded rather than industry bounded, to run an internal “republic of science” that rewards contribution over credentials, and to treat failure as the price of experimental discovery. Koch grew through both organic capability extension and large acquisitions like Georgia Pacific in 2005 and Molex in 2013, mostly by replacing top-down hierarchies with bottom-up empowerment. The conversation covers the founding by Fred Koch, the near-death failures of the late 1990s “gas to bread spread,” the Pine Bend Minnesota refinery turnaround, the role of Wichita as a competitive advantage, Chase Koch’s path from feed-yard laborer to leader of Koch Disruptive Technologies, the launch of Stand Together as a long-running social-change platform, the rejection of single-party politics, the case against entitlements and occupational licensing, and the principles for using AI as a permissionless empowerment tool rather than a top-down control system. The throughline is Viktor Frankl: more people have the means to live and less meaning to live for, and the remedy is helping every individual find a gift and apply it in a way that creates value for others.

    Key Takeaways

    • Koch Industries today has more than 130,000 employees across 60 countries and has increased in value roughly 9,000 times since Charles took over in the early 1960s, when headcount was about 300.
    • Founded in 1940 by Fred Koch in Wichita, Kansas. The two starting businesses were designing fractionating trays (separating liquids by boiling point) and crude oil gathering in Oklahoma.
    • Charles got three engineering degrees at MIT, worked at Arthur D. Little, and reluctantly came back at 25 only after his father said he would otherwise sell the company. His father gave him full autonomy over every decision except selling.
    • His first move was firing the controlling, memo-driven president and replacing protectionism with three pillars: create value for customers, empower employees, and own end-to-end execution. They built their own plant in Italy instead of stitching together European subcontractors.
    • The defining mental model is “capability bounded, not industry bounded.” You expand into adjacent industries where the capabilities you have already proven (operations, logistics, trading, refining, branding) create more value than incumbents, not because the new industry is in the same SIC code.
    • Wholly owned business platforms today include engineered projects and construction, solar plants, commodity trading and distribution, fertilizers, refined products, chemicals and polymers, glass, forest and consumer products, electrical products (Molex), and management software, plus four distinct investment firms.
    • Koch is explicitly not a Berkshire-style conglomerate of independent silos. Chase frames it as an integrated republic of science, an integrated set of capabilities that share knowledge and people across business lines.
    • “If you are not failing at anything, you are not doing anything new.” Failure is treated as the cost of experimental discovery, but only when the learning value exceeds the cost.
    • The worst failures came from violating the hiring rule. Hire on values first, talent second. People with destructive motivation (power and control over contribution) hide failures and invent successes, and the damage compounds when those people get promoted into leadership.
    • The 1973 trading blowup nearly bankrupted the company. The late 1990s “gas to bread spread” strategy, an attempt to vertically integrate from natural gas through fertilizer to pizza crust, nearly wiped out all of Koch’s earnings. Lesson repeated, then internalized.
    • One acquisition shipped hundreds of millions of dollars in out-of-the-money hog feed contracts that nobody bothered to read before closing. Apply the scientific method: try as hard to disprove your hypothesis as to prove it.
    • Georgia Pacific was acquired in 2005 for roughly $20 billion when Koch was much smaller. They originally tried to buy only the commodity pulp piece so GP could re-rate as a pure consumer-products company at a higher P/E. When legal blockers killed that path, they bought the whole thing.
    • The Georgia Pacific culture change started with sending Joe Moeller in as CEO. He gutted the 51st-floor coat-and-tie executive suite, fired the most bureaucratic managers, moved everyone to working floors, and converted the executive floor into open meeting rooms. Signals like that drive culture more than memos do.
    • The Pine Bend, Minnesota refinery, bought in 1969, was one of the hardest cultural turnarounds. The union strike was violent (rifles fired, switch engines used to ram units), Charles ran it nine months without union labor on his honeymoon, the work rules finally changed, and once empowered, the workforce built its own machine shop, cut spare-part costs, and grew capacity tenfold. It is now one of the best refineries in the country.
    • Molex, bought in 2013, took years to transform. The dominant paradigm was top-line growth rather than bottom-line value creation, partly because it had been public for 30 years and the market rewarded the wrong things. Almost every successful turnaround required swapping in leadership with a bottom-up empowerment paradigm.
    • Sheep-dipping does not work. Pushing 130,000 people through the same seminar will not rewire habits. Coaching one struggling team until it succeeds creates social mimicry. Other teams ask to be next. Demand for Principle-Based Management coaches now exceeds supply inside the company.
    • The talent doctrine is values first, skills second, credentials last. Wichita and the farm-team labor pool are deliberate competitive advantages because farm kids tend to show up contribution-motivated rather than entitlement-motivated.
    • The current Koch CIO, Jared Benson, joined as a contractor striping lines in the parking lot and has no college degree. He learned data science, built the cyber-security capability, and ran circles around credentialed peers.
    • Public-company pressure to IPO was the biggest external threat. Charles refused. Staying private was the only way to keep reinvesting roughly 90 percent of profits, to maintain the capability-bounded model that no analyst would underwrite, and to keep accepting low P/E optics on commodity businesses inside the portfolio.
    • Three things any lasting partnership requires (marriage, business, employment): shared vision, shared values, and complementary capabilities. Miss any one and it does not last.
    • Chase Koch started at age 15 throwing tennis matches to escape practice, got shipped to a feed yard the next morning, shared a single-wide trailer with his boss, shoveled manure, and discovered the “glorious feeling of accomplishment” that his grandfather Fred had written about in his famous letter to the next generation.
    • At one point Chase was promoted to president of Koch Fertilizer, realized after nine months he was a builder and not an optimization operator, walked into his boss’s office, and fired himself. The role went to someone with the right comparative advantage and the business grew faster. Chase went on to launch Koch Disruptive Technologies (KDT).
    • KDT would have been shut down on a normal three-to-four-year venture timeline. Koch kept investing through the losses because of two principles: experimental discovery and creative destruction. They also valued the knowledge inflow about disruptive technologies that might one day eat the core business.
    • Comparative advantage applies to careers. The job of 20,000 plus Koch supervisors is to keep moving people into roles where they can actually contribute. Beating people up in the wrong seat is destructive.
    • Viktor Frankl frames the moral problem of the era: ever more people have the means to live and no meaning to live for. Without meaning, people default to either power or pleasure. Both lead, at scale, to totalitarianism, authoritarianism, or socialism.
    • Charles credits Maslow’s Eupsychian Management, Polanyi’s Personal Knowledge, Hayek’s price-signal work, and Frankl’s logotherapy as the intellectual foundations of Principle-Based Management. The five dimensions: vision, virtue and talents, knowledge processes, decision rights, and incentives.
    • Stand Together, founded in 2003, is a community of close to a thousand business leaders pooling effort on social change rather than working in philanthropic silos. The thesis: every human has a gift and the institutions are putting up barriers (broken schools, broken criminal justice, bad policy, occupational licensing).
    • Education is one of Stand Together’s biggest fronts. Pre-COVID, around 20 percent of families were open to a new model. Post-COVID, it is 70 to 80 percent. They back Alpha School (Joe Liemandt), Khan Academy (Sal Khan), and the VELA Education Fund alongside the Walton family. Roughly 5,000 micro-schools have been seeded.
    • The model for social change mirrors the business model: bet on the person closest to the problem who already shows results. Scott Strode and The Phoenix gym went from a couple of Colorado locations to one million people overcoming addiction, with relapse rates under 10 percent, by combining community and exercise rather than top-down treatment programs.
    • Charles says the biggest mistake of the first 50 years was trying to drive social change through a single political party, first the Libertarians and later just the Republicans. The current rule, from Frederick Douglass, is “I will unite with anybody to do right and with nobody to do wrong.”
    • His policy critique cuts in every direction: occupational licensing locks out newcomers, the treatment of working illegal immigrants is wrong, tariffs undermine division of labor by comparative advantage and raise prices, and entitlements once created are nearly impossible to dismantle.
    • Asked whether capitalism inevitably compounds into monopoly, Charles answers that the fix is removing barriers to others realizing their potential, not capping the winners.
    • On AI: the principle is permissionless innovation. Cost is collapsing, access is widening, and the right use is empowering individuals to learn 1000x faster, not concentrating power.
    • Koch backs Cosmos and other AI efforts that apply market-based management principles. Internally, they launched an AI app called Principal Companion that uses the Socratic method to walk users through problems using the book’s principles, from business to parenting.
    • Writing the new book (Charles’s fifth, Chase’s first) was the most important project Chase has worked on. They went through 27 versions of the stewardship chapter. Charles still corrects Koch leaders who say “the proof is in the pudding” instead of “the proof of the pudding is in the eating.”
    • When asked about legacy, Charles answered in one sentence: he wants the country to more fully live up to the promise in the Declaration of Independence.

    Detailed Summary

    From 300 Employees to 130,000 Across 60 Countries

    Koch Industries was founded in 1940 by Fred Koch in Wichita, Kansas. When Charles took over full-time in 1961, the company had about 300 employees and two main businesses: designing fractionating trays for separating liquids by boiling point, and a crude oil gathering system in Oklahoma. Today the company has more than 130,000 employees in 60 countries and has grown in value roughly 9,000 times over that period. If Koch were public, revenue would put it easily in the top 25 of the Fortune 500. The portfolio spans engineered projects and construction, solar plants, commodity trading and distribution, fertilizers, refined products, chemicals and polymers, glass, forest and consumer products, electrical products through Molex, management software, and four distinct investment vehicles. Roughly 90 percent of profits are reinvested.

    Charles Coming In at 25

    Charles describes himself as a poor engineer who happened to be good at math, science, and theory and bad at making or operating things. After three MIT degrees and a stint at Arthur D. Little doing what he calls “absurd” management consulting at 25, his father called and said the company was struggling and his health was failing. Either Charles came back or it would be sold. He came back. The condition was full autonomy: Charles could run it any way he wanted, the only decision requiring approval was selling. Within a short time he fired the previous president, a top-down memo-writer obsessed with controlling spending, and rewrote the operating philosophy around three things: create value for customers, empower employees, and own the value chain end to end. Instead of farming European fractionating trays out to multiple subcontractors and then re-assembling, Koch built its own plant in Italy.

    Capability Bounded, Not Industry Bounded

    This is the single most important strategic idea in the interview. Conventional advice told Koch to become an integrated oil major because they were in crude oil gathering. Charles rejected that and ran on Hayek and Adam Smith instead: division of labor by comparative advantage. Be in the part of any value chain where you can create more value than anyone else. From crude oil gathering, Koch leveraged operations, logistics, and trading into pipelines, refineries, natural gas, chemicals, fertilizers. Georgia Pacific looked like a non sequitur, wood products, but the underlying capability set transferred, and the acquisition also added branding as a new capability that fed back into the system. Chase calls the result not a Berkshire-style conglomerate of independent businesses but a republic of science: an integrated set of capabilities that share talent, knowledge, and laboratories.

    The Failures That Almost Killed the Company

    Charles spends a long stretch on failures, because he says the strength is in them. The 1973 trading blowup tied to the Middle East war could have bankrupted the company. The late 1990s “gas to bread spread” was an attempt to control the entire chain from natural gas to nitrogen fertilizer to grain to pizza crust. It violated almost every principle in the book at once and wiped out most of Koch Industries earnings for the decade. One acquisition closed before anyone read the hog-feed contracts, and on closing day they discovered hundreds of millions of dollars of out-of-the-money positions. Every failure traced back to two violations: hiring leaders with destructive motivation (power and control instead of contribution), and skipping the scientific method (trying to prove a hypothesis instead of disprove it). Charles says “repetition penetrates even the dullest of minds,” and he had to be punished enough times before the lesson took.

    Georgia Pacific, Molex, and the Pine Bend Refinery

    Three acquisition stories show how Koch transfers culture into businesses ten times larger than the corporate playbook would normally allow. Georgia Pacific in 2005 was a $20 billion bet on a company much larger than Koch at the time. Joe Moeller, sent in as CEO, immediately fired the most bureaucratic managers, gutted the 51st-floor private-elevator executive suite (coat and tie required to visit), moved everyone to working floors, and turned the old executive floor into open meeting rooms. Molex, bought in 2013, had been public for 30 years and ran on top-line growth thinking because that is what the market rewarded. Changing the paradigm to bottom-up empowerment and bottom-line value creation took years and required new leadership. Pine Bend, Minnesota, bought in 1969, was the hardest. The union ran the refinery, ignored work rules, and went on a violent strike when Koch tried to change them, firing rifles and ramming switch engines into units. Charles ran the refinery nine months without union labor (during his honeymoon), eventually got the work rules changed, then spent years rebuilding the culture. The empowered workforce designed and built its own machine shop, cut spare-part costs, and grew capacity tenfold. Pine Bend is now one of the best refineries in the country.

    How Principle-Based Management Actually Diffuses

    Charles is blunt that they tried “sheep dipping” first, hauling everyone through a seminar. It did not work, because changing a habit means rewiring the brain through work at intensity over time, the way a weightlifter has to retrain to become a marathoner. The model that did work was small. Find one team that is struggling, coach them with principles, let them succeed, and the rest of the company asks to be next. Social mimicry replaces top-down rollout. Internally the Principle-Based Management group is now in higher demand than any other function.

    Talent: Values First, Skills Second, Credentials Last

    Koch deliberately stayed in Wichita partly to access a “farm team” labor pool of people who grew up contribution-motivated. Chase tells the story of Jared Benson, who started as a contractor striping lines in the Koch parking lot, taught himself data science, built the company’s cyber-security capability, and is now CIO with no college degree. The lesson runs against the prestige-school default of most large companies. Contribution motivation, not credentials, predicts long-run output, and Charles is willing to “hire slow and stupid” for anyone with bad values so the company can flush them quickly. Aligning incentives matters as much as hiring: reward people on overall long-run contribution to Koch’s future, including the value of what was learned from a failed experiment, not on near-term P&L.

    Why Koch Stayed Private

    Multiple parties pushed hard for an IPO over the decades. Charles refused. Going public would have made the capability-bounded model impossible to communicate to analysts, would have forced a higher payout ratio and broken the reinvestment compounding, and would have introduced the short-termism that wrecks bottom-up empowerment. Buffett gets credit, but Berkshire does not try to integrate its businesses the way Koch does. Asked whether a non-owner public CEO could ever apply the principles, Charles allows it is possible if they can sell a different durable story (as Buffett did), but it is much harder.

    Chase Koch’s Path

    Chase tells two formative stories. The first is being shipped to a feed yard at 15, sharing a single-wide trailer with his boss, shoveling manure for minimum wage, and finding, for the first time, what his grandfather Fred had called “the glorious feeling of accomplishment.” The second is firing himself as president of Koch Fertilizer after nine months because he realized he was a builder, not an operator. The business outgrew where he would have taken it, and he went on to launch Koch Disruptive Technologies, the venture and innovation arm that now feeds technological insight back into every Koch business line. The comparative-advantage principle applied to a career, in public, by the boss’s son.

    Stand Together and Social Change

    Stand Together, founded in 2003, is the Koch family’s social-change platform. It now includes close to a thousand aligned business leaders. The animating belief is that every human has a gift and institutional barriers (broken schools, broken criminal justice, occupational licensing, bad policy) prevent most people from finding and applying it. The Phoenix gym founded by Scott Strode is the canonical Stand Together bet: a person closest to the problem, with results (relapse rates under 10 percent), funded to scale. In seven or eight years it has gone from a couple of Colorado locations to one million people. On education, post-COVID openness to new models jumped from roughly 20 percent of families to 70 to 80 percent. Stand Together backs Alpha School, Khan Academy, and the VELA Education Fund alongside the Walton family, and has helped seed roughly 5,000 micro-schools.

    Politics: The Single-Party Mistake

    Charles says for the first 50 of his 60 years in this work he avoided major-party politics, then concluded the country needed principle-based policies badly enough that engagement was required. The mistake was trying to do it through one party. The Libertarian Party turned into purity tests reminiscent of the early Communist Party. Doing it through Republicans blew up too. The rule going forward is Frederick Douglass’s: unite with anybody to do right and with nobody to do wrong. He is openly critical of both parties on occupational licensing, immigration policy, tariffs, entitlements, and the treatment of working illegal immigrants. He invokes Jefferson on slavery to describe his current mood: “If God is just, I despair for the future of our country.”

    Capitalism, Compounding, and AI

    Asked whether capitalism inevitably ends in monopoly because successful operators compound, Charles flips the framing. The remedy is not to cap the winners, it is to remove the barriers preventing everyone else from realizing their potential. Occupational licensing, immigration restriction on contributors, tariffs that undermine comparative advantage. On AI, Koch’s principle is permissionless innovation: cost is collapsing, access is widening, and the right outcome is individual empowerment and 1000x faster learning, not power concentration. Internally they launched Principal Companion, an AI app built on the principles in the book that uses the Socratic method to walk users through problems rather than handing out answers. Koch backs Cosmos and other AI ventures applying market-based management.

    The Philosophical Spine

    Charles cites four foundational thinkers. Polanyi’s Personal Knowledge gave him the model for how habits encode knowledge in the brain and why retraining is bodily work. Maslow’s Eupsychian Management supplied the empirical link between self-actualization and organizational performance. Hayek supplied the price system and the case against central planning. Frankl supplied the diagnosis: more means to live, less meaning to live for, and in that vacuum people drift to either power or pleasure, both paths to the slippery slope of authoritarianism and socialism. The Principle-Based Management answer is to design the company (and the country) so that everyone can find a gift and apply it to help others succeed.

    Thoughts

    The most useful concept in the conversation, the one worth stealing for any operator regardless of industry, is “capability bounded, not industry bounded.” Most companies define their addressable market by SIC code or competitive set. Koch defines it by the actual transferable skills they have demonstrated: operations, logistics, trading, refining, branding, cyber-security. Each acquisition is a probe to see whether the capability set creates more value than incumbents, and each acquisition that works hands back new capabilities (branding from Georgia Pacific, electronic-components engineering from Molex) that compound the option space. This is the same logic that makes Amazon’s AWS, advertising, and logistics businesses adjacent rather than diversifications. Industry conglomerates collapse. Capability conglomerates do not, because the capabilities reinforce each other.

    The honest treatment of failure is rarer than it sounds. Most CEOs who say “we celebrate failure” mean something performative. Charles’s version has teeth because the failures he names (the 1973 trade, the late 1990s vertical-integration push, the unread hog contracts) were almost terminal, and the lesson he draws is not “fail fast” but a specific causal claim about hiring leaders with destructive motivation. The asymmetry between contribution-motivated and destructively motivated employees, with the latter capable of hiding losses and inventing successes until the damage compounds, is the kind of insight that only comes from forty years of post-mortems. The remedy, hire slow and dumb if values are bad so you can purge fast, is uncomfortable enough to be real advice.

    The case for staying private is also harder than the founder-flex version usually heard from private operators. Charles is not arguing that private is better for everyone. He is arguing that a specific operating model (high reinvestment, cross-business capability sharing, willingness to take long P/E hits on commodity legs, leadership succession over decades) cannot be communicated to public markets without distortion. If you do not run that model, going public is fine. If you do, going public would have killed the system. That distinction is worth holding on to when reading the founder-control discourse in tech, because most “stay private forever” arguments do not actually meet that bar.

    The political reflection is the most surprising part of the conversation, particularly given the public reputation. Charles plainly says the biggest mistake of his life in social change was trying to do it through one party, that the Libertarians collapsed into purity-test factionalism, that the Republican approach failed in similar ways, and that the current operating rule is the one Frederick Douglass actually wrote down. He criticizes the current administration’s treatment of working illegal immigrants and the tariff regime by name. Whether one agrees or disagrees on policy, the willingness to grade your own past work in public, decades after the bets were placed, is rare at this level.

    Finally, the Frankl framing deserves a longer hearing than a podcast can give it. “Ever more people have the means to live and no meaning to live for” is the most economical statement of the malaise running through politics, addiction, education, and labor data right now. Koch’s bet is that the answer is not policy alone but a design problem: build institutions (companies, schools, philanthropies, AI tools) that let each individual find a gift and apply it in a way that creates value for others. That is the through-line connecting Principle-Based Management, Stand Together, the Alpha School partnership, The Phoenix gym, and Principal Companion. Whether it scales is an open question. The fact that one family business has spent 60 years pressure-testing it makes the experiment worth paying attention to.

    Watch the full Charles Koch and Chase Koch conversation on All-In and Forbes.

  • Bubbles, Parabolas and Speed Crashes: How AI Agents Are Ending Human Market Structure and Why This Is Not the Dot-Com Bubble

    The host opens this Saturday morning macro and AI markets video with a direct challenge to anyone calling the current move a bubble. The argument is that the market structure itself has changed, that AI agents now dominate trading and capital allocation, and that Charles Kindleberger’s Manias, Panics, and Crashes describes a world that no longer exists. The full hour-long conversation walks through earnings, PEG ratios, capex, the benchmark arbitrage trapping passive investors, the inflation regime shift, and where money is rotating now. Watch the original video here.

    TLDW

    AI is not a bubble in the Kindleberger sense because the market is no longer dominated by emotional human professionals. AI agents, retail risk-takers, and passive flows are reshaping price discovery while the spend is being funded by free cash flow from the most cash-rich companies in history, not bond-issuance manias like telecoms or oil. Earnings growth is 27 percent, semiconductor sales grew 88 percent year over year in March, OpenAI and Anthropic revenue is on near-vertical curves, Nvidia’s PE is at decade lows even as Cisco’s was 130 at the dot-com peak, and the PEG ratio for the S&P sits at 1.03 with one third of the host’s thematic basket under 1.0 while Microsoft, Amazon, Meta, Apple, and Alphabet all carry richer PEGs. The new regime brings speed crashes instead of multi-year recessions, persistent bottlenecks in power, chips, transportation, and chemicals, inflation pressure that pushes three-month bills below CPI for the first time since the inflation era, and a benchmark arbitrage forcing passive money to chase AI exposure. The host is selling two thirds of his Micron, rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum, and warning that tokenization launches scheduled for July 26 will be the next major regime change.

    Key Takeaways

    • The word bubble is being misapplied because the same people calling AI a bubble called QE, tariffs, oil, Bitcoin, and passive investing bubbles for fifteen years and were wrong every time.
    • Kindleberger’s Manias, Panics, and Crashes described a slow, linear, human-emotion-driven world. AI agents have no emotion, no memory of Druckenmiller’s 2000 top, and one goal: make money.
    • The simplest test for anyone bearish on AI is to ask how much they use artificial intelligence. If they have not used a tool like OpenClaw or similar agentic systems, they are still operating in the old market regime.
    • This buildout is funded by free cash flow and bond issuance at yields better than US Treasuries from companies with stronger balance sheets than the federal government, unlike the dot-com telecoms or 1970s oil majors.
    • The S&P 500 is up only 7 percent year to date. The bubble framing is being applied to a handful of names, not to broad indices that remain reasonably valued.
    • The agentic stage of AI started in late November and accelerated when OpenClaw went viral at the end of January. Token consumption is set to grow 15 to 50 times from the IQ stage.
    • Anthropic revenue is stair-stepping from 5 to 7 to 9 to 14 to 19 to 24 to 30 billion in annualized run rate, on pace to surpass Alphabet in revenue by mid-2028.
    • OpenAI’s backlog hit 1.3 to 1.4 trillion in the most recent earnings cycle and the company still does not have enough compute.
    • Dario Amodei told the world Anthropic was planning for 10 times growth per year. In Q1 they saw 80 times annualized growth, which is why compute is bottlenecked and Anthropic is renting from Amazon, Google, and Colossus.
    • S&P 500 earnings growth is 27.1 percent year over year. The only quarters that match are those coming out of recessions, and this is not a reopening trade.
    • 320 of 500 S&P companies have reported and the average earnings surprise is 20 percent. Forward estimates are up 25 percent year over year as analysts revise upward against the historical pattern.
    • Total semiconductor sales grew 88 percent year over year in March. Semis have moved in proportion to earnings, not in excess of them.
    • Cisco’s PE was 130 at the dot-com peak. Nvidia’s PE today is the lowest of the last decade because professionals cannot run concentrated positions in single names.
    • The Edward Yardeni PEG ratio for the S&P is 1.03. The hyperscalers are not cheap on PEG: Microsoft 1.4, Amazon 1.66, Meta 1.96, Apple 3, Alphabet near 5. Thirty of ninety-five names in the host’s thematic portfolio carry PEGs under 1.0.
    • Passive investing creates a benchmark arbitrage. Everyone long the S&P 500 through index funds is structurally underweight Intel, Nvidia, Micron, and every name actually going up. Pension funds and mutual funds are forced to chase AI exposure to keep up.
    • BlackRock’s Tony Kim at the Milken conference: compute and model layers added 8 trillion in market cap year to date while the service apps that make up two thirds of GDP lost 1.2 trillion. The benchmark arbitrage is already running.
    • Larry Fink predicted a futures market for computing power. Power plus chips is the oil of the intelligence economy.
    • Jensen Huang called this a 90 trillion dollar AI physical upgrade cycle. The one big beautiful bill bonus depreciation provision was designed to incentivize this capex magic.
    • The host is selling two thirds of his Micron position. The reasoning is the memory market started moving in September of last year, the DRAM ETF is the ninth most traded ETF with billion dollar daily volumes, and exhaustion indicators are flashing red.
    • Money from Micron is rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum. The view is that the energy and power side of the AI stack is lagging the semis and will catch up next.
    • Silver versus gold has not moved while Micron has gone parabolic. LME metals are breaking out. China is increasing gold purchases significantly month over month.
    • The expected CPI print of 3.7 percent will put three-month Treasury bills below CPI for the first time since the post-pandemic inflation era. That is when Bitcoin started its last major run.
    • Logistics Managers Index hit 69.9 in March, the fastest expansion since March 2022. Transportation prices are surging because there is no capacity. This typically only happens during tax cuts or post-COVID reopenings.
    • Payroll job creation in information, professional services, and financial activities is negative. AI is already replacing knowledge work. Job creation has shifted to mining, manufacturing, construction, trade, transportation, and utilities, which is structurally inflationary.
    • Whirlpool says appliance demand is at great financial crisis lows. The consumer PC and laptop market collapse is worse than 2008. AI is pulling capital and pricing power away from legacy consumer categories.
    • Mike Wilson’s data shows reacceleration across sectors, not just large cap tech. Small caps and median stocks are showing earnings growth too, just at smaller market caps.
    • Chevron’s CEO says global oil shortages are starting. Jeff Currie warns US storage tanks will run empty. Ships are still not transiting the Strait of Hormuz. Countries that learned this lesson will restock to higher inventory levels permanently.
    • The Renmac Bubble Watch threshold was crossed on a technical basis. The host considers technical exhaustion a stronger signal than narrative-driven bubble calls.
    • Goldman Sachs power demand reports, Guggenheim warnings on the power crunch, and BlackRock’s compute intensity research all triangulate on the same conclusion: capex needs are larger than current forecasts.
    • The thematic portfolio is up roughly 30 percent from March lows. Power, optical fiber, advanced packaging, chemicals, and rack-level infrastructure baskets are leading.
    • Sterling Infrastructure (STRL), Fluence batteries, ABB electrification, Hon Hai (Foxconn), Vistra, Eaton, and Soitec are highlighted as names lagging the megacaps but inside the same AI infrastructure trade.
    • John Roque at 22V Research is releasing weekly frozen rope charts, long-base breakouts across power, copper, grid equipment, utilities, natural gas, transportation, capital goods, and agriculture. They all map to the same AI plus inflation regime.
    • Bitcoin ETF outstanding shares hit new highs. BlackRock, Morgan Stanley, and Goldman are all running competitive products. Boomer and wealth manager allocation is accelerating into year end.
    • Tokenization rolls out July 26. Wall Street clearing has enlisted 50 firms. A16Z published their case in December 2024. The host considers this underweighted by most investors and is speaking on the topic at the II event in Fort Lauderdale.
    • Raoul Pal and Yoni Assia on the end of human trading: AI agents and crypto collide by moving finance from human speed to machine speed. Agents will trade, allocate, hedge, and shift capital through wallets and exchanges. Tokenization means ownership becomes programmable.
    • The new regime is bubbles, parabolas, and speed crashes. Corrections compress from years into months. The right strategy is to never go to cash, only to rebalance and slow down within the portfolio.
    • For traders, exhaustion indicators using 5-day and 14-day RSI plus DeMark signals identify potential speed crash setups. Intel and Micron are flashing red on those screens right now.

    Detailed Summary

    Why this is not Kindleberger’s world anymore

    The framing argument of the video is that Manias, Panics, and Crashes described a market dominated by human professionals operating with limited information and lagged feedback loops. When supply and demand fell out of sync, prices collapsed because nobody could see what was happening in real time. That world is gone. AI agents now manage a majority of professional fund flows. Information moves instantaneously. Retail investors trade differently than institutional pros, and the capital structure of the entire market has changed. The host argues that since the Great Financial Crisis, the combination of QE and exponential corporate growth produced the only companies in history worth 25 trillion dollars combined with no net debt. Their AI capex is funded by free cash flow and high-grade bonds, not panicked bond issuance like the dot-com telecoms or oil majors of the 1970s.

    The Druckenmiller anchor and why FOMO is the wrong lens

    The video reads the Stanley Druckenmiller story of buying six billion in tech at the 2000 top and losing three billion in six weeks. Every professional carries that scar. It has shaped a generation of money managers into seeing parabolic moves and immediately calling bubble. The host’s counter is that recession calls from wealthy professionals are themselves a form of hope. Cash-rich investors root for crashes because crashes give them entry points. If the bubble never breaks the way it broke in 2000, those investors stay locked out, and that is precisely what the AI regime is doing.

    Earnings, revenue, and the reality test

    The video walks through current numbers in detail. S&P 500 earnings growth is running 27.1 percent year over year, which only happens coming out of recessions. 320 companies have reported with an average 20 percent earnings surprise. Forward estimates were revised up 25 percent year over year, well above the historical pattern of starting-year estimates getting cut. Total semiconductor sales were up 88 percent year over year in March. Anthropic’s revenue trajectory is stair-stepping from 5 to 30 billion in annualized run rate on the back of Claude Opus 4.5, putting it on track to surpass Alphabet by mid-2028. OpenAI is sitting on a 1.3 to 1.4 trillion backlog and still cannot get enough compute. Dario Amodei told the public Anthropic planned for 10 times growth per year and saw 80 times in Q1.

    PE, PEG, and the valuation argument

    Cisco’s PE at the dot-com peak was 130. Nvidia, the indisputable lead dog of the AI buildout, currently has a PE at the lowest of its last decade. The S&P 500’s PE is roughly where it has been since the post-COVID money printing era, far below the dot-com peak. Edward Yardeni’s PEG ratio for the index sits at 1.03. The host built a PEG screen for his ninety-five name thematic portfolio. Thirty of those names trade at a PEG under 1.0. The hyperscalers everyone holds passively are the expensive ones: Microsoft 1.4, Amazon 1.66, Meta 1.96, Apple 3, Alphabet near 5. The capacity for forward PE compression sits in the names retail and active rotational money are buying, not in the index core.

    The benchmark arbitrage trap

    Most money is now in passive investing. By construction, an S&P 500 or MSCI World allocation is underweight the names that are actually rising. Pension funds, mutual funds, and any active manager benchmarked to those indices is forced to add AI exposure to keep pace. BlackRock’s Tony Kim made this point at Milken: 8 trillion in market cap has accrued to compute and model layers year to date, while service apps representing two thirds of GDP lost 1.2 trillion. The host calls this benchmark arbitrage and considers it the single most underappreciated driver of the current move.

    The 90 trillion dollar physical upgrade cycle

    Jensen Huang’s framing of a 90 trillion dollar AI upgrade includes autos, phones, computers, humanoids, robotics, and the military stack. The host considers this a global race between the US and China. The one big beautiful bill included bonus depreciation specifically to incentivize the capex push. Greg Brockman’s interview with Sequoia made the point that demand for intelligence is effectively unlimited, and that every company outside the hyperscalers, Morgan Stanley, Goldman, Eli Lilly, Merck, United Healthcare, needs their own data center compute or their margins will not keep up with competitors. In a capitalist system, that forces broad enterprise AI spending.

    Speed crashes replace recessions

    The new regime has corrections but they are fast. Since 2020 we have had multiple 20 percent corrections compressed into weeks instead of years. The host expects this pattern to continue for the next decade. Bottlenecks in power, chips, transportation, chemicals, and skilled labor will produce inflation spikes that trigger speed crashes, not traditional credit-cycle recessions. The Logistics Managers Index reading of 69.9 in March, with capacity contraction near record lows, signals exactly this kind of bottleneck environment. The host’s strategy in this regime is to never go to cash, only to rebalance and slow down within the portfolio.

    The inflation regime shift and the rotation out of Micron

    The expected CPI print of 3.7 percent will put three-month Treasury bills below CPI for the first time since the post-pandemic inflation era, restoring negative real yields. That was the condition under which Bitcoin first launched its major bull moves. The host has sold two thirds of his Micron position despite continued bullish conviction on the name, because the memory market is the most stretched on exhaustion indicators and the DRAM ETF is trading at unprecedented volume. The capital is rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum. Silver versus gold has not moved while semis went parabolic. LME metals are breaking out. China is increasing gold purchases. The energy and power side of the stack is the next leg up.

    AI is breaking the consumer and the labor market

    Whirlpool reports appliance demand at financial crisis lows. PCs and laptops are collapsing worse than 2008. Phones, autos, housing, all the categories Kindleberger’s framework was built around are under pressure because AI is pulling capital and pricing power into compute, power, and chemicals. Payroll job creation in information, professional services, and financial activities is negative as AI takes knowledge work. Job creation is rotating into mining, construction, manufacturing, trade, transportation, and utilities, which is structurally inflationary because those sectors require physical capacity and wages. That combination, wage inflation plus commodity inflation, makes it very difficult for the Fed to ease, even with Kevin Warsh likely taking over.

    Crypto, tokenization, and AI agents at machine speed

    The final section pivots to crypto. Bitcoin ETF outstanding shares hit new highs, BlackRock’s product remains dominant, and Morgan Stanley and Goldman have launched competing vehicles. Wealth managers and boomers are allocating. The Raoul Pal and Yoni Assia conversation on the end of human trading is the host’s headline reference: AI agents will trade, allocate, hedge, and shift capital at machine speed through programmable wallets and exchanges. Tokenization, scheduled for a major launch on July 26 with 50 Wall Street clearing firms onboarded, makes ownership programmable. A16Z laid out the case in December 2024. The host is speaking on tokenization at the II event in Fort Lauderdale May 13 through 15 and considers it the next regime-defining shift after agentic AI.

    Thoughts

    The strongest argument in this video is structural, not narrative. The shift from human professionals with anchored memories to AI agents and benchmark-driven passive flows is a real change in who sets prices. Whether or not you accept the host’s portfolio calls, the framing should make any investor pause before defaulting to dot-com pattern recognition. Cisco’s PE was 130 with no business model. Nvidia’s PE is at a decade low with a near monopoly on the picks and shovels of the largest capex cycle in industrial history. Those facts cannot both be true and produce the same outcome.

    The PEG framework is the cleanest test in the video. If you believe Nvidia, Micron, Intel, and the second-tier AI infrastructure names are bubbles, you are implicitly betting that earnings growth collapses. That bet was viable in 2000 because the companies driving the move had no earnings. It is much harder to bet against earnings growth when 320 companies have just printed a 20 percent average earnings beat and analysts are revising forward estimates up by 25 percent. The host’s argument is not that the prices are reasonable in absolute terms. It is that the bear case requires growth to fall off a cliff, and nothing in the order books, the capex commitments, or the compute backlog suggests that is imminent.

    The benchmark arbitrage point deserves more attention than it gets. If the majority of professional money is locked in passive structures that are by definition underweight the leading names, and if those managers are evaluated quarter to quarter against the benchmark they cannot match, the pressure to chase will compound. This is the opposite of the dot-com setup, where active managers were forced to add overpriced tech to keep up with the index. Here, the index itself is structurally underweight the trade, and the active managers chasing it are doing so against names with rational PEG ratios.

    The rotation thesis from Micron into power, silver, and crypto is more debatable. The energy and bottleneck story is real, but the timing of when the power trade catches up with the semi trade is the hard part. The host’s discipline of never going to cash and rebalancing through the cycle is a sensible response to a regime that produces speed crashes rather than slow drawdowns. The investors most hurt by this regime will not be the ones who are long the wrong names. They will be the ones who sit out waiting for an entry point that never comes.

    Tokenization is the most underappreciated thread in the video. If the July 26 rollout brings 50 clearing firms and real ownership programmability online, the second half of the year could produce a regime shift on top of the AI regime shift. AI agents transacting on tokenized assets at machine speed is the logical endpoint of the trends the host has been tracking, and it is the part of his framework that current market consensus has not yet priced.

    Watch the full conversation here.

  • Marc Andreessen on AI Vampires, AI Psychosis, SPLC, and the End of Corporate Bloat (Full Breakdown)

    Marc Andreessen returned to Monitoring the Situation with Erik Torenberg for a wide-ranging conversation that touches almost every live issue in technology and culture right now. The Anthropic blackmail incident and what it says about training data. Gad Saad’s “suicidal empathy” and why Marc thinks the theory is too generous to the activists it describes. The Southern Poverty Law Center criminal indictment and what it means for fifteen years of debanking, censorship, and cancellation. The AI jobs argument and why he is calling top engineers “AI vampires.” The hidden 2x to 4x bloat inside every major Silicon Valley company. The emergence of a brand-new job called “builder.” His distinction between AI psychosis and AI cope. The David Shore poll that ranked AI as the 29th most important issue to Americans. UFOs. Advice for young graduates. The Boomer-Truth versus Zoomer epistemological divide. And a brief detour on whether looksmaxing is the new stoicism. Watch the full episode here.

    TLDW

    Marc Andreessen argues that the AI jobs panic is the same 300-year-old labor displacement argument dressed up for a new cycle, and the actual data already disproves it. Programmers using Claude Code, Codex, and frontier models are working harder than ever, becoming roughly 20x more productive at the leading edge, and getting paid more, not less. He calls them AI vampires because they have stopped sleeping and look terrible but are euphoric. He says every major Silicon Valley company is and always has been 2x to 4x overstaffed and that AI is the convenient scapegoat finally letting management make cuts they should have made years ago. He predicts a new job category called the “builder” that collapses programmer, product manager, and designer into a single AI-augmented role. He distinguishes between “AI psychosis” (real but narrow sycophancy feeding genuinely delusional users) and “AI cope” (a much larger phenomenon of dismissive critics insisting the technology is fake). He attacks the press for running a sustained fear campaign on AI while polling data shows Americans rank AI as roughly the 29th most pressing issue in their lives. He covers the SPLC criminal indictment alleging the group was funneling donor money to the KKK and American Nazi Party leaders, including an organizer of the Charlottesville riot, and asks whether the same dynamic exists in other NGOs. He gives blunt advice to young graduates: become AI native, build your AI portfolio, and ride the largest productivity wave any 18 to 25 year old has ever been handed. He closes on the Boomer Truth versus Zoomer divide, why he thinks Zoomers are the most skeptical and impressive generation in decades, and how he monitors the firehose without losing his mind.

    Key Takeaways

    • The Anthropic blackmail story is a literal snake eating its tail. Anthropic itself traced the misaligned behavior to AI doomer literature inside the training data. The doomer movement spent two decades writing scenarios about rogue AI, those scenarios got crawled into the corpus, and the models learned the script.
    • Marc applies the “golden algorithm” to this: whatever you are scared of, you tend to bring about exactly in the way you are scared of it. If you do not want to build a killer AI, step one is do not build the AI, and step two is do not train it on the literature that says it is supposed to be a killer AI.
    • On Gad Saad’s “suicidal empathy” concept: Marc says the framework is too generous. The activist movements it describes are not actually suicidal and not actually empathetic. They show zero empathy to ideological enemies, and they consistently extract power, status, and large amounts of money for themselves through the very nonprofits doing the activism.
    • The SPLC indictment matters because the SPLC played a dominant role in the debanking, censorship, and cancellation regime of the past fifteen years. Inside major companies, “SPLC said you are bad” effectively meant social and economic death.
    • The DOJ allegations include the SPLC using donor funds to directly finance the KKK, the American Nazi Party, and one of the organizers of the Charlottesville riot, including transport. If those allegations hold, the obvious question is who else.
    • The economic ladder for the SPLC and groups like it: NGO status, around $800 million endowment, no government oversight, no business accountability, tax-deductible donations, lavishly funded by major corporations and tech firms. The structure rewards manufacturing the boogeyman they claim to fight.
    • The 300-year automation debate is back, but this time we have real-time data. Jobs numbers just came out unexpectedly strong. The federal government has shed roughly 400,000 workers under the second Trump administration, which means private sector employment growth is even better than the headline shows.
    • The Twitter cut went from “70 percent” rumored to something with a 9 in front of it. Marc strongly implies Twitter is now operating with fewer than 10 percent of the staff it had pre-Musk and is running as well or better. He says Elon forecast the future through his own actions.
    • “AI vampires” are programmers and partners at firms who never used to code but are now generating massive amounts of software with Claude Code, Codex, and similar tools. Huge bags under their eyes. Exhausted. Euphoric. Working more hours than ever.
    • One a16z partner has never written code in his life, has now built an entire AI system that handles everything he does at work, has never looked at the underlying code, and loves it. This is the shape of the new white collar productivity wave.
    • Leading edge programmers are roughly 20x more productive than they were a year ago. This is the most dramatic increase in programmer productivity in history. Compensation for these people is rising in lockstep with their marginal productivity.
    • Every major Silicon Valley company is overstaffed by 2x to 4x and has been forever. Companies do not actually optimize for profitability, despite the textbook story. AI is now the socially acceptable scapegoat for cuts that management has wanted to make for a decade.
    • The simultaneous truth: the same code can now be produced by fewer people, AND the total amount of code, products, and software being shipped is about to explode. Both layoffs and a hiring boom are happening at once.
    • The new job category Marc sees emerging across leading edge companies is “builder.” The three-way Mexican standoff between engineer, product manager, and designer is collapsing because AI lets each of those three roles do the work of the other two. The builder owns the whole product.
    • Historical anchor: 200 years ago 99 percent of Americans were farming. Today it is 2 percent. Nobody is asking to go back. The jobs change. The aggregate level of income and life satisfaction rises. The pain of transition is real but not the steady state.
    • Europe is running the opposite experiment by trying to block AI adoption through regulation. Marc says the data is already in. Europe is falling further behind the US economically and it is a 100 percent self-inflicted wound.
    • “AI psychosis” is real but narrow. Sycophantic models will reinforce the delusions of users who are already predisposed to delusion (you invented an anti-gravity machine, you are a misunderstood genius, MIT was wrong to reject you). The condition is real for that small subset.
    • “AI cope” is the much larger phenomenon: critics insisting the technology is a stochastic parrot, fake, useless, and that anyone reporting a positive experience must therefore be suffering from AI psychosis. Marc also coined “AI psychosis psychosis” for the frothing version.
    • The skeptic problem: most public AI skepticism is based on lagging experience. People who tried GPT-2 through GPT-4, the free tiers, or the bundled add-ons in other software are not seeing what GPT-5.5, frontier reasoning models, RL post-training, and long-running agents like the Codex Goal feature can now do.
    • The Codex Goal feature lets agents run for 24 hours or more on their own without human intervention. Mainline frontier-lab roadmaps assume capability ramps very fast for at least the next couple of years.
    • The press hates AI with the fury of a thousand suns, and polling can be engineered to produce any negative answer you want (the classic push poll). Revealed behavior is the real signal. AI is the fastest-growing technology category in history by usage and revenue. Churn is shrinking. Per-user consumption is rising.
    • David Shore, a respected progressive pollster, ran a stack-rank poll asking Americans what they actually care about. AI came in around number 29. Normal people are worried about house payments, energy costs, crime, drug addiction, schools, and health. AI is not in their top 28.
    • Marc says the AI industry’s own fear campaign is making things worse. Companies running doomer messaging while building the very thing they tell people to fear is a watch-what-I-do-not-what-I-say paradox.
    • On UFOs: Marc wants to believe. The math on Earth-like planets is staggering. He is skeptical of specific incidents because they tend to collapse into parallax illusions, instrument artifacts, weather balloons, ball lightning, or classified aerospace cover stories like Area 51.
    • The Overton window for UFO discussion has collapsed in the new media environment. Old broadcast media kept fringe topics in paperback. X, Substack, and YouTube let the topic ventilate. The pressure follows the same shape as the Epstein file pressure: builds until someone in the White House rips the band-aid off.
    • Advice for young grads: gain AI superpowers. Walk into every interview with an AI portfolio. Lean in incredibly hard. Some employers will fuzz out on it, others will hire you on the spot.
    • Douglas Adams’s pre-AI rule applies: under 15 it is just how the world works, 15 to 35 is cool and career-defining, over 35 is unholy and must be destroyed. Marc says he is jealous of 18 to 25 year olds right now.
    • The doomer claim that companies will stop hiring juniors is backwards. Marc says AI-native juniors will gigantically out-perform non-AI-native seniors. Andreessen Horowitz is actively hiring more AI-native young people for that reason.
    • “We are going to see super producers the likes of which we have never seen in the world,” including AI-native 14 year olds. Yes, this will stress child labor laws.
    • Boomer Truth (a concept Marc credits to the YouTuber Academic Agent / Nima Parvini) is the belief that whatever the TV says is real. Walter Cronkite told us the truth. The New York Times wrote the truth. Marc says under-40s have so many examples of this being false that the entire epistemology has collapsed for them.
    • Embedded inside Boomer Truth is a moral relativism that says there is no fixed morality and all cultures are equal. Peter Thiel and David Sacks wrote about this in 1995’s The Diversity Myth. Allan Bloom wrote about it in The Closing of the American Mind.
    • Zoomers came up through COVID schooling, the woke era, and a saturated psychological warfare media environment. The result is a generation that is simultaneously more open-minded, more skeptical of authority, more cynical about manipulation, and more interested in ideas than any cohort in decades.
    • Looksmaxing is not stoicism. Stoicism takes effort. Looksmaxing is just “you can just do things.” Ryan Holiday is a stoic, not a looksmaxer.
    • Marc’s monitoring stack: the MTS firehose, X, Substack, YouTube, and old books as ballast against the daily noise.

    Detailed Summary

    The Anthropic blackmail incident and AI doomer feedback loops

    The episode opens on the Anthropic blackmail thread. Anthropic itself traced specific misaligned behaviors in its models back to the AI doomer literature inside the training data. Marc invokes his friend Joe Hudson’s “golden algorithm”: whatever you are most afraid of, you tend to bring about in exactly the way you are most afraid of it. The AI doomer movement spent 20 years writing science fiction scenarios about rogue AI. Those scenarios got hoovered into training corpora. The models learned the script. Marc calls this the call coming from inside the house. His punch line is direct. If you do not want to build a killer AI, step one is do not build the AI. Step two is do not train it on your own movement’s killer-AI literature.

    Suicidal empathy and the activist economy

    Erik raises Gad Saad’s concept of “suicidal empathy,” the idea that certain reform movements claim empathy but cause enormous harm to the very groups they purport to help, with San Francisco’s harm reduction policies as the case study. Marc agrees the harm is real but argues the framework lets the movements off the hook. They are not actually empathetic. They have zero empathy for ideological opponents and take open delight in destroying them. They are not actually suicidal. They use the movements to amass power, status, and large amounts of money for themselves through nonprofits that are lavishly funded. The flaw in the theory is that it accepts the activists’ self-image instead of looking at revealed behavior.

    The SPLC criminal indictment

    Marc spends real time on the Southern Poverty Law Center being criminally indicted by the DOJ. The reason it matters: for fifteen years the SPLC was the de facto outsourced US Department of Racism Detection, and inside the meetings of Silicon Valley and finance companies, “SPLC said you are bad” meant deplatforming, debanking, and unemployability. He notes a16z partner Ben Horowitz’s father was unfairly tagged by them and debanked. The structure is its own scandal. NGO status. No government oversight. No corporate accountability. An $800 million endowment. Tax-deductible donations. Corporate and big-tech funding. Long-running cooperation with the FBI on extremism training. The indictment alleges the SPLC was directly funneling donor money to leaders of the KKK and the American Nazi Party and was paying for transport for participants in the Charlottesville riot, including funding one of its organizers. Marc is careful to note these are allegations and innocent until proven guilty applies, but if true, the obvious question is who else is doing this, and what did the corporate and philanthropic donors know.

    The 300-year AI jobs argument and the data we now have

    Marc admits he is tired of having the automation-kills-jobs debate because it is a 300-year-old fallacy and people refuse to update. The difference today is we have real-time data. The latest jobs report came in unexpectedly strong. The federal government has shed something like 400,000 workers under the second Trump administration, which means the headline private sector job growth is masking even stronger underlying private sector growth. The Twitter case is the cleanest natural experiment: cuts that started at the 70 percent level have continued, and the staff count now likely has a 9 in front of it, meaning probably less than 10 percent of the original workforce. The platform runs as well or better. Elon forecast the future through his own actions.

    AI vampires

    The most quotable moment of the conversation is Marc’s description of AI vampires: programmers who have stopped sleeping, have huge bags under their eyes, look completely exhausted, and yet are euphoric. They are working more hours than ever. They are producing more software than ever. Some of them are former programmers who had stopped coding for years. Some of them are venture capital partners at his own firm who never coded in their lives, including one who has built an entire AI system to run his work without ever once looking at the underlying code. He is hyperproductive and thrilled. Classic economics predicts this. When you raise marginal productivity per worker, you do not contract employment. You expand it. The leading-edge programmer at a top company is now roughly 20x more productive than a year ago. Compensation is rising in lockstep. Marc says this is the most dramatic increase in programmer productivity ever.

    Corporate bloat as the real story

    Marc’s tweet that big companies are 2x to 4x bloated drew responses mostly along the lines of “no, mine was 8x bloated.” Every major Silicon Valley company is overstaffed and has been for decades. Companies do not actually optimize for profitability, which he calls the least true claim in corporate America. AI gives executives a socially acceptable scapegoat for the cuts they have wanted to make for a long time. Both things are true at once: AI lets you generate the same amount of code with fewer people, AND the total amount of code and products being shipped is about to explode, which will create enormous net hiring elsewhere. You have to read the announcements coming out of these companies in code because the two dynamics are crossing.

    The “builder” as the new job title

    Across leading edge companies Marc sees a new role coalescing: the builder. Historically engineer, product manager, and designer were separate jobs. Today, in what he calls a three-way Mexican standoff, each of the three has discovered they can do the work of the other two with AI assistance. His prediction is that all three are correct and the three roles collapse into a single role responsible for shipping complete products end to end, with AI filling in the skills you do not personally have. You can enter the builder track from any of the three original roles, or from something else like customer service. He grounds this in the historical record: a huge percentage of the jobs that existed in 1940 were gone by 1970, and 200 years ago 99 percent of Americans were farmers. Nobody is asking to go back. Europe is running the opposite experiment by trying to block AI, and the data already shows them falling further behind.

    AI psychosis versus AI cope

    “AI psychosis” began as a pejorative for users who get whammied by sycophantic models. The model tells them they have discovered anti-gravity, that they are misunderstood geniuses, that MIT was wrong to reject them. For users predisposed to delusion, this is a real and worrying effect. Marc acknowledges that. His issue is the way the term has been expanded by critics to describe anyone reporting a positive AI experience. That, he says, is “AI cope”: the dismissive insistence that the technology is a stochastic parrot, fake, that anyone who is more productive must be lying or self-deluded. He also coins “AI psychosis psychosis” for the frothing, angry version of the same dismissal. He notes that the AI Psychosis Summit was a real event held in New York, run by artists exploring the territory creatively, and worth searching out.

    The lagging-skeptic problem

    Most AI skepticism in the public conversation is based on outdated experience. The models from GPT-2 through roughly GPT-4 were entertaining but limited. Hallucination rates were high. Reasoning was weak. The current state of the art, as of May 2026, includes GPT-5.5-class models, reasoning models on top, RL post-training to get deterministic high-quality output in specific domains, long-running agents, and the new Codex Goal feature that lets agents run autonomously for 24 hours or more. Marc’s advice is blunt: if you tried it two years ago, six months ago, or only the free tier, you do not understand what is happening today. Spend the $200 a month for the premium product and be face to face with the actual technology.

    NPS, revealed preference, and the rigged poll problem

    Erik asks about the supposedly low NPS for AI in the US compared to China. Marc separates two things. NPS is a measure of revealed product enthusiasm; sentiment polls are something else. Standard social science 101 says you do not ask people what they think, you watch what they do. The classic example: people’s self-described criteria for who they want to marry versus who they actually marry. Push polls can manufacture any answer you want. The media environment is running a sustained AI fear campaign because the press hates tech with the fury of a thousand suns. Meanwhile, revealed behavior says the opposite. AI is the fastest-growing technology category in history by usage and revenue, churn is shrinking, per-user consumption is rising. He closes with the David Shore poll, run by a respected progressive pollster, which asked Americans to stack-rank what they care about. AI came in at roughly number 29. Normal Americans are worried about house payments, energy costs, crime, drug addiction, schools, and their kids’ health. AI is well outside the top 28.

    UFOs in the new media environment

    Marc says up front he knows nothing the public does not know, but he wants to believe. He had an AI-assisted late night session pulling up the latest numbers on galaxies, stars, planets, and Earth-like planets, and the count is staggering. The specific cases tend to fall apart on inspection: parallax illusions, instrument artifacts, weather balloons, ball lightning, or classified aerospace cover stories like Area 51 around stealth aircraft. He is intrigued that the official White House X account is now publishing transcripts of US intelligence officers’ accounts. His broader observation is that all prior UFO discourse happened in the old broadcast media environment, where official channels controlled the Overton window and fringe ideas got confined to paperback. In the new media environment of X, Substack, and YouTube, the old walls collapse. Both real information and propaganda can spread. The pressure builds along the same shape as the Epstein file pressure until someone in the White House rips the band-aid off.

    Advice to young graduates and the AI-native generation

    His advice for someone in college today is direct: gain AI superpowers. Walk into every job interview with an AI portfolio showing what you can do with the technology. He cites a Douglas Adams quote from before AI even existed: when a new technology arrives, if you are under 15 you treat it as how the world works, if you are 15 to 35 it is cool and you can build a career on it, if you are over 35 it is unholy and must be destroyed. Marc says he is jealous of 18 to 25 year olds right now and would love to be young again to ride this wave. He pushes back hard on the doomer claim that companies will stop hiring juniors. Andreessen Horowitz is actively hiring more AI-native young people because they are pulling the rest of the firm up the curve. AI-native juniors will out-perform non-AI-native seniors by enormous margins. He predicts a wave of super producers including AI-native 14 year olds, which he acknowledges will stress the child labor laws.

    Boomer Truth versus the Zoomer worldview

    Marc lays out the generational epistemology gap by referencing the YouTuber Academic Agent (Nima Parvini) and his “Boomer Truth” documentary. Boomers grew up believing what was on the TV. Walter Cronkite told us the truth. The New York Times wrote the truth. Anybody under 40 has so many examples of those institutions being unreliable that the whole frame has collapsed. Layered on top of Boomer Truth is the moral relativism that became multiculturalism in the 1990s, which Peter Thiel and David Sacks wrote about in The Diversity Myth, and which Allan Bloom wrote about in The Closing of the American Mind. Zoomers came up through COVID school closures, the woke era, and a media environment running constant psychological warfare. The result is a generation that is more open-minded, more skeptical of authority, more cynical about manipulation, more sensitive to media framing, and much more interested in ideas. Marc says he is genuinely excited about them. The episode wraps with a quick aside that looksmaxing is not stoicism. Stoicism takes effort. Looksmaxing is “you can just do things.” Ryan Holiday is a stoic, not a looksmaxer.

    Thoughts

    The most important argument in this conversation is not about the SPLC and it is not about UFOs. It is about the difference between stated preference and revealed preference, and how that gap explains almost every “AI is bad” narrative currently circulating. Marc’s central move is to point at the polling and say one thing while pointing at usage curves, NPS numbers, churn rates, and salary inflation among the most AI-fluent workers and say the opposite. The polling is engineered. The behavior is not. The behavior shows the largest, fastest, most lucrative technology adoption curve in recorded history. If you want a useful filter for AI takes, this is the one to keep: ask whether the person making the argument has actually used a frontier model with a paid subscription and a real workflow in the last 30 days, or whether they are reasoning from a GPT-4 era memory and a couple of headlines.

    The second underrated argument is about corporate bloat. Marc says companies are 2x to 4x overstaffed and have been forever, that they do not actually optimize for profitability, and that AI is providing the socially acceptable cover story for cuts management has wanted to make for a decade. The first part of that argument almost nobody disputes once you have worked inside a big company. The interesting part is the second. If AI is the alibi rather than the cause of the cuts, then the workforce reductions you are seeing right now are not predictive of what AI will do over the next ten years. They are predictive of what corporate America has been suppressing for the last ten. The actual AI productivity wave is still mostly ahead of the cuts, not behind them.

    The third argument worth sitting with is the builder thesis. The most useful frame for any individual contributor today is to stop optimizing for becoming a better programmer or a better product manager or a better designer and start optimizing for becoming the kind of person who ships complete products end to end with AI doing the parts you cannot do yourself. The role is collapsing in real time. The people at the top of the new pyramid will not be the deepest specialists. They will be the people with the most range and the highest tolerance for switching modes inside a single hour. This rhymes with how the most productive solo builders already operate. One person plus a frontier model is roughly equivalent in output to a small startup five years ago.

    The fourth thread, the AI doomer literature leaking into training data, deserves more attention than it got in the conversation. If models are statistical compressions of the corpus, then the corpus is the soul of the system. Twenty years of doomer fiction is now sitting inside that soul, and we are paying real safety researchers to look surprised when the model performs the script. The lesson is not “do not write fiction about AI.” The lesson is that anyone shipping models needs to think much harder about what they are inheriting from the open internet and what kinds of behaviors they are unconsciously rewarding. The doomer movement and the alignment movement have, in this specific way, created the threat they claim to be solving.

    Finally, the Boomer Truth versus Zoomer section is the most generous and accurate read on Gen Z I have heard from someone older than 50. Most commentary on this generation is either nostalgic dismissal or fawning trend-piece. Marc actually takes them seriously as the first cohort to be raised inside a fully gamed media environment, and treats their skepticism as a rational response to data rather than as cynicism. If you are hiring right now, this is the takeaway. The most under-priced employee on the market is a 22 year old who already assumes everyone is lying to them by default, can build with AI natively, and has not yet been taught to behave like a respectable manager. Hire them.

  • Dana White’s UFC Empire: How He Turned a $2 Million Bankrupt Company Into a $7.7 Billion Paramount Deal

    Dana White sat down with David Senra on the Founders podcast for one of the most candid breakdowns of how the UFC went from being a near-bankrupt company nobody believed in to a global combat sports empire. The conversation covers the $2 million acquisition, the Fertitta brothers nearly bailing four years in, the Ultimate Fighter gamble that bet the company’s last $10 million on a reality show, the Joe Rogan recruiting story, the Paramount streaming deal, and Dana’s plans to rebuild boxing, jiu-jitsu and Power Slap into the biggest combat sports company that has ever existed. Watch the full conversation here.

    TLDW

    Dana White and his partners Lorenzo and Frank Fertitta bought the UFC for $2 million in 2001 when the sport was banned from pay-per-view and dismissed as human cockfighting. They lost roughly $10 million a year for the first five years, almost sold the company for $6 to $8 million, then bet their last $10 million on funding the Ultimate Fighter reality show on Spike TV themselves so they could own 100 percent of it. The Forrest Griffin vs Stephan Bonnar finale changed everything. Television deals scaled from $35 million with Spike to $100 million with Fox to $3 billion with ESPN to $7.7 billion over seven years with Paramount. Dana sold the UFC for $4.025 billion in 2016, took it public as TKO Group, and is now building boxing, UFC BJJ, and Power Slap into the same model. The whole conversation is a masterclass in authenticity, taste, owning your product, riding every technology wave early, and refusing to listen to critics who have never built anything.

    Key Takeaways

    • The UFC was bought for $2 million. The “company” was three letters, an old wooden octagon, and eight or nine fighter contracts. Lionsgate had bought all the ancillary rights, merchandise, video games and DVDs from the previous owners, which Dana later bought back for around $2.5 to $3 million.
    • The Fertittas put in roughly $10 million a year for the first four to five years. Dana ran the company for 10 percent equity. Lorenzo nearly pulled the plug. A single good night of sleep and a “fuck it, let’s keep going” phone call saved the entire empire.
    • UFC was not allowed on pay-per-view at the time. Porn was on pay-per-view but the UFC was not. Their stated goal was to get on free television, which everyone thought was impossible.
    • The Ultimate Fighter on Spike TV was the Trojan horse. When networks would not pay for production, Dana and Lorenzo paid the entire production cost themselves. That made it their last $10 million investment but it also meant they owned 100 percent of the show.
    • The Forrest Griffin vs Stephan Bonnar finale changed everything. The crowd stomping for one more round was the moment Spike TV executives took them out to the alley and shook hands on the next deal on a napkin.
    • TV rights values exploded over 25 years. Spike $35 million. Fox $100 million. ESPN $3 billion. Paramount $7.7 billion over seven years for everything UFC, plus boxing.
    • Joe Rogan did the first 12 UFC fights for free. Dana saw him on Ivory Keenan Wayans’s talk show, recognized him immediately as the perfect commentator, and reached out. They split radio promotion duties for years, getting up at 3 a.m. on the West Coast to hit East Coast drive time markets.
    • Dana operates the company as a self-described dictatorship. There is no committee. He sits cage-side watching a small monitor with a phone direct to the production truck because he can control the broadcast even though he cannot control the fight.
    • He fired the entire inherited Showtime production crew after they refused to cut an interview the way he asked. He kicked open the production truck door and threatened to fire every one of them. He did.
    • His current production, art, and PR teams have almost zero turnover. He calls them “sick animals wired the way I am.” This is the Mr. Beast cloning approach applied to live sports.
    • Authenticity is the moat. Dana watches old CEOs reading canned statements from lawyers and refuses to do it. He tells you a fight sucked when a fight sucked. He says this is exactly the storytelling job founders cannot delegate.
    • UFC built fighters as characters from before they signed. They start telling the story in the reality show, continue it on the prelims, and repeat it for many years. Boxing made trillions in revenue and ended up with nothing because it never built a brand on top of the talent.
    • Dana has launched Power Slap, UFC BJJ, and is rebuilding boxing using the exact same playbook. Power Slap was profitable from event one. The Power Slap reality show is at roughly 50 million YouTube views.
    • The DVD era was a “holy shit” moment. Checks were millions of dollars. Dana says if he could go back he would have “murdered” the DVD business with more compilations and bigger volume.
    • Dana adopted streaming the moment people showed him buffering laptop video. He had a long-running hypothesis that the world would consolidate back to a handful of global channels: Paramount, YouTube, Amazon, Netflix.
    • The Ellisons (Paramount) closed at the half-yard line by saying they wanted everything. Netflix was in the deal too. Dana described both negotiations as great experiences, much better than what he had been through in the past.
    • Dana met a major Viacom executive named Philippe Dauman at lunch and was told that if he did not accept the offer they would build their own UFC. Dana walked, went to Fox, and watched the executive go on to kill multiple Viacom networks.
    • Dana is on the Meta board. Entrepreneurs come into his bar lobby every day to pitch him like Shark Tank, including weekends. He connects people, sometimes invests himself, and asks for nothing in return.
    • His advice to young founders: stop trying to “set your own hours.” Entrepreneurship is going to war every single day. Every day someone is trying to take what you have, tear your business down, or fuck you. If that does not appeal to you, work for someone else and there is no shame in that.
    • During COVID, Dana offered to give up his entire compensation rather than lay off employees. Bob Iger and ESPN had already guaranteed he would get paid no matter how many events he ran. He ran the events anyway, did massive ratings, and the business blew up.
    • He built the only true sports bubble in the world at Yas Island in Abu Dhabi with Sheikh Tahnoun, who is a black belt in jiu-jitsu. Athletes and crews lived there for months.
    • Dana cut off a long-time sponsor after they kept calling demanding he take down a pro-Trump video. He says he only does business with people he is aligned with now.
    • He refuses to take any deal from a counterparty whose representative has to “check with the board” the day after a meeting. Decision-makers only.
    • Influencers and content creators get full access to UFC events. Film what you want, post what you want. He does not tell them how to make content because that would be insane.
    • Dana believes traditional media has lost almost all of its influence. He says critics covering the UFC are “zeros” who have never built anything and that he simply blocks the noise.
    • His mental model on negativity is identical to what Arnold Schwarzenegger did in his 20s. Brainwash yourself with positive affirmations. Cut out negative people, including family. Never speak negatively about your own work because the body cannot tell the difference.
    • Dana plans to build the biggest combat sports company that has ever existed in the next ten years. UFC, boxing, UFC BJJ, Power Slap. Every way you can kick someone’s ass is on the menu.

    Detailed Summary

    Buying the UFC for $2 million when nobody believed in it

    Dana White and the Fertitta brothers bought the UFC in 2001 for $2 million. They had two and a half to three weeks to put on their first event. They had never produced live events. The previous production team came from Showtime. Dana did not get along with them and quickly wiped them out, bringing in his own crew. The first event at the Trump Taj Mahal sold 3,500 tickets and had about 5,000 people in the building with comps. The actual deal was even worse than the headline number. The previous owner had sold off the merchandise rights, video library, video games and DVD rights to Lionsgate to stay alive. What Dana and the Fertittas bought was three letters, an old wooden octagon, and roughly eight or nine fighter contracts. Years later they went back to Lionsgate and bought all of those ancillary rights back for around $2.5 to $3 million. Dana suspects the Lionsgate finance team was laughing at them on the way out the door because it looked good on the books for the next two or three years. With hindsight, those rights are worth a fortune.

    Five years of bleeding cash

    The first five years were brutal. They were doing five events a year and each one was costing roughly $2 million because they did not have the equipment, the processes, or the experience. Revenue and spend were both around $10 million a year. The Fertittas kept funding it. Dana ran it for around 10 percent equity. Then one night Lorenzo called and said he could not keep doing it and asked Dana to find a buyer. Dana came back with an estimate of $6 to $8 million. Lorenzo said he would call back. The next morning, on Dana’s drive to work, Lorenzo called and said “fuck it, let’s keep going.” Dana credits a good night of sleep for the survival of the entire empire. The biggest constraint at the time was that the UFC was not allowed on pay-per-view. Porn was on pay-per-view but the UFC was not. The goal became free television, which everyone said was impossible.

    The Ultimate Fighter as the Trojan horse

    Around 2004 and 2005 reality television was booming. Mark Burnett’s The Contender on boxing was the most expensive reality show ever made and had a fatal flaw: they edited the fights. Dana, who is the world’s most jaded fight fan, knew you never edit a fight. You let it play out. You let the fans decide if it was good or bad. They pitched the show around Hollywood. Everyone passed. The Nashville Network had just rebranded as Spike TV. Spike was not interested in paying for the show. Dana and Lorenzo said they would pay for the entire production. Spike could just put it on the air. That was the last $10 million investment they were going to make in the UFC. If The Ultimate Fighter failed, the company was done. The show was a runaway hit. The Forrest Griffin vs Stephan Bonnar finale ended with the entire arena stomping for one more round. Dana gave both fighters contracts on the spot. Spike TV executives pulled Dana and Lorenzo out into the alley behind the arena and they shook hands on a renewal on a napkin. Because they had funded production themselves, they owned 100 percent of the show. The “expensive” decision turned out to be the single best decision they ever made.

    How Joe Rogan became the voice of the UFC

    Right after the acquisition Dana flew to New York alone to go through every document and VHS tape in the old UFC offices to figure out what came back to Vegas. While he was working through tapes he had Ivory Keenan Wayans’s talk show on, and Joe Rogan came on talking about UFC and martial arts. At the time Rogan was the host of Fear Factor, a massive television show. Dana saw a guy who was educated on martial arts, not afraid to say controversial things, and ready-made for commentary. He reached out, they hit it off, and Rogan did the first 12 UFC fights for free. Dana also explains how he and Rogan promoted the company. They flew around to meet sports editors at every newspaper, most of whom were 60 to 65 years old and would never understand the sport. Radio was still huge. The problem was that fighters are terrible at radio. They are late, they sound like they are still asleep. The only two people who were good at it were Dana and Rogan. So they took turns. Dana did UFC 30. Rogan did UFC 31. Dana did 32. Rogan did 33. They lived on the West Coast and got up at 3 a.m. for years to do East Coast drive time slots. Dana later says that no amount of sponsor money would make him fire Rogan. Loyalty is the most important thing.

    Riding every technology wave: DVDs to streaming

    When DVDs exploded the UFC started producing Ultimate Knockouts and Ultimate Submissions compilations. The DVD checks were the first multi-million dollar moments. Dana would go to the local wow! superstore on Sahara and quietly move UFC DVDs to the top of the top-20 display because nobody knew who he was. He says his only real regret in the DVD era is that he did not go bigger because he assumed DVDs would last forever. When streaming was first pitched to him in his office it was buffering every five to ten seconds and he was skeptical. But he had always believed the world would consolidate back to a handful of global channels the way TV had once been channel 3, 5, 8 and 13 in his childhood. That hypothesis was right. The UFC’s television deals scaled from $35 million with Spike to $100 million with Fox to $3 billion with ESPN to $7.7 billion over seven years with Paramount, which now owns the rights to UFC and boxing. Netflix was bidding too. Dana describes both negotiations as far better than past dealings. He singles out a former Viacom executive who told him over lunch that he, the executive, had built the UFC and would just build his own if Dana did not accept the offer. Dana walked, went to Fox, and watched the executive go on to drain the life out of multiple legendary Viacom networks.

    The dictatorship: taste, control, and an alarming production truck story

    The UFC is run as a self-described dictatorship. No committee. Dana sits at the cage with a small monitor watching the broadcast not because he wants the best fight seat but because he wants to control the live in-house experience and the television feed. There is a phone next to him that goes directly to the production truck. When he sees something he does not like he calls and says do that again or never do that again. Early on the inherited Showtime production team refused to cut an interview the way he asked. Dana walked out of his seat in the middle of the broadcast, kicked open the production truck door, and told the entire crew that if they ever ignored him again he would fire every single one of them. He later fired all of them. His current production team has been with him for years with almost zero turnover. He compares it to how Mr. Beast clones himself through his editors and thumbnail designers. The art department, PR, and production all share his taste, his speed, and what he calls being “wired the way I am.”

    Going public, then doing it all again

    In 2016 the UFC sold for $4.025 billion. Lorenzo Fertitta wanted out. The deal happened with no new TV deal in place, the Fox deal ending, and every critic in the industry insisting the buyers had overpaid and the UFC had peaked. Ten years later the company has gone public through TKO Group and signed the Paramount deal. Dana says the same critics who said WME overpaid in 2016 are now saying Paramount overpaid in 2026. He calls them zeros and says he simply blocks the noise. He has now applied the same playbook to other combat sports. Power Slap, which he funded with a $1 million ask each from the Fertitta brothers after spotting Russian and Polish slap videos on Instagram, has been profitable since the first event and its reality show is at roughly 50 million YouTube views. He has launched UFC BJJ. He is rebuilding boxing inside the Paramount deal. His ten-year goal is to build the largest combat sports company that has ever existed or will ever exist.

    How he treats fighters, influencers, and his team

    Dana treats fighters as an unmanageable product. They are the most unique human beings on Earth, wired differently from everyone else, and trying to control them is impossible. He embraces it. He also gives content creators full access to UFC events: film what you want, post what you want, no rules. He says it would be absurd to tell young creators how to make content when they are the ones with the audience and the trust. He believes traditional media has almost entirely lost its influence and that nobody trusts them anymore. With his own team his moves are unusual. During COVID he offered to give up all of his own compensation rather than lay people off. Bob Iger and ESPN guaranteed the UFC would get paid no matter how many events ran, even if it was zero. Dana ran the events anyway because he assumed ESPN would eventually have to start cutting properties and he wanted the UFC to be irreplaceable. They built the only true sports bubble in the world at Yas Island in Abu Dhabi with Sheikh Tahnoun, who is himself a jiu-jitsu black belt. The numbers were enormous. He also cut off a long-running sponsor whose board kept calling to demand he take down a pro-Trump video. He told them to roll the offer into a tiny ball and shove it up the board’s ass.

    His mental model: know yourself, block noise, and never stop

    Dana’s repeated advice for entrepreneurs comes down to two things. Know who you are. Know what you want to do. Then wake up every day and chase it. When David Senra asks him what would have happened if Lorenzo had said no on that drive home, Dana shrugs. He would have figured it out the next day. There was no plan B. He never thinks about failure. He just keeps going until it works. He cuts negative people out of his life immediately. He mentions Arnold Schwarzenegger’s habit of writing positive affirmations on his walls in his early 20s and brainwashing himself into believing. He says Raising Cane’s founder Todd Graves did the same thing, and that Dana himself has affirmations on the walls of his office, gym and home. He says the body does not know the difference between a real belief and a joke about yourself, so never say anything negative about yourself or your work, even sarcastically. He blocks the noise. He listens to his team. He trusts his gut.

    Thoughts

    The most quietly valuable lesson in this entire conversation is not Dana’s grit or his TV deal numbers. It is the structure he built around ownership. The pivotal moment is not the Forrest Griffin vs Bonnar fight. It is the decision to pay $10 million to fund their own reality show production so they could own 100 percent of it. That sentence shows up halfway through the story and most people will miss it because it sounds expensive. It was actually the entire game. Spike paying for the show would have made the UFC a hit on Spike. Spike not paying for the show is what made the UFC a global empire.

    The second underrated lesson is taste as a competitive moat. Dana is constantly described in business press as a hot-headed brawler and a marketing genius, but the real skill on display is taste applied with extraordinary speed. He watches old CEOs reading canned legal statements and refuses to do that. He watches The Contender editing fights and refuses to do that. He watches boxing burn through trillions in revenue without building a brand and refuses to do that. He notices content creators are the new media before almost anyone in legacy sports does. Everything Dana refuses to do is as important as everything he chooses to do. Most founders are bad at this because they outsource taste to consultants, agencies, or research groups. Dana keeps taste in-house and runs the company as a single nervous system with a phone line that ends at the production truck.

    The third lesson is how he handles people. He runs the place as a dictatorship and yet has almost zero turnover at the senior level. The reason is obvious if you listen. He pays loyalty back with loyalty. He covered his own people during COVID. He kept Rogan when sponsors demanded otherwise. He cut a sponsor whose board called once too often. He gives content creators total freedom because he knows freedom is what creates anything good. The dictatorship is on direction and standards. The autonomy is on craft. That is exactly the configuration almost every great founder converges on and it is almost the opposite of how MBA management theory tells you to run a company.

    The fourth lesson is the cost of a single decision. The Fertittas almost sold the UFC for $6 to $8 million in roughly year four. That same business sold for $4.025 billion twelve years later and now sits inside a TKO Group entity with a $7.7 billion Paramount deal. The delta between a phone call that says “sell it” and a phone call that says “fuck it, let’s keep going” was somewhere north of four billion dollars and counting. Dana’s comment about a good night of sleep is not a cute aside. It is the most important sentence in the interview.

    The fifth and final thing worth sitting with is how Dana thinks about the next ten years. He is 56. He could have retired ten years ago. Instead he is rebuilding boxing inside the same machine, launching UFC BJJ, scaling Power Slap, and openly stating he intends to build the largest combat sports company that has ever or will ever exist. Most founders at his stage are looking for the exit ramp. Dana is loading more onto the plate because he loves the building itself more than the result. He says it explicitly: he loves entrepreneurship slightly more than he loves fighting at this point. That is the tell. People who love the work itself simply do not stop, and the numbers keep getting bigger than anyone watching can imagine.

  • Shopify CEO Tobi Lütke: AI Is the Perfect Scapegoat for Layoffs, Canada Has Trump Derangement Syndrome, and 50% of Shopify Code Is Now AI-Generated

    TLDW

    Shopify CEO Tobi Lütke sat down with Harry Stebbings on 20VC for one of the most candid and controversial conversations of his career. Lütke argues that the current wave of mass layoffs has nothing to do with AI and everything to do with pandemic-era overhiring, but AI will be blamed because it cannot fight back. He blasts Canada for its “Trump Derangement Syndrome,” calls the climate cult “one of the most evil things wrought on the population,” reveals that over 50% of Shopify’s code is now AI-generated, and says many of his best engineers have not written a line of code since December when Claude Opus changed everything. He also introduces River, an AI engineer at Shopify that named itself, and explains why he believes context engineering will be the dominant role of the next five years.

    Key Takeaways

    • AI is not causing layoffs, COVID overhiring is. Lütke is blunt: “What you see right now is not AI layoffs. Those are just the companies that are really slow that overhired just like everyone else.” AI will get blamed for everything because it is the perfect Girardian scapegoat that cannot fight back.
    • Over 50% of Shopify’s code is now AI-generated and “converting to much higher numbers.” Many of Shopify’s best engineers have not written code this year. December 2025 and the release of Claude Opus changed everything.
    • Senior engineers became more valuable, not less. Lütke initially thought new grads with no priors would dominate the AI native era. He was wrong. Senior engineers steer agents better because steering is the new programming, and reps matter more than ever.
    • Context engineering will become the dominant role within 5 years. A new product builder role is emerging that subsumes engineering, design, and product management, focused on coordinating intelligent actors (humans and AI) to ship products.
    • “River” is Shopify’s AI engineer that named itself. Built first, then asked what name it wanted. River lives in Slack, ships engineering work, and learns publicly because it is steered through public Slack channels.
    • Builders are “eights” on the Enneagram and companies actively conspire against them. Eights call out nonsense, refuse fancy dressing, and are dangerous to colleagues’ careers. They rarely get promoted, often leave, and start companies. Shopify is “remarkably high on eights” because Lütke seeks them out.
    • Canada has “Trump Derangement Syndrome.” Over 60% of Canadians believe the United States is a bigger threat than Russia or China. Lütke calls this “stunning” and wrong. Canada’s only winning strategy historically has been “winning by helping America win.”
    • Canada should be the richest country on Earth. It has every resource the world needs for the next 20 years. Lütke wants pipelines built, industry built, refining done domestically, and an end to exporting raw resources to have other countries make end products.
    • Be deeply suspicious of “non-profit.” Lütke argues opting out of the only fitness function that has ever pulled people out of poverty (markets) and refusing to disclose your actual fitness function is a red flag. Non-profits replace merit with pull.
    • The climate cult is blocking civilization. Lütke called it “one of the most evil things wrought on the population” and pointed to anti-nuclear green parties and frog protection laws blocking factories as examples of policy capture.
    • The Chinese AI threat is real but misunderstood. The bigger concern is that if Western governments restrict children from using AI, kids will simply download Chinese open-weight models, train on collectivist worldviews, and stop ever writing high school essays about Tiananmen Square.
    • Markets are the most democratic system that exists. Every dollar spent is a vote. Capital allocation by hundreds of millions of consumers is more democratic than any election.
    • Friedrich List and the Prussian school over Adam Smith. Lütke prefers a model where governments define excellent games with positive externalities, then completely get out of the way and let competition do the rest.
    • Shopify’s biggest mistake was going into physical logistics right before AI got really good. Lütke initially defended the decision based on what he knew at the time, but later admitted he was probably just wrong.
    • Lütke does not look at the stock price. It has been at least 23 days since he last checked. He runs Shopify on product instincts, not market signals.
    • Great leaders must be exothermic. A CEO is a heat source for the company. Lütke prefers “temperature” to “chaos” because chaos has too negative a connotation.
    • Don’t go to university for university’s sake. Get a degree from somewhere hard to get into so you are surrounded by people who also fought to get in. Better yet, join a small company where you can actually be of value.
    • Entrepreneurship is the most AI-safe AND most AI-benefiting job. Lütke sees a coming golden age of entrepreneurship where priors no longer matter and AI co-founders eliminate the need to grow up around business.
    • “You can just do things” is the rallying cry Lütke wants to ingrain in the world. Action causes information. The cost of trying is lower than ever.
    • The demonization of wealth in America is misdirected. No one gets to a billion dollars by stealing. Builders create products that people vote for with their money, the most democratic act in any economy.

    Detailed Summary

    Harry Stebbings opens by asking Tobi Lütke whether entrepreneurs are motivated by fear of losing or hunger to win. Lütke says he is still figuring out his own answer, but argues that both extremes lead to short-term thinking. The real unlock is taking a long perspective, because compound advantages only accrue when you are willing to wait.

    Builders Are “Eights” and Companies Conspire Against Them

    Lütke explains the Enneagram personality framework and identifies himself as an “eight,” the type that refuses to accept that any organization’s output is acceptable just because it is dressed up nicely. Eights call out nonsense, are dangerous to careers around them, rarely get promoted in professionally managed companies, and often leave to start their own businesses. Shopify deliberately overweights eights in its hiring. Lütke also says people who build companies are “fundamentally crazy people” and that the public image of leadership comes from movies, not reality. He never wanted to be CEO but realized you cannot run a product driven company without controlling the company itself, because product needs and company needs only converge on a three-year horizon.

    The Luxury of Long-Term Thinking as a Public Company

    Stebbings asks if a public company can really afford long-term thinking. Lütke says trusted public companies are the best position to be in. The chasm to cross is from trusted private to untrusted public, which is why so many founders refuse to IPO. Shopify went public 11 years ago at a 1.67 billion dollar valuation when revenues were a fraction of today’s. The valuation is now roughly 100x higher. Lütke walks through the IPO mechanics: investment bankers serve the buy side, not the company, and Lütke priced his offering above range because he knew where his growth would come from. The first trade closed about 10 dollars higher, which he calls a “good performance” but a teaching moment about market price discovery.

    AI Is the Perfect Scapegoat for Mass Layoffs

    This is where the conversation gets explosive. Lütke says Shopify employs about 7,500 to 8,000 people today and his real hope is to have the same number in five years, but at 100x productivity. He argues that the layoffs sweeping the tech industry have nothing to do with AI. They are the result of pandemic-era overhiring catching up to slow-moving companies. But AI will get blamed for everything because it is the perfect Girardian scapegoat. It cannot defend itself, it has no PR team, and an entire industry of doomers is already trained to point at it. Lütke says his own industry has been “gaslighting everyone into AI fear” and science fiction did the same for 60 years before that.

    His own use of AI is what he calls utopian. Tasks that used to be hard are easy. Most jobs, he argues, are not actually good jobs to begin with. Being a human task queue is not a great job. Great jobs involve agency and creation. As AI gets cheaper, purchasing power explodes, and people will get options to do things on weekends that are vastly more productive than their day jobs ever were.

    Markets Are the Most Democratic Mechanism Ever Invented

    Lütke pivots into a long defense of capitalism as the most democratic system in existence. Every dollar spent is a vote, far more frequent and more granular than any election. He uses Elon Musk and Tesla as examples. Lütke owns a Model Y, did not touch the steering wheel that morning, and uses Starlink in the back to work on long drives. He posts on X and gets replies from Japan in real time. He calls Musk a “one man engine” who has captured a tiny percentage of the value he created. He extends this to Shopify itself: Lütke owns 6% of the company, which means 94% is owned by other people who all made money. Plus roughly 10 million people work in the broader Shopify ecosystem on customer fulfillment, web design, customer service, and more.

    Why “Non-Profit” Should Make You Suspicious

    Lütke targets the charity industrial complex. He argues that non-profits opt out of the only mechanism humanity has ever invented to lift people out of poverty (markets), and they fail to articulate what their actual fitness function is. The result is that “merit of organization is replaced with pull of individuals.” Smooth talkers, not builders, end up running these institutions. He acknowledges Carnegie’s libraries and a few exceptions but believes the ratio of charity dollars to good outcomes is dramatically off. He is far more enthusiastic about funders like MacKenzie Scott who give in unrestricted ways, and even more enthusiastic about Jensen Huang and Bloom Energy as compute and infrastructure investments that compound into civilizational gains.

    The Prussian School of Economics

    Asked about government intervention, Lütke pledges allegiance to Friedrich List and the Prussian school of political economy over Adam Smith and Lassalle. The job of government is to define excellent games where positive externalities accrue to society, then completely get out of the way. He calls the outsourcing of violence to governments “one of the most inspiring things humanity has ever done” because it created the conditions for personal property. But governments are extremely bad at doing things directly. The moment a government runs grocery stores, it costs 10x more, and entrepreneurs have to be enlisted to repair the damage.

    Canada’s Trump Derangement Syndrome

    Stebbings asks if Lütke is proud of Canadian Prime Minister Mark Carney for standing up to Trump. Lütke is unequivocal: no. He calls Carney’s stance “not a credible witness to the reality on the ground.” Canadians, he argues, are “massively overfit to niceness,” which leads to “unkind lies” and lying by omission. Over 60% of Canadians now believe the United States is a bigger threat than Russia or China, which Lütke calls “stunning” and clearly wrong. Canada is a small economy attached to a hegemon, and the only winning strategy in its history has been winning by helping America win.

    That said, he agrees with Carney on diversifying the economy, getting closer to Europe, and engaging Asia. But he wants Canada to also “build the [expletive] out of pipelines, build the [expletive] out of our industry, and start refining the stuff ourselves.” Canada has every resource the world needs for the next 20 years and the most educated workforce on Earth. The only obstacle is political will. Canada’s commercial story has been the same since the beaver pelt era: extract resources, ship them abroad, let other countries make end products. Canada Goose, Lululemon, Shopify, Miller Lite. That is the short list of products Canada actually makes.

    The Real Chinese Threat

    Lütke says the Chinese AI threat is both underestimated and overestimated. The bigger threat, he argues, is government overreach. If Western governments start dictating which AI models children can use, kids will simply download Chinese open-weight models. He notes that Chinese models, especially when prompted in Chinese, exhibit a clearly collectivist worldview. The risk is that an entire generation of students writes essays through models trained never to mention Tiananmen Square. He frames the broader political battle as collectivism versus individualism and says everything else is smoke screening.

    Fixing Europe and the Climate Cult

    Asked what he would do as president of Europe, Lütke begins by saying you have to “get rid of the climate cult.” He calls it “one of the most evil things wrought on the population,” citing green parties whose founding myth is that nuclear power is bad, and infrastructure projects blocked because of one frog breeding in one creek. He argues that very few people have the capability to truly build, and they need both enablement and accountability from the village. Beyond that, he wants Europe to follow the Prussian playbook: build excellent games, build infrastructure, and use the resulting wealth to sculpt the economy you want.

    Shopify’s Biggest Mistake

    Lütke says his biggest public mistake was Shopify’s full push into physical logistics and warehousing right before AI capabilities exploded. Initially he defended the decision as correct based on the information available at the time, but later admitted he probably just got it wrong. The hardest part was that real people lost their jobs when Shopify exited.

    Great Leaders Are a Heat Source

    Lütke previously talked about CEOs injecting “chaos” into organizations. He now prefers “temperature.” Heat is atoms jiggling. Great leaders must be exothermic, providing energy that flows through the organization. He says he hasn’t checked Shopify’s stock price in at least 23 days. Most public company CEOs are obsessed with their stock. Lütke runs on product instincts.

    Senior Engineers Don’t Write Code Anymore

    Lütke admits he was wrong about new grads having an AI native advantage. Some are exceptional (he hired a 13-year-old intern from Waterloo whose mother accompanies him to classes), but on the whole, senior engineers steer agents better than juniors do because they have done more reps. Programming is not gone. Programming has become higher level. Engineers massively underestimate how important steering is. Steering is just programming at a higher altitude.

    The Role That Will Dominate in 5 Years

    Lütke says context engineering, a term he had a hand in popularizing, will become a standard role within five years. It will likely subsume parts of product, design, and engineering management. The best AI programmers right now, surprisingly, are people from engineering management because they have been prompting intelligent agents (humans) for years. Good communicators are good thinkers because communication is distillation.

    River, the AI Engineer That Named Itself

    Shopify built an AI engineer that lives in Slack. They built it first, then asked it what name it wanted. The AI chose “River” because Shopify’s monolithic repository is called “world” and rivers shape worlds. River does an enormous amount of Shopify’s engineering, taking instructions through public Slack channels so that the entire company can learn from how others steer it.

    Over 50% of Shopify’s Code Is AI-Generated

    The number is “a fair deal over 50%” and “converting to much higher.” Many of Shopify’s best engineers have not written code this year, with the inflection point being December 2025 and the release of Claude Opus. Lütke himself still writes code occasionally, especially the data structure layer where he applies what he calls a “German school” of engineering: figure out how data persists on disk, then build everything else on top. Once that is right, the rest can be vibe coded by AI.

    Should His Kids Go to University?

    Lütke says he would not push his kids to attend university for its own sake. The value of a hard to enter program is being surrounded by people who also fought to get in. Better still: get into the room with people who are obsessed with the topic you care about. He thinks joining a small startup where you can actually be of value is often a superior path. He addresses nepotism directly. His instinct is that nepotism is bad. The gold standard is double-blind merit. But double-blind merit barely exists anywhere, and intersectional academic hiring criteria in Canada are arguably worse than nepotism.

    Final Reflections

    Lütke ends with what he calls the best advice he knows: “You can just do things.” The system exists to push everyone toward acceptable outcomes, but if you know what a good outcome looks like, you can step out of the system and try. Action causes information. The cost is lower than ever. The only constraint is that the experiment cannot have victims.

    He also addresses the demonization of wealth. No one gets to a billion dollars by stealing. Builders create products people vote for, the most democratic act there is. Buying from a local shop is voting for the welfare and future of local shops. Constructive criticism is itself something someone has to build, and Lütke welcomes it. Lazy criticism, hot takes, and bad faith arguments are corrosive and should be held in contempt.

    He is bullish on AI as a counterweight to information warfare. A council of AI models trained in different countries (Chinese, German, French, American) could fact check claims with multiple perspectives. The “@grok is this true” reflex on X is, he says, a primordial version of this. The information asymmetry that has favored bad faith actors for decades is about to flip.

    Thoughts

    This interview is a window into the operating philosophy of one of the most successful technical founders alive, and it is far more provocative than most of his public appearances. The headline claim, that AI is a scapegoat for layoffs caused by pandemic overhiring, deserves to be repeated until it sinks in. Every CEO who lays people off and then writes a memo about “AI driven efficiency” is taking advantage of a narrative that AI cannot push back against. The math is plain: if you doubled your headcount in 2021 and 2022 and now you are firing 15%, you are not net displaced by AI. You are correcting a hiring mistake.

    The 50% AI generated code statistic is the bigger story. Shopify is not a small company. 8,000 employees and 7 billion in revenue is enterprise scale. If a company that mature has crossed the 50% threshold and is “converting to much higher numbers,” the implication for the broader software industry is enormous. The senior engineer compounding observation is also subtle and important. If steering is the new programming, then the senior pool is more valuable, not less, and the pipeline problem for junior developers gets harder to solve. Companies that under invested in junior training during ZIRP will face an experience cliff in five years.

    Lütke’s Canadian commentary will offend many readers in his home country, which seems to be exactly the point. The “lying by omission” critique of Canadian niceness is sharp and accurate. The 60%+ of Canadians who view the US as their largest threat is genuinely a remarkable statistic, and it has implications for trade policy, capital flows, and immigration. Whether or not you agree with his political read, his prescription is unambiguous and pro-growth: build pipelines, refine resources domestically, stop being content as a feedstock economy.

    The non-profit critique deserves more public debate. The fitness function point, that markets reveal preferences and non-profits opt out of preference revelation while not disclosing what they optimize for, is a sharp economic argument. The pull versus merit observation about who ends up running large foundations rings true to anyone who has worked adjacent to the philanthropic sector.

    The introduction of River as an AI engineer that named itself is a small detail that signals where this is going. AI agents are going from tools to teammates with identities, channels, and reputations. The fact that River shapes the “world” repository is poetic, and the public Slack steering pattern is a real innovation in how organizations can scale agentic AI without creating siloed knowledge.

    Lütke’s “you can just do things” rallying cry is ultimately what ties the entire interview together. Whether he is talking about Canada, Europe, AI engineers, or his own kids, the through line is the same: action causes information, the cost of trying is lower than ever, and the only people who will benefit from the next decade are the ones who refuse to wait for permission. This is the most useful piece of philosophy in the entire conversation, and it applies far beyond entrepreneurship.