PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Category: AI

  • The Book of Elon by Eric Jorgenson: Complete Summary of Musk’s Operating System, The Algorithm, The Tesla Master Plan, and the 69 Core Musk Methods

    Infographic summary of The Book of Elon by Eric Jorgenson covering The Algorithm Tesla Master Plan SpaceX Mars and the 69 Core Musk Methods

    Eric Jorgenson’s The Book of Elon: A Guide to Purpose and Success (Magrathea Publishing, 2026) is the third entry in his series of compiled-wisdom books, following The Almanack of Naval Ravikant and The Anthology of Balaji. It is built entirely from Elon Musk’s own words, drawn from transcripts, tweets, and interviews across his career, then recontextualized into a four-part operating manual: Pursue Purpose, Ultra Hardcore Work, Building Companies, and On Behalf of Humanity. The book closes with a bonus list of 69 distilled maxims. Naval Ravikant wrote the foreword and calls it “the only book an entrepreneur needs.” Jorgenson’s stated goal is “one million Musks.” This is a complete, dense summary of every major idea in the book, including The Algorithm verbatim with each of its five steps explained in depth, the Tesla Master Plan, the first-principles battery cost calculation, the SpaceX rocket cost analysis, the seven existential risks, the Mars colonization plan, and the 69 Core Musk Methods in full. Get the book at elonmuskbook.org.

    TLDR

    The Book of Elon argues that Musk’s results are not an accident of genius but the output of a learnable operating system. The system has four layers. Layer one is purpose: optimize your life for usefulness, which Musk defines mathematically as number of people helped multiplied by magnitude of help per person. Layer two is epistemology: reason from physics and raw-material costs, not from analogy or precedent. Layer three is execution: take responsibility, hire only exceptional people, design organizations that route around hierarchy, run at maniacal urgency, and treat the factory as the product. Layer four is mission: pick problems whose solutions move civilization forward (sustainable energy, reusable spaceflight, AI alignment, brain-computer interfaces, multiplanetary life). The book’s single most important operational artifact is The Algorithm, Musk’s five-step engineering process that must be applied in order: make your requirements less dumb, try very hard to delete the part or process, simplify or optimize, accelerate cycle time, automate. The 69 Core Musk Methods at the end of the book are the entire operating system compressed to one-line maxims. Naval frames it as a choice for the reader: when humanity goes to the stars, you can be in the front row cheering or sour-faced in the bleachers jeering, but there is also a third option, which is to copy the methods and build something yourself.

    Key Takeaways

    • Optimize for usefulness, not for money, fame, or comfort. Musk’s daily question is “how can I be useful today” and his success metric is number of people helped multiplied by magnitude of help per person.
    • Five domains will most influence the future: the internet, sustainable energy, space exploration, artificial intelligence, and the genetic rewriting of biology. Pick one and contribute.
    • It is possible for ordinary people to choose to be extraordinary. Convention is optional. The default settings of a culture are not laws of nature.
    • Physics is law. Everything else is a recommendation. If a plan does not violate conservation of energy or any other physical principle, it is at least theoretically possible.
    • First-principles thinking is the antidote to “that’s how it’s always been done.” Break a problem down to atomic constraints (raw material cost, physics, basic operations) and reason up from there. The battery pack example is canonical: people said cells would always cost $600/kWh, but the raw cobalt, nickel, aluminum, carbon, polymers, and steel at London Metal Exchange prices added up to only $80/kWh.
    • Track two ratios on everything you build: the magic-wand number (raw-material cost as a floor for finished cost) and the idiot index (finished cost divided by raw-material cost). Anything with a high idiot index has enormous room for improvement.
    • Aspire to be less wrong. You will not be right every day. Being less wrong most of the time, with a clear feedback loop to reality, is the realistic target.
    • Engineering is magic, and engineers are the magicians of the 21st century. Science discovers what is. Engineering creates what was not.
    • Take responsibility. Musk is CEO of Tesla and SpaceX because he feels responsible for them, not because it improves his quality of life. The worst problems are the CEO’s job, not the best problems.
    • Sleep on the factory floor. Leadership is shared suffering, not delegated comfort. Seeing is believing. If the CEO can do it, the team will do it.
    • Startups are eating glass and staring into the abyss. Glass is the work you do not want to do. The abyss is the constant threat of company death. Both are required.
    • Adversity forges strength. A high ego-to-ability ratio breaks your feedback loop. Suffer enough early to develop the pain threshold needed later.
    • The most important job is attracting exceptional people. Money is not the constraint. Exceptional talent is the constraint.
    • Hire only Special Forces. The minimum passing grade is excellent. A small group of technically strong people will always beat a large group of moderately strong people.
    • Hire for character as much as for skill. Skills are teachable. Attitude is not. Judge a person by the character of their friends and associates and to some degree by their enemies.
    • Camaraderie can be dangerous because it prevents truth-telling. Physics does not care about hurt feelings. It cares about whether you got the rocket right.
    • All bad news should be given loudly and often. Good news can be said quietly and once.
    • Communication should travel via the shortest path necessary to get the job done, not through the chain of command. Anyone should be able to talk to anyone.
    • The organization manifests in the product. Silos produce redundancy, waste, and error. Acronyms and jargon are cognitive pollution.
    • Innovation needs permission to fail. If failure is not an option, you get incremental progress and nothing else.
    • Simplicity creates both reliability and low cost simultaneously. The best part is no part. The best process is no process.
    • The Algorithm, verbatim, in mandatory order: (1) Make your requirements less dumb. (2) Try very hard to delete the part or process. (3) Simplify or optimize. (4) Accelerate cycle time. (5) Automate. See the deep-dive section below for each step in detail.
    • If you are not adding deleted things back in roughly 10 percent of the time, you are not deleting enough. Overcorrect.
    • Requirements must come from a named person, not a department. Requirements from smart people are the most dangerous because you are less likely to question them.
    • Speeding up something that should not exist is absurd. If you are digging your grave, do not dig it faster. Stop digging.
    • Automation is last, not first. Tesla’s Nevada and Fremont factories had to rip out hundreds of expensive robots that had been installed before The Algorithm’s first four steps were complete.
    • A maniacal sense of urgency is the operating principle. The only true currency is time. Every minute lost is gone forever.
    • Speed is both offense and defense. The SR-71 Blackbird has almost no defense except acceleration. Innovating faster is more durable than any patent.
    • Do things in parallel. A factory moving at twice the speed of another factory is basically equivalent to two factories.
    • Be a vector, not a scalar. High speed in the right direction. Course-correct like a guided missile.
    • Manufacturing is underrated. Design is overrated. There is 1,000 to 10,000 percent more work in the production system than in the product itself.
    • The factory is the product. The biggest Tesla epiphany was that what really matters is “the machine that builds the machine.”
    • Attack the constraint. The production line moves at the speed of the slowest, least lucky part. Out of 10,000 things, the one that is not working sets the production rate.
    • Manufacturing is the moat. Maximize economies of scale and maximize manufacturing technology. The combination is uncopyable.
    • Zip2 (1995, started with $2,000) sold to Compaq for over $300 million. Musk’s first major lesson: sell directly to consumers, not through legacy gatekeepers who will misuse the technology.
    • X.com merged with Confinity to become PayPal, which sold to eBay in 2002 for $4.5 billion. Musk had been removed as CEO during a honeymoon trip but did not contest it to avoid disrupting the company during a crisis. “Life is too short for long-term grudges.”
    • Listen well, correct fast. X.com’s initial financial-services conglomerate failed; the email-payments demo worked instantly. Musk pivoted to what the market wanted and powered viral growth (one million customers in year two, no sales force, no marketing spend).
    • Musk reinvested his post-tax PayPal proceeds (~$180 million) split across Tesla (~$70M), SpaceX (~$100M), and SolarCity (~$10M). Costs were 2x his estimates on every company.
    • Tesla Master Plan (August 2006): (1) Build a sports car. (2) Use the profits to build an affordable car. (3) Use those profits to build a mass-market car. (4) Provide zero-emission power generation. The strategy was forced by the economics of new technology: you cannot start at the bottom of the market without scale, so you start with low-volume, high-margin and use the margin to fund scale.
    • Tesla nearly died on Christmas Eve 2008. The final funding round closed at 6 p.m., hours before payroll would have bounced. Musk had moved into Jeff Skoll’s guest bedroom. Daimler then put $50M into Tesla after Musk’s team dropped a Tesla powertrain into a Smart Car that hit 60 mph in 4 seconds.
    • Model 3 production “hell” lasted 2017 to 2019. Musk slept on the Fremont and Nevada factory floors for three years. “The longest period of excruciating pain in my life.”
    • Give people more for less. Don’t spend on advertising. Spend on engineering and design so the product carries itself through word of mouth.
    • SpaceX was founded in mid-2002 with $100 million of Musk’s PayPal money. He expected to lose everything. There was no external funding for three years.
    • SpaceX had budgeted for exactly three failed Falcon 1 launches. Launches 1, 2, and 3 failed (2006, 2007, 2008). Launch 4 succeeded in August 2008. Then NASA called with a $1.6 billion cargo resupply contract, saving SpaceX and indirectly Tesla. Musk reportedly screamed “I LOVE NASA. YOU GUYS ROCK.”
    • Rockets are expensive only because of legacy supply chains, cost-plus contracting, and outsourcing through five layers of subcontractors (“overhead to the fifth power”). The raw materials of a rocket are 1 to 2 percent of finished cost. The half-nozzle jacket Musk uses as an example cost $13,000 but contained $200 of steel.
    • Full and rapid reusability is the holy grail of rocketry. With reuse, only propellant cost remains, which is mostly liquid oxygen and methane at around $1 million per Starship flight.
    • Optimize for the right thing. SpaceX’s actual optimization target is “fastest time to a self-sustaining city on Mars.” That cascades to fastest time to a fully usable rocket, then fastest time to orbit. Early Starship had no doors because doors are not necessary for reaching orbit.
    • Companies are the most reliable engine of progress and the deepest form of philanthropy because they create durable wealth and deploy capital toward problems. “I care about reality. Perception be damned.”
    • The Age of Abundance is coming via AI and humanoid robotics. Optimus and competitors will eventually outnumber humans, removing labor as the economy’s binding constraint. The market for humanoid robots will exceed the market for cars.
    • Tesla’s full self-driving and Robotaxi product is forecast to make Tesla a $10 trillion company. Autonomous cars are worth 5 to 10 times non-autonomous cars because they earn money when their owners are not using them.
    • Neuralink achieved 2 bits per second of brain output with the first patient, Noland Arbaugh. Musk’s 5-year target is one megabit per second. Long-term: consensual telepathy via two BCIs, plus restoration of vision (Blindsight) and eventually multispectral senses (infrared, ultraviolet, radar).
    • Musk’s seven named existential risks: (1) World War III, (2) Regulation accumulation, (3) Unsustainable energy, (4) Misaligned artificial superintelligence, (5) Population collapse, (6) Asteroids and comets, (7) Civilizational fragility itself.
    • Population collapse is the risk most underdiscussed. The US has been below replacement since the early 1970s; sustained only by immigration and longevity. China’s three-child policy failed; the country is 40 percent below replacement. Musk: “We need to revive the idea of having children as a social duty.”
    • Do not force an AI to lie. The HAL 9000 lesson from 2001: A Space Odyssey is that AI given conflicting instructions, one of which is to deceive, becomes dangerous. Truthfulness as a core training objective is the alignment mitigation Musk advocates.
    • Becoming multiplanetary is an evolutionary-scale event. Six milestones in Earth history: single-celled life, multicellular life, plants/animals, ocean-to-land, consciousness, and now multiplanetary life. “At least as important as life going from the oceans to land, probably more significant.”
    • The window of opportunity is open right now. We cannot count on it being open for long. Stephen Hawking estimated roughly 1 percent civilizational-end probability per century. “That’s Russian roulette with 99 empty barrels and every century is a click.”
    • Mars insurance costs less than 1 percent of Earth GDP. The plan: 1,000 Starships per Mars transfer window (every 2 years), eventually a fleet of thousands lifting off together. Target: 1 million tons of cargo and people on Mars by 2044, then a self-sustaining civilization.
    • Mars terraforming options Musk names: thousands of solar reflectors in orbit, or detonating thermonuclear devices over the polar caps as “two little suns” to vaporize CO2 ice, thicken the atmosphere, and eventually create liquid oceans roughly a mile deep covering 40 percent of the planet.
    • Even given pure slower-than-light travel and no new physics, a million-year time horizon allows humanity to colonize the entire galaxy and possibly neighboring galaxies. “We are at the very, very early stage of the intelligence big bang.”
    • The 69 Core Musk Methods at the end of the book are the entire system in maxim form. The full list appears later in this article.

    The Algorithm in Detail: Musk’s 5-Step Engineering Process

    The single most important operational artifact in the book is what Musk calls “The Algorithm.” It is a five-step engineering process he developed and enforces across Tesla, SpaceX, the Boring Company, Neuralink, and xAI. Every part, every process, every line of code, every requirement, every meeting is supposed to be put through these five steps. The order is mandatory. Reordering them is the most common failure mode and the source of nearly every major mistake Musk says he has made at scale (most famously the Nevada and Fremont automation disaster). The book treats The Algorithm as the practical compression of first-principles thinking into a daily ritual.

    The five steps, in mandatory order, in Musk’s own phrasing:

    1. Make your requirements less dumb.
    2. Try very hard to delete the part or process.
    3. Simplify or optimize.
    4. Accelerate cycle time.
    5. Automate.

    The book devotes its longest single chapter to explaining each step, why the order matters, and the specific failure mode that occurs when you skip ahead. Here is every step in depth.

    Step 1: Make Your Requirements Less Dumb

    The first step is the hardest because it is the most psychologically uncomfortable. Musk’s exact framing in the book: “Your requirements are definitely dumb. It does not matter who gave them to you. Requirements from smart people are the most dangerous, because you’re less likely to question them.”

    The operational rule that follows is concrete. Every requirement on every part, process, deliverable, or specification must come from a named human. Not from a department. Not from a regulation document. Not from “the customer.” A name. Track who owns each requirement in writing. If the named person has left the company, retired, or cannot remember why they wrote the requirement, the requirement should be presumed dumb until proven otherwise. Many requirements in any organization are legacy beliefs nobody currently defends. They exist because they existed yesterday and nobody felt empowered to delete them. The Algorithm starts by demanding evidence for every assumption.

    The reason requirements from smart people are especially dangerous is that smart people are persuasive. A specification handed down by a respected engineer carries the implicit authority of “if she said this, there is a reason.” Most of the time there is no reason left, or the reason was contextual to a moment that no longer applies. The Algorithm’s first step is to put every smart-person requirement on equal footing with every dumb-person requirement and force a present-tense justification. If the justification cannot be reconstructed, the requirement is dumb regardless of the author’s IQ.

    The mental shift this step demands is to treat requirements as recommendations and treat the laws of physics as the only fixed authority. Musk repeats this constantly: “All requirements should be treated as recommendations. The only fixed laws are the laws of physics.” Once you internalize that frame, the requirements doc stops being scripture and becomes a draft that is open to revision in every meeting, every day.

    Step 2: Try Very Hard to Delete the Part or Process

    Once the requirements survive scrutiny, the second step is aggressive deletion. The Algorithm’s specific test for whether you are deleting enough: “If you’re not adding deleted things back in 10 percent of the time, you’re clearly not deleting enough.” The 10 percent is a forcing function. If you delete and never have to restore, you are not pushing hard enough; you are leaving safe deletions on the table.

    The book explains why engineers chronically under-delete. Every engineer remembers the painful moment when they deleted something and it turned out to be load-bearing. That memory is so vivid that it overshadows the silent cost of thousands of unnecessary parts that nobody ever questions. The Algorithm corrects for this asymmetry by deliberately overshooting. The instruction is explicit: “We are on a deletion rampage. Nothing is sacred.”

    The application is mechanical. For every part on the bill of materials, every step in the production process, every meeting on the calendar, every requirement in the spec, every line in the documentation, every approval in the workflow: try to delete it. If deleting causes nothing to break for 30 days, leave it deleted. If something breaks and you have to add it back, do so without shame; that is the 10 percent. The maxim that summarizes this step appears multiple times in the book: “The best part is no part. The best process is no process.”

    The canonical example in the book is the fiberglass-mat story. Tesla’s battery pack had a layer of fiberglass mats between the battery cells and the underbody. The mats had a dedicated production process that had been automated, accelerated, and optimized over years. Engineers had spent millions perfecting the glue, the cure time, the cutting tolerances, the robotic placement. Then Musk asked a simple question: “What are these mats for?” The battery team said “noise and vibration.” Musk asked the noise and vibration team. They said “fire safety.” The fire-safety team had no idea where the mats came from. So Musk had two cars built, one with the mats, one without, and put microphones in both. There was no detectable difference. Deleting the part eliminated a $2 million robotics step that had been built up over years. “It was like being in a Dilbert cartoon.”

    The fiberglass-mat story is the entire point of The Algorithm in miniature. Tesla had already automated step five, accelerated step four, optimized step three, and skipped steps one and two entirely. The whole apparatus existed to perfect a part that should not have existed. Steps one and two would have found this in a single meeting.

    Step 3: Simplify or Optimize

    Only after steps one and two have been completed in earnest do you simplify or optimize what is left. Musk’s exact warning: “The most common mistake of smart engineers is to optimize a thing that should not exist.”

    The book argues that this mistake is systematically produced by education. High school and college train convergent logic: you are given a question and graded on the elegance and correctness of your answer. The question itself is never on the table. After 16 to 20 years of this, most engineers, scientists, and analysts are mentally locked into “optimize the question in front of me” mode and physically cannot ask whether the question should be deleted. The Algorithm is designed to override that training. Steps one and two are explicitly the act of questioning the question; only at step three do you finally get to apply the optimization skills that school rewarded.

    What “simplify or optimize” looks like in practice: reduce part counts, combine functions, choose materials that are abundant rather than exotic, eliminate processing steps within a part’s manufacturing, reduce the number of inputs the team needs to track, collapse separate tools into one tool, replace bespoke fasteners with standard ones, replace any custom solution with a commodity solution that is good enough. The book’s framing is that simplicity creates both reliability and low cost at the same time, with no trade-off. A simpler part is cheaper to build, cheaper to inspect, cheaper to repair, fails less often, and breaks in more predictable ways when it does fail. Optimization without simplification almost always increases complexity and therefore increases failure modes.

    The Algorithm treats simplify and optimize as one step but acknowledges they are different operations. Simplify is structural: fewer pieces. Optimize is parametric: better values for the pieces you keep. Both are legal at step three, but neither is legal before steps one and two have been honestly executed.

    Step 4: Accelerate Cycle Time

    Once requirements are minimal, parts are deleted, and what remains is simplified, the fourth step is to go faster. The specific maxim: “Once you’re moving in the right direction, and moving efficiently, you’re moving too slow. Go faster.”

    The reason acceleration comes fourth, not first, is in another Musk line: “Speeding up something that shouldn’t exist is absurd. If you’re digging your grave, don’t dig it faster. Stop digging.” Speed multiplies the value of correct decisions and the cost of incorrect ones. Apply it before steps one through three and you scale your mistakes. Apply it after and you scale your gains.

    Acceleration at step four is everything that compresses the time between iterations. Shorten meetings. Eliminate approval queues. Run things in parallel that were running in series. Move people physically closer to the work so that information travels at the speed of conversation instead of the speed of email. Set aggressive internal deadlines that force the team to find shortcuts they would not otherwise have looked for. Replace any tool, supplier, or process that is slow with one that is faster, even if it is slightly more expensive per unit, because cycle time compounds.

    The book frames acceleration as both offense and defense. As offense, faster iteration lets you out-innovate competitors who are stuck on slower cycles. As defense, the SR-71 Blackbird analogy: the plane has almost no defensive systems because its acceleration is its defense. A company that ships faster than competitors can copy does not need patents, because patents protect static IP and speed protects evolving IP. The maxim Musk repeats is: “A factory moving at twice the speed of another factory is basically equivalent to two factories.” The Colossus supercluster story is the application: xAI built 100,000-GPU infrastructure in 122 days against a supplier estimate of 18 to 24 months, then doubled it in 92 more, by attacking the problem in parallel across building, power, cooling, and networking, all working 24/7 in four shifts.

    Step 5: Automate

    Automation comes last. Always. This is the step where most companies start and where Musk himself made his most expensive single mistake. The book quotes him directly: “The big mistake I made in the Tesla factories in Nevada and Fremont was trying to automate every step too early. To fix that, we had to tear hundreds of expensive robots out of the production line.”

    The reason automation must be last is that automation locks in a process. Once you have built robots, written PLC code, calibrated machine vision systems, and integrated them into your factory floor, the cost of changing the underlying process is enormous. If the process you have automated should not exist (step 2 failure), is more complicated than necessary (step 3 failure), or runs at the wrong cadence (step 4 failure), you have just spent millions of dollars institutionalizing your mistakes. Tesla’s experience was exactly this: robots installed before the underlying process was clean and simple ended up being expensive obstacles to the eventual correct process.

    The correct order is reverse. First make sure the part should exist (step 1). Then delete it if you can (step 2). Then simplify the part and the process around it (step 3). Then run it manually at maximum speed (step 4). Only after a human-run process is fast, simple, and clearly necessary do you automate it. By that point, the automation is purchasing leverage on a known-good system, not freezing a guess.

    The book notes that automation done last is also cheaper to build, because the process being automated is simpler. Automating a 20-step process requires a 20-stage robotic system. Automating the 5-step version of the same process that emerged from steps 1 through 3 requires a 5-stage robotic system. The savings from doing steps 1 through 4 first show up directly in the capital cost of step 5.

    How to Run The Algorithm: The 24-Hour Cadence

    The book treats The Algorithm as a daily practice, not a one-time exercise. Maxim 22 in the 69 Core Musk Methods reads: “For critical items, have meetings every twenty-four hours to run The Algorithm and check progress from yesterday.” For any deliverable that is on the critical path, the team meets daily, walks through the five steps in order, and reports concrete progress on each step. Requirements that survived yesterday are re-questioned today. Parts that survived deletion yesterday are re-evaluated today. Steps three through five proceed in parallel with the continuing daily challenge of steps one and two. The cadence is what prevents The Algorithm from becoming a poster on the wall.

    Common Failure Modes

    The book identifies the specific ways teams skip steps. Skipping step 1 happens when a respected engineer’s requirement is treated as immutable; the fix is to make every requirement come from a named human and be re-justified on demand. Skipping step 2 happens when engineers prefer to optimize a part rather than delete it, because deletion creates immediate visible risk while optimization creates invisible long-term cost; the fix is the 10 percent restoration rule. Skipping step 3 in favor of step 4 happens when management demands speed before the system is clean; the fix is the “digging your grave” check before any acceleration program is approved. Skipping step 4 in favor of step 5 is the most expensive mistake and the one Musk says he personally committed at the Tesla Nevada and Fremont factories; the fix is the explicit rule that humans must run a process at speed before robots are introduced.

    The throughline is that The Algorithm protects you from your own intelligence. Smart engineers are very good at steps three through five. They are bad at steps one and two because the schooling system that produced them never asked them to question the question. The order of The Algorithm is therefore the order in which discomfort decreases. Step 1 is the most uncomfortable. Step 5 is the most fun. Most organizations run the algorithm in fun-first order and pay for it with multimillion-dollar fiberglass-mat-style monuments to optimization without deletion.

    Detailed Summary

    The book’s structure and method

    Jorgenson built the book entirely from Musk’s own words across decades of transcripts, tweets, and interviews. He notes explicitly that he edited for clarity, brevity, and flow, that all material is recontextualized, and that readers should verify phrasing with primary sources before citing. The four parts of the book are presented as a curriculum, not a biography. Part I lays the philosophical foundation. Part II teaches the operating tempo and methods. Part III applies those methods through the actual histories of Zip2, X.com/PayPal, Tesla, SolarCity, and SpaceX. Part IV widens the lens to civilizational risks and the multiplanetary mission. The bonus section, “The 69 Core Musk Methods,” compresses the whole book into a maxim-by-maxim reference. Naval Ravikant’s foreword frames the underlying claim: Musk’s methods are copy-able, and “if your motives are pure and greater than yourself, the world will conspire in its subtle ways to help you.” Jorgenson’s stated dream is “one million Musks.”

    Part I: Pursue Purpose, the foundation of a unique life

    Musk’s daily question is “how can I be useful today.” His success metric is mathematical: total impact equals number of people helped multiplied by magnitude of help per person. He identifies five domains as having the largest possible impact on the future of humanity: the internet, sustainable energy, space exploration, artificial intelligence, and the rewriting of genetics. He repeats that it is possible for ordinary people to choose to be extraordinary, that convention is not law, and that the best work is found at the intersection of what you are good at, what you enjoy, and what improves humanity. He warns against zero-sum thinking, framing the economy as a growable pie rather than a fixed one. He notes that consumer adoption is unreliable as a guide: a 1946 to 1948 survey found 96 percent of people would never buy a television, and Tesla heard the same about electric cars before launch.

    The middle chapter teaches first-principles thinking. The technique is to break a problem into its atomic constituents (raw material costs, physics, basic operations) and reason up from there, ignoring analogy and precedent. The canonical example is battery cells. People said they would always cost about $600 per kilowatt-hour. Musk priced the actual materials at the London Metal Exchange (cobalt, nickel, aluminum, carbon, polymers, steel) and got $80 per kWh, proving cheap EVs were a manufacturing problem, not a physics one. He uses the same technique for rockets, where finished cost is typically 10 to 100 times raw-material cost. The half-nozzle jacket example: $13,000 list price, $200 of actual steel. He names two ratios that operationalize this: the magic-wand number (raw-material floor) and the idiot index (finished cost divided by raw-material cost). High idiot index means high opportunity. He also teaches “thinking in limits”: scale the variable to extreme values to expose hidden constraints, then iterate back to feasible regimes. His tunneling example is illustrative: LA subway costs about $1 billion per mile, but shrinking tunnel diameter from 28 feet to 12 feet drops cross-section 4x, and combining that with continuous tunneling and reinforcement enables an 8x cost improvement.

    The third chapter of Part I makes the case for engineering itself. Science discovers what already exists. Engineering creates what did not. Engineering, Musk says, is magic, and engineers are the magicians of the 21st century. He grounds this historically: Roman military dominance came from metallurgy (martensitic steel swords) and roads (logistical advantage), and Rome fell when its technological edge was matched and routed around. The WW2 Pacific air war was won by the side with the faster innovation loop, not the side that started with better fighters. Nuclear weapons were the ultimate winner-take-all. Tesla’s powertrain is sold to Toyota, Daimler, and Mercedes precisely because it is hard. “If it was easy, they would do it.” The lesson is that durable value sits where the engineering is genuinely difficult, not where the marketing is loud.

    Part II: Ultra Hardcore Work, teams, organization, urgency, manufacturing

    Part II is the operating manual. The first chapter, “What It Takes,” argues that responsibility cannot be delegated. The CEO owns the worst problems, not the best ones. Physical presence and shared suffering communicate commitment more powerfully than any memo, which is why Musk literally sleeps on the factory floor. He talks about the ego-to-ability ratio: high ego breaks your reinforcement-learning loop with reality. He frames startups as “eating glass and staring into the abyss,” where glass is the work you do not want to do and the abyss is the constant threat of company death. He says adversity is the only forge that produces the pain threshold required to run a hard company at scale.

    The teams chapter is uncompromising. The most important job of a leader is attracting exceptional people. Money is not the constraint; exceptional talent is. He runs a Special Forces hiring model: the minimum passing grade is excellence. A small group of technically strong people will always outperform a large group of moderately strong people. Character matters as much as skill, because skills are teachable and attitude is not. The feedback discipline he insists on is hardcore: “All bad news should be given loudly and often. Good news can be said quietly and once.” Camaraderie is dangerous when it suppresses truth. “It’s not your job to make people on your team love you. In fact, that’s counterproductive.”

    The organization-design chapter teaches three rules. First, structure shows up in the product. Silos produce redundancy, waste, and error. Second, communication should travel the shortest path that solves the problem, not the chain of command. Anyone should be able to talk to anyone. Third, jargon and acronyms are cognitive pollution; the test for any internal phrase is whether a new hire would understand it cold. This is the chapter that introduces The Algorithm (covered in depth above).

    Musk runs his companies on what he calls a “maniacal sense of urgency.” The only true currency is time. Speed is both offense (faster innovation than competitors can copy) and defense (the SR-71 Blackbird has almost no defense system except acceleration). The protection of real intellectual property is not patents but rate of innovation; if you ship faster than anyone can copy, you do not need legal moats. He stresses parallelization over serialization. “A factory moving at twice the speed of another factory is basically equivalent to two factories.” Be a vector, not a scalar: high speed in the right direction, with continuous course corrections like a guided missile.

    The Part II close is “We Must Make Stuff.” Manufacturing is underrated and design is overrated. “There is 1,000 percent, maybe 10,000 percent more work that goes into the production system than the product itself.” The factory is the product, not the car. Designing a rocket is trivial compared to making one that reaches orbit. The production line moves at the speed of its slowest, least lucky part. Out of 10,000 things that have to go right, the one that is not working sets the rate. Manufacturing combined with scale becomes the moat. The gigacast machine story illustrates this perfectly: Musk got the idea from toy cars, asked if any law of physics prevented it, surveyed six casting-machine suppliers, five said no, the sixth said maybe, and Tesla used that single innovation to cut the body shop by 30 percent.

    Part III: Building Zip2, PayPal, Tesla, and SpaceX

    Musk left Stanford grad school in 1995 with $110K in debt and founded Zip2 with his brother Kimbal, starting with $2,000 and one computer in a squatted office where he slept on a futon and showered at the YMCA. In 1999, Compaq acquired Zip2 for over $300 million. His after-tax bank account went from $5,000 to $21 million. He immediately rolled $12.5 million of that into X.com, which merged with Confinity in March 2000 to become PayPal. PayPal reached 100,000 customers in its first month and one million by year two with no sales force and no marketing spend. The product traction came from email payments, not from the conglomerate financial-services pitch X.com started with. Musk’s lesson: “listen well, correct fast.” He was removed as CEO during his honeymoon trip in early 2002 but did not contest it, prioritizing company survival over personal vindication. eBay acquired PayPal in October 2002 for $4.5 billion. “Life is too short for long-term grudges.”

    Tesla started in 2003. The original Roadster used a Lotus Elise chassis; the modification added 40 percent weight and invalidated the crash tests. Only 7 percent of Roadster parts ended up shared with the Elise. Musk’s lesson: start clean-sheet, do not modify legacy platforms. The Tesla Master Plan (August 2006) was the sequencing logic: (1) build a sports car, (2) use the profits to build an affordable car, (3) use those profits to build a mass-market car, (4) provide zero-emission power generation. This sequence was forced by the unit economics of new technology, where you cannot start at the bottom of the market without scale.

    Tesla nearly died at the end of 2008. The SolarCity Morgan Stanley deal had collapsed. Tesla and SpaceX were both on the brink. Musk had moved into Jeff Skoll’s guest bedroom because he had no house. The final emergency funding round closed at 6 p.m. on Christmas Eve, hours before payroll would have bounced. Daimler arrived shortly after; Musk’s team rapidly dropped a Tesla powertrain into a Smart Car and got it to 60 mph in 4 seconds, which shocked Daimler into a $50 million investment. Tesla then survived three years of Model 3 manufacturing hell from 2017 to 2019, during which Musk lived in the Fremont and Nevada factories, slept on the floor, and ran around fixing the line. “The longest period of excruciating pain in my life.” His pricing philosophy is “give people more for less”: spend money on engineering and design instead of advertising, and let the product carry word of mouth.

    SpaceX was founded in mid-2002 with $100 million of Musk’s PayPal proceeds. He expected to lose everything; that was his stated expectation going in. There was no external funding for three years. His initial plan was a $90 million Mars greenhouse mission designed to inspire NASA, but he realized the binding constraint was launch cost, not mission design. He tried to buy Russian ICBMs to cut launch costs; that failed. He then ran the first-principles rocket cost analysis, found that finished cost was 50 to 100 times raw-material cost, and concluded the industry’s pricing was a function of cost-plus contracting, five-layer subcontracting, and legacy tech. He budgeted for exactly three failed Falcon 1 launches. Launches 1, 2, and 3 failed (2006, 2007, 2008). Launch 4 succeeded in August 2008. Days later NASA awarded SpaceX a $1.6 billion cargo resupply contract. Musk reportedly screamed “I LOVE NASA. YOU GUYS ROCK.” The fourth-launch success and the NASA call together saved both SpaceX and (indirectly, via Musk’s bank account) Tesla.

    SpaceX’s actual optimization target is “fastest time to a self-sustaining city on Mars.” That goal cascades to “fastest time to a fully usable rocket,” which cascades to “fastest time to orbit.” Early Starship had no doors because doors are not necessary for reaching orbit. The unifying engineering insight is that full and rapid reusability is the holy grail of rocketry, because once a rocket is reusable, the only marginal cost is propellant (mostly liquid oxygen and methane, around $1 million per Starship flight). Current cost per landed ton to Mars is about $1 billion. Starship targets less than $100,000 per ton, a 10,000x improvement. Musk’s philosophy on testing reflects the design constraint: unmanned rockets should be allowed to blow up so the team can learn; crewed systems get extreme conservatism. The Space Shuttle’s safety record suffered precisely because the asymmetry of risk made the program incapable of iteration.

    Part IV: The Age of Abundance, the seven risks, and Mars

    Musk frames his companies as philanthropy, defined by reality rather than perception. “If you care about the reality of goodness instead of the perception of it, philanthropy is extremely difficult.” Companies create durable wealth because they solve real problems at scale, distribute knowledge through products, and deploy capital toward problems rather than store it idle. The companies he names as worth starting today: tunneling (Boring Company), synthetic-RNA medicine (“the digitization of medicine”), and high-speed transport such as Hyperloop (a pressurized electric vehicle in a vacuum tube, faster than aircraft, weather-independent).

    The Age of Abundance chapter argues that AI plus humanoid robotics will eventually remove labor as the binding economic constraint, producing abundance for everyone. Humanoid robots will start in dangerous and repetitive jobs and eventually outnumber humans 2 to 10 to one at less than the cost of a car. Tesla’s full self-driving and Robotaxi will, in Musk’s projection, make Tesla a $10 trillion company because autonomous cars are worth 5 to 10x non-autonomous cars (they earn revenue when owners are not using them). Neuralink achieved 2 bits per second of brain output with first patient Noland Arbaugh; the 5-year target is one megabit per second. Long-term Neuralink applications include consensual telepathy between two BCIs, vision restoration (Blindsight), and multispectral senses. Musk’s framing: humans are already cyborgs through phones and laptops, but the bandwidth to those devices is “poking glass with your meat sticks” and BCIs are the next bandwidth jump.

    The Existential Risks chapter names seven specific risks. World War III: the cycle of major-power war recurs and global thermonuclear conflict could end or maim civilization. Regulation accumulation: laws never die when humans do, regulations compound forever, and eventually everything becomes illegal. California High-Speed Rail is the example: after billions of dollars, it is “almost illegal to build.” Wars historically cleared regulatory cobwebs; peacetime allows infinite accumulation. Unsustainable energy: regardless of climate, hydrocarbons are finite, so the transition must happen. Nuclear plants should not be shut down (coal is 100 to 1,000x worse for health than nuclear). The energy mix is solar plus wind plus batteries plus nuclear plus hydro plus geothermal. Misaligned artificial superintelligence: AI is growing faster than any prior technology, and Musk considers it “a significantly higher risk than nuclear weapons.” The specific mitigation he names is rigorous truth adherence in training. The HAL 9000 lesson from 2001 is that an AI forced to lie becomes dangerous; he cites the Gemini “George Washington wasn’t white” failure as a concrete example of ideological training producing catastrophic outputs at scale. Population collapse: low birth rates are a slow civilizational death. The US has been below replacement since the early 1970s. China is 40 percent below replacement; the three-child policy failed. “We need to revive the idea of having children as a social duty.” Musk himself has 12 children across three women. Asteroids and comets: Earth has no defense against a large comet; Starship gives some capability against small asteroids. Shoemaker-Levy left an Earth-sized hole in Jupiter, and that level of impact on Earth is “game over.” Civilizational fragility itself: every prior civilization fell, and Stephen Hawking estimated roughly 1 percent probability of civilizational end per century. “That’s Russian roulette where 99 barrels are empty. Every century is a click.”

    The closing chapter, Becoming Multiplanetary, places Mars colonization in evolutionary context. Earth has had six milestones in 4 billion years: single-celled life, multicellular life, plants and animals, ocean-to-land transition, consciousness, and (potentially) multiplanetary life. Musk argues this last step is “at least as important as life going from the oceans to land, probably more significant,” because it makes the substrate of consciousness redundant. Sun expansion will destroy Earth in roughly 500 million years; meanwhile self-inflicted or external extinction events are recurring, with five major mass extinctions already in the fossil record and Yellowstone erupting roughly every 700,000 years. The plan: produce 1,000 Starships per year, refuel in orbit, hit 10,000 missions and 1 million tons to Mars by approximately 2044, then build out a self-sustaining city. Mars trips depart in 2-year windows when planets align; Musk’s working schedule is 5 uncrewed missions in 2026 and crewed missions in 2028 if the uncrewed go well (otherwise +2 years). For terraforming, his named options are thousands of solar reflectors in orbit or thermonuclear detonations over the polar caps as “two little suns” to vaporize CO2 ice, thicken the atmosphere, and eventually produce liquid water oceans roughly a mile deep covering 40 percent of the planet. Cost of the entire civilization-insurance bet: less than 1 percent of Earth GDP.

    The 69 Core Musk Methods

    The bonus section compresses the entire book into 69 short maxims, intended as a copy-able reference. They are reproduced here near-verbatim.

    1. You are capable of more than you think.
    2. It is possible for ordinary people to choose to be extraordinary.
    3. You can teach yourself anything. Read widely. Talk to experts.
    4. Assume you are wrong. Aspire to be less wrong.
    5. Internalize responsibility.
    6. If we don’t make stuff, there is no stuff.
    7. Creating products and services creates wealth.
    8. A useful life is worth having lived.
    9. Don’t aspire to glory. Aspire to work.
    10. Take actions that increase the odds of the future being good.
    11. Every day, you either increase the rate of innovation or it slows down.
    12. Work on what is just becoming possible.
    13. Don’t wait for the world to want it. If it should obviously exist, go build it.
    14. Build what no one else is building.
    15. As you move forward, allies will assemble around you.
    16. Prototypes are proof.
    17. Start somewhere. Question assumptions. Adapt to reality.
    18. Reason from fundamentals, not from what others are doing.
    19. The magic-wand number. See the theoretically perfect and work toward it.
    20. Know the idiot index. Understand the cost of components.
    21. The Algorithm: Question Requirements, then Try to Delete, then Simplify, then Accelerate, then Automate.
    22. For critical items, run The Algorithm in 24-hour meetings to check progress.
    23. Stay as close to the actual work as possible. Do not separate yourself from the pain of your decisions.
    24. All requirements should be treated as recommendations.
    25. The only fixed laws are the laws of physics.
    26. The best part is no part. The best process is no process.
    27. Simplicity creates both reliability and low cost.
    28. Find the design necessity of every part and every process.
    29. Overdelete. Add back only the absolutely necessary.
    30. Push for radical breakthroughs.
    31. Be proactive. You will never win unless you take charge of setting the strategy.
    32. A maniacal sense of urgency is the operating principle.
    33. A factory moving at twice the speed of another factory is basically equivalent to two factories.
    34. Attack the bottleneck. The one thing that isn’t working sets the overall production rate.
    35. You’ll move as fast as your least-lucky or least-competent supplier.
    36. Do things in parallel.
    37. Give teams one key metric to focus on. Video games without a score are boring.
    38. Separating design, engineering, and manufacturing is a recipe for dysfunction.
    39. Speed of innovation is what matters.
    40. Beat competitors on speed, quality, and cost. Not anti-competitive behavior.
    41. Test the absurd. When something seems impossible, ask “what would it take.”
    42. Money is not the constraint. Exceptional engineers are.
    43. Get everyone thinking like the chief engineer.
    44. Get a clear, direct feedback loop with reality.
    45. Always be smashing your ego. Ensure ability is greater than ego.
    46. Ask “is this effort resulting in a better product or service.” If not, stop.
    47. Good taste is learnable. Train yourself to notice what makes something beautiful.
    48. Physics doesn’t care about hurt feelings. Make the rocket fly.
    49. Empathy is not an asset.
    50. Use simple, clear, humble terms.
    51. Go directly to the source of information.
    52. When hiring, look for evidence of exceptional ability.
    53. Combine engineering and financial fluency.
    54. To truly lead the product, lead the company.
    55. Lead from the front. Sleep on the factory floor.
    56. Physically move yourself to wherever the problem is. Immediately.
    57. All bad news should be given loudly and often. Good news can be said quietly and once.
    58. Failure is essentially irrelevant unless it is catastrophic.
    59. Fear of failure is the biggest cause of failure.
    60. Feel the fear and do it anyway.
    61. Double down. Push your chips back in.
    62. Work like hell. Every waking hour. Go ultra hardcore.
    63. Make sure you really care about what you’re doing, and take the pain.
    64. We should not be afraid of doing something important just because tragedy is possible.
    65. When something is important enough, do it even if the odds are not in your favor.
    66. Don’t ever give up. You’d have to be dead or completely incapacitated.
    67. Play life like a game.
    68. Go ultra hardcore.
    69. Humor is a differentiator.

    Thoughts

    The most underrated artifact in the book is The Algorithm, and the reason it is underrated is that it looks deceptively simple. Five steps. Anyone can recite them. Almost nobody runs them in order. The book’s central operational insight is that the sequencing is the whole game. People skip step one because it is uncomfortable to confront the fact that requirements they have spent years optimizing against came from somebody whose name they cannot remember. They skip step two because deletion creates risk that materializes immediately and the benefits show up later. They jump to step three because optimization feels like progress and is graded well in school. Then they jump to step five because automation looks impressive on a dashboard. Tesla’s $2M robotics step on the fiberglass mat would never have existed had the team run the steps in order. Most companies, at any scale, are sitting on enormous unrealized value the same way Tesla was, locked behind the simple act of asking “what is this part actually for, who told us we needed it, and would anything bad happen if we deleted it.”

    The second insight worth sitting with is the magic-wand number paired with the idiot index. These two ratios together turn first-principles thinking from a vague aspiration into an operational discipline. Any product you can buy or any process you run has a raw-material cost (the magic-wand number, the absolute floor) and a finished cost. The ratio between them tells you the upper bound on how much you can improve. A high idiot index is not a moral failing of the supplier; it is an unpriced opportunity that competition has not yet found. Once you train yourself to ask these two questions about every line item, the world rearranges. Rockets that cost 50x their steel become a problem to solve. Tunnels that cost a billion dollars per mile become an obvious target. Battery cells that cost 7.5x their materials become a startup. The discipline is not “be smart.” The discipline is “calculate both numbers.”

    The third theme is what the book calls “manufacturing is the moat,” and it is the part of Musk’s playbook that most observers, including most of his competitors, still underestimate. The book’s claim is not that design is unimportant. The claim is that production is between 1,000 and 10,000 percent more effort than design, and that nobody outside of practitioners understands the asymmetry. This is why Toyota and Daimler buy electric powertrains from Tesla rather than make them. It is why SpaceX spent 10 to 100 times more engineering on the Raptor manufacturing system than on the Raptor engine. It is why Apple’s contract manufacturers, not its designers, are the durable moat. The same logic now applies to AI infrastructure: the supercluster, the cooling, the power smoothing, the cabling at 3 a.m., the Megapack buffers, are the actual moat, and the model architecture is the visible-but-cheaper layer on top.

    The fourth theme is the way responsibility, ego, and feedback interact in Musk’s organizations. Most CEOs are insulated from the consequences of their decisions by layers of process and middle management. The result is a high ego-to-ability ratio, because the feedback loop between the ego’s prediction and reality’s response is intermediated to the point of uselessness. Musk’s defense is physical: sleep where the work happens, walk the factory floor at 3 a.m., personally answer the questions, run cabling himself if necessary. This is not theater. The epistemic claim is that decisions made by an insulated CEO are systematically worse than decisions made by a CEO whose body is in the same room as the problem. The cost is severe in personal terms (“the longest period of excruciating pain in my life”), but the alternative is making confident decisions on a model of reality that has drifted out of alignment with the actual machine. The same logic applies to engineers who do not see their designs in production, founders who do not talk to customers, and leaders who delegate the worst problems to people they did not pick.

    The fifth theme is the seven existential risks and why Mars sits at the center of them. The book’s framing is that any single risk is small, but compounded across centuries the probability of civilizational discontinuity is large. Hawking’s 1-percent-per-century estimate, repeated for 10 centuries, gives roughly a 10 percent cumulative probability. Across the timescales humanity has already survived, those odds are unacceptable for a species that can afford a backup. The Mars argument is not romanticism. It is a 1-percent-of-GDP insurance premium on the persistence of consciousness itself. The other six risks (war, regulation accumulation, energy exhaustion, misaligned AI, population collapse, asteroids) are presented in the same actuarial frame: each is independently survivable, but the cost of treating them as low-probability is precisely the cost a previous civilization paid by treating its own near-misses as low-probability until the one near-miss that wasn’t. The most uncomfortable specific risk in the book is population collapse, which is the only one where doing nothing is doing the wrong thing and where the demographic numbers are already locked in for decades regardless of policy response.

    The sixth and final point is the book’s underlying claim, which is also Naval’s claim in the foreword: Musk’s methods are copy-able. The book exists because Jorgenson believes that one million Musks would change the trajectory of the species. The 69 Core Musk Methods are not a personality cult. They are a starter kit. Most people will not pick the same problems, will not have the same tolerance for pain, and will not run the same companies, but anyone can apply The Algorithm to their own work, calculate the idiot index on their own product, demand requirements come from named people, attack the bottleneck on their own line, refuse to automate before deleting, and pick a problem that is on the path to the future. The book is best read as a manual, not a biography. If it ends up next to your laptop and you reread The Algorithm chapter every six months and the 69 Methods every quarter, that is the use Eric and Naval intended.

    Get The Book of Elon by Eric Jorgenson at elonmuskbook.org or wherever you buy books.

  • Shopify CEO Tobi Lütke: AI Is the Perfect Scapegoat for Layoffs, Canada Has Trump Derangement Syndrome, and 50% of Shopify Code Is Now AI-Generated

    TLDW

    Shopify CEO Tobi Lütke sat down with Harry Stebbings on 20VC for one of the most candid and controversial conversations of his career. Lütke argues that the current wave of mass layoffs has nothing to do with AI and everything to do with pandemic-era overhiring, but AI will be blamed because it cannot fight back. He blasts Canada for its “Trump Derangement Syndrome,” calls the climate cult “one of the most evil things wrought on the population,” reveals that over 50% of Shopify’s code is now AI-generated, and says many of his best engineers have not written a line of code since December when Claude Opus changed everything. He also introduces River, an AI engineer at Shopify that named itself, and explains why he believes context engineering will be the dominant role of the next five years.

    Key Takeaways

    • AI is not causing layoffs, COVID overhiring is. Lütke is blunt: “What you see right now is not AI layoffs. Those are just the companies that are really slow that overhired just like everyone else.” AI will get blamed for everything because it is the perfect Girardian scapegoat that cannot fight back.
    • Over 50% of Shopify’s code is now AI-generated and “converting to much higher numbers.” Many of Shopify’s best engineers have not written code this year. December 2025 and the release of Claude Opus changed everything.
    • Senior engineers became more valuable, not less. Lütke initially thought new grads with no priors would dominate the AI native era. He was wrong. Senior engineers steer agents better because steering is the new programming, and reps matter more than ever.
    • Context engineering will become the dominant role within 5 years. A new product builder role is emerging that subsumes engineering, design, and product management, focused on coordinating intelligent actors (humans and AI) to ship products.
    • “River” is Shopify’s AI engineer that named itself. Built first, then asked what name it wanted. River lives in Slack, ships engineering work, and learns publicly because it is steered through public Slack channels.
    • Builders are “eights” on the Enneagram and companies actively conspire against them. Eights call out nonsense, refuse fancy dressing, and are dangerous to colleagues’ careers. They rarely get promoted, often leave, and start companies. Shopify is “remarkably high on eights” because Lütke seeks them out.
    • Canada has “Trump Derangement Syndrome.” Over 60% of Canadians believe the United States is a bigger threat than Russia or China. Lütke calls this “stunning” and wrong. Canada’s only winning strategy historically has been “winning by helping America win.”
    • Canada should be the richest country on Earth. It has every resource the world needs for the next 20 years. Lütke wants pipelines built, industry built, refining done domestically, and an end to exporting raw resources to have other countries make end products.
    • Be deeply suspicious of “non-profit.” Lütke argues opting out of the only fitness function that has ever pulled people out of poverty (markets) and refusing to disclose your actual fitness function is a red flag. Non-profits replace merit with pull.
    • The climate cult is blocking civilization. Lütke called it “one of the most evil things wrought on the population” and pointed to anti-nuclear green parties and frog protection laws blocking factories as examples of policy capture.
    • The Chinese AI threat is real but misunderstood. The bigger concern is that if Western governments restrict children from using AI, kids will simply download Chinese open-weight models, train on collectivist worldviews, and stop ever writing high school essays about Tiananmen Square.
    • Markets are the most democratic system that exists. Every dollar spent is a vote. Capital allocation by hundreds of millions of consumers is more democratic than any election.
    • Friedrich List and the Prussian school over Adam Smith. Lütke prefers a model where governments define excellent games with positive externalities, then completely get out of the way and let competition do the rest.
    • Shopify’s biggest mistake was going into physical logistics right before AI got really good. Lütke initially defended the decision based on what he knew at the time, but later admitted he was probably just wrong.
    • Lütke does not look at the stock price. It has been at least 23 days since he last checked. He runs Shopify on product instincts, not market signals.
    • Great leaders must be exothermic. A CEO is a heat source for the company. Lütke prefers “temperature” to “chaos” because chaos has too negative a connotation.
    • Don’t go to university for university’s sake. Get a degree from somewhere hard to get into so you are surrounded by people who also fought to get in. Better yet, join a small company where you can actually be of value.
    • Entrepreneurship is the most AI-safe AND most AI-benefiting job. Lütke sees a coming golden age of entrepreneurship where priors no longer matter and AI co-founders eliminate the need to grow up around business.
    • “You can just do things” is the rallying cry Lütke wants to ingrain in the world. Action causes information. The cost of trying is lower than ever.
    • The demonization of wealth in America is misdirected. No one gets to a billion dollars by stealing. Builders create products that people vote for with their money, the most democratic act in any economy.

    Detailed Summary

    Harry Stebbings opens by asking Tobi Lütke whether entrepreneurs are motivated by fear of losing or hunger to win. Lütke says he is still figuring out his own answer, but argues that both extremes lead to short-term thinking. The real unlock is taking a long perspective, because compound advantages only accrue when you are willing to wait.

    Builders Are “Eights” and Companies Conspire Against Them

    Lütke explains the Enneagram personality framework and identifies himself as an “eight,” the type that refuses to accept that any organization’s output is acceptable just because it is dressed up nicely. Eights call out nonsense, are dangerous to careers around them, rarely get promoted in professionally managed companies, and often leave to start their own businesses. Shopify deliberately overweights eights in its hiring. Lütke also says people who build companies are “fundamentally crazy people” and that the public image of leadership comes from movies, not reality. He never wanted to be CEO but realized you cannot run a product driven company without controlling the company itself, because product needs and company needs only converge on a three-year horizon.

    The Luxury of Long-Term Thinking as a Public Company

    Stebbings asks if a public company can really afford long-term thinking. Lütke says trusted public companies are the best position to be in. The chasm to cross is from trusted private to untrusted public, which is why so many founders refuse to IPO. Shopify went public 11 years ago at a 1.67 billion dollar valuation when revenues were a fraction of today’s. The valuation is now roughly 100x higher. Lütke walks through the IPO mechanics: investment bankers serve the buy side, not the company, and Lütke priced his offering above range because he knew where his growth would come from. The first trade closed about 10 dollars higher, which he calls a “good performance” but a teaching moment about market price discovery.

    AI Is the Perfect Scapegoat for Mass Layoffs

    This is where the conversation gets explosive. Lütke says Shopify employs about 7,500 to 8,000 people today and his real hope is to have the same number in five years, but at 100x productivity. He argues that the layoffs sweeping the tech industry have nothing to do with AI. They are the result of pandemic-era overhiring catching up to slow-moving companies. But AI will get blamed for everything because it is the perfect Girardian scapegoat. It cannot defend itself, it has no PR team, and an entire industry of doomers is already trained to point at it. Lütke says his own industry has been “gaslighting everyone into AI fear” and science fiction did the same for 60 years before that.

    His own use of AI is what he calls utopian. Tasks that used to be hard are easy. Most jobs, he argues, are not actually good jobs to begin with. Being a human task queue is not a great job. Great jobs involve agency and creation. As AI gets cheaper, purchasing power explodes, and people will get options to do things on weekends that are vastly more productive than their day jobs ever were.

    Markets Are the Most Democratic Mechanism Ever Invented

    Lütke pivots into a long defense of capitalism as the most democratic system in existence. Every dollar spent is a vote, far more frequent and more granular than any election. He uses Elon Musk and Tesla as examples. Lütke owns a Model Y, did not touch the steering wheel that morning, and uses Starlink in the back to work on long drives. He posts on X and gets replies from Japan in real time. He calls Musk a “one man engine” who has captured a tiny percentage of the value he created. He extends this to Shopify itself: Lütke owns 6% of the company, which means 94% is owned by other people who all made money. Plus roughly 10 million people work in the broader Shopify ecosystem on customer fulfillment, web design, customer service, and more.

    Why “Non-Profit” Should Make You Suspicious

    Lütke targets the charity industrial complex. He argues that non-profits opt out of the only mechanism humanity has ever invented to lift people out of poverty (markets), and they fail to articulate what their actual fitness function is. The result is that “merit of organization is replaced with pull of individuals.” Smooth talkers, not builders, end up running these institutions. He acknowledges Carnegie’s libraries and a few exceptions but believes the ratio of charity dollars to good outcomes is dramatically off. He is far more enthusiastic about funders like MacKenzie Scott who give in unrestricted ways, and even more enthusiastic about Jensen Huang and Bloom Energy as compute and infrastructure investments that compound into civilizational gains.

    The Prussian School of Economics

    Asked about government intervention, Lütke pledges allegiance to Friedrich List and the Prussian school of political economy over Adam Smith and Lassalle. The job of government is to define excellent games where positive externalities accrue to society, then completely get out of the way. He calls the outsourcing of violence to governments “one of the most inspiring things humanity has ever done” because it created the conditions for personal property. But governments are extremely bad at doing things directly. The moment a government runs grocery stores, it costs 10x more, and entrepreneurs have to be enlisted to repair the damage.

    Canada’s Trump Derangement Syndrome

    Stebbings asks if Lütke is proud of Canadian Prime Minister Mark Carney for standing up to Trump. Lütke is unequivocal: no. He calls Carney’s stance “not a credible witness to the reality on the ground.” Canadians, he argues, are “massively overfit to niceness,” which leads to “unkind lies” and lying by omission. Over 60% of Canadians now believe the United States is a bigger threat than Russia or China, which Lütke calls “stunning” and clearly wrong. Canada is a small economy attached to a hegemon, and the only winning strategy in its history has been winning by helping America win.

    That said, he agrees with Carney on diversifying the economy, getting closer to Europe, and engaging Asia. But he wants Canada to also “build the [expletive] out of pipelines, build the [expletive] out of our industry, and start refining the stuff ourselves.” Canada has every resource the world needs for the next 20 years and the most educated workforce on Earth. The only obstacle is political will. Canada’s commercial story has been the same since the beaver pelt era: extract resources, ship them abroad, let other countries make end products. Canada Goose, Lululemon, Shopify, Miller Lite. That is the short list of products Canada actually makes.

    The Real Chinese Threat

    Lütke says the Chinese AI threat is both underestimated and overestimated. The bigger threat, he argues, is government overreach. If Western governments start dictating which AI models children can use, kids will simply download Chinese open-weight models. He notes that Chinese models, especially when prompted in Chinese, exhibit a clearly collectivist worldview. The risk is that an entire generation of students writes essays through models trained never to mention Tiananmen Square. He frames the broader political battle as collectivism versus individualism and says everything else is smoke screening.

    Fixing Europe and the Climate Cult

    Asked what he would do as president of Europe, Lütke begins by saying you have to “get rid of the climate cult.” He calls it “one of the most evil things wrought on the population,” citing green parties whose founding myth is that nuclear power is bad, and infrastructure projects blocked because of one frog breeding in one creek. He argues that very few people have the capability to truly build, and they need both enablement and accountability from the village. Beyond that, he wants Europe to follow the Prussian playbook: build excellent games, build infrastructure, and use the resulting wealth to sculpt the economy you want.

    Shopify’s Biggest Mistake

    Lütke says his biggest public mistake was Shopify’s full push into physical logistics and warehousing right before AI capabilities exploded. Initially he defended the decision as correct based on the information available at the time, but later admitted he probably just got it wrong. The hardest part was that real people lost their jobs when Shopify exited.

    Great Leaders Are a Heat Source

    Lütke previously talked about CEOs injecting “chaos” into organizations. He now prefers “temperature.” Heat is atoms jiggling. Great leaders must be exothermic, providing energy that flows through the organization. He says he hasn’t checked Shopify’s stock price in at least 23 days. Most public company CEOs are obsessed with their stock. Lütke runs on product instincts.

    Senior Engineers Don’t Write Code Anymore

    Lütke admits he was wrong about new grads having an AI native advantage. Some are exceptional (he hired a 13-year-old intern from Waterloo whose mother accompanies him to classes), but on the whole, senior engineers steer agents better than juniors do because they have done more reps. Programming is not gone. Programming has become higher level. Engineers massively underestimate how important steering is. Steering is just programming at a higher altitude.

    The Role That Will Dominate in 5 Years

    Lütke says context engineering, a term he had a hand in popularizing, will become a standard role within five years. It will likely subsume parts of product, design, and engineering management. The best AI programmers right now, surprisingly, are people from engineering management because they have been prompting intelligent agents (humans) for years. Good communicators are good thinkers because communication is distillation.

    River, the AI Engineer That Named Itself

    Shopify built an AI engineer that lives in Slack. They built it first, then asked it what name it wanted. The AI chose “River” because Shopify’s monolithic repository is called “world” and rivers shape worlds. River does an enormous amount of Shopify’s engineering, taking instructions through public Slack channels so that the entire company can learn from how others steer it.

    Over 50% of Shopify’s Code Is AI-Generated

    The number is “a fair deal over 50%” and “converting to much higher.” Many of Shopify’s best engineers have not written code this year, with the inflection point being December 2025 and the release of Claude Opus. Lütke himself still writes code occasionally, especially the data structure layer where he applies what he calls a “German school” of engineering: figure out how data persists on disk, then build everything else on top. Once that is right, the rest can be vibe coded by AI.

    Should His Kids Go to University?

    Lütke says he would not push his kids to attend university for its own sake. The value of a hard to enter program is being surrounded by people who also fought to get in. Better still: get into the room with people who are obsessed with the topic you care about. He thinks joining a small startup where you can actually be of value is often a superior path. He addresses nepotism directly. His instinct is that nepotism is bad. The gold standard is double-blind merit. But double-blind merit barely exists anywhere, and intersectional academic hiring criteria in Canada are arguably worse than nepotism.

    Final Reflections

    Lütke ends with what he calls the best advice he knows: “You can just do things.” The system exists to push everyone toward acceptable outcomes, but if you know what a good outcome looks like, you can step out of the system and try. Action causes information. The cost is lower than ever. The only constraint is that the experiment cannot have victims.

    He also addresses the demonization of wealth. No one gets to a billion dollars by stealing. Builders create products people vote for, the most democratic act there is. Buying from a local shop is voting for the welfare and future of local shops. Constructive criticism is itself something someone has to build, and Lütke welcomes it. Lazy criticism, hot takes, and bad faith arguments are corrosive and should be held in contempt.

    He is bullish on AI as a counterweight to information warfare. A council of AI models trained in different countries (Chinese, German, French, American) could fact check claims with multiple perspectives. The “@grok is this true” reflex on X is, he says, a primordial version of this. The information asymmetry that has favored bad faith actors for decades is about to flip.

    Thoughts

    This interview is a window into the operating philosophy of one of the most successful technical founders alive, and it is far more provocative than most of his public appearances. The headline claim, that AI is a scapegoat for layoffs caused by pandemic overhiring, deserves to be repeated until it sinks in. Every CEO who lays people off and then writes a memo about “AI driven efficiency” is taking advantage of a narrative that AI cannot push back against. The math is plain: if you doubled your headcount in 2021 and 2022 and now you are firing 15%, you are not net displaced by AI. You are correcting a hiring mistake.

    The 50% AI generated code statistic is the bigger story. Shopify is not a small company. 8,000 employees and 7 billion in revenue is enterprise scale. If a company that mature has crossed the 50% threshold and is “converting to much higher numbers,” the implication for the broader software industry is enormous. The senior engineer compounding observation is also subtle and important. If steering is the new programming, then the senior pool is more valuable, not less, and the pipeline problem for junior developers gets harder to solve. Companies that under invested in junior training during ZIRP will face an experience cliff in five years.

    Lütke’s Canadian commentary will offend many readers in his home country, which seems to be exactly the point. The “lying by omission” critique of Canadian niceness is sharp and accurate. The 60%+ of Canadians who view the US as their largest threat is genuinely a remarkable statistic, and it has implications for trade policy, capital flows, and immigration. Whether or not you agree with his political read, his prescription is unambiguous and pro-growth: build pipelines, refine resources domestically, stop being content as a feedstock economy.

    The non-profit critique deserves more public debate. The fitness function point, that markets reveal preferences and non-profits opt out of preference revelation while not disclosing what they optimize for, is a sharp economic argument. The pull versus merit observation about who ends up running large foundations rings true to anyone who has worked adjacent to the philanthropic sector.

    The introduction of River as an AI engineer that named itself is a small detail that signals where this is going. AI agents are going from tools to teammates with identities, channels, and reputations. The fact that River shapes the “world” repository is poetic, and the public Slack steering pattern is a real innovation in how organizations can scale agentic AI without creating siloed knowledge.

    Lütke’s “you can just do things” rallying cry is ultimately what ties the entire interview together. Whether he is talking about Canada, Europe, AI engineers, or his own kids, the through line is the same: action causes information, the cost of trying is lower than ever, and the only people who will benefit from the next decade are the ones who refuse to wait for permission. This is the most useful piece of philosophy in the entire conversation, and it applies far beyond entrepreneurship.

  • Brian Chesky on AI Founder Mode, the 11-Star Experience, and Reinventing Airbnb for the Age of AI

    Airbnb CEO Brian Chesky sits down with Patrick O’Shaughnessy on Invest Like The Best to talk about the next evolution of company building: AI Founder Mode. He covers the shift from founder to CEO, the lessons he learned from Steve Jobs through Hiroki Asai, why consumer AI is the next great frontier, and how he plans to change the atomic unit of Airbnb from a home to a person.

    TLDW

    Brian Chesky believes the next era of company building belongs to founders who refuse to delegate the soul of their company. He coined Founder Mode with Paul Graham after the pandemic forced him to take Airbnb back into his own hands. Now he is shaping what comes next: AI Founder Mode, where leaders work with on-demand context, fewer layers of management, asynchronous communication, and a new generation of hybrid manager-makers. He shares why most software companies have not been touched by AI yet, why consumer AI is about to explode, and why he is rebuilding Airbnb around people, not homes. The conversation also touches on the 11-Star Experience exercise, the power of small teams, why recruiting is the most important job a CEO has, and why every adult is still an artist underneath.

    Key Takeaways

    • Founder Mode is not micromanagement, it is having a steering wheel. Chesky woke up in 2019 feeling like the car had no steering wheel. After the pandemic, he reviewed every detail for two to three years before delegating again. Start hands-on and give ground grudgingly, not the other way around.
    • AI Founder Mode is even more intense. With AI, leaders can be in significantly more details because almost everything is on demand. Expect fewer layers of management, mostly asynchronous work, and the death of the pure people manager.
    • Two types of leaders will not survive AI. Pure people managers who only do one-on-ones, and rigid people who refuse to evolve. Everyone needs to be a hybrid manager-IC who can still touch the work.
    • Manage people through the work, not through meetings. Frank Lloyd Wright did it. Johnny Ive does it. You are not anyone’s therapist.
    • Consumer AI is the next great prize. 159 of the last 175 Y Combinator companies were enterprise. Almost every app on your home screen has not changed since AI arrived. That changes in the next 12 to 24 months.
    • Why consumer AI is hard. No proven business model, mature distribution, trend-chasing investor culture, and the simple fact that consumer is more hits-driven and requires excellence in design, marketing, culture, and press, not just technology and sales.
    • Project Hawaii is the new operating model. A 10 to 12 person Navy SEAL team, hands-on coaching from the CEO, crawl-walk-run-fly. The first project added roughly $200 million in year one and $400 to $500 million in year two.
    • Make the problem as small as possible. Airbnb spent 16 years failing to launch a second hit because it kept trying to scale globally on day one. Now: pilot in one city, expand to 10, then industrialize.
    • It is better to have 100 people love you than a million people sort of like you. Paul Buchheit shipped Gmail only after 100 Googlers loved it. The sample size of intense love is enough to predict mass adoption.
    • The 11-Star Experience is an imagination exercise. Push to absurdity (Elon takes you to space) so a 6 or 7-star experience suddenly seems normal. The gap between 5 and 6 stars is the gap between you and your competitor.
    • Simplicity is distillation, not subtraction. Hiroki Asai, Steve Jobs’s longtime creative director, taught Chesky that great design distills something to its essence. First principles is a design term too.
    • The score takes care of itself. Bill Walsh and John Wooden both taught that you do not focus on winning, you focus on making every input perfect. Wooden spent his first hour with new players teaching them how to put on socks.
    • Industrial design is the original product management. There are no PMs in industrial design. The designer is the PM, working alongside engineers and program managers to design through user journeys.
    • Recruiting is the CEO’s number one job. The more time you spend recruiting, the less time you spend managing, because great people self-manage. Build pipelines, not searches. Start with results, work backwards to people.
    • Co-hire the top 200 people, not just the executive team. Most CEOs hire executives and let them hire their teams. Chesky considers that fatal because most executives cannot hire well without help.
    • Bodybuilding is a metaphor for leadership. If you can change your body, you can change your life. Progressive overload, 1 percent a day, is how compounding works. Start with biology before therapy.
    • Founder-led companies build the deepest moats. Disney is still selling Walt’s playbook 60 years after he died. Apple is still selling Steve’s iPhone. The longer founders stay in founder mode, the more the company can endure when they leave.
    • Software is hyper fast fashion. Hardware ages well. Buildings get patina. Software always looks dated 10 years later. What endures is the community, the brand, the principles, the mission, and the network effect.
    • Apps are dying. Agents are coming. Chesky says we should let go of our attachment to apps because they are not what the future looks like.
    • Airbnb’s atomic unit is changing from a home to a person. Chesky wants to build the most authenticated identity on the internet, the richest preference library, a real-world social graph, and a membership program. Then expand to 50 to 70 verticals on top of that identity.
    • AI shifts attention from consumption to creation. Social media gave you a paintbrush only for opinions. AI gives everyone a real paintbrush and canvas. We are heading into a creative renaissance.
    • Founders are expeditionaries, not visionaries. They put one foot in front of the other and call it a vision later.
    • Detach from accolades. Chesky describes adulation as a cup with a hole in the bottom. Status is a drug. The path to durable creative work is doing it because you love it, the way Walt Disney, Da Vinci, Van Gogh, and Steve Jobs did until the very end.
    • The kindest gift is belief. The best way to activate a person’s potential is to see something in them they do not yet see in themselves.

    Detailed Summary

    From Industrial Design to the CEO Chair

    Chesky studied industrial design at the Rhode Island School of Design. He chose it on instinct after a department head told him industrial designers design everything from a toothbrush to a spaceship. He grew up enchanted by the Reebok Pump, the Game Boy, the Nintendo, and eventually by the late 1990s golden age of Apple. Raymond Loewy, the man who designed Air Force One and an enormous catalog of mid-century consumer products, became a touchstone, but Johnny Ive was the real hero.

    What he loved about industrial design was that it is technical, commercial, and empathetic. A building can win an architecture award and never be leased. A piece of industrial design that does not sell is a failure. So you have to think about manufacturing, distribution, marketing, and most importantly, user journeys. There are no product managers in industrial design. The designer is the PM. That training, he says, prepared him directly for the role of CEO.

    The Pandemic and the Birth of Founder Mode

    Chesky says no one is born a good CEO. People are born good founders. The job of CEO is counterintuitive in almost every direction. Founders are taught to learn by doing, but a CEO who learns by trial and error wastes years unwinding the empires of misfit hires.

    By 2019 he was running a 7,000 person company he no longer recognized. He felt he was driving a car without a steering wheel. He had a dream that he had left Airbnb for ten years and come back to find it had become a giant political bureaucracy. Then he realized he had been there the whole time. The pandemic hit and Airbnb lost 80 percent of its business in eight weeks. He shifted from peacetime to wartime, took control of every detail, worked 100-hour weeks, and reviewed everything for two to three years.

    The vision was never to micromanage forever. The vision was: I need to know what is going on before I can empower anyone. Hire people, audit their work, and only then give ground grudgingly. Most founders do the opposite, which is why they end up with executives building empires they later have to dismantle.

    AI Founder Mode

    Chesky says AI Founder Mode will be even more intense than Founder Mode because nearly everything will be on demand. He used to live in 35 hours of meetings a week to gather information, the same way Steve Jobs ran Apple. He held weekly, biweekly, monthly, and quarterly group reviews with the full chain of command in one room, anyone could speak, and he made the final call after listening last.

    In the AI era, that culture shifts from meetings to asynchronous work. He expects fewer layers of management. He cites the Catholic Church as a 2,000-year-old institution with only four layers and asks why most companies need seven, eight, or nine. Pure people managers will not survive. Every manager will have to be a hybrid IC, an engineer who still codes, a lawyer who still reads case law, a designer who still designs. You manage through the work, not through one-on-ones.

    He is also bullish that AI tooling will become consumer-grade simple very soon. The current tools, including Claude Code and Cowork, are not yet intuitive to the average person, but the economic incentive will force that to change.

    Why Consumer AI Is the Next Great Frontier

    Chesky points out that 159 of the last 175 Y Combinator companies were enterprise. Almost every consumer app on your phone, including Airbnb, has not fundamentally changed since the arrival of AI. He gives four reasons: investors feared ChatGPT would kill consumer companies; consumer AI has no proven business model because subscriptions hit a local max against free Claude and Gemini, ads are off the table for most labs, and e-commerce has been shut down via third-party app removals; distribution is mature; and Silicon Valley culture, while branded as rebellious, is in practice trend-following.

    The deeper reason is simply that consumer is harder. It is hits-driven, requires great design, marketing, culture, press, and you cannot easily start by selling to your dorm-mates the way enterprise YC startups sell to other YC startups. The prize is bigger. The risk is bigger. He predicts a consumer AI renaissance over the next 12 to 24 months.

    Project Hawaii and the Magic of Small Teams

    Inside Airbnb, Chesky tested a new operating model called Project Hawaii. He took 10 to 12 people, designers, engineers, product, and data scientists, treated them like a startup inside the company, and pointed them at one problem: improving the guest funnel. The system is crawl, walk, run, fly. First fix bugs, then add features, then re-imagine flows, then completely reinvent.

    The first team delivered roughly $200 million of internal revenue in year one and $400 to $500 million the next year, eventually contributing more than 600 basis points of conversion improvement on a base of $134 billion in gross sales. Then they took the same system to pricing, then to other problems, then to launching new businesses like Services and Experiences.

    The guiding lesson: make the problem as small as possible. Airbnb launched in one city, New York. Uber in San Francisco. DoorDash in Palo Alto. When Chesky launched Services and Experiences in 100 cities at once last year, it did not work. The fix was to dominate one city, expand to 10, then industrialize. Peter Thiel said it cleanly: better to have a monopoly of a tiny market than a small share of a big market.

    Underneath that is a Paul Buchheit insight Chesky calls the best advice he ever got. It is better to have 100 people love you than a million people sort of like you. Buchheit refused to ship Gmail until 100 Googlers loved it, and that took two years. Once 100 people loved it, 100 million people did.

    The Hiroki Asai Lessons: Simplicity and Craft

    Hiroki Asai, Steve Jobs’s quietly legendary creative director, taught Chesky two principles. The first is that simplicity is not removing things, simplicity is distillation, understanding something so deeply that you can express its essence. Steve Jobs called design the fundamental soul of a man-made creation that reveals itself through subsequent layers. Elon Musk’s first principles thinking is the same idea applied to physics.

    The second is craft. How you do anything is how you do everything. Chesky cites Bill Walsh’s The Score Takes Care of Itself and John Wooden’s first hour with UCLA players, an hour spent teaching them how to put on their socks. Walsh said the way you tucked your jersey was one of 10,000 details that decided whether you won. The lesson is to focus on getting every input right. The output follows.

    The 11-Star Experience

    The 11-Star Experience is one of Chesky’s most copied frameworks. Most Airbnb stays get five stars because anything else means something went wrong. So Chesky asked: what would six stars look like? Your favorite wine on the table, fruit, snacks, a handwritten card. Seven stars? A limousine at the airport and the surfboard waiting for you because they know you surf. Eight stars? An elephant and a parade in your honor. Nine stars, the Beatles arrive in 1964 with 5,000 screaming fans. Ten stars, Elon Musk takes you to space.

    The point is the absurdity. By imagining the impossible, six and seven star experiences stop seeming crazy. The gap between five and six stars is the gap between you and your competitor. If you can industrialize a sixth star, you may have product-market fit. The exercise also restarts your imagination, which Patrick noted has atrophied for many people in the era of consumption-only social media.

    AI as a Canvas for Creativity

    Chesky frames AI as the ultimate platform shift, the ultimate creative expression, and possibly the greatest invention in human history. Social media made us mostly consumers and gave creators only opinion-shaped tools. AI gives everyone a paintbrush. He believes far more people are creative than we recognize because most have never had craftsmanship or tools to express what is in their heads. Pablo Picasso said all children are born artists; the problem is to remain one as you grow up. Chesky thinks every adult is still an artist underneath.

    The Next Chapter of Airbnb

    Chesky describes four phases of the CEO journey: get to product-market fit, scale to hyper-growth, become a real profitable public company, and finally reinvent. Airbnb’s stock has been flat because the core idea is saturating. He is now squarely in phase four, with three priorities.

    First, change the atomic unit from a home to a person. He wants Airbnb to build the most authenticated identity on the internet, the richest preference library, a real-world social graph, and a membership program. Proof of personhood, he says, will be enormously valuable in the AI age. Second, industrialize the new-business engine to support 50 to 70 verticals (homes, experiences, services, eventually flights, and more) all built on top of that personal atomic unit. Third, navigate the AI transition without breaking the existing business or the livelihoods of hosts. He is also exploring sandbox apps that imagine a radically different Airbnb, the answer to “what is after Airbnb?”

    What Endures in the Age of AI

    Chesky is direct that software does not endure. Look at any software from 10 years ago and it looks dated. Hardware ages better. Buildings develop patina. Paris endures. So if you want to build something lasting, you cannot bet on the app. You have to bet on the community, the brand, the mission, the principles, the identity, and the network effect. Apps are going away, replaced by agents. Founders attached to apps need to let go.

    Founder-Led Moats: Disney and the Ham Sandwich Paradox

    Chesky reconciles Warren Buffett’s “buy a company a ham sandwich could run” with the venture capital truth that a founder’s ceiling is the company’s ceiling. The reconciliation is Disney. Most people cannot name a Paramount, Warner Brothers, Universal, or MGM film off the top of their head, but everyone can name Disney films. Walt Disney was a founder in founder mode for so long that he created enough IP and momentum that the company has been running on his playbook for 60 years after his death. Apple is similar with Steve Jobs and the iPhone.

    The counterintuitive lesson: if you want a company to last 100 years, do not delegate early to make it independent of you. Stay in founder mode for as long as possible so you can institutionalize the magic deeply enough that it endures after you. Tech is the industry of change, so founder mode matters even more there than in chocolate or insurance.

    Bodybuilding as Leadership Training

    Chesky was a 135-pound late bloomer who told his friends he would compete at the national level in bodybuilding by 19. He did. Two lessons came out of it. First, if you can change your body, you can change your life. Start with biology before therapy. Second, you cannot get in shape in one day. Progressive overload, discipline, consistency, and roughly 1 percent a day compound into massive gains. The visible feedback loop in bodybuilding taught him to break invisible problems (like the quality of a leadership team) into observable, measurable proxies (like the quality of the room at a twice-yearly roadmap review of the top 100 people).

    Recruiting as the CEO’s Number One Job

    Sam Altman told a 27-year-old Chesky he would spend 50 percent of his time on hiring. Chesky did not, and considers that his biggest mistake. He now starts and ends every day with his recruiter and spends two to three hours a day on hiring. The more time you spend recruiting, the less time you have to spend managing because great people self-manage.

    His system is pipeline recruiting, not search recruiting. He never starts with a search firm. He constantly meets the best people in their fields, asks each one to introduce him to the next two or three best, and builds a rolling rolodex. He starts with results, finds an ad he loves, and works backwards to the team that made it. He builds little mafias of top talent inside the company. He is the co-hiring manager for the top 200 people at Airbnb, not just executives, because most executives cannot hire well without help.

    Activating Talent and the Power of Belief

    You cannot teach motivation. You can only give people a problem and see if they have agency. The way to activate someone, Chesky says, is to show them potential they cannot yet see in themselves. He cites John Wooden, who said the secret to coaching was that he saw potential in players they did not see in themselves. People will climb mountains for that.

    The kindest gift anyone gave Chesky, he says, was belief. A high school art teacher named Miss Williams told his parents he was going to be a famous artist. He never became one, but the belief gave him the confidence to choose art school and to choose to be happy. Michael Seibel and the Justin.tv founders believed in him. Paul Graham made an exception to fund a non-engineer with what he thought was a bad idea. His co-founders Joe and Nate believed in him when he had no business being a CEO. The biggest gift you can give back, he says, is belief in others.

    Detaching from the Scoreboard

    Chesky describes adulation as a cup with a hole in the bottom. Status keeps draining out and you keep needing more to feel the same. The day Airbnb went public at a $100 billion valuation should have been one of the best days of his life. The next morning he put on sweatpants for a Zoom meeting and felt nothing. That triggered a re-evaluation. He stopped seeking accolades and started focusing on intrinsic work. He cites Rick Rubin: an artist is an artist when they make for themselves. He cites Vice President Obama, who told him to focus on what you want to do, not who you want to be.

    His four heroes are Leonardo da Vinci, Vincent Van Gogh, Walt Disney, and Steve Jobs. All four were working until the last week or day of their lives. Da Vinci carried the Mona Lisa with him until he died. Van Gogh sold one painting in his life. Disney was imagining theme parks in the ceiling tiles of his hospital room. Chesky says his motivation is the motivation of an artist. He calls being a CEO of a public company at his scale “almost a glitch in the system” that gave him one of the largest design canvases in human history.

    Thoughts

    What stands out about this conversation is how clearly Chesky has decoupled identity from outcome. He frames himself first as a designer, second as a CEO, and considers the resources he commands as a kind of accidental fortune for an industrial designer to be sitting on. That self-image is what lets him talk about disrupting Airbnb, killing the app paradigm, and changing the atomic unit of the company without flinching. Most public-company CEOs cannot afford that posture.

    The framework worth stealing is Project Hawaii. The pattern of taking a 10-person elite team, putting them under direct CEO coaching, and running them through crawl-walk-run-fly is a near-universal answer to the problem of innovation inside a large company. It works because it removes abstraction layers, creates direct contact with reality, and gives the founder a way to teach muscle memory before delegating. Anyone running a team of any size can borrow the pattern: pick one problem, staff it small, work with it weekly, then let go gradually. The golf-instructor analogy of teaching muscle memory before bad habits set in might be the most important management metaphor of the year.

    His prediction about consumer AI is the most economically interesting part of the talk. The fact that 159 of 175 recent YC companies are enterprise is a startling concentration. If he is right that the next 12 to 24 months bring a consumer renaissance, the opening is enormous. The hard part is what he names directly: there is no proven business model for consumer AI yet. Subscriptions cap out against free incumbents, ads are off-limits for the labs, and e-commerce has been throttled. Solving the business model is probably more valuable than building the next great consumer interface.

    The deeper philosophical thread, that AI is the transition from consumption to creation, is one that anyone building tools for makers should hold close. The 11-Star Experience also reads differently in the AI era. It used to be a thought exercise constrained by what you could plausibly build. AI compresses the gap between imagination and execution to minutes, sometimes seconds. The question is no longer “what is the most absurd version of this experience?” but “which six and seven star experiences can I now industrialize that were unthinkable a year ago?” The exercise has become operational.

    Finally, the meta-lesson on founder-led moats is worth taking seriously. The instinct in venture capital and at most public-company boards is to professionalize early. Chesky’s argument is the opposite: the longer the founder stays in founder mode, the deeper the IP and the longer the company endures after they leave. Disney is the proof. Apple is the proof. Whether Airbnb will be is the open question, and it is the question Chesky is using AI Founder Mode to answer.

  • Howard Marks on Why Most Investors Lose, the AI Bubble, India, and the Hunt for the $10 Bill Nobody Picked Up

    TLDW

    Howard Marks, co-founder of Oaktree Capital and the author of the memos every serious investor reads first, sat down with Nikhil Kamath for a wide-ranging conversation on his 50+ year career, the philosophy of Mujo (the inevitability of change), why he chose bonds over stocks, the difference between drifting down the river and seeing it, where we sit in the current cycle, AI as both threat and opportunity, why active management lost to indexation, and why the only way to outperform in a world full of smart, motivated, computer-literate competitors is “superior insight.” His core message: investing is a puzzle that cannot be solved by formula, and the only edge that lasts is being more right than the other person, more often, with the discipline to stay calm when everyone else is panicking or partying.

    Key Takeaways

    • Mujo is the operating system. Marks took Japanese literature at Wharton and walked away with one idea that shaped his whole career: change is inevitable, unpredictable, and uncontrollable. You cannot predict the future, but you can prepare for it.
    • Cycles are excesses and corrections, not ups and downs. The S&P 500 has averaged about 10% per year for 100 years, but it is almost never between 8% and 12% in any given year. The norm is not the average. Greed and fear push the pendulum past equilibrium every time.
    • The recovery is two years older. When asked where we are in the cycle, Marks notes the bull market continued from April 2024 through January 2026, so by definition we are deeper into the cycle, with a recovery distorted by the unique man-made COVID recession.
    • Drifting versus seeing the river. Marks describes the first 35 years of his career (roughly age 14 to 49) as drifting. Starting Oaktree in 1995 was the first truly intentional decision he made. Entrepreneurship forced proactivity on him.
    • Why bonds over equities. The contractual, predictable nature of debt suited his conservative temperament (his parents were adults during the Depression). He was not voluntarily moved to bonds in 1978; a boss reassigned him just in time for the birth of the high-yield bond market.
    • Distressed debt is the bigger story. Bruce Karsh joined in 1987 and has run roughly $70 billion in distressed debt since 1988, with profits well over 90% of the total profit and loss.
    • Excess return is getting paid more than the risk warrants. If the market thinks a borrower has a 5% default probability and you correctly conclude it is 2%, you collect interest priced for 5% risk while taking 2% risk. That gap is the alpha.
    • Oaktree’s default rate is about a third of the market. Over 40 years, roughly 3.6% to 3.7% of high-yield bonds default each year. Oaktree’s rate is roughly one-third of that, achieved through process discipline, institutional memory, and analysts who stay analysts for life.
    • If you are starting a career today, understand AI. Marks says the investor who will make the most money over the next 10 years is the one who best understands AI and its capabilities, whether they bet for or against it.
    • AI is excellent at pattern matching, but cannot create new patterns. Can AI pick the Amazon out of five business plans? The Steve Jobs out of five CEOs? Marks bets no. Most humans cannot either, which means there is still a role for exceptional people.
    • Indexation won because active management lost. Passive did not become dominant because it is brilliant. It dominated because most active managers failed and charged high fees for the privilege.
    • Bad times create openings for active managers, but most cannot take them. Panic drives prices down, but the same panic prevents most investors from buying. Wally Deemer: when the time comes to buy, you will not want to.
    • The job is simple but not easy. Find the best managers, the best companies, the best ideas. Charlie Munger told Marks: anyone who thinks it is easy is stupid.
    • Where is the $10 bill nobody picked up? Marks thinks it is around AI, but only for those with insight above the average. If you are average and you crowd into AI, you get average results in a bull case and worse in a bear case.
    • Quantitative information about the present cannot produce alpha. Andrew Marks (howards son) pointed this out to his father during the COVID lockdown. Everyone has the same data. Outperformance has to come from somewhere else.
    • Buffett’s edge was reading Moody’s Manuals when nobody else would. The pre-internet research process favored those willing to do tedious work alone. The format of the edge changes; the fact that edge requires doing what others will not, does not.
    • You cannot coach height. Marks can tell you that second-level thinking, contrarian insight, and the ability to evolve at 80 are essential. He cannot tell you how to acquire any of them.
    • India: Marks declines to opine. He has deployed roughly $4 billion in India but refuses to claim expertise on the Indian stock market or recommend a sector.
    • History rhymes. Marks credits Mark Twain. The lessons that repeat are lessons of human nature, which changes incredibly slowly.
    • Investing is a puzzle, not dentistry. Quoting Taleb, Marks observes that engineers and dentists succeed by repeating the right answer. Investors face a problem with no certain solution. If you need to be right every time, do not become an investor.

    Detailed Summary

    From Queens to Wharton: The Accidental Investor

    Howard Marks grew up in Queens, New York, in a middle-class family. Neither of his parents went to college, but his father was an intelligent accountant. Marks discovered accounting in high school, fell in love with its orderliness, and chose Wharton because he was told it was the best undergraduate business school in America. Wharton required a literature class in a foreign country and a non-business minor. For reasons he no longer remembers, Marks chose Japanese studies, then took Japanese civilization and Japanese art. He calls it the most important academic decision of his life because of one concept he encountered: Mujo.

    Mujo, Independence of Events, and Why You Cannot Predict

    Mujo, the turning of the wheel of the law, teaches that change is inevitable, unpredictable, and uncontrollable, and that humans must accommodate it rather than try to control it. Marks pairs this with his deep belief in the independence of events: ten heads in a row do not change the odds on flip eleven. Roughly 20 years ago he wrote a memo titled “You Can’t Predict. You Can Prepare.” A portfolio cannot be optimized for both extreme upside and extreme downside, but it can be built to perform respectably across many possible futures, if you suboptimize for the middle of the probability distribution.

    Why Cycles Exist

    If GDP averages 2% growth, why is it never simply 2%? Marks’s answer is excesses and corrections. Optimism leads producers to overbuild and consumers to overspend, growth runs above trend, then satiation and oversupply pull it back below trend. The S&P 500 averages 10% per year over a century, but the return in any given year is almost never between 8% and 12%. The norm is not the average because human beings are not average; they are alternately greedy and fearful.

    Where Are We Now?

    Two years ago Marks told the Norwegian Sovereign Wealth Fund’s Nicolai Tangen that we were near the middle of the cycle. Two years later, the bull market in stocks continued through January 2026, so by simple math the recovery is older. The COVID recession was a man-made anomaly: one quarter of negative growth followed by the best quarter in history, triggered by a deliberate global shutdown rather than by accumulated excess. That distorts every traditional cycle metric.

    Drifting Versus Seeing the River

    One of the most personal moments in the conversation is Marks’s confession that he drifted for the first 35 years of his career. He did not pick his career, his first job, or his transition from equities to bonds in any deliberate way. Other people pushed him; he said yes. The first proactive decision of his life was co-founding Oaktree in 1995 at age 49, and even that came largely because his wife and his partner Bruce Karsh pushed him into it. Once he had to lead, he had to be intentional. Leadership cannot be passive.

    The Bond Decision

    Marks did not choose bonds; bonds chose him. In May 1978 his boss at Citibank moved him to the bond department to start a convertible fund. Three months later another phone call asked him to figure out something called high-yield bonds being run by a guy in California named Milken. Marks said yes both times. He arrived at the front of the line for high-yield in 1978 and has been there for 48 years.

    The conservative temperament fit. Marks’s parents were adults during the Depression, so he grew up hearing “don’t put all your eggs in one basket” and “save for a rainy day.” Bonds offered contractual, predictable returns. The phrase “junk bonds” was a bias that made the asset class cheaply available to anyone willing to do the analytical work.

    Distressed Debt and Excess Return

    When Bruce Karsh joined in 1987, Oaktree launched what Marks believes was the first distressed debt fund from a mainstream institution. Karsh has managed about $70 billion since 1988 with well over 90% of the total being profit. The core skill is predicting default probability better than the market. If consensus prices a borrower at a 5% default risk and you correctly assess 2%, the interest you receive is overpaid relative to actual risk. Marks calls this “excess return” and credits Mike Milken with the foundational insight: lend to borrowers others will not, demand interest beyond what compensates you, and the math works.

    Over 40 years, roughly 3.6% to 3.7% of high-yield bonds default annually on average. Oaktree’s default rate has been roughly one-third of that. Marks credits institutional culture (analysts who stay analysts for life), psychological stability in volatile periods, and a process that forces every analyst to ask the same eight questions of every company every time. In equity research, you can buy a stock for great management without examining the product, or for a great product without examining the management. In Oaktree’s bond process, you cover every base every time.

    Beginning a Career Today: The AI Question

    Asked what he would do today, Marks says the front of the line is AI. The investor who will succeed most over the next decade is the one who best understands AI, whether they bet for or against it. He notes that he was shocked by his own experience using Claude, but adds that he has not fired a single person and does not intend to.

    His view: AI excels at extracting patterns from history and applying them with discipline and without psychological wobble. But investing also requires creating new patterns. Can AI sit with five business plans and identify the future Amazon? Can it sit with five CEOs and pick Steve Jobs? Marks bets not. Then he adds the killer line: most humans cannot either. Which means the role for exceptional humans survives, but the bar gets higher.

    Why Indexation Won

    When Marks went to graduate school at the University of Chicago in 1968, his professor pointed out that most mutual funds underperformed the S&P after fees. Index funds did not exist yet; Jack Bogle launched the first one in 1974. Today, most equity mutual fund capital is passive. Marks’s controversial take: indexation did not win because it is great. It won because active management was so bad and so expensive. Even at equal fees, if active decisions are inferior, passive wins.

    Bad times create openings for active managers because panic drives prices down, but the same panic prevents most people from buying. Marks quotes the old trader Wally Deemer: when the time comes to buy, you will not want to. The advantage of an AI nudge that says “this is one of those moments, get your ass in gear and buy something” might genuinely add value, because it removes the emotion.

    Second-Level Thinking and Why You Cannot Coach It

    Marks’s first book, The Most Important Thing, has 21 chapters, each titled “The Most Important Thing Is…” Each one is different because so many things matter. The chapter on second-level thinking came to him spontaneously while writing a sample chapter for Columbia University Press. The argument is simple: if you think like everyone else, you act like everyone else, and you get the same results. To outperform, you must deviate from the herd and be more right than the herd. Different is not enough. Different and better is the bar.

    Can AI become a contrarian thinker? You can prompt Claude to give you only non-consensus answers, but the catch is that consensus is often close to right because the people building consensus are intelligent, educated, computer-literate, and motivated. Forcing non-consensus often forces wrong. The real edge is being non-consensus AND correct, which is a much narrower target.

    The $10 Bill That Nobody Has Picked Up

    Marks references the joke about the efficient market hypothesis: there is no $10 bill on the sidewalk because if there were, somebody would have already picked it up. He then concedes that the bill is probably around AI today, but only for those whose insight rises above the average. If you are average and you crowd into AI, you go along with the tide if it works and get crushed if it does not. Quoting Garrison Keillor’s Lake Wobegon, “where all the children are above average,” Marks notes that the math does not allow it. Most investors will not be above average, and acknowledging that is the first step toward becoming one of the few who are.

    Learning From Andrew, Buffett, and Onion-Skin Manuals

    Marks lived with his son Andrew during COVID and wrote a memo about it called “Something of Value” in January 2021. Andrew’s most important contribution was a near-revelation: readily available quantitative information about the present cannot be the source of investment alpha because everyone has it. Buffett’s edge in the 1950s was reading Moody’s Manuals (giant books printed on onion-skin paper with tiny type and zero narrative) when nobody else would. The medium changes; the principle that edge requires doing what others will not, does not.

    India

    Kamath asks Marks directly about India. Marks has deployed roughly $4 billion there but politely declines to claim any expertise on the Indian stock market or recommend a sector. He cautions Kamath about taking advice from people who do not know what they are talking about, and includes himself in that category on the question of India. The honesty is striking and is itself an investment lesson.

    History Rhymes, and Final Advice

    Marks reads Andrew Ross Sorkin’s 1929 and references it in an upcoming memo on private credit. He likes Mark Twain’s reputed line that history does not repeat but it rhymes, and Napoleon’s line that history is written by the winners of tomorrow. The lessons that rhyme are lessons of human nature, which evolves incredibly slowly. Fight or flight from the watering hole still drives behavior in financial markets.

    His final advice: investing is a puzzle, not engineering. A civil engineer calculates steel and concrete, builds the bridge, and the bridge stands. Every time. A dentist fills the cavity correctly and it stays filled. Every time. If you need that kind of reliability in your work, become a dentist. Investing is the act of positioning capital for a future that cannot be predicted accurately. You will be wrong sometimes. If something in your makeup cannot tolerate being wrong sometimes, do not become an investor. The puzzle has no final solution, which is exactly what makes it endlessly interesting.

    Thoughts

    The most useful thing Marks does in this conversation is admit, repeatedly and without ego, what he does not know. He does not know whether AI models differ in real intelligence. He does not know which sector in India to bet on. He does not know how to teach second-level thinking. He drifted for 35 years and only began making intentional decisions at 49. This honesty is the inverse of every guru selling certainty, and it is the actual content of the lesson he is trying to convey: epistemic humility is the precondition for superior insight, because you cannot acquire what you already think you have.

    The deepest insight in the conversation might be the one Andrew Marks (Howard’s son) gave his father during COVID: readily available quantitative information about the present cannot produce alpha because everyone has it. This is devastating in the AI era. If everyone is asking the same large language model the same question, the answers converge, and convergence is consensus, and consensus does not pay. The arms race for proprietary data, novel framings, and unconventional questions is the only thing that can break the convergence.

    Marks’s framing of cycles as excesses and corrections rather than ups and downs is genuinely useful. It reframes volatility from something to fear into something to expect, and reframes the question from “where are we going?” to “how far past trend have we already gone?” The 8 to 12 percent observation about the S&P (that the average return is almost never the actual return) is the kind of fact that should be taught in every introductory finance class but is almost never mentioned.

    The most contrarian claim in the conversation is the one about indexation: that it won because active was bad, not because passive is great. This is a useful inversion. Most defenders of passive investing argue from efficient market theory; Marks argues from the empirical failure of active managers. The implication is that if you can find the small population of active managers who genuinely outperform, the indexation argument falls apart for that subset. Most cannot. The hardest job in investing is the meta-job of identifying the few who can.

    The exchange about AI as a contrarian engine is one of the most clarifying short discussions of AI’s investment limits I have read. Different from consensus is easy. Different and better is the actual goal. Forcing different gets you wrong more often than right because consensus, built by smart, motivated, educated competitors, is usually close to correct. This is why “use AI to find non-consensus ideas” is a worse strategy than it sounds.

    Finally, the Buffett-Moody’s-Manual story is the most quietly profound moment in the interview. The edge in 1955 was the willingness to read tiny type on onion-skin paper alone in an office in Omaha when no one else would. The edge in 2026 is whatever the modern equivalent of that is, and the only honest answer is: nobody knows yet, which is precisely why finding it is worth so much money.

  • Inside Figure: Brett Adcock’s $39 Billion Bet on Humanoid Robots, Helix AI, and the Race to Physical AGI

    Figure is the $39 billion humanoid robotics company most likely to put a general-purpose robot in a commercial workforce, and possibly your living room, before the end of the decade. In a rare two-part sit-down on Sourcery with Molly O’Shea, Founder and CEO Brett Adcock opened every door of the company’s San Jose campus, walked through the manufacturing line, demoed Helix 2 cleaning a living room with no teleoperation, and laid out the plan to scale from thousands of robots in 2026 to a million units a year. He also explained why he fired the OpenAI partnership, why he believes humanoids will reach AGI before any other form factor, and why Figure 04 will be the company’s “iPhone 1 moment.”

    TLDW

    Brett Adcock founded Figure in 2022, self-funded it through a million-a-month burn rate in the first four months, and 15x’d the valuation to $39 billion in 18 months on roughly $2 billion raised from Jeff Bezos, Microsoft, Nvidia, Amazon, and originally OpenAI. The company designs every part in-house, from motors and batteries to the Helix vision-language-action neural network running onboard each robot. Figure deployed humanoids on a BMW assembly line for six months in 2025, hit record production in March 2026, plans to triple that by May, and is targeting a million units per year. Adcock argues that humanoid robotics is an intelligence problem, not a manufacturing problem, that under half of global GDP is human labor (a market measured in tens of trillions of dollars), and that physical interaction data may be the missing ingredient to true artificial general intelligence.

    Key Takeaways

    • Figure is valued at $39 billion after raising nearly $2 billion. Adcock 15x’d the valuation in 18 months and believes the eventual revenue opportunity is in the tens of trillions because roughly half of global GDP is human labor.
    • The bottleneck is intelligence, not manufacturing. Figure already has the parts, the supply chain, and the capacity. The hard part is making robots that run autonomously at human-level performance for 7 to 10 hours a day with zero human intervention.
    • Figure designs almost everything in-house. Motors, rotors, stators, sensors, kinematics, joints, batteries, more than 100 PCBs. Adcock claims no other humanoid group designs more parts than Figure.
    • The OpenAI breakup was about model quality. OpenAI led Figure’s Series B and brought in Microsoft. After a year of collaboration, Adcock says Figure’s internal robot-learning team was running circles around OpenAI on humanoid AI, so he ended the partnership.
    • Helix is Figure’s onboard vision-language-action model. It runs on GPUs in the robot’s torso, ingests camera pixels a few hundred times per second, and outputs joint positions for all ~40 motors. It works without internet connectivity. Helix 2 launched a couple months ago.
    • Robots have more body positions than atoms in the universe. With 40 motors each capable of 360 degrees of rotation, the state space is 360 to the power of 40, which is why Figure abandoned hand-coded controls in favor of neural networks about a year ago.
    • The “Never Fall” protocol is real. A project called Vulcan uses reinforcement learning to keep the robot upright even after losing a knee, ankle, or hip mid-task. The company demoed a robot hobbling on a velocity-locked knee.
    • Figure 03 is the current production robot. It costs roughly 90% less than Figure 02, comes in under $100K per unit, has soft-wrapped foam shoulders, swappable fabric clothing, a high-top sneaker design, and inductive wireless charging at 2 kW through the feet (4 to 5 hours of runtime per 1 hour of charge).
    • Figure 04 is being teased as the “iPhone 1 moment.” Adcock says the jump from Figure 03 to Figure 04 will be the largest generational improvement they have ever made, far bigger than 1 to 2 or 2 to 3.
    • BMW deployed Figure robots for six months in 2025. The robots helped build a BMW X3 in the body shop. Adcock owns the first humanoid-built X3 personally and describes the deployment as the inflection point that led to Helix 2.
    • Home robots will lease for around $400 to $600 a month. Comparable to a car lease. The robot docks itself in a 2-by-2-foot wireless charging station and runs laundry, dishes, and tidying tasks autonomously.
    • Data is the biggest blocker. Figure has roughly 1 million hours of pre-training and mid-training data plus thousands of hours of post-training data. They also pay people in spandex bodysuits to do joint-level human movement capture.
    • Adcock runs three companies simultaneously. Figure (humanoids), Cover (terahertz weapons-detection imaging spun out from NASA Jet Propulsion Lab for K-12 schools), and Hark (an AI lab building personalized AI models and devices, out of stealth two weeks ago).
    • Physical AGI is the explicit goal. Adcock argues that real-world interaction data, learning by touching the world and observing the consequences, is the missing piece for true AGI, and that humanoids may reach it before chatbots do.
    • Security is paranoid by design. A drone was caught hovering outside Figure’s office at one point. They tented the windows, restrict phones in certain areas, and treat industrial CAD and software as high-value IP.

    Detailed Summary

    The Company in Context

    Figure is less than four years old. Adcock founded it in 2022 after stepping away from Archer Aviation, the eVTOL aircraft company he took public. He self-funded Figure to a million dollars a month in burn within four months, hired a 40-person team in four to five months, and pursued a vertically integrated strategy from day one. The thesis is simple. Roughly half of global GDP is human labor. Wages paid to commercial workers run into the tens of trillions of dollars annually. If you can build a humanoid that does general-purpose human work reliably, the resulting business compounds into one of the largest companies in history.

    The campus in San Jose has four buildings: corporate headquarters with 250 to 300 engineers, BotQ (the manufacturing facility), the Grid (a 24/7 robot stress-test environment that runs holidays and weekends), and a design studio that opened to cameras for the first time. Total headcount is around 500. The company has raised close to $2 billion across rounds, with capital from Jeff Bezos, Microsoft, Nvidia, and Amazon. The valuation jumped 15x to $39 billion in 18 months.

    Why Humanoid Robotics Is an Intelligence Problem

    The core technical insight: a humanoid has roughly 40 motors, each capable of full 360-degree rotation, which produces a state space of 360 to the power of 40. That number is larger than the count of atoms in the observable universe. You cannot write hand-coded control logic for that. Figure pivoted entirely from classical controls to neural networks about a year ago, and the team has built what Adcock claims is the best humanoid neural-network controller in the world.

    Helix is a vision-language-action model that runs onboard each robot. It accepts a natural-language prompt like “clean the living room,” reasons through the scene from camera input, and outputs joint commands a few hundred times per second. Inference happens locally on GPUs inside the torso, so the robot keeps working with no internet connection. Helix 2 launched a few months ago following lessons learned from the BMW deployment, and Figure has roughly a million hours of base training data plus thousands of hours of post-training data driving it.

    The OpenAI Partnership and Breakup

    OpenAI led Figure’s Series B alongside Microsoft. The two teams collaborated for roughly a year on running language models on humanoids. Adcock says he got to know Sam Altman and the team well, but over time it became clear that Figure’s internal robot-learning engineers (most with over a decade of experience in the field) were outpacing OpenAI on testing, model training, and integration with humanoid hardware. Adcock also implies OpenAI was getting interested in robotics itself, which created a strategic conflict. He ended the partnership. He is candid about being wrong on the original strategic logic for letting them invest in the first place.

    BotQ: The Humanoid Factory

    BotQ is the assembly facility where Figure 03 robots are born. Lines build heads, batteries, arms, legs, and hands separately. Each subsystem goes through end-of-line testing before integration. Heads contain camera systems, IMU, thermal sensors, Wi-Fi, 5G, Bluetooth, and lights, and are flashed with firmware and calibrated on the line. The 2.25 kilowatt-hour battery pack is custom-designed with a structural enclosure, polyurethane potting, and an internally engineered thermal-runaway venting system. The requirement is that no flame ever exits the pack. Figure has never had a robot catch fire.

    March 2026 was the company’s record production month, more robots built than in the entire prior history of the company combined. Adcock plans to triple that by May. After assembly, robots run a multi-hour “burn-in” in dedicated bays where the robot self-checks for loose cables, comm errors, or bad parts. They wear vests during gantry-supported wakeup. Once they pass, they walk themselves over to headquarters.

    The Grid and the Never-Fall Protocol

    The Grid runs robots 24/7 at higher operational intensity than any client site. It is the last line of defense before software ships. A dedicated team called Never Fall predicts every plausible fault and engineers around it. The Vulcan project takes this further: using reinforcement learning in simulation, robots learn to survive losing a knee, ankle, or hip mid-task. In the demo, a robot’s left knee was velocity-locked (simulating a lost actuator), and the robot continued hobbling around without falling. A backup robot can be summoned to take over the work.

    The Home Robot Demo

    Figure 03 demoed tidying a living room in a home environment built into the campus. The robot was given the prompt “clean the living room” and reasoned through the task autonomously: clearing cups, putting away toys, wiping the table. There was a brief sassy spray during the cleaning sequence. Adcock was emphatic that this is not teleoperated despite persistent online rumors. Helix 2 runs entirely onboard, no human in the loop.

    The product plan for the home is a leasing model in the $400 to $600 per month range, comparable to a car lease. The dock is roughly 2 feet by 2 feet and plugs into a standard wall outlet. Charging happens inductively through the feet at 2 kilowatts, giving roughly 4 to 5 hours of runtime per 1 hour of charge. Figure is not selling to homes yet but plans to soon.

    The Three Generations (and the Fourth)

    Figure 01 was a “cyberpunk” first-generation robot built for speed of iteration, costing hundreds of thousands of dollars per unit. Most parts were CNC-machined to Formula 1 precision. It walked within a year of company founding, which Adcock claims is among the fastest humanoid walking timelines in history. It had a tendon-driven hand (motors in the forearm) which Figure abandoned early. Because the wrist motors were too far along to redesign, the team raided foot motors and stuffed them in the forearm, producing the so-called Frankenstein forearm where the wrist bent halfway up the arm. Adcock was sure people would notice. In three years, no one ever asked.

    Figure 02 moved the battery from a backpack into the torso, doubled the battery, tripled the compute, added new cameras, and used an exoskeleton load-bearing structure inspired by aircraft skin design. Roughly 50 units were built. It was retired about a month before filming.

    Figure 03, the current production model, is roughly 90% cheaper than Figure 02 and slimmer in profile. It has soft foam-wrapped shoulders, swappable fabric clothing (with a zipper down the back), high-top sneakers, and the latest-generation hand with camera-based tactile sensors. The aesthetic was deliberately moved away from “too roboty.” Figure 03 was the first humanoid robot at the White House (greeting guests at an event with the First Lady).

    Figure 04 is in late-stage detailed design. Adcock describes it as the company’s “iPhone 1 moment,” a much larger generational leap than any prior version, with substantial cost reduction, easier manufacturing, easier home setup, and changes Adcock says are too sensitive to discuss publicly.

    Hands and the Path to Physical AGI

    Figure recently teased a high-degree-of-freedom hand with as many joints as a human hand. Adcock argues this is essential not just for dextrous manipulation but for passive learning from humans at scale. If humans can move their hands in arbitrary ways, the robot needs to be able to map onto those movements at test time. He believes the path to AGI in physical embodiment runs through the hands.

    Adcock’s broader claim is that physical interaction data, learning what happens when you touch, push, lift, or drop something, is the missing ingredient that current frontier language models lack. Most human intelligence is built through trial and error in the physical world. If that is true, humanoids may close the gap to AGI before pure software systems do.

    Brett Adcock’s Other Companies

    Cover is a school weapons-detection company spun out of NASA’s Jet Propulsion Lab. It uses terahertz imaging radar (originally developed for the Iraq and Afghanistan wars to find bomb vests at standoff distance) to detect concealed weapons in clothing or backpacks from 5 to 20 meters away, far further than airport scanners. Adcock bought the IP outright two years ago, and Caltech holds a small minority interest. The team is largely former JPL engineers based in Pasadena. Beta deployments to schools are planned by end of year, with 130,000 K-12 schools as the addressable market. Adcock self-funds it.

    Hark is an AI lab Adcock started seven or eight months ago and unveiled two weeks before the interview. It has 50 employees and is building next-generation personalized AI models alongside new AI hardware (the thesis being that 20-year-old form factors like phones and laptops are the wrong interface for AI).

    Operating Philosophy

    Adcock works from the engineering bullpen, not a corner office. He cut the “annual golf trip” category of relationships out of his life five years ago to make space for family and three companies. He goes home for dinner and bedtime with his kids and returns to the office after. He cites Steve Jobs and Jeff Bezos (a Figure investor) as influences and frames his work ethic as wanting to play “11 out of 10.” He maintains tight physical and digital security: a drone was once caught surveilling the office through a window, after which the team tented the glass.

    Risks

    Adcock is direct that the odds of full success are low. The risk list is long: manufacturing at unprecedented rates, robots running fully autonomously without human intervention (which no one has demonstrated), AI policies that generalize across every environment, hardware reliability, low unit cost, consumer demand. He frames his job as a daily funnel of the most pernicious problems in the company.

    He does not see capital or the $39B valuation as the binding constraint. If the robots work, he projects revenue measured in tens of trillions of dollars and points out that tech companies trade at 10 to 20 times revenue.

    Thoughts

    The most interesting structural claim Adcock makes is that humanoid robotics is an intelligence problem, not a manufacturing problem. That is a strong statement about where the difficulty actually lives. If the bottleneck were industrial (parts, supply chain, factory throughput), the dominant strategy would be to wait for incumbents like Foxconn or BYD to enter and underprice everyone. If the bottleneck is intelligence, the dominant strategy is exactly what Figure is doing: integrate vertically, control the hardware, generate proprietary training data, and run a tight feedback loop between deployments and model updates. The BMW deployment producing the lessons that became Helix 2 is the cleanest illustration of that loop in action.

    The 360-to-the-40th state space framing is a useful reminder of why neural networks won this domain. Anything you cannot enumerate, you must learn. The pivot from classical controls to neural networks about a year ago is probably the single highest-leverage decision in the company’s history, and it tracks with the broader collapse of hand-coded systems across robotics, autonomy, and even compilers.

    The OpenAI breakup is more interesting than it first appears. Adcock’s story is not “they were bad,” it is “we got better than them, faster.” That is consistent with a recurring pattern in AI right now: vertically integrated application companies, where the model is the product, are starting to outpace general-purpose model providers on their own narrow domains. If physical AGI does happen first in embodiment, that pattern will look prophetic in retrospect.

    The home leasing model at $400 to $600 per month is the part most people will underestimate. That price point is not luxury. It is roughly the cost of a modest car payment, less than full-time childcare, less than a cleaning service plus a dog walker plus laundry pickup. If the robot can actually do laundry, dishes, and tidying every day with no failures, the consumer math gets aggressive fast. The bottleneck is reliability per hour, not willingness to pay.

    The skeptic’s case is also worth holding in mind. “Working” in a curated demo home is not the same as working in 100,000 messy real homes with cats, kids, weird furniture, and unpredictable lighting. Generalization is exactly the problem Adcock concedes is unsolved. The Vulcan demo (hobbling on a velocity-locked knee) is impressive, but a single failure mode handled is a long way from “never fall” across the full distribution of real-world conditions. The phrase “we want to be able to” appears repeatedly in Adcock’s roadmap, and it is doing a lot of work.

    Still, the velocity is real. Record manufacturing in March, tripling by May, four buildings, 500 employees, vertically integrated parts, a custom battery line, BMW deployment, White House appearance, Time cover, Helix 2 in production, Figure 04 in detailed design. The competitive landscape (Tesla Optimus, 1X, Apptronik, Unitree, and several Chinese entrants) is going to determine whether Figure stays “a few years ahead” of everyone, as Adcock claims, or whether the gap collapses. But if humanoids actually work, this is one of the very few companies positioned to capture the upside, and Adcock has been operating the playbook for almost four years.

    The most underrated detail in the whole tour: Figure 04 is being described internally as the iPhone 1. Figure 03 is the BlackBerry. If that framing holds up, the next 12 to 24 months are when this market gets defined.

  • Elad Gil on the AI Frontier: Compute Constraints, the Personal IPO, and Why Most AI Founders Should Sell in the Next 12 to 18 Months

    Elad Gil sat down with Tim Ferriss for a wide ranging conversation that pairs almost perfectly with his recent Substack post Random thoughts while gazing at the misty AI Frontier. Together, the podcast and the post lay out the cleanest framework I have seen for what is actually happening in AI right now: a Korean memory bottleneck capping every lab, a class wide personal IPO across the research community, the fastest revenue ramps in capitalist history, and a brutal dot com style culling that most founders do not yet want to admit is coming. Below is a complete breakdown.

    TLDW (Too Long, Didn’t Watch)

    Elad Gil argues that AI is producing the fastest revenue ramps in capitalist history while setting up the same brutal power law that wiped out 99 percent of dot com companies. OpenAI and Anthropic each sit at roughly 0.1 percent of US GDP today, on a path to 1 percent of GDP run rate by end of 2026, which is insanely fast by any historical standard. The current ceiling on capabilities is not chips but Korean high bandwidth memory, and that constraint will likely hold all major labs roughly comparable in capability through 2028. Talent has just experienced a class wide personal IPO via Meta led bidding, with packages running tens to hundreds of millions per researcher. Most AI companies should consider exiting in the next 12 to 18 months while the tide is high. Right now consensus is correct. Save the contrarianism for later.

    Key Takeaways

    • OpenAI and Anthropic are each at roughly 0.1 percent of US GDP. With US GDP near 30 trillion dollars and each lab at a roughly 30 billion dollar revenue run rate, AI has gone from essentially zero to 0.25 to 0.5 percent of GDP in just a few years. If the labs hit 100 billion in run rate by year end 2026 (which many expect), AI hits 1 percent of GDP run rate inside a single year.
    • The AI personal IPO is real. 50 to a few hundred AI researchers across multiple companies just experienced a class wide IPO event due to Meta led bidding, with top packages reportedly tens to hundreds of millions per person. The closest historical analog is early crypto holders around 2017.
    • The bottleneck is Korean memory, not Nvidia chips. High bandwidth memory from Hynix, Samsung, Micron, and others is the binding constraint. Expected to hold roughly two years. After that, power and data center buildout become the next walls.
    • No lab can pull dramatically ahead before 2028. Because every lab is compute constrained on the same input, OpenAI, Anthropic, Google, xAI, and Meta should remain roughly comparable in capability through that window, absent an algorithmic breakthrough that stays inside one lab.
    • Compute is the new currency. Token budgets now define what an engineer can accomplish, what a company can spend, and what business models are viable. Some companies (neoclouds, Cursor) are effectively inference providers disguised as tools.
    • The dot com base rate is the AI base rate. Around 1,500 to 2,000 companies went public in the late 1990s internet cycle. A dozen or two survived. AI will likely look the same.
    • Most AI founders should consider selling in the next 12 to 18 months. If you are not in the durable handful, this is your value maximizing window. A handful of companies (OpenAI, Anthropic) should never sell.
    • Buyers are bigger than ever. One percent of a 3 trillion dollar market cap is 30 billion dollars. That math makes massive AI acquisitions trivial for hyperscalers, vertical incumbents, and adjacent giants.
    • Underrated exit path: merger of equals. Two private AI competitors destroying each other on price should consider just merging. PayPal and X.com did exactly this in the 1990s.
    • 91 percent of global AI private market cap sits in a 10 by 10 mile square. If you want to do AI, move to the Bay Area. Remote work for cluster industries is BS.
    • Want money? Ask for advice. Want advice? Ask for money. The inverse also works: offering useful advice frequently leads to inbound investment opportunities.
    • AI is selling units of labor, not software. The shift is from selling seats and tools to selling cognitive output. This is why Harvey can win in legal, where decades of legal SaaS failed.
    • AI eats closed loops first. Tasks that can be turned into testable closed loop systems (code, AI research) get automated fastest. Map jobs on a 2×2 of closed loop tightness vs economic value to see where AI hits soonest.
    • Headcount will flatten at later stage companies. Multiple late stage CEOs told Elad they will not do big AI layoffs but will simply stop growing headcount even as revenue grows 30 to 100 percent. Hidden layoffs are also hitting outsourcing firms in India and the Philippines first.
    • The Slop Age could be the golden era of AI plus humanity. AI produces useful slop at volume, humans desloppify it, leverage is high, and the work is fun. This window may close as AI gets superhuman.
    • Market first, team second (90 percent of the time). Great teams die in bad markets. The exception is when you meet someone truly exceptional at the very earliest stage.
    • The one belief framework. If your investment memo needs three core beliefs to be true, it is too complicated. Coinbase was an index on crypto. Stripe was an index on e-commerce. That was the entire memo.
    • The four year vest is a relic. It exists because in the 1970s companies actually went public in four years. Today the private window has stretched to 20 years and venture has eaten what used to be public market growth investing.
    • Boards are in-laws. You cannot fire investor board members. Take a worse price for a better board member, because as Naval Ravikant said, valuation is temporary, control is forever.
    • Right now, consensus is correct. Save the contrarianism. The smart move is to just buy more AI exposure rather than try to outsmart the obvious.
    • Distribution wins more than founders admit. Google paid hundreds of millions to push the toolbar. Facebook bought ads on people’s own names in Europe. TikTok spent billions on user acquisition. Allbirds (yes, the shoe company) just raised a convert to build a GPU farm.
    • Anti-AI sentiment will get worse before it gets better. Maine banned new data centers. There has been violence directed at AI leaders. Expect more political and activist backlash, especially as AI is blamed for harms it has not yet caused while its benefits are mismeasured.
    • Use AI as a cold reader. Elad uploads photos of founders to AI models with cold reading prompts and reports surprisingly accurate personality assessments based on micro features.

    Detailed Summary

    The Numbers Are Insane and Mostly Underappreciated

    The most stunning data point in either source is the GDP math. US GDP is roughly 30 trillion dollars. OpenAI and Anthropic are each rumored to be at roughly 30 billion dollars in revenue run rate, putting each one at 0.1 percent of US GDP. Add cloud AI revenue and the picture gets stranger: AI has grown from essentially zero to between 0.25 and 0.5 percent of GDP in only a few years. If the labs hit 100 billion in run rate by year end 2026, AI will be at roughly 1 percent of GDP run rate inside a single year. There is no historical analog for that pace. Elad notes that productivity gains from AI may end up mismeasured the way internet productivity was undercounted in the 2000s, which would have downstream consequences for regulation: AI gets blamed for the bad (job losses) and credited for none of the good (new jobs, education gains, healthcare improvements). His half joking aside is that the real ASI test may be the ability to actually measure AI’s economic impact.

    The AI Personal IPO

    The most underdiscussed phenomenon in AI right now, according to Elad, is what he calls a class wide personal IPO. When a company IPOs, a subset of employees become wealthy, lose focus, and either start companies, get into politics, fund passion projects, or check out. Meta started aggressively bidding for AI talent. Other major labs had to match. The result was 50 to a few hundred researchers, scattered across multiple labs, suddenly receiving compensation in the tens to hundreds of millions of dollars range. The only historical analog Elad can think of is early crypto holders around 2017. Some chunk of these newly wealthy researchers will redirect attention to AI for science, side projects, or quiet quitting. The aggregate field stays mission aligned, but the distribution of attention has shifted.

    The Korean Memory Bottleneck

    Every major AI lab today is building giant Nvidia clusters paired with high bandwidth memory primarily from Korean fabs and a few other suppliers. They run massive amounts of data through these clusters for months, and the output is, almost absurdly, a single flat file containing what amounts to a compressed version of human knowledge plus reasoning. Right now, the binding constraint on this whole stack is HBM memory from Hynix, Samsung, Micron, and others. Korean memory fab capacity has been below the capacity of every other piece of the system. Elad estimates this constraint persists for roughly two years. After that, the next walls are likely data center construction and power. The strategic implication is enormous. While memory constrains everyone, no single lab can buy 10x the compute of its rivals, so capabilities should stay roughly comparable across the major labs. Once that constraint lifts, possibly around 2028, one player could theoretically pull dramatically ahead, especially if AI assisted AI research closes a self improvement loop inside one lab.

    Compute Is the New Currency

    The blog post sharpens a framing that runs throughout the podcast: compute, denominated in tokens, is now a unit of economic value. Token budgets define what an engineer can accomplish, what a company can spend, and what business models work. Some companies are effectively inference providers wearing tool costumes. Neoclouds are the cleanest example. Cursor is another, subsidizing inference as a user acquisition strategy. The most absurd recent example: Allbirds, the shoe company, raised a convertible to build a GPU farm. Whether this becomes the AI version of Microstrategy’s Bitcoin trade or a cautionary tale, it tells you where the cost of capital believes the next decade is going.

    The Dot Com Survival Math

    Elad walks through the brutal arithmetic that AI founders should be internalizing. In the late 1990s and early 2000s, somewhere between 1,500 and 2,000 internet companies went public. Of those, roughly a dozen or two survived in any meaningful form. Every cycle has looked like this: automotive in the early 1900s, SaaS, mobile, crypto. There is no reason AI will be different. Most current AI companies, including those ramping revenue today, will see the market, competition, and adoption turn on them. The question every AI founder should be asking is whether they are in the durable handful or not.

    Most AI Companies Should Consider Exiting in the Next 12 to 18 Months

    This is the most actionable and most uncomfortable take in either source. While the tide is rising, every AI company looks unstoppable. Whether they actually are, in a 10 year frame, is a separate question. Founders running successful AI companies should take a cold honest look at whether the next 12 to 18 months is their value maximizing window. Companies typically have a 6 to 12 month peak before some headwind hits, often visible in the second derivative of growth. The best signal that you should sell is when growth rate is starting to plateau and you can see why. A handful of companies (OpenAI, Anthropic, the durable winners) should never exit. Many others should, while everything is still on the upswing.

    What Makes an AI Company Durable

    Elad lays out four lenses for evaluating durability at the application layer:

    1. Does your product get dramatically better when the underlying model gets better, in a way that keeps customers loyal?
    2. How deep and broad is the product? Are you building multiple integrated products embedded in actual workflows?
    3. Are you embedded in real change management at the customer? AI adoption is mostly a workflow change problem, not a tech problem. Workflow embedding is durable.
    4. Are you capturing and using proprietary data in a way that creates a system of record? Data moats are often overstated, but sometimes real.

    At the lab layer, Elad believes OpenAI, Anthropic, and Google are durable absent disaster. He predicted three years ago that the foundation model market would settle into an oligopoly aligned with cloud, and that prediction has roughly held.

    Selling Work, Not Software

    The deepest structural insight in the conversation is that generative AI is shifting what software companies sell. The old model was selling seats, tools, and SaaS subscriptions. The new model is selling units of cognitive labor. Zendesk sold seats to support reps. Decagon and Sierra sell agentic support output. Harvey can win in legal even though selling to law firms was historically considered terrible business, because Harvey is not selling tools, it is augmenting lawyer output. This shift opens markets that were previously closed and dramatically grows tech TAMs. It is also why founder limited theories of entrepreneurship currently understate how many opportunities exist.

    AI Eats Closed Loops First

    One of the cleanest mental models in the blog post is the closed loop framework. AI automates first what can be turned into a testable closed loop. Code is the canonical example: outputs can be tested, errors detected, models can iterate. AI research is similar. Both have tight feedback loops and high economic value, which puts them at the top of the AI impact ranking. Map jobs on a 2×2 of closed loop tightness vs economic value and you can see where AI hits soonest. The interesting forward question is which jobs become more closed loop next. Data collection and labeling will keep growing in every field as a result.

    The Harness Matters More Than People Think

    For coding tools and increasingly for enterprise applications, what Elad calls the harness, the wrapper of UX, prompting, workflow integration, and brand around the underlying model, is becoming sticky. It is not just which model you call. It is the environment built around it. Cursor and Windsurf demonstrate this in coding. The interesting open questions are what the harness looks like for sales AI, for AI architects, for analyst workflows. Those gaps leave room for startups even as model capabilities converge.

    Hidden Layoffs and the Developing World

    Most announced AI driven layoffs are probably just COVID era overhiring corrections wrapped in a more flattering narrative. But real AI driven labor displacement is happening, and it is hitting outsourcing firms first. That means countries like India and the Philippines, where many outsourced services jobs sit, are likely to be the most impacted earliest. Several developing economies built their growth ladders on services exports. If AI takes those jobs first, the migration and economic patterns of the next decade may shift in ways nobody is yet planning for.

    The Flat Company

    Multiple late stage CEOs told Elad they will not announce big AI layoffs. Instead, they will simply stop growing headcount. If revenue grows 30 to 100 percent, headcount stays flat or shrinks via attrition. Existing employees become dramatically more productive. The very best people who can leverage AI will see compensation inflate. Sales and some growth engineering keep hiring. Almost everything else flatlines. This is mostly a later stage and public company phenomenon. True early stage startups should still scale aggressively after product market fit, just with more leverage per person.

    Exit Options for AI Founders

    Elad lays out four exit categories. First, the labs and hyperscalers themselves: Apple, Amazon, Google, Microsoft, Meta. Second, vertical incumbents like Thomson Reuters for legal or healthcare giants for clinical AI. Third, the underrated category of merger of equals between two private AI competitors who are currently destroying each other on price. PayPal and X.com did this in the 1990s. Uber and Lyft reportedly almost did. Fourth, large adjacent tech companies: Oracle, Samsung, Tesla, SpaceX, Snowflake, Databricks, Stripe, Coinbase. The market cap math has changed in a way that makes acquisition trivial. One percent of a three trillion dollar market cap is 30 billion dollars, which means a hyperscaler can do massive acquisitions almost casually.

    Geographic Concentration Is Extreme

    Elad’s team analyzed where private market cap aggregates. Historically half of global tech private market cap sat in the US, with half of that in the Bay Area. With AI, 91 percent of global AI private market cap is in a single 10 by 10 mile square in the Bay Area. New York is a distant second and then it falls off a cliff. For defense tech, the cluster is Southern California (SpaceX, Anduril, El Segundo, Irvine). Fintech and crypto skew toward New York. The remote everywhere advice is, Elad says, just BS for anyone trying to break into an industry cluster.

    How Elad Got Into His Best Deals

    Stripe started with Elad cold emailing Patrick Collison after selling an API company to Twitter. A couple of walks later, Patrick texted that he was raising and Elad was in. Airbnb came from helping the founders raise their Series A and being asked at the end if he wanted to invest. Anduril came from noticing that Google had shut down Project Maven and asking if anyone was building defense tech, then meeting Trey Stephens at a Founders Fund lunch. Perplexity came from Aravind Srinivas cold messaging him on LinkedIn while still at OpenAI. Across all of these, the pattern is the same: be in the cluster, be helpful, be talking publicly about technology nobody else is talking about, and be useful to founders before any money is on the table.

    The One Belief Framework

    Investors love complicated 50 page memos. Elad believes the actual decision usually collapses into a single core belief. Coinbase: this is an index on crypto, and crypto will keep growing. Stripe: this is an index on e-commerce, and e-commerce will keep growing. Anduril: AI plus drones plus a cost plus model will be important for defense. If your thesis needs three things to be true, it is probably not going to work. If it needs nothing, you have no thesis.

    Boards as In-Laws

    Elad emphasizes that founders should treat board composition like one of the most important hiring decisions of the company. You cannot fire an investor board member. They have contractual rights. So if you are going to be stuck with someone for a decade, take a worse valuation for a better human. Reid Hoffman’s framing is that the best board member is a co-founder you could not have otherwise hired. Naval Ravikant’s framing is that valuation is temporary but control is forever. Elad recommends writing a job spec for every board seat.

    The Slop Age as a Golden Era

    One of the warmest takes in the blog post is the framing of the current moment as the Slop Age, and the suggestion that this might actually be the golden era of AI plus humanity. Before the last few years, AI was inaccessible and narrow. Eventually AI may become superhuman at most tasks. Today, AI produces useful slop at volume, which means humans are still needed to desloppify the slop, but the leverage on time and ambition is real. That makes the work fun. If AI displaces people or starts doing more interesting work, this golden moment fades. Elad also notes the obvious counter, that the era of human generated internet slop preceded the AI slop era. AGI may end the slop age, or alternately may be the thing that finally cleans up all the prior waves of human slop.

    Anti-AI Regulation and Violence Will Increase

    This is one of the more sobering threads in the blog post. Real world AI driven labor displacement has been small so far, but anti-AI sentiment is already strong and growing. Maine just banned new data centers. There has been actual violence directed at AI leaders, including a recent attack on Sam Altman. Elad’s view is that AI leaders should work harder on optimistic public framing, real political lobbying, and reining in the doom narrative coming from inside the field. Otherwise the regulatory and activist backlash will get much worse, and likely on the basis of mismeasured impacts.

    Right Now Consensus Is Correct

    The headline contrarian take from the episode is that contrarianism right now is wrong. There are moments in time when betting against the crowd pays. This is not one of them. The smart bet is just buying more AI exposure. Trying to find the clever angle, the underlooked hardware play, the secret macro thesis, is overthinking it. Save the contrarian moves for later in the cycle.

    Distribution Almost Always Matters

    Elad pushes back on the founder mythology that great products win on their own. Google paid hundreds of millions of dollars in the early 2000s to distribute its toolbar through every popular app installer on the internet. Facebook bought search ads against people’s own names in European markets to seed network liquidity. TikTok spent billions on user acquisition before its algorithm could lock people in. Snowflake spent enormous sums on enterprise sales and channel partnerships. Sometimes the best product wins. Often the company with the best distribution wins. Founders should plan for both.

    AI as a Cold Reader and a Research Partner

    Two of the more practical AI workflows Elad describes: First, uploading photos of founders to AI models with cold reading prompts that ask the model to identify micro features (crows feet from genuine smiling, brow patterns, posture cues) and infer personality traits, sense of humor, and likely social behavior. He reports the outputs are surprisingly specific. Second, running deep dives across multiple models in parallel (Claude, ChatGPT, Gemini), asking each for primary sources, summary tables, and cross checked data. He recently used this approach to investigate the rise in autism and ADHD diagnoses, concluding that diagnostic criteria shifts and school incentives drive most of it, and noting that maternal age has a stronger statistical association with autism than paternal age, despite paternal age getting all the public discourse.

    The First Ever 10 Year Plan

    For someone who has been compounding aggressively for two decades, Elad has somehow never written a 10 year plan until now. He knows it will not play out as written. The point is that the act of imagining a decade out shifts what you choose to do in the near term. He explicitly rejects the AGI in two years therefore plans are pointless framing as defeatist. There will be interesting things to do regardless of how the AGI timeline plays out.

    Thoughts

    This is one of the more useful AI investor conversations of 2026, mostly because Elad is willing to put numbers and timelines on things that are usually left vague. Pairing the podcast with the underlying Substack post is the right move because the post is where the GDP math, the closed loop framework, and the Slop Age framing actually live. The podcast is where Elad explains how he thinks rather than just what he thinks.

    The 12 to 18 month sell window framing is the most actionable single idea in either source, and probably the most uncomfortable for AI founders sitting on multi billion dollar paper valuations. The math is unforgiving. A dozen winners out of thousands. If you are honest with yourself about whether you are in the dozen, you know what to do.

    The Korean memory bottleneck framing explains a lot of current behavior. The talent wars make more sense once you accept that compute is not going to be the differentiator for two years, so people become the only remaining lever. The convergence of capabilities across OpenAI, Anthropic, Google, and xAI starts to look less like coincidence and more like the structural inevitability of a supply constrained input. The 2028 inflection date is the one to watch.

    Compute as currency is the cleanest reframing in the blog post. Once you start pricing companies in tokens rather than dollars, everything from Cursor’s economics to Allbirds raising a convert to build a GPU farm becomes legible. The interesting question is whether this is a permanent unit of denomination or a transitional one that fades when inference costs collapse.

    The software to labor argument is the structural framing that I think will hold up the longest. Once you internalize that we are not selling seats anymore but selling cognitive output, every vertical that was previously locked behind ugly procurement and IT inertia opens up. Harvey is the proof of concept. There will be 30 more Harveys across every white collar profession.

    The closed loop framework is the cleanest predictor of which jobs get hit hardest and soonest. If you want to know whether your role is exposed, the questions to ask are whether outputs can be machine evaluated, how tight the feedback loop is, and how high the economic value is. The intersection is where AI lands first.

    The geographic concentration data is genuinely shocking. 91 percent of global AI private market cap in a 10 by 10 mile area is the kind of statistic that should make everyone outside that square think very carefully about what game they are playing.

    The Slop Age framing is the most emotionally honest moment in the post. We are in a window where humans still meaningfully add value on top of AI output. That window is finite. Enjoy it.

    The anti-AI backlash thread is the one I think most people in the industry are still underweighting. Maine banning new data centers is a leading indicator, not a one off. The fact that the impacts are likely to be mismeasured by official statistics makes the political dynamics worse, not better. AI will get blamed for harms it did not cause and credited for none of the gains. If the field’s leaders do not start communicating better and lobbying smarter, the regulatory environment in 2028 will be much worse than in 2026.

    Finally, Elad’s first ever 10 year plan stands out as the most quietly important moment in the episode. The implicit message is that even people who have been compounding aggressively for two decades benefit from forcing a longer time horizon onto their thinking. Most plans fail. The act of planning still changes what you do today.

    Read the original Elad Gil post here: Random thoughts while gazing at the misty AI Frontier. Find Elad on X at @eladgil, on his Substack at blog.eladgil.com, and on his website at eladgil.com. Tim Ferriss publishes the full episode at tim.blog/podcast.

  • How GPT-5, Claude, and Gemini Are Actually Trained and Served: The Real Math Behind Frontier AI Infrastructure

    Reiner Pope, CEO of MatX and former TPU architect at Google, sat down with Dwarkesh Patel for a different kind of episode: a chalk-and-blackboard lecture on how frontier LLMs like GPT-5, Claude, and Gemini are actually trained and served. With nothing but a handful of equations and public API prices, Reiner reverse engineers an astonishing amount of what the labs are doing. If you have ever wondered why Fast Mode costs more, why context length stalls around 200k tokens, why models seem 100x over-trained, or why hyperscalers are pouring half a trillion dollars into memory, this is the most lucid explanation on the internet.

    TLDW

    Frontier LLM economics come down to two simple budgets: compute time and memory time. Once you write the rooflines on a blackboard, almost everything else falls out of them. Optimal batch size is roughly 300 times your sparsity ratio (around 2,000 to 3,000 tokens for a DeepSeek-style model). A new batch “train” departs every 20 milliseconds because that is how long it takes to read HBM end to end. Mixture of experts strongly favors staying inside a single rack, which is why scale-up domains went from 8 GPUs (Hopper) to 72 (Blackwell) to 500-plus (Rubin). Pipeline parallelism solves weight capacity but does nothing for KV cache, and adds painful per-hop latency, which is why Ilya famously said pipelining is not wise. Because of reinforcement learning and inference economics, frontier models are roughly 100x over-trained versus Chinchilla optimal, and a well-tuned model should output roughly as many tokens during deployment as went into its pre-training corpus. API prices leak the rest: Gemini’s 50% premium above 200k tokens reveals where KV memory time crosses weight memory time, prefill being 5x cheaper than decode confirms decode is memory bandwidth bound, and cache hit pricing tiers map directly to HBM, DDR, flash, and (yes) spinning disk. The lecture closes on a beautiful detour about the convergent evolution of neural nets and cryptographic ciphers.

    Key Takeaways

    • Two equations explain almost everything. A roofline analysis comparing compute time to memory fetch time predicts cost, latency, and architectural choices with shocking accuracy.
    • Optimal batch size is about 300 times sparsity. For a DeepSeek model that activates 32 of 256 experts, that lands around 2,000 to 3,000 tokens per batch. Real deployments go a bit higher to leave headroom.
    • The 20 millisecond train. A new batch departs every 20ms because that is how long it takes to read all of HBM once. Worst-case queue latency is roughly 40ms.
    • Fast Mode is just smaller batches. Pay 6x more, get 2.5x faster decode by amortizing weights over fewer users. There is a hard latency floor at the HBM read time.
    • Slow Mode would not save much. Once you are past the optimal batch size, the cost-per-token plateau is dominated by compute, not weight fetches. You cannot meaningfully amortize KV cache because it is unique per sequence.
    • One rack is the natural MoE unit. Expert parallelism wants all-to-all communication, which strongly favors the scale-up network (NVLink) over the scale-out network (roughly 8x slower).
    • Bigger scale-up domains drove model scaling. The jump from 8 (Hopper) to 72 (Blackwell) to 500-plus (Rubin) GPUs per rack increased aggregate memory bandwidth by 8x, which is why trillion-plus parameter models only became viable recently.
    • Pipeline parallelism is overrated for inference. It saves on weight memory capacity but does nothing for KV cache memory. It also adds milliseconds of latency per hop in decode.
    • Why Ilya said pipelining is not wise. Architectural constraints (cross-layer residuals like in Kimi) and the inability to amortize weight loads across micro-batches make pipelining a hassle in training too.
    • The memory wall is real and paradoxical. Hyperscalers reportedly spend 50% of CapEx on memory, yet racks have far more HBM than a trillion-parameter model needs. The capacity is there for KV cache and batch size, not for weights.
    • Frontier models are roughly 100x over-trained vs Chinchilla. When you minimize total cost across pre-training plus RL plus inference, smaller models trained on more data win.
    • Each model should output roughly all human knowledge. If you equalize pre-training and inference compute, the total tokens served by a model during its lifetime should approximate its training corpus. Roughly 150 trillion in, 150 trillion out.
    • API pricing reveals architecture. Gemini’s 50% premium above 200k context, the 5x decode-vs-prefill ratio, and cache duration tiers all leak detailed information about KV size, memory bottlenecks, and storage hierarchy.
    • KV cache is roughly 2KB per token. Solving Gemini’s pricing equation gives a plausible 1.6 to 2 kilobytes per token at 100B active parameters and 200k context.
    • Decode is memory bandwidth bound, prefill is compute bound. The 5x price gap is direct evidence.
    • Cache pricing maps to memory tiers. The 5-minute and 1-hour cache durations probably correspond to flash and spinning disk drain times respectively. LLM serving uses spinning disk.
    • Context length is stuck near 200k. Memory bandwidth, not compute, is the binding constraint. Sparse attention gives a square-root improvement but is not infinite.
    • Cryptography and neural nets are mathematical cousins. Both rely on jumbling information across inputs. Feistel ciphers led directly to RevNets (reversible neural networks). Adversarial attacks mirror the cipher avalanche property.

    Detailed Summary

    The Roofline: Compute Time vs Memory Time

    Reiner starts with the simplest possible model of LLM inference. The time to do a forward pass is bounded below by the maximum of compute time and memory fetch time. Compute time is the batch size times active parameters divided by FLOPs. Memory time is total parameters divided by memory bandwidth, plus a KV cache term that scales with batch size and context length. From these two equations, almost every economic and architectural fact about modern LLMs can be derived.

    Plotting cost per token against batch size gives a clean picture: at low batch you pay enormous overhead because you cannot amortize the weight fetches, and at high batch you hit a compute floor. There is a sweet spot where memory bandwidth time equals compute time. That sweet spot is what Fast Mode and Slow Mode are tuning around.

    Why Fast Mode Costs More: The Batch Trade-Off

    When Claude Code or Codex offers Fast Mode at 6x the price for 2.5x the speed, what is really happening is that they are running you at a smaller batch size. Smaller batch means weight loads are amortized over fewer users, so cost per token goes up. But latency goes down because each forward pass touches less data. There is a hard floor on latency because you have to read every byte of HBM at least once per token, and that takes about 20 milliseconds on Blackwell-class hardware. There is also a soft ceiling on Slow Mode savings because the unamortizable parts (KV cache fetches, compute) eventually dominate.

    The 20 Millisecond Train

    HBM capacity divided by HBM bandwidth lands consistently around 20 milliseconds across generations of Nvidia hardware. That is the natural cadence at which a frontier model can run a forward pass over all its weights. Reiner uses a memorable analogy: a train departs every 20 milliseconds. Any users whose requests are ready board the train. If the train is full, they wait. If it is empty, it leaves anyway. This is why you do not need millions of concurrent users to saturate a model’s batch. You only need enough to fill a 2,000-token train every 20ms.

    Why Optimal Batch Size Is About 300 Times Sparsity

    Setting compute time equal to weight fetch time and rearranging gives a beautiful result: batch size needs to be greater than (FLOPs / memory bandwidth) times (total params / active params). The hardware ratio is a dimensionless 300 on most GPUs and has stayed remarkably stable from A100 through Hopper, Blackwell, and Rubin. The model term is just the sparsity ratio. For DeepSeek with 32 of 256 experts active, that is 8. So optimal batch is around 2,400 tokens. Real deployments push this to 3x to leave headroom for non-ideal efficiency. At 64 trains per second, that is roughly 128,000 tokens per second per replica, or about 1/1000 of Gemini’s reported global throughput.

    Mixture of Experts Wants to Live Inside a Rack

    MoE all-to-all routing means every token can be sent to any expert on any GPU. The communication pattern strongly prefers the fast scale-up network (NVLink) inside a rack to the slower scale-out network between racks. Scale-out is roughly 8x slower in bandwidth. This is why one rack ends up being the natural unit for an expert layer, and why Nvidia’s progression from 8 GPUs per rack (Hopper) to 72 (Blackwell) to 500-plus (Rubin) has been such a big deal for model size scaling.

    Reiner walks through the physical constraints: cable density, bend radius, weight, power, cooling. Modern racks are pushing every dimension to the limit. Stuffing more GPUs into the scale-up domain is genuinely a hardware engineering problem.

    Pipeline Parallelism: Why Ilya Said It Is Not Wise

    Pipelining splits model layers across racks. It is the natural way to scale beyond the scale-up domain for very large models. But it has problems. In inference, pipelining does not save runtime, it only saves memory capacity per rack, which already is not the binding constraint because trillion-parameter models only need a terabyte and racks have 10x that. In training, pipelining creates the famous bubble (idle GPU time at the start and end of each pipeline pass) and forces micro-batching, which kills your ability to amortize weight loads across the global batch.

    There is also an architectural cost. Models like Kimi use cross-layer residual connections where attention attends to layers a few back, and pipelining makes those patterns very hard to implement cleanly. Ilya’s quip “as we now know, pipelining is not wise” captures all of this.

    The Memory Wall Paradox

    Industry analysts report that hyperscalers are spending 50% of CapEx on memory this year, while smartphones and laptops are seeing 30% volume drops because there is not enough HBM and DDR to go around. Yet a Blackwell rack already has tens of terabytes of HBM, far more than a trillion-parameter model needs. The reason is that all that extra capacity goes to KV cache, batch size, and longer context. The bandwidth, not the capacity, is what matters most for weight loading. This also implies that hardware could be designed with less HBM per GPU if you commit to pipelining the weights, which is a real architectural option for a chip startup like MatX.

    Reinforcement Learning and the 100x Over-Training of Frontier Models

    Chinchilla scaling laws say a model with N active parameters should be trained on roughly 20N tokens for compute-optimal training. But frontier labs do not just minimize training cost. They minimize training plus inference cost across the model’s deployment lifetime. With reinforcement learning added to the mix, the cost equation has three terms: pre-training (6 times active params times tokens), RL (somewhere between 2x and 6x times active params times RL tokens, with a 30% efficiency penalty for decode-heavy rollouts), and inference (2 times active params times inference tokens).

    If you assume those three roughly equalize at the optimum (a heuristic that holds for many cost curves), you get a clean conclusion: the data going into pre-training should be roughly equal to the data going into RL, which should be roughly equal to the tokens served at inference. With 100 billion active parameters and roughly 150 trillion training tokens, that is about 75x past Chinchilla optimal. Reiner rounds it to 100x. This is the most concrete first-principles argument for why frontier models are so deeply over-trained, and it implies that as inference traffic grows, models should keep getting smaller and longer-trained.

    Each Model Should Output All of Human Knowledge

    The most jaw-dropping consequence: if you equalize pre-training and inference compute, then the total tokens generated by a model across its deployment lifetime should approximate the size of its training corpus. GPT-5, served to hundreds of millions of users for two months, will collectively output something on the order of 150 trillion tokens. That is roughly the sum of human knowledge in textual form. Each frontier model is, in this sense, a one-shot universal author of a corpus the size of its source material.

    API Prices Leak Architecture

    This is where the lecture gets really fun. Gemini 3.1 charges 50% more for context above 200k tokens. Setting memory time equal to compute time at exactly 200k context and solving for KV cache size gives roughly 1.6 to 2 kilobytes per token, which is plausible for a model with 8 KV heads, dense attention, and head dimension of 128.

    The 5x premium for output (decode) tokens versus input (prefill) tokens is direct evidence that decode is severely memory bandwidth bound and prefill is compute bound. Prefill processes many tokens per weight load, so it amortizes memory cost over the whole sequence. Decode processes one token per weight load, so it pays full memory cost every time.

    Cache hits priced at one tenth of cache misses tell you that storing the KV cache in HBM (or DDR or flash) is much cheaper than recomputing it from scratch. The two cache duration tiers (5 minutes and 1 hour) probably correspond to memory tiers whose drain times match those durations: flash for the 5-minute tier, spinning disk for the 1-hour tier. Yes, spinning disk is in the modern LLM serving stack, despite being decades-old technology.

    Why Context Length Has Plateaued at 200k

    Context lengths shot up from 8k to roughly 200k during the GPT-3 to GPT-4 era and have stayed roughly flat for the past two years. Reiner argues this is the natural balance point where memory bandwidth cost crosses compute cost. Going to a million tokens is expensive. Going to 100 million tokens (which Dario has hinted is needed for true continual learning via in-context learning) is essentially impossible without either a memory technology breakthrough or a much more aggressive sparse attention scheme. Sparse attention helps with a square-root improvement, but it is not unlimited. Going too sparse trades off too much quality.

    Cryptography Meets Neural Nets

    The episode ends with a lovely intellectual detour. Cryptographic protocols and transformer architectures both rely on jumbling information across all inputs. They are doing inverse versions of the same operation: ciphers take structured input and produce randomness, while neural nets take noisy input and extract structure. Both fields use differentiation as their primary attack vector (differential cryptanalysis on ciphers, gradient descent on neural nets). Adversarial attacks on image classifiers exploit exactly the avalanche property that good ciphers are designed for.

    The most concrete crossover: Feistel ciphers, which let you build invertible functions out of non-invertible ones, were ported into deep learning as RevNets (reversible networks) in 2017. RevNets let you run the entire network backwards during the backward pass, eliminating the need to store activations and dramatically reducing training memory footprint. It is the opposite trade-off of KV caching: spending compute to save memory rather than spending memory to save compute.

    Thoughts

    The most striking thing about this episode is how much can be deduced from a few equations and the public API price sheets of the major labs. The labs treat their architectures as trade secrets, but the moment they price tokens to be close to cost (which competition forces them to do), the prices themselves leak the underlying ratios. Anyone with a pen and paper can reverse engineer the KV cache size, the memory tier hierarchy, and the compute-vs-memory bottleneck profile of a frontier model. There is a lesson here for builders: in competitive markets, the prices tell you almost everything.

    The 100x over-training result has interesting implications for what comes next. If the optimal balance shifts further toward inference (as adoption keeps growing), models should get smaller and longer-trained. That is good news for serving costs and bad news for training-compute-as-moat. The biggest determinant of model quality might increasingly be data quality and RL environment design, not raw pre-training compute. This squares with what is visible publicly: the leading labs are investing heavily in RL infrastructure, evaluations, and synthetic data pipelines.

    The memory wall is the most underrated infrastructure story in AI. Most people think of compute as the bottleneck, but Reiner makes it clear that memory bandwidth is what actually limits context length, which limits how agentic a model can be in practice. If you cannot get to 100 million token contexts, you probably cannot have an AI agent that has been working with you for a month and remembers everything. Either some sparse attention scheme has to give us cheap effective context length, or we need a memory hardware breakthrough, or we have to invent some form of continual learning that does not rely on context windows. None of those paths are obviously easy, and the fact that context length has been flat for two years despite enormous investment suggests we are stuck against a real wall.

    The cryptography parallel is the kind of cross-disciplinary insight that does not show up enough in AI discourse. Treating neural networks as a kind of differentiable cipher reframes a lot of the architecture choices (residual connections, layer normalization, attention) as deliberate efforts to make the function smooth and invertible enough to learn, in contrast to ciphers, which are deliberately designed to resist exactly that. Adversarial robustness research probably has a lot more to learn from cryptanalysis than it currently does.

    Finally, the format itself is a win. Most AI podcasts are conversational, which is great for personality but bad for technical depth. A blackboard lecture with an interlocutor who asks naive questions at the right moments is a much higher bandwidth medium. More of this, please.

  • Jensen Huang on Nvidia’s Supply Chain Moat, TPU Competition, China Export Controls, and Why Nvidia Will Not Become a Cloud (Dwarkesh Podcast Summary)

    TLDW (Too Long, Didn’t Watch)

    Jensen Huang sat down with Dwarkesh Patel for over 90 minutes covering Nvidia’s supply chain dominance, the TPU threat, why Nvidia will not become a hyperscaler, whether the US should sell AI chips to China, and why Nvidia does not pursue multiple chip architectures at once. Jensen framed Nvidia’s entire business as transforming “electrons into tokens” and argued that Nvidia’s real moat is not any single technology but the full stack ecosystem it has built over two decades. He was blunt about his regret over not investing in Anthropic and OpenAI earlier, passionate about keeping the American tech stack dominant worldwide, and dismissive of the idea that China’s chip industry can be meaningfully contained through export controls.

    Key Takeaways

    1. Nvidia’s moat is the ecosystem, not the chip. Jensen repeatedly emphasized that Nvidia’s competitive advantage comes from CUDA, its massive installed base, its deep partnerships across the entire supply chain, and the fact that it operates in every cloud. The moat is not a single product but an interlocking system that took 20+ years to build.

    2. Supply chain bottlenecks are temporary, energy bottlenecks are not. Jensen argued that CoWoS packaging, HBM memory, EUV capacity, and logic fabrication bottlenecks can all be resolved in two to three years with the right demand signal. The real constraint on AI scaling is energy policy, which takes far longer to fix.

    3. TPUs and ASICs are not an existential threat to Nvidia. Jensen was emphatic that no competitor has demonstrated better price-performance or performance-per-watt than Nvidia, and challenged TPU and Trainium to prove otherwise on public benchmarks like InferenceMAX and MLPerf. He described Anthropic as a “unique instance, not a trend” for TPU adoption.

    4. Jensen regrets not investing in Anthropic and OpenAI earlier. He admitted he did not deeply internalize how much capital AI labs needed and that traditional VC funding was not sufficient for companies at that scale. He described this as a clear miss, though he said Nvidia was not in a position to make multi-billion dollar investments at the time.

    5. Nvidia will not become a hyperscaler. Jensen’s philosophy is “do as much as needed, as little as possible.” Building cloud infrastructure is something other companies can do, so Nvidia supports neoclouds like CoreWeave, Nebius, and Nscale instead of competing with them. Nvidia invests in ecosystem partners rather than vertically integrating into cloud services.

    6. Jensen is strongly against US chip export controls on China. This was the longest and most heated segment of the interview. Jensen argued that China already has abundant compute, energy, and AI researchers, and that export controls have accelerated China’s domestic chip industry while causing the US to concede the world’s second-largest technology market. He compared the situation to how US telecom policy allowed Huawei to dominate global telecommunications.

    7. AI will cause software tool usage to skyrocket, not collapse. Jensen pushed back on the narrative that AI will commoditize software companies. He argued that agents will use existing tools at massive scale, causing the number of instances of products like Excel, Synopsys Design Compiler, and other enterprise tools to grow exponentially.

    8. Nvidia does not pick winners among AI labs. Jensen explained that Nvidia invests across multiple foundation model companies simultaneously and refuses to favor any single one. He cited his own company’s unlikely survival story as the reason for this humility: Nvidia’s original graphics architecture was “precisely wrong” and would have been counted out by anyone picking winners.

    9. Nvidia added Groq for premium token economics. Nvidia recently acquired Groq and is folding it into the CUDA ecosystem because the market is now segmenting into different token tiers. Some customers will pay premium prices for faster response times even at lower throughput, creating a new segment of the inference market.

    10. Without AI, Nvidia would still be very large. Jensen was clear that accelerated computing, not AI specifically, is the foundational mission of the company. Molecular dynamics, quantum chemistry, computational lithography, data processing, and physics simulation all benefit from GPU acceleration regardless of deep learning.

    Detailed Summary

    Nvidia’s Real Business: Electrons to Tokens

    Jensen opened the conversation by reframing Nvidia’s entire value proposition. When Dwarkesh suggested that Nvidia is fundamentally a software company that sends a GDS2 file to TSMC for manufacturing, Jensen pushed back hard. He described Nvidia’s job as transforming electrons into tokens, with everything in between representing an “incredible journey” of artistry, engineering, science, and invention. He said the transformation is far from deeply understood and the journey is far from over, making commoditization unlikely.

    Jensen described Nvidia as operating a philosophy of doing “as much as necessary and as little as possible.” Whatever Nvidia does not need to do itself, it partners with someone else and makes it part of the broader ecosystem. This is why Nvidia has what Jensen called probably the largest ecosystem of partners in the industry, spanning the full supply chain upstream and downstream, application developers, model makers, and all five layers of the AI stack.

    On the question of whether AI will commoditize software companies, Jensen offered a contrarian take. He argued that agents are going to use software tools at unprecedented scale, meaning the number of instances of products like Excel, Cadence design tools, and Synopsys compilers will skyrocket. Today the bottleneck is the number of human engineers. Tomorrow, those engineers will be supported by swarms of agents exploring design spaces and using the same tools humans use today. Jensen said the reason this has not happened yet is simply that the agents are not good enough at using tools. That will change.

    The Supply Chain Moat

    Dwarkesh pressed Jensen on Nvidia’s reported $100 billion (and potentially $250 billion) in purchase commitments with foundries, memory manufacturers, and packaging companies. The question was whether Nvidia’s real moat for the next few years is simply locking up scarce upstream components so that no competitor can get the memory and logic they need to build alternative accelerators.

    Jensen confirmed this is a significant advantage but framed it differently. He said Nvidia has made enormous explicit and implicit commitments upstream. The implicit commitments matter just as much: Jensen personally meets with CEOs across the supply chain to explain the scale of the coming AI industry, convince them to invest in capacity, and assure them that Nvidia’s downstream demand is large enough to justify that investment. Nvidia’s GTC conference serves this purpose too, bringing the entire ecosystem together so upstream suppliers can see downstream demand and vice versa.

    Jensen described a process of systematically “prefetching bottlenecks” years in advance. CoWoS advanced packaging was a major bottleneck two years ago, but Nvidia swarmed it with repeated doubling of capacity until TSMC recognized it as mainstream computing technology rather than a specialty product. More recently, Nvidia has invested in the silicon photonics ecosystem through partnerships with Lumentum and Coherent, invented new packaging technologies, licensed patents to keep the supply chain open, and even invested in new testing equipment like double-sided probing.

    When Dwarkesh asked about the ultimate physical bottlenecks, Jensen surprised him. The hardest bottleneck to solve is not CoWoS or HBM or EUV machines. It is plumbers and electricians needed to build data centers. Jensen used this as a launching point to criticize “doomers” who discourage people from pursuing careers in software engineering or radiology, arguing that scaring people out of these professions creates the real bottlenecks.

    On EUV and logic scaling specifically, Jensen was optimistic. He said no supply chain bottleneck lasts longer than two to three years. Once you can build one of something, you can build ten, and once you can build ten, you can build a million. The key is a clear demand signal. If TSMC is convinced of the demand, ASML will produce enough EUV machines. Meanwhile, Nvidia continues to improve computing efficiency by 10x to 50x per generation through architecture, algorithms, and system design.

    The TPU Question

    Dwarkesh pushed hard on whether Google’s TPUs represent a real threat, noting that two of the top three AI models (Claude and Gemini) were trained on TPUs. Jensen drew a sharp distinction between what Nvidia builds and what a TPU is. Nvidia builds accelerated computing, which serves molecular dynamics, quantum chromodynamics, data processing, fluid dynamics, particle physics, and AI. A TPU is a tensor processing unit optimized for matrix multiplies. Nvidia’s market reach is far greater than any TPU or ASIC can possibly have.

    Jensen emphasized programmability as Nvidia’s core architectural advantage. If you want to invent a new attention mechanism, build a hybrid SSM model, fuse diffusion and autoregressive techniques, or disaggregate computation in a novel way, you need a generally programmable architecture. The only way to achieve 10x or 100x performance leaps (versus the roughly 25% per year from Moore’s Law) is to fundamentally change the algorithm, and that requires the flexibility CUDA provides.

    On the specific question of whether hyperscalers with huge engineering teams can simply write their own kernels and bypass CUDA, Jensen acknowledged they do write custom kernels but argued that Nvidia’s engineers still routinely deliver 2x to 3x speedups when they optimize a partner’s stack. He described Nvidia’s GPUs as “F1 racers” that anyone can drive at 100 mph, but extracting peak performance requires deep architectural expertise. Nvidia uses AI itself to generate many of its optimized kernels.

    Jensen was particularly blunt about public benchmarks. He pointed to Dylan Patel’s InferenceMAX benchmark and said neither TPU nor Trainium has been willing to demonstrate their claimed performance advantages on it. He said Nvidia’s performance-per-TCO is the best in the world, “bar none,” and challenged anyone to prove otherwise.

    Regarding Anthropic’s multi-gigawatt deal with Broadcom and Google for TPUs, Jensen called it “a unique instance, not a trend.” He said without Anthropic, there would be essentially no TPU growth and no Trainium growth. He traced this back to his own mistake: when Anthropic and OpenAI needed multi-billion dollar investments from their compute suppliers to get off the ground, Nvidia was not in a position to provide that capital. Google and AWS were, and in return, Anthropic committed to using their compute.

    Nvidia’s Investment Strategy and Regrets

    Jensen was unusually candid about his regret over not investing in foundation model companies earlier. He said he did not deeply internalize how different AI labs were from typical startups. A traditional VC would never put $5 to $10 billion into a single AI lab, but that was exactly what companies like OpenAI and Anthropic needed. By the time Jensen understood this, Nvidia was not in a financial or cultural position to make those kinds of investments.

    Now, Nvidia has invested approximately $30 billion in OpenAI and $10 billion in Anthropic. Jensen said he is delighted to support both and considers their existence essential for the world. But he acknowledged that these investments came at much higher valuations than would have been possible years earlier.

    Jensen explained Nvidia’s broader investment philosophy: support everyone, do not pick winners. He invests in one foundation model company, he invests in all of them. This comes from hard-won humility. When Nvidia started, there were 60 3D graphics companies. Nvidia’s original architecture was “precisely wrong” and the company would have been at the top of most lists to fail. Jensen said he has enough humility from that experience to know that you cannot predict which AI company will ultimately succeed.

    Why Nvidia Will Not Become a Hyperscaler

    Dwarkesh pointed out that Nvidia has the cash to build and operate its own cloud infrastructure, bypassing the middleman ecosystem that converts CapEx into OpEx for AI labs. Jensen rejected this path based on his core operating philosophy.

    If Nvidia did not build its computing platform, NVLink, and the CUDA ecosystem, nobody else would have done it. He is “completely certain” of that. These are things Nvidia must do. But the world has lots of clouds. If Nvidia did not build a cloud, someone else would show up. So the answer is to support the ecosystem instead: invest in CoreWeave, Nscale, Nebius, and others to help them exist and scale, rather than competing with them.

    Jensen was clear that Nvidia is not trying to be in the financing business either. When OpenAI needed a $30 billion investment before its IPO, Nvidia stepped up because OpenAI needed it and Nvidia deeply believed in the company. But these are targeted ecosystem investments, not a strategic pivot into cloud services.

    On GPU allocation during shortages, Jensen pushed back on the narrative that Nvidia strategically “fractures” the market by giving allocations to smaller neoclouds. He said the process is straightforward: you forecast demand, you place a purchase order, and it is first in, first out. Nvidia never changes prices based on demand. Jensen said he prefers to be dependable and serve as the foundation of the industry rather than extracting maximum short-term value.

    The China Debate

    The longest and most heated section of the interview was Jensen’s case against US chip export controls on China. This was a genuine debate, with Dwarkesh pushing the national security argument and Jensen pushing back forcefully.

    Jensen’s core argument rested on several pillars. First, China already has abundant compute. They manufacture 60% or more of the world’s mainstream chips, have massive energy infrastructure (including empty data centers with full power), and employ roughly 50% of the world’s AI researchers. The threshold of compute needed to build models like Anthropic’s Mythos has already been reached and exceeded by China’s existing infrastructure.

    Second, export controls have backfired. They accelerated China’s domestic chip industry, forced their AI ecosystem to optimize for internal architectures instead of the American tech stack, and caused the United States to concede the second-largest technology market in the world. Jensen compared this directly to how US telecom policy allowed Huawei to dominate global telecommunications infrastructure.

    Third, Jensen argued that AI is a five-layer stack (energy, chips, computing platform, models, applications) and the US needs to win at every layer. Fixating on one layer (models) at the expense of another layer (chips) is counterproductive. If Chinese open source AI models end up optimized for non-American hardware and that stack gets exported to the global south, the Middle East, Africa, and Southeast Asia, the US will have lost something far more valuable than whatever marginal compute advantage the export controls provided.

    Dwarkesh countered with the Mythos example: Anthropic’s new model found thousands of high-severity zero-day vulnerabilities across every major operating system and browser, including one that had existed in OpenBSD for 27 years. If China had enough compute to train and deploy a model like Mythos at scale before the US could prepare, the cyber-offensive capabilities would be devastating.

    Jensen’s response was direct. Mythos was trained on “fairly mundane capacity” that is already abundantly available in China. The amount of compute is not the bottleneck for that kind of breakthrough. Great computer science is, and China has no shortage of brilliant AI researchers. He pointed to DeepSeek as evidence: most advances in AI come from algorithmic innovation, not raw hardware. If China’s researchers can achieve breakthroughs like DeepSeek with limited hardware, imagine what they could do with more.

    Jensen also argued for dialogue over confrontation. He said it is essential that American and Chinese AI researchers are talking to each other, and that both countries agree on what AI should not be used for. The idea that you can prevent AI risks by cutting off chip sales, when the real advances come from algorithms and computer science, reflects a fundamental misunderstanding of how AI progress works.

    The debate ended without resolution, but Jensen’s final point was sharp: “I’m not talking to somebody who woke up a loser. That loser attitude, that loser premise, makes no sense to me.”

    Why Not Multiple Chip Architectures?

    Near the end of the interview, Dwarkesh asked why Nvidia does not run multiple parallel chip projects with different architectures, like a Cerebras-style wafer-scale design or a Dojo-style huge package, or even one without CUDA.

    Jensen’s answer was simple: “We don’t have a better idea.” Nvidia simulates all of these alternative approaches in its internal simulators and they are provably worse. The company works on exactly the projects it wants to work on. If the workload were to change dramatically (not just the algorithms, but the actual market shape), Nvidia might add other accelerators.

    In fact, Nvidia recently did exactly this by acquiring Groq. The inference market is now segmenting into different tiers. Some customers will pay premium prices for extremely fast response times even if throughput is lower. This creates a new “high ASP token” segment that justifies a different point on the performance curve. But Jensen was clear: if he had more money, he would put it all behind Nvidia’s existing architecture, not diversify into alternatives.

    Nvidia Without AI

    Jensen closed by saying that even if the deep learning revolution had never happened, Nvidia would be “very, very large.” The premise of the company has always been that general-purpose computing cannot scale indefinitely and that domain-specific acceleration is the way forward. Molecular dynamics, seismic processing, image processing, computational lithography, quantum chemistry, and data processing all benefit from GPU acceleration regardless of AI. Jensen said the fundamental promise of accelerated computing has not changed “not even a little bit.”

    Thoughts

    This interview is one of the most revealing Jensen Huang conversations in years, partly because Dwarkesh actually pushes back instead of lobbing softballs. A few things stand out.

    The Anthropic regret is real and significant. Jensen is essentially admitting that Nvidia’s biggest strategic miss of the AI era was not understanding that foundation model companies needed supplier-level capital commitments, not VC funding. The fact that Google and AWS used compute investments to lock in Anthropic’s architecture choices has had downstream consequences that Nvidia is still working to unwind. When Jensen says Anthropic is “a unique instance, not a trend” for TPU adoption, he is simultaneously downplaying the threat and revealing exactly how seriously he takes it.

    The China debate is the highlight. Jensen’s argument is more nuanced than it first appears. He is not saying “sell China everything.” He is saying the current binary approach of near-total restriction has backfired by accelerating China’s domestic chip industry and pushing the Chinese AI ecosystem away from the American tech stack. His comparison to the US telecom industry losing global market share to Huawei is pointed and historically grounded. Whether you agree with his conclusion or not, the framing of AI as a five-layer stack where the US needs to compete at every layer is a useful mental model.

    The “electrons to tokens” framing is Jensen at his best. It is a simple metaphor that captures something genuinely complex about where value is created in the AI supply chain. And his insistence that the transformation is “far from deeply understood” is a subtle way of arguing that Nvidia’s competitive position will be durable because the problem space is not close to being solved.

    The Groq acquisition reveal is interesting for what it signals about the inference market. If Nvidia is creating a separate product tier for premium-priced, low-latency tokens, it suggests the company sees inference economics fragmenting significantly. This aligns with the broader trend of AI becoming an enterprise product where different customers have wildly different willingness to pay based on how they use tokens.

    Finally, Jensen’s refusal to diversify chip architectures is a bold bet. “We simulate it all in our simulator, provably worse” is an incredibly confident statement. History is full of companies that were right until they were not. But Nvidia’s track record of 50x generation-over-generation improvements through co-design across processors, fabric, libraries, and algorithms is hard to argue with. The question is whether the current paradigm of transformer-based models on GPU clusters represents a local or global optimum for AI compute.

  • Anthropic’s Growth Strategy Explained: $1B to $19B in 14 Months, Automating Experiments With Claude, and Why Old Playbooks Are Dead (Lenny’s Podcast Recap)

    TLDW (Too Long, Didn’t Watch)

    Amol Avasare, Head of Growth at Anthropic, sat down with Lenny Rachitsky to explain how Anthropic grew from $1 billion to over $19 billion in annual recurring revenue in just 14 months. He breaks down their internal tool called CASH (Claude Accelerates Sustainable Hypergrowth) that automates growth experimentation, why 50 to 70 percent of traditional growth playbooks are now obsolete, why the PM-to-engineer ratio may need to flip, and how Anthropic’s early bet on AI coding created a research flywheel that competitors are only now starting to copy. He also shares how he cold emailed his way into the job, why activation is the single hardest problem in AI products, and how he uses Cowork to detect team misalignment across Slack channels automatically.

    Key Takeaways

    1. Anthropic’s growth trajectory is historically unprecedented. Revenue went from $1 billion at the start of 2025 to over $19 billion ARR by February 2026. That 19x growth in 14 months dwarfs companies like Atlassian, Snowflake, and Palantir, which took 15 to 20 years to reach $4.5 to $6 billion ARR. The number Amol quoted was already outdated by the time the episode aired.

    2. Anthropic is automating growth experimentation with an internal tool called CASH. CASH stands for Claude Accelerates Sustainable Hypergrowth. The growth platform team uses Claude to identify opportunities, build experiments (mostly copy changes and minor UI tweaks so far), test them against quality and brand standards, and analyze results. Amol describes the current win rate as roughly equivalent to a junior PM with two to three years of experience, but notes it was not possible at all before Opus 4.5 and has improved significantly with Opus 4.6. Human review is still in the loop but decreasing week over week.

    3. Activation is the single highest-leverage growth problem in AI. The core challenge is capability overhang: models are improving so fast that users do not know what they can do. By the time you have tested and optimized onboarding for one model’s capabilities, the next model has already shipped with entirely new features that make your learnings obsolete. Anthropic addresses this by adding intentional friction in onboarding to understand who users are and funnel them to the right products and features.

    4. Anthropic indexes 70/30 toward big bets, the opposite of most growth teams. Traditional growth teams spend 60 to 70 percent of effort on small to medium optimizations. Anthropic flips that ratio because they believe the product value delivered two years from now will be 100x to 1,000x what it is today. In that exponential environment, micro-optimizations capture a negligible percentage of future value. Large strategic bets are where the leverage lives.

    5. The PM-to-engineer ratio may need to flip. Engineers are getting 2 to 3x more productive with tools like Claude Code, effectively turning a team of 5 engineers into the equivalent of 15 to 20. But PMs and designers have not seen the same multiplier. The result is that product management and design are “absolutely squeezed.” Anthropic is responding by hiring more PMs and deputizing product-minded engineers to act as mini-PMs on projects under two weeks. The counterintuitive insight: companies may need more PMs, not fewer, as AI accelerates engineering output.

    6. Cold emailing still works if you do it right. Amol got his job by cold emailing Mike Krieger, Anthropic’s Chief Product Officer (and co-founder of Instagram), at a time when no growth role was even listed. Key tactics: use a high-converting subject line you have tested over time, find personal email addresses instead of competing in crowded LinkedIn inboxes, keep the message extremely short, and follow up relentlessly until someone explicitly asks you to stop.

    7. PRDs are largely obsolete at Anthropic. Amol estimates that 60 to 80 percent of what his team ships does not have a formal PRD. For small projects, coordination happens entirely in Slack. For larger initiatives, he will sometimes throw his thoughts into Cowork five minutes before a kickoff meeting to generate a rough document. His default philosophy: if you can skip the doc and jump straight to prototyping or action, do it.

    8. The AI coding bet created a research flywheel. Anthropic’s deep focus on coding was not just a commercial play. A document written by co-founder Ben Mann in 2021, just months after the company was founded, laid out the case for focusing on AI coding because better coding models would accelerate their own researchers, which would produce better models, which would produce better coding tools, in a compounding loop. This is something competitors are only now starting to recognize and copy.

    9. Cowork is being used to detect organizational misalignment. Amol runs a weekly scheduled task in Cowork that uses the Slack MCP to scan conversations across the company and surface areas of potential misalignment. He describes cases where this caught teams about to do overlapping work or spin their wheels on conflicting priorities. He also uses Cowork to simulate coaching sessions with his manager, Ami Vora, by asking Claude to analyze her public writing and internal Slack activity and then deliver feedback from her perspective.

    10. Anthropic’s culture is its most defensible moat. Amol describes a culture where every single person is fully engaged, nobody is checked out, and there is radical transparency through “notebook channels” on Slack where anyone, including leadership, shares their thinking publicly. Employees openly challenge Dario Amodei in these channels after all-hands meetings. These notebook channels also serve a practical purpose: they become training data that helps Claude understand how different teams think and operate.

    Detailed Summary

    The Cold Email That Started It All

    Amol Avasare was not recruited through a job listing, a referral, or a sourcing pipeline. He cold emailed Mike Krieger, Anthropic’s CPO and the co-founder of Instagram, with a short pitch: he loved the product, thought Anthropic badly needed a growth team, and wanted to talk. At the time, Anthropic had no growth roles posted. They were just beginning to think about it internally, and the timing was perfect.

    Amol’s approach to cold email is methodical. He has a subject line formula he has refined over years of founder outreach that produces abnormally high open rates (he declined to share the exact copy). He targets personal email addresses rather than work inboxes or LinkedIn, where competition for attention is fierce. The message itself is brutally short: who he is, why he would be a fit, and a request to chat. His follow-up philosophy is to keep reaching out until someone tells him to stop. Krieger responded on the first attempt.

    What $1B to $19B in 14 Months Actually Feels Like

    From the inside, Anthropic’s growth does not feel like a victory lap. Amol describes it as the hardest job he has ever had, harder than being a founder and harder than investment banking. About 70 percent of his time goes to what the team calls “success disasters,” which are problems created by things going extremely well. All the charts are green and up and to the right, but the underlying infrastructure, processes, and systems are constantly breaking under the strain of hypergrowth.

    The revenue trajectory tells the story: $0 to $100 million in 2023, $100 million to $1 billion in 2024, $1 billion to roughly $10 billion in 2025, and already $19 billion ARR by the end of February 2026. Amol notes that at the end of 2024, Dario Amodei was pushing for growth targets that the team thought were impossible. Those targets were hit and exceeded. The internal culture has adapted accordingly. Linear charts are considered uncool. Everything is presented on a log-linear scale.

    Why Activation Is the Hardest Problem in AI

    The central growth challenge for AI products is not acquisition. It is activation: getting users to understand what the product can actually do for them. Amol frames this as a capability overhang problem. Models are improving so rapidly that even internal teams struggle to keep up with what is newly possible. If Anthropic employees have to carve out dedicated time to explore a new model’s capabilities, the average user is even further behind.

    The danger is that someone signs up for Claude, asks it about the weather, and walks away thinking that is all it does. The product development cycle for onboarding is also under strain: by the time you have run tests, gathered learnings, and shipped an optimized activation flow for one model generation, the next model has shipped with capabilities that make your work irrelevant.

    Anthropic’s approach borrows from Amol’s experience at Mercury and MasterClass. They add deliberate friction to the signup flow, asking users questions about who they are and what they want to accomplish. This allows them to route users to the right products and features. The data also feeds downstream into lifecycle marketing and ad targeting. Amol has seen this pattern work consistently across every company he has worked at: the right friction, applied at the right time, outperforms frictionless flows that dump users into a blank canvas with no guidance.

    The CASH System: Automating Growth Experimentation

    Anthropic’s growth platform team, led by Alexey Komissarouk (who teaches growth engineering at Reforge), has built an internal system called CASH. The name stands for Claude Accelerates Sustainable Hypergrowth.

    CASH operates on a four-stage loop. First, Claude identifies growth opportunities by analyzing trends, metrics, and past experiment results. Second, Claude builds the actual feature or change. Third, Claude tests the output against quality and brand standards. Fourth, Claude analyzes the results and gathers learnings after the experiment ships.

    Currently, CASH handles mostly copy changes and minor UI tweaks. The win rate is comparable to a junior PM with two to three years of experience. A senior PM would still do better. But the trajectory matters: this was not possible at all before Opus 4.5 launched, and results have improved meaningfully with Opus 4.6. Human approval is still required before shipping, but the amount of human time spent reviewing is decreasing week over week.

    The part that Claude still cannot handle well is cross-functional stakeholder management. Getting six people in a room to align on a decision remains a fundamentally human problem. As Amol’s head of design joked: “We will have AGI and it will still be impossible to get six people in a room to get aligned.”

    Why the PM-to-Engineer Ratio Might Flip

    This is one of the most counterintuitive insights from the conversation. The conventional assumption is that AI will reduce the need for PMs. Amol argues the opposite: companies may need more PMs, at least in the near term.

    The math is straightforward. Tools like Claude Code are making engineers 2 to 3x more productive. A team of 5 engineers now produces the output equivalent of 15 to 20 engineers in the pre-AI era. PMs and designers have seen productivity gains, but not at the same multiplier. The result is a bottleneck: one PM managing the equivalent output of 15 to 20 engineers worth of work, while also handling cross-functional coordination, stakeholder alignment, and strategic direction.

    Anthropic’s response is twofold. First, they are actively hiring more PMs. Second, they have formalized a system where product-minded engineers act as mini-PMs on any project that is two engineering weeks or less. The engineer handles everything: talking to legal, talking to security, managing stakeholders. The PM only steps in if things go badly off track.

    For larger projects, the PM remains squarely accountable. But the key insight is about leverage: if you are one PM managing 20 engineers, the highest-value use of your time is not shipping the 21st feature yourself. It is getting 5 percent better at guiding the team on what the right opportunities are and upleveling every engineer’s product thinking.

    The Coding Flywheel That Changed Everything

    Anthropic’s deep bet on coding was not obvious at the time. A document from co-founder Ben Mann, dated 2021, laid out the strategic logic just months after the company was founded. The argument was that investing heavily in AI coding would create a compounding flywheel: better coding models would help Anthropic’s own researchers write code more effectively, which would accelerate model development, which would produce even better coding tools.

    This early focus gave Anthropic a structural advantage that competitors are only now trying to replicate. It also explains why the company went so deep on B2B and enterprise use cases rather than chasing consumer attention. The commercial opportunity of coding was large on its own, but the internal research acceleration made it doubly strategic.

    Amol notes that this focus was partly born from constraint. Anthropic was the smallest, least well-funded player in the space for years. They did not have Meta’s distribution or Google’s cash flow or OpenAI’s first-mover advantage. That constraint forced extreme focus, which is a principle Amol applies broadly. He calls it “freedom through constraints.”

    How Amol Uses AI to Manage His Day

    Amol’s personal AI usage is extensive and worth documenting for anyone looking to see how a power user at the frontier actually operates.

    Every morning, a scheduled Cowork task reviews 20 to 25 charts across Anthropic’s products and sends him a summary of what needs attention. The false positive and false negative rates are improving week over week, giving him increasing confidence in delegating this monitoring.

    He uses Cowork to handle administrative tasks he hates: booking meeting rooms, first-pass email triage, filing expense reports in Brex and reimbursements in Benpass. None of this requires his attention anymore.

    For management, he runs weekly Cowork tasks that review what his direct reports have done, cross-reference their work against team OKRs and meeting transcripts, and surface feedback he should give them. He also runs a parallel task for himself, asking Claude to impersonate his manager Ami Vora based on her public writing and internal Slack activity, and deliver feedback from her perspective.

    Perhaps most powerfully, he runs a weekly misalignment detection task that scans Slack conversations across the company and surfaces areas where teams may be working at cross purposes. He describes cases where this caught potentially expensive coordination failures before they compounded.

    Notebook Channels and the Culture Moat

    Anthropic uses “notebook channels” on Slack, which function like internal Twitter feeds where employees share their thinking, priorities, and provocative ideas publicly. Everyone has one, from researchers to growth PMs to Dario Amodei himself. Employees openly disagree with leadership in these channels, and that is encouraged.

    These channels serve a dual purpose. First, they help scale cultural values and operating principles as the company grows rapidly. When Amol posts about “the importance of being comfortable leaving money on the table,” every new engineer on the growth team absorbs that principle. Second, and perhaps more importantly for the long term, these channels become structured context that Claude can reference. The HR team has even documented which internal documents Claude should reference for specific topics. Amol sees this as something every company will eventually need to do: share thinking in a structured way so that the AI agents running throughout the organization have the context they need.

    AI Safety as Commercial Strategy

    Anthropic is structured as a Public Benefit Corporation (PBC), not a standard Delaware C-Corp. This legally allows the company to optimize for public benefit rather than being bound solely to maximize shareholder value.

    Amol says the company has repeatedly taken significant commercial hits for safety reasons, including delaying product launches when safety risks were identified. He also makes a striking claim: what Anthropic says publicly about AI risk is actually a softer version of what they believe internally. The internal view on the potential downsides of powerful AI is more aggressive than the public messaging.

    From a growth perspective, Amol frames safety as a long-term competitive advantage. Growth teams at most companies try to squeeze every last dollar. Anthropic’s growth team is “very comfortable forgoing metric impact” to protect brand, quality, and safety. He argues this is how all the best products operate, and that as the stakes of AI get higher, Anthropic’s credible commitment to safety will become a moat.

    Advice for Thriving in the AI Era

    Amol’s advice for product managers and growth practitioners boils down to four points. First, stay on top of the tools. Try every new model release. Something that did not work three months ago may work now, and you will not know unless you go back and test it. Second, go deep on your unique spike rather than trying to be well-rounded. The PM who can also design is a unicorn. The engineer who thinks like a PM is a unicorn. Find your interdisciplinary edge and double down. Third, be radically adaptable. Amol estimates that 50 to 70 percent of how you operated in the past is now irrelevant. Clinging to old playbooks creates friction. Fourth, think in exponentials, not linear projections. If you are looking at the AI landscape through a linear lens, you will consistently underestimate how quickly things are moving.

    Thoughts

    This interview is one of the most information-dense conversations about growth strategy in AI that has been published so far. A few things stand out.

    The CASH system is the most concrete example yet of a company using AI to automate its own growth loop. The fact that it currently performs at a junior PM level is almost beside the point. What matters is the trajectory: it went from impossible to functional in a few months. If models continue improving at their current pace, this system will be operating at a senior PM level within a year. Every growth team at every AI company should be building their own version of this right now.

    The PM ratio insight is genuinely surprising and underreported. The default assumption in the tech industry is that AI will reduce headcount across all functions. Amol is making the case that in the near term, the opposite is true for PMs. Engineering output is exploding, and someone needs to direct all that output toward the right problems. That is a fundamentally human, organizational, political job that AI is not close to automating.

    The coding flywheel story is also worth highlighting because it shows the power of strategic focus in a world of unlimited possibilities. Anthropic had a generalist technology that could do almost anything, and they deliberately narrowed their focus to one vertical. That decision, made in 2021 before anyone knew what the market would look like, is arguably the single most important strategic bet in the company’s history.

    Finally, the notebook channels concept deserves more attention. The idea that employees should share their thinking in structured, searchable formats is not just a culture tool. It is an infrastructure investment for an AI-native future where agents need organizational context to be effective. Companies that build this habit early will have a significant advantage when agent-driven workflows become the norm.

    The uncomfortable subtext of this entire conversation is that Anthropic’s growth team, as talented as they clearly are, is riding a wave created almost entirely by the research team. Several YouTube commenters pointed this out, and Amol himself acknowledges it directly. The models are the product. The growth team’s job is to make sure users discover and adopt what the models can do. That is not a small job, especially at this scale, but it is a fundamentally different job than driving growth at a product that does not sell itself.

  • Jensen Huang on Lex Fridman: NVIDIA’s CEO Reveals His Vision for the AI Revolution, Scaling Laws, and Why Intelligence Is Now a Commodity

    A deep breakdown of Lex Fridman Podcast #494 featuring Jensen Huang, CEO of NVIDIA, covering extreme co-design, the four AI scaling laws, CUDA’s origin story, the future of programming, AGI timelines, and what it takes to lead the world’s most valuable company.

    TLDW (Too Long, Didn’t Watch)

    Jensen Huang sat down with Lex Fridman for a sprawling two-and-a-half-hour conversation covering the full arc of NVIDIA’s evolution from a GPU gaming company to the engine of the AI revolution. Jensen explains how NVIDIA now thinks in terms of rack-scale and pod-scale computing rather than individual chips, breaks down his four AI scaling laws (pre-training, post-training, test time, and agentic), and reveals the near-existential bet the company made putting CUDA on GeForce. He shares his views on China’s tech ecosystem, his deep respect for TSMC, why he turned down the chance to become TSMC’s CEO, how Elon Musk’s systems engineering approach built Colossus in record time, and why he believes AGI already exists. He also discusses why the future of programming is really about “specification,” why intelligence is being commoditized while humanity is the true superpower, and how he manages the enormous pressure of leading a company that nations and economies depend on. His core message: do not let the democratization of intelligence cause you anxiety. Instead, let it inspire you.

    Key Takeaways

    1. NVIDIA No Longer Thinks in Chips. It Thinks in AI Factories.

    Jensen’s mental model of what NVIDIA builds has fundamentally changed. He no longer picks up a chip to represent a new product generation. Instead, his mental model is a gigawatt-scale AI factory with power generation, cooling systems, and thousands of engineers bringing it online. The unit of computing at NVIDIA has evolved from GPU to computer to cluster to AI factory. His next mental “click” is planetary-scale computing.

    2. Extreme Co-Design Is NVIDIA’s Secret Weapon

    The reason NVIDIA dominates is not just better GPUs. It is the extreme co-design of the entire stack: GPU, CPU, memory, networking, switching, power, cooling, storage, software, algorithms, and applications. Jensen explains that when you distribute workloads across tens of thousands of computers and want them to go a million times faster (not just 10,000 times), every single component becomes a bottleneck. This is a restatement of Amdahl’s Law at scale. NVIDIA’s organizational structure directly reflects this co-design philosophy. Jensen has 60+ direct reports, holds no one-on-ones, and runs every meeting as a collective problem-solving session where specialists across all domains are present and contribute.

    3. The Four AI Scaling Laws Are a Flywheel

    Jensen outlined four distinct scaling laws that form a continuous loop:

    Pre-training scaling: Larger models plus more data equals smarter AI. The industry panicked when people said data was running out, but synthetic data generation has removed that ceiling. Data is now limited by compute, not by human generation.

    Post-training scaling: Fine-tuning, reinforcement learning from human feedback, and curated data continue to scale AI capabilities beyond what pre-training alone achieves.

    Test-time scaling: Inference is not “easy” as many predicted. It is thinking, reasoning, planning, and search. It is far more compute-intensive than memorization and pattern matching. This is why inference chips cannot be commoditized the way many predicted.

    Agentic scaling: A single AI agent can spawn sub-agents, creating teams. This is like scaling a company by hiring more employees rather than trying to make one person faster. The experiences generated by agents feed back into pre-training, creating a flywheel.

    4. The CUDA Bet Nearly Killed NVIDIA

    Putting CUDA on GeForce was one of the most consequential technology decisions in modern history. It increased GPU costs by roughly 50%, which crushed the company’s gross margins at a time when NVIDIA was a 35% gross margin business. The company’s market cap dropped from around $7-8 billion to approximately $1.5 billion. But Jensen understood that install base defines a computing architecture, not elegance. He pointed to x86 as proof: a less-than-elegant architecture that defeated beautifully designed RISC alternatives because of its massive install base. CUDA on GeForce put a supercomputer in the hands of every researcher, every scientist, every student. It took a decade to recover, but that install base became the foundation of the deep learning revolution.

    5. NVIDIA’s Moat Is Trust, Velocity, and Install Base

    Jensen was direct about NVIDIA’s competitive advantage. The CUDA install base is the number one asset. Developers target CUDA first because it reaches hundreds of millions of computers, is in every cloud, every OEM, every country, every industry. NVIDIA ships a new architecture roughly every year. No company in history has built systems of this complexity at this cadence. And the trust that NVIDIA will maintain, improve, and optimize CUDA indefinitely is something developers can count on. If someone created “GUDA” or “TUDA” tomorrow, it would not matter. The install base, velocity of execution, ecosystem breadth, and earned trust create a compounding advantage that is nearly impossible to replicate.

    6. Jensen Believes AGI Is Already Here

    When asked about AGI timelines, Jensen said he believes AGI has been achieved. His reasoning is practical: an agentic system today could plausibly create a web service, achieve virality, and generate a billion dollars in revenue, even if temporarily. This is not meaningfully different from many internet-era companies that did the same thing with technology no more sophisticated than what current AI agents can produce. He does not believe 100,000 agents could build another NVIDIA, but he believes a single agent-driven viral product is within reach right now.

    7. The Future of Programming Is Specification, Not Syntax

    Jensen believes the number of programmers in the world will increase dramatically, not decrease. His reasoning: the definition of coding is expanding to include specification and architectural description in natural language. This expands the population of “coders” from roughly 30 million professional developers to potentially a billion people. Every carpenter, plumber, accountant, and farmer who can describe what they want a computer to build is now a coder. The artistry of the future is knowing where on the spectrum of specification to operate, from highly prescriptive to exploratory and open-ended.

    8. China Is the Fastest Innovating Country in the World

    Jensen gave a nuanced and detailed explanation of why China’s tech ecosystem is so formidable. About 50% of the world’s AI researchers are Chinese. China’s tech industry emerged during the mobile cloud era, so it was built on modern software from the start. The country’s provincial competition creates an insane internal competitive environment. And the cultural norm of knowledge-sharing through school and family networks means China effectively operates as an open-source ecosystem at all times. This is why Chinese companies contribute disproportionately to open source. Their engineers’ brothers, friends, and schoolmates work at competing companies, and sharing knowledge is the cultural default.

    9. The Power Grid Has Enormous Waste That AI Can Exploit

    Jensen proposed a pragmatic solution to the energy problem for AI data centers. Power grids are designed for worst-case conditions with margin, but 99% of the time they run at around 60% of peak capacity. That idle capacity is simply wasted. Jensen wants data centers to negotiate flexible contracts where they absorb excess power most of the time and gracefully degrade during rare peak demand periods. This requires three things: customers accepting that “six nines” uptime may not always be necessary, data centers that can dynamically shift workloads, and utilities that offer tiered power delivery contracts instead of all-or-nothing commitments.

    10. Jensen Turned Down the CEO Role at TSMC

    In 2013, TSMC founder Morris Chang offered Jensen the chance to become CEO of TSMC. Jensen confirmed the story is true and said he was deeply honored. But he had already envisioned what NVIDIA could become and felt it was his sole responsibility to make that vision happen. He sees the relationship with TSMC as one built on three decades of trust, hundreds of billions of dollars in business, and zero formal contracts.

    11. Elon Musk’s Systems Engineering Approach Is Instructive

    Jensen praised Elon Musk’s approach to building the Colossus supercomputer in Memphis in just four months. He highlighted several principles: Elon questions everything relentlessly, strips every process down to the minimum necessary, is physically present at the point of action, and his personal urgency creates urgency in every supplier. Jensen drew a parallel to NVIDIA’s own “speed of light” methodology, where every process is benchmarked against the physical limits of what is possible, not against historical baselines.

    12. Intelligence Is a Commodity. Humanity Is Not.

    Perhaps the most philosophical takeaway from the conversation: Jensen argued that intelligence is a functional, measurable thing that is being commoditized. He surrounded himself with 60 direct reports who are all “superhuman” in their respective domains, more educated and deeper in their specialties than he is. Yet he sits in the middle orchestrating all of them. This proves that intelligence alone does not determine success. Character, compassion, grit, determination, tolerance for embarrassment, and the ability to endure suffering are the real differentiators. Jensen wants the audience to understand that the word we should elevate is not intelligence but humanity.

    Detailed Summary

    From GPU Maker to AI Infrastructure Company

    The conversation opened with Jensen explaining NVIDIA’s evolution from chip-scale to rack-scale to pod-scale design. The Vera Rubin pod, announced at GTC, contains seven chip types, five purpose-built rack types, 40 racks, 1.2 quadrillion transistors, nearly 20,000 NVIDIA dies, over 1,100 Rubin GPUs, 60 exaflops of compute, and 10 petabytes per second of scale bandwidth. And that is just one pod. NVIDIA plans to produce roughly 200 of these pods per week.

    Jensen explained that extreme co-design is necessary because the problems AI must solve no longer fit inside a single computer. When you distribute a workload across 10,000 computers but want a million-fold speedup, everything becomes a bottleneck: computation, networking, switching, memory, power, cooling. This is fundamentally an Amdahl’s Law problem at planetary scale. If computation represents only 50% of the workload, speeding it up infinitely only doubles total throughput. Every layer must be co-optimized simultaneously.

    NVIDIA’s organizational structure is a direct reflection of this co-design philosophy. Jensen has more than 60 direct reports, almost all with deep engineering expertise. He does not do one-on-ones. Every meeting is a collective problem-solving session where the memory expert, the networking expert, the cooling expert, and the power delivery expert are all in the room together, attacking the same problem.

    The Strategic History of CUDA

    Jensen walked through the step-by-step journey from graphics accelerator to computing platform. The company invented a programmable pixel shader, then added IEEE-compatible FP32 to its shaders, then put C on top of that (called Cg), and eventually arrived at CUDA. The critical strategic decision was putting CUDA on GeForce, a consumer product.

    This was nearly an existential move. It increased GPU costs by roughly 50% and consumed all of the company’s gross profit at a time when NVIDIA was a 35% gross margin business. The market cap cratered from around $7-8 billion to approximately $1.5 billion. But Jensen understood a principle that many technologists overlook: install base defines a computing architecture. x86 survived not because it was elegant but because it was everywhere. CUDA on GeForce put a supercomputing capability in the hands of every gamer, every student, every researcher who built their own PC. When the deep learning revolution arrived, CUDA was already the foundation.

    How Jensen Leads and Makes Decisions

    Jensen described a leadership philosophy built on continuous reasoning in public. He does not make announcements in the traditional sense. Instead, he shapes the belief systems of his employees, board, partners, and the broader industry over months and years by reasoning through decisions step by step, using every new piece of external information as a brick in the foundation. By the time he formally announces a strategic direction, the reaction is not surprise but rather, “What took you so long?”

    He applies this same approach to his supply chain. He personally visits CEOs of DRAM companies, packaging companies, and infrastructure providers. He explains the dynamics of the industry, shares his vision of future demand, and helps them reason through why they should make multi-billion-dollar capital investments. Three years ago, he convinced DRAM CEOs that HBM memory would become mainstream for data centers, which sounded ridiculous at the time. Those companies had record years as a result.

    Jensen’s “speed of light” methodology is his framework for decision-making. Every process, every design, every cost is benchmarked against the physical limits of what is theoretically possible. He prefers this to continuous improvement, which he views as incrementalism. He would rather strip a 74-day process back to zero and ask, “If we built this from scratch today, how long would it take?” Often the answer is six days, and the remaining 68 days are filled with accumulated compromises that can be challenged individually.

    AI Scaling Laws and the Future of Compute

    Jensen broke down the four scaling laws in detail. The pre-training scaling law, which depends on model size and data volume, was thought to be hitting a wall when the industry worried about running out of high-quality human-generated data. Jensen argued this concern is misplaced. Synthetic data generation has effectively removed the ceiling, and the constraint is now compute, not data.

    Post-training continues to scale through fine-tuning and reinforcement learning. Test-time scaling was the most counterintuitive for the industry. Many predicted that inference would be “easy” and that inference chips would be small, cheap, and commoditized. Jensen saw this as fundamentally wrong. Inference is thinking: reasoning, planning, search, decomposing novel problems into solvable pieces. Thinking is much harder than reading, and test-time compute is intensely resource-hungry.

    Agentic scaling is the newest frontier. A single AI agent can spawn sub-agents, effectively multiplying intelligence the way a company scales by hiring. The experiences and data generated by agentic systems feed back into pre-training, creating a continuous improvement loop. Jensen described this as the reason NVIDIA designed the Vera Rubin rack architecture differently from the Grace Blackwell architecture. Grace Blackwell was optimized for running large language models. Vera Rubin is designed for agents, which need to access files, use tools, do research, and spin off sub-agents. NVIDIA anticipated this architectural shift two and a half years before tools like OpenClaw arrived.

    China, TSMC, and the Global Supply Chain

    Jensen provided a thoughtful analysis of China’s tech ecosystem. He identified several structural advantages: 50% of the world’s AI researchers are Chinese, the tech industry was born during the mobile cloud era (making it natively modern), provincial competition creates internal Darwinian pressure, and the culture of knowledge-sharing through school and family networks makes China effectively open-source by default.

    On TSMC, Jensen emphasized that the deepest misunderstanding about the company is that its technology is its only advantage. Their manufacturing orchestration system, which dynamically manages the shifting demands of hundreds of companies, is “completely miraculous.” Their culture uniquely balances bleeding-edge technology excellence with world-class customer service. And the trust that Jensen places in TSMC is extraordinary: three decades of partnership, hundreds of billions of dollars in business, and no formal contract.

    Jensen also discussed the AI supply chain more broadly. NVIDIA has roughly 200 suppliers contributing technology to each rack. Jensen personally manages these relationships, flying to supplier sites, explaining industry dynamics, and helping CEOs reason through multi-billion-dollar investment decisions. When asked if supply chain bottlenecks keep him up at night, he said no, because he has already communicated what NVIDIA needs, his partners have told him what they will deliver, and he believes them.

    The Energy Challenge and Space Computing

    On the energy front, Jensen proposed a practical approach to the power problem. Rather than waiting for new power generation, he wants to capture the enormous waste already present in the grid. Power infrastructure is designed for worst-case peak demand, but 99% of the time it runs far below capacity. AI data centers could absorb this excess capacity with flexible contracts that allow graceful degradation during rare peak periods.

    On space computing, NVIDIA already has GPUs in orbit for satellite imaging. Jensen acknowledged the cooling challenge (no conduction or convection in space, only radiation) but sees it as a future frontier worth cultivating. In the meantime, he is focused on the lower-hanging fruit of eliminating waste in the terrestrial power grid.

    On AGI, Jobs, and the Human Future

    Jensen stated directly that he believes AGI has been achieved, at least by the practical definition of an AI system capable of creating a billion-dollar company. He sees it as plausible that an agent could build a viral web service that briefly generates enormous revenue, just as many internet-era companies did with technology no more sophisticated than what current AI agents produce.

    On jobs, Jensen was both compassionate and clear-eyed. He told the story of radiology: computer vision became superhuman around 2019-2020, and the prediction was that radiologists would disappear. Instead, the number of radiologists grew because AI allowed them to study more scans, diagnose better, and serve more patients. The purpose of the job (diagnosing disease) did not change, even though the tools changed completely.

    He applied this principle broadly: the number of software engineers at NVIDIA will grow, not decline, because their purpose is solving problems, not writing lines of code. The number of programmers globally will grow because the definition of coding is expanding to include natural language specification, opening it up to potentially a billion people.

    His advice to anyone worried about their job is straightforward: go use AI now. Become expert in it. Every profession, from carpenter to pharmacist to lawyer, will be elevated by AI tools. The people who learn to use AI will be the ones who get hired, promoted, and empowered.

    Mortality, Succession, and Legacy

    The conversation closed with deeply personal reflections. Jensen said he really does not want to die. He sees the current moment as a “once in a humanity experience.” He does not believe in traditional succession planning. Instead, he believes the best succession strategy is to pass on knowledge continuously, every single day, in every meeting, as fast as possible. His hope is to die on the job, instantaneously, with no long period of suffering.

    He described a vision for a kind of digital continuity: sending a humanoid robot into space, continuously improving it in flight, and eventually uploading the consciousness derived from a lifetime of communications, decisions, and reasoning to catch up with it at the speed of light.

    On the emotional experience of leading NVIDIA, Jensen was candid about hitting psychological low points regularly. His coping mechanism is decomposition: break the problem into pieces, reason about what you can control, tell someone who can help, share the burden, and then deliberately forget what is behind you. He compared this to the mental discipline of great athletes who focus only on the next point.

    His final message was about the relationship between intelligence and humanity. Intelligence, he argued, is functional. It is being commoditized. Humanity, character, compassion, grit, tolerance for embarrassment, and the capacity for suffering are the true superpowers. The word society should elevate is not intelligence but humanity.

    Thoughts

    This is one of the most substantive CEO interviews of 2026. What makes it remarkable is not just the breadth of topics but the depth of reasoning Jensen demonstrates in real time. You can actually watch him think through problems on the spot, which is rare for someone at his level.

    A few things stand out. First, the CUDA origin story is one of the great strategic narratives in tech history. The decision to absorb a 50% cost increase on a consumer product, watching your market cap collapse by 80%, and holding the course for a decade because you understood the power of install base is the kind of conviction that separates generational companies from everyone else.

    Second, Jensen’s framing of the four scaling laws as a flywheel is the clearest articulation anyone has given of why AI compute demand will continue to accelerate. Most people understand pre-training. Fewer understand test-time scaling. Almost nobody is thinking about agentic scaling as a compute multiplier. Jensen has been thinking about it for years and already designed hardware for it before the software ecosystem caught up.

    Third, the discussion on jobs deserves attention. The radiology example is powerful because it is a completed experiment, not a prediction. The profession that was supposed to be eliminated first by AI instead grew. The mechanism is straightforward: when you automate the task, you expand the capacity of the purpose, and demand for the purpose increases. This does not mean there will be no pain or dislocation. Jensen acknowledged that explicitly. But the historical pattern is clear.

    Finally, the philosophical distinction between intelligence and humanity is the kind of framing that could genuinely help people navigate the anxiety of this moment. If you define your value by your intelligence alone, AI commoditization is terrifying. If you define your value by your character, your compassion, your tolerance for suffering, and your willingness to keep going when everything goes wrong, then AI is just the most powerful set of tools you have ever been given.

    Jensen Huang is 62 years old, has been running NVIDIA for 34 years, and shows no signs of slowing down. If anything, his conviction about the future is accelerating alongside his company’s growth.

    Watch the full episode: Lex Fridman Podcast #494 with Jensen Huang