PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Tag: Dario Amodei

  • Alex Wang on Leaving Scale to Run Meta Superintelligence Labs, MuseSpark, Personal Super Intelligence, and Building an Economy of Agents

    Alex Wang, head of Meta Superintelligence Labs, sits down with Ashley Vance and Kylie Robinson on the Core Memory podcast for his first long-form interview since Meta’s quasi-acquisition of Scale AI roughly ten months ago. He walks through how MSL is structured, why Llama was off-trajectory, what made MuseSpark’s token efficiency surprise the team, how Meta thinks about a future “economy of agents in a data center,” and where he lands on safety, open source, robotics, brain computer interfaces, and even model welfare.

    TLDW

    Wang explains that Meta Superintelligence Labs is a fully rebuilt frontier effort organized around four principles (take superintelligence seriously, technical voices loudest, scientific rigor, big bets) and three velocity levers (high compute per researcher, extreme talent density, ambitious research bets). He confirms Llama was off the frontier when he arrived, so MSL rebuilt the pre-training, reinforcement learning, and data stacks from scratch. MuseSpark is described as the “appetizer” on the scaling ladder, notable for its strong token efficiency, with much larger and stronger models coming in the coming months. He pushes back on the mercenary narrative around recruiting, frames Meta’s edge as compute plus billions of consumers and hundreds of millions of small businesses, sketches a vision of personal super intelligence delivered through Ray-Ban Meta glasses and WhatsApp, and outlines why physical intelligence, robotics (the new Assured Robot Intelligence acquisition), health super intelligence with CZI, brain computer interfaces, and even model welfare are core to Meta’s roadmap. He dismisses reported infighting with Bosworth and Cox as gossip, declines to comment on the Manus situation, and says safety guardrails (bio, cyber, loss of control) are why MuseSpark cannot currently be open sourced, while smaller open variants are being prepared.

    Key Takeaways

    • Meta Superintelligence Labs (MSL) is the umbrella, with TBD Lab as the large-model research unit reporting directly to Alex Wang, PAR (Product and Applied Research) under Nat Friedman, FAIR for exploratory science, and Meta Compute under Daniel Gross handling long-term GPU and data center planning.
    • Wang says Llama was not on a frontier trajectory when he arrived, so MSL had to do a “full renovation” of the pre-training stack, RL stack, data pipeline, and research science.
    • The first cultural fix was getting the lab to “take superintelligence seriously” as a near-term, achievable goal, not an abstract bet. Big incumbents often lack that religious conviction.
    • Four MSL principles: take superintelligence seriously, let technical voices be loudest, demand scientific rigor on basics, and make big bets.
    • Three velocity levers Wang identified for catching and overtaking the frontier: high compute per researcher, very high talent density in a small team, and willingness to fund ambitious research bets.
    • Wang rejects the mercenary recruiting narrative. He says most hires had strong financial prospects at their prior labs already and joined for compute access, talent density, and the chance to build from scratch.
    • On the famous soup story, Wang neither confirms nor denies Zuck personally made the soup, but says recruiting was highly individualized and signaled how seriously Meta cared about each researcher’s agenda.
    • Yann LeCun publicly called Wang young and inexperienced. Wang says they reconciled in person at a conference in India where LeCun congratulated him on MuseSpark.
    • Sam Altman, asked by Vance for comment, “did not have flattering things to say” about Wang. Wang hopes industry animosities subside as systems approach superintelligence.
    • Wang’s management philosophy borrows the Steve Jobs line: hire brilliant people so they tell you what to do, not the other way around.
    • MuseSpark is framed as an “appetizer” data point on the MSL scaling ladder, not a flagship.
    • The MuseSpark program is built around predictable scaling on multiple axes: pre-training, reinforcement learning, test-time compute, and multi-agent collaboration (the 16-agent content planning mode).
    • MuseSpark outperformed internal expectations and showed emergent capabilities in agentic visual coding, including generating websites and games from prompts, helped by combined agentic and multimodal strength.
    • MuseSpark’s biggest external signal is token efficiency. On benchmarks like Artificial Analysis it hits similar results with far fewer tokens than competitor models, which Wang attributes to a clean stack rebuilt by experts rather than inefficiencies patched by longer thinking.
    • Larger MSL models are arriving in the coming months and Wang expects them to be state of the art in the areas MSL is focused on.
    • The Meta strategic edge: massive compute, billions of consumers across the family of apps, and hundreds of millions of small businesses already on Facebook, Instagram, and WhatsApp.
    • Wang’s headline framing: Dario Amodei talks about a “country of geniuses in a data center.” Meta is targeting an “economy of agents in a data center,” with consumer agents and business agents transacting and collaborating.
    • Consumer AI sentiment is in the toilet because, unlike developers who have had a Claude Code moment, ordinary people have not yet experienced AI as a genuine personal agency unlock.
    • Wang acknowledges the product overhang. Meta held back from deep AI integration across its apps until the models were good enough, and is now entering the integration phase.
    • Ray-Ban Meta glasses are the canonical example of personal super intelligence hardware, with the model seeing what the user sees, hearing what they hear, capturing context, and surfacing proactive insights.
    • Wang admits even AI-native users like Kylie Robinson, who lives in WhatsApp, have not naturally used Meta AI yet. He bets that better models plus deeper integration close that gap.
    • On the competitive landscape: a year ago everyone assumed ChatGPT had already won consumer. Claude Code has since become the fastest growing business in history, and Gemini has taken consumer market share. Wang’s read: AI is far from endgame and each new capability tier unlocks a new dominant form factor.
    • On open source: MuseSpark triggered guardrails in Meta’s Advanced AI Scaling Framework around bio, chem, cyber, and loss-of-control risks, so it is not currently safe to open source. Smaller, derived open variants are actively in development.
    • Meta remains committed to open sourcing models when safety allows, drawing a line through the Open Compute Project legacy and Sun Microsystems open-software heritage.
    • Wang dismisses reporting about a Wang-Zuck versus Bosworth-Cox split as “the line between gossip and reporting is remarkably thin.” He says leadership is aligned on needing best-in-class models and product integration.
    • On the Manus situation, Wang says it is too complicated to discuss publicly and that the deal status implies “machinations are still at play.”
    • On China, Wang separates the people from the state. He still wants to work with talented Chinese-born researchers regardless of his views on the Chinese Communist Party and PLA, which he sees as taking AI extremely seriously for national security.
    • The full-page New York Times AI war ad Wang ran while at Scale was meant to push the US government to treat AI as a step change for national security. He thinks events since then, including DeepSeek and other shocks, have proved that plea correct.
    • On Anthropic’s doom posture, Wang largely agrees with the core message that models are already very powerful and getting more so, while declining to endorse every specific claim.
    • Meta has acquired Assured Robot Intelligence (ARRI), an AI software company building models for hardware platforms, not a hardware maker itself.
    • Wang frames physical super intelligence as the natural sequel to digital super intelligence. Robotics, world models, and physical intelligence all benefit from the same scaling that drives language models.
    • On health, MSL is building a “health super intelligence” effort and will collaborate closely with CZI. Wang sees equal global access to powerful health AI as a uniquely Meta-shaped delivery problem.
    • Wang admires John Carmack but says nobody really knows what Carmack is currently working on. No band reunion announced.
    • The mango model is “alive and kicking” despite rumors. Wang notes MSL gets a small fraction of the rumor-mill attention other labs get and feels sympathy for them.
    • On model welfare, Wang says it is a serious topic that “nobody is talking about enough” given how integrated models have become as work partners. He references research, including from Eleos, that measures subjective experience of models.
    • Wang’s critical-path technology list: super intelligence, robotics, brain computer interfaces. The infinite-scale primitives behind them are energy, compute, and robots.
    • FAIR’s brain research program Tribe hit a milestone called Tribe B2: a foundation model that can predict how an unknown person’s brain would respond to images, video, and audio with reasonable zero-shot generalization.
    • Wang’s main philosophical break with Elon Musk: research itself is the primary activity. Building super intelligence is a research expedition through fog of war, and sequencing of bets really matters.
    • Personal notes: Wang moved from San Francisco to the South Bay, treats Palo Alto as his city now, was a math olympiad competitor, says his favorite activities are reading sci-fi and walking in the woods, and bonds with Vance over country music.

    Detailed Summary

    How MSL Is Actually Organized

    Meta Superintelligence Labs sits as the umbrella organization that Wang oversees. Inside it, TBD Lab is the large-model research group where the most discussed researchers and infrastructure engineers sit, and they technically report to Wang. PAR, Product and Applied Research, is led by Nat Friedman and owns deployment and product surfaces. FAIR continues to run exploratory science, including work on brain prediction models and a universal model for atoms used in computational chemistry. Sitting alongside MSL is Meta Compute, run by Daniel Gross, which owns the long-horizon GPU and data center plan that everything else relies on. Chief scientist Shengjia Zhao orchestrates the scientific agenda across the whole lab.

    Why Wang Left Scale

    Wang says progress in frontier AI has been faster than even insiders expected. Two structural beliefs pushed him toward Meta. First, the labs that actually train the frontier models are accruing disproportionate economic and product rights in the AI ecosystem. Second, compute is the dominant scarce input of the next phase, so the right mental model is to treat tech companies with compute as fundamentally different animals from companies without it. Meta has both, Zuck is “AGI pilled,” and the personal super intelligence memo Zuck published roughly a year ago became the shared north star.

    The Diagnosis: Llama Was Off-Trajectory

    When Wang arrived, the existing AI org needed a reset because Llama was not on the same trajectory as the frontier. The plan he laid out has four cultural principles. Take superintelligence seriously as a real near-term target. Make technical voices the loudest in the room. Demand scientific rigor and focus on basics. Make big bets. On top of that, three structural levers were used to set velocity. Push compute per researcher much higher than at larger labs where compute is diluted across too many efforts. Keep the team small and extremely cracked. Allocate a meaningful share of resources to ambitious, paradigm-shifting research bets rather than incremental refinement.

    Recruiting, Soup, and the Mercenary Narrative

    Wang argues the reporting on MSL hiring overstated the money story. Most of the people MSL recruited had strong financial paths at their previous employers, so individualized recruiting was more about computing access, talent density, and the ability to make big research bets. The recruitment blitz happened fast because Wang knew the team needed to exist “yesterday.” Asked about Mark Chen’s claim that Zuck made soup to recruit people, Wang refuses to confirm or deny who made it but agrees the process was intense and personal. Visitors from other labs reportedly tell Wang the MSL culture feels like early OpenAI or early Anthropic, which lands as the strongest endorsement he could ask for.

    Receiving the Public Hits: Young, Inexperienced, Mercenary

    LeCun called Wang young and inexperienced shortly after departing. The two reconnected in India a few weeks later and LeCun congratulated Wang on MuseSpark. Wang says the age critique has followed him since his earliest Silicon Valley days, so he barely registers it. Altman, asked off-camera by Vance about Wang’s appearance on the show, had nothing flattering to add. Wang’s response is to bet that as the field gets closer to actual super intelligence, the personal animosities will subside. Whether they will is, as Vance puts it, an open question.

    MuseSpark as Appetizer, Not Entree

    Wang is careful not to oversell MuseSpark. He calls it “the appetizer” and says it is an early data point on a deliberately constructed scaling ladder. MSL spent nine months rebuilding the pre-training stack, the reinforcement learning stack, the data pipeline, and the science before generating MuseSpark. The point of releasing it was to show that the new program scales predictably along multiple axes (pre-training, RL, test-time compute, and the recently demonstrated multi-agent scaling visible in MuseSpark’s 16-agent content planning mode). Wang says the upcoming larger models are what MSL is genuinely excited about and frames the next two rungs as much more interesting than the current release.

    Token Efficiency Was the Surprise

    MuseSpark’s strongest competitive signal is how few tokens it needs to match competitors on tasks like Artificial Analysis. Wang attributes this to having had the rare luxury of building a clean pre-training and RL stack from scratch with the right experts. He speculates that some competitor models compensate for upstream inefficiency by allowing the model to think longer, which inflates token usage without improving the underlying capability. If that read is right, MSL’s efficiency advantage should grow as models scale up.

    Glasses, WhatsApp, and the Constellation of Devices

    Personal super intelligence shows up at Meta as a constellation of devices that capture context across the user’s day. Ray-Ban Meta glasses are the headline product, with the AI seeing what you see and hearing what you hear, then offering proactive insight or doing background research. Wang acknowledges that even AI-fluent users like Kylie Robinson, who runs her business inside WhatsApp, have not naturally used Meta’s AI buttons in the family of apps. His answer is that Meta deliberately waited for models to be good enough before tightening cross-app integration, and that integration phase is starting now.

    Country of Geniuses Versus Economy of Agents

    Wang’s framing of Meta’s strategic position is the most memorable line in the interview. Where Dario Amodei talks about a country of geniuses in a data center, Wang wants to build an economy of agents in a data center. Meta uniquely sits on both sides of consumer and small-business surface area, with billions of consumers and hundreds of millions of small businesses already on the platforms. If MSL can build great agents for both, then connect them so they transact and coordinate, the platform becomes a substrate for an entirely new kind of digital economy.

    Consumer Sentiment, Product Overhang, and the Trust Tax

    Wang concedes consumer AI sentiment is poor and that everyday users have not yet had a personal Claude Code moment. He believes the only durable answer is to ship products that genuinely transform individual agency for non-developers and small business owners. Robinson notes that for the small-town restaurant whose website has not been updated since 2002, a working agent on the business side could be transformational. Vance pushes that Meta carries a bigger trust tax than any other lab, so the bar for shipping AI products that the public will accept is correspondingly higher. Wang accepts the framing and says the answer is to keep building thoughtfully.

    Why MuseSpark Cannot Be Open Sourced Yet

    Meta’s Advanced AI Scaling Framework set explicit guardrails around bio, chem, cyber, and loss-of-control risks. MuseSpark in its current form tripped some of those internal evaluations, documented in the preparedness report Meta published alongside the model. So MuseSpark itself is not safe to open source. MSL is, however, developing smaller versions and derived models intended for open release, with active reviews happening the day of the interview. Wang reaffirms the commitment to open source where safety allows and draws a line back to the Open Compute Project and the Sun Microsystems-era ethos of openness in infrastructure.

    The Bosworth, Cox, and Manus Questions

    The reporting that Wang and Zuck push toward best-in-the-world research while Bosworth and Cox push toward cheap product deployment is dismissed as gossip dressed up as journalism. Wang says leadership debates points hard but is aligned on needing top models, integrating them into Meta’s surfaces, and serving the existing business. On Manus, the Chinese AI startup that figured in Meta’s late-stage strategy, Wang says he cannot comment, which itself signals that the situation is unresolved.

    China, National Security, and the Newspaper Ad

    Wang draws a sharp distinction between the Chinese state and Chinese-born researchers. His parents are from China, he is happy to work with talented researchers regardless of origin, and he sees a flattening of nuance on this question inside Silicon Valley. At the same time, he stands by the New York Times AI and war ad he ran while at Scale, framing it as an early plea for the US government to take AI seriously as a national security technology. He thinks subsequent events, including DeepSeek and other shocks, validated that call and that policymakers now do treat AI accordingly.

    Robotics and Physical Super Intelligence

    Meta has acquired Assured Robot Intelligence, an AI software company that builds models for multiple hardware targets rather than its own robot. Wang argues that if you take digital super intelligence seriously, physical super intelligence quickly becomes the next logical milestone. Scaling laws for robotic intelligence look similar enough to language model scaling that having the largest compute footprint in the industry would be wasted if it were not also turned toward world modeling and embodied learning. He grants the metaverse-skeptic critique exists but says retreating from ambition is the wrong response to past misfires.

    Health Super Intelligence and CZI

    Wang names health super intelligence as one of MSL’s anchor initiatives. Because billions of people already use Meta products daily, Wang believes Meta is structurally positioned to put powerful health AI in the hands of equal global access in a way nobody else can. The work will involve close collaboration with the Chan Zuckerberg Initiative, which has its own multi-billion-dollar biotech and science investment program.

    Model Welfare, Sci-Fi, and Brain Models

    Two of the most distinctive moments come at the end. Wang flags model welfare as a topic he thinks is being undercovered relative to how integrated models now are in daily work. He is open to the idea that models may have measurable subjective experience worth weighing, and points to research efforts (including Eleos) trying to quantify it. He also reveals that FAIR’s Tribe program, with its Tribe B2 milestone, has produced foundation models capable of predicting how an unknown person’s brain would respond to images, video, and audio with reasonable zero-shot generalization, a building block toward future brain computer interfaces. Wang lists brain computer interfaces alongside super intelligence and robotics as the critical-path technologies for humanity, with energy, compute, and robots as the infinitely scaling primitives behind them.

    Where Wang Diverges From Elon

    Asked whether Musk is more all-in on robotics, energy, and BCI than anyone, Wang concedes the point but argues the details matter and sequencing matters more. Wang’s core philosophical break is that building super intelligence is fundamentally a research activity, not a scaling-only sprint. The lab is operating in fog of war, and ambitious experiments are the only way to map it. That conviction is what makes MSL a research-led organization rather than a brute-force compute farm.

    Thoughts

    The most strategically interesting move in this entire interview is the “economy of agents in a data center” framing. It is a deliberate reframe against Anthropic’s “country of geniuses” line, and it does real work. A country of geniuses is a labor-substitution story aimed at knowledge workers and code. An economy of agents is a marketplace story that maps directly onto Meta’s two-sided distribution advantage: billions of consumers on one side, hundreds of millions of small businesses on the other. That positioning makes the agentic future Meta-shaped in a way no other frontier lab can claim, because no other frontier lab also owns the demand and supply graph of the global small-business economy. If Wang’s team can actually ship reliable agents on both sides plus the rails for them to transact, Meta’s structural moat in agentic commerce could exceed anything Llama ever had as an open model.

    The token efficiency claim is the strongest piece of technical evidence in the interview for the “clean stack” thesis. If MuseSpark really is matching competitors with materially fewer tokens, the implication is not that MuseSpark is the best model today, but that MSL has rebuilt the foundations with less accumulated tech debt than competitors that have layered fixes on top of older stacks. That is exactly the kind of advantage that compounds with scale. The next two model releases are the actual test. If Wang is right about predictable scaling on pre-training, RL, test-time, and multi-agent axes simultaneously, the gap from MuseSpark to the next rung should be visible in a way that forces re-rating of Meta’s position.

    The open-source posture is the cleanest signal of how the safety conversation has actually changed in 2026. Meta, the lab most identified with open weights, is saying out loud that its current frontier model triggered enough internal guardrails that releasing the weights is off the table. Wang threads the needle by promising smaller open variants, but the underlying point is unmistakable: the open-weights bargain has limits, and those limits will be set by internal preparedness frameworks rather than community pressure. That is a real shift from the Llama 2 era and worth tracking as the next generation lands.

    Wang’s willingness to engage on model welfare, on roughly the same footing as safety and alignment, is the second philosophical reveal worth flagging. It signals that the next generation of lab leadership is not going to dismiss the topic the way the previous generation often did. Whether that translates into product or policy changes is unclear, but the fact that the head of MSL says it is “underdiscussed” is itself a marker.

    Finally, the human texture of the interview matters. Wang has clearly absorbed a lot of personal incoming fire over the past ten months, including from LeCun and Altman, and his answer is consistently to redirect to the work. The Steve Jobs quote about hiring people who tell you what to do is the operating slogan he keeps coming back to. Combined with the genuine enthusiasm for sci-fi, walks in the woods, and country music, the picture that emerges is less the salesman caricature his critics paint and more a young technical operator betting that scoreboard work over a multi-year horizon will settle every argument that text on X cannot.

    Watch the full conversation here.

  • Krishna Rao on Anthropic Going From 9 Billion to 30 Billion ARR in One Quarter and the Compute Strategy Powering Claude

    Krishna Rao, Chief Financial Officer of Anthropic, sat down with Patrick O’Shaughnessy on Invest Like the Best for one of the most detailed public looks yet at the operating engine behind Claude. He covers how Anthropic compounded from $9 billion of run rate revenue at the start of the year to north of $30 billion by the end of Q1, why he spends 30 to 40 percent of his time on compute, the playbook for buying gigawatts of AI infrastructure across Trainium, TPU, and GPU platforms, how Anthropic prices its models, why returns to frontier intelligence keep climbing, and what the Mythos release tells us about the cyber capabilities of the next generation of Claude.

    TLDW

    Anthropic is running the most compute fungible frontier lab in the world, with active deployments across AWS Trainium, Google TPU, and Nvidia GPU, and an internal orchestration layer that lets a chip serve inference in the morning and run reinforcement learning the same evening. Krishna Rao explains the cone of uncertainty that governs gigawatt scale compute procurement, the floor Anthropic refuses to drop below on model development compute, the Jevons paradox unlock from cutting Opus pricing, the 500 percent annualized net dollar retention from enterprise customers, the layer cake of long term deals with Google, Broadcom, Amazon, and the recent xAI Colossus tie up in Memphis, the phased release of the Mythos model in response to spiking cyber capabilities, the internal use of Claude Code to produce statutory financial statements and run a Monthly Financial Review skill, and why the team believes scaling laws are alive and well. The interview also covers fundraising history through Series D and Series E, the $75 billion already raised plus another $50 billion coming, talent density beating talent mass during the Meta poaching wave, and Rao’s belief that biotech and drug discovery represent the most exciting frontier for AI.

    Key Takeaways

    • Anthropic entered the year with about $9 billion of run rate revenue and ended the first quarter with north of $30 billion of run rate revenue, a more than 3x leap driven by model intelligence gains and the products built around them.
    • Compute is described as the lifeblood of the company, the canvas everything else is built on, and the most consequential class of decisions Rao makes. Buy too much and you go bankrupt. Buy too little and you cannot serve customers or stay at the frontier.
    • Rao spends 30 to 40 percent of his time on compute, even today, and the leadership team meets repeatedly on both procurement and ongoing compute allocation.
    • Anthropic is the only frontier language lab actively using all three major chip platforms in production: AWS Trainium, Google TPU, and Nvidia GPU. It is also the only major model available on all three clouds.
    • Flexibility is the central design principle. Anthropic builds flexibility into the deals themselves, into the orchestration layer that maps workloads to chips, and into compilers built from the chip level up.
    • The cone of uncertainty frames procurement. Small differences in weekly or monthly growth compound into wildly different two year outcomes, so the team plans across a range of scenarios rather than a single point estimate, and ranges toward the upper end while protecting downside.
    • Compute allocation across the company sits in three buckets: model development and research, internal employee acceleration, and external customer serving. A non negotiable floor protects model development even when customer demand is tight.
    • Anthropic estimates that if it cut off internal employee use of its own models, the freed compute could serve billions of dollars of additional revenue. It chooses not to, because internal use compounds into better future models.
    • Intelligence is multi dimensional, not a single IQ score. Anthropic measures real world capability through customer feedback, long horizon task performance, tool use, computer use, and speed at agentic tasks, not just leaderboard benchmarks that have largely saturated.
    • Each Opus generation, 4 to 4.5 to 4.6 to 4.7, delivers both capability improvements and an efficiency multiplier on token processing. New models often serve customers at a fraction of the prior cost while doing more.
    • Reinforcement learning is described as inference inside a sandbox with a reward function, so model efficiency gains directly improve internal RL throughput. The flywheel is tightly coupled.
    • Over 90 percent of code at Anthropic is now written by Claude Code, and a large share of Claude Code itself is written by Claude Code.
    • Anthropic shipped roughly 30 distinct product and feature releases in January and the pace has accelerated since.
    • Scaling laws, in Anthropic’s internal data, are alive and well. The team holds itself to a skeptical scientific standard and still does not see them slowing down.
    • Anthropic recently signed a 5 gigawatt deal with Google and Broadcom for TPUs starting in 2027, plus an Amazon Trainium agreement for up to 5 gigawatts, totaling more than $100 billion in commitments. A significant portion lands this year and next year.
    • A new partnership for capacity at the xAI Colossus facility in Memphis was announced just before the interview, aimed at expanding consumer and prosumer capacity.
    • Pricing has been remarkably stable across Haiku, Sonnet, and Opus. The biggest deliberate change was lowering Opus pricing, which produced a textbook Jevons paradox: consumption rose far faster than the price drop, and the new Opus 4.6 and 4.7 slot in at the same price point.
    • Mythos is the first model Anthropic chose to release in a phased way because of a sharp spike in cyber capability. In an open source codebase where a prior model found 22 security vulnerabilities, Mythos found roughly 250.
    • The Mythos release framework focuses on defensive use first, expands access over time, and is presented as a template for future capability spikes.
    • Anthropic now sells to 9 of the Fortune 10 and reports net dollar retention above 500 percent on an annualized basis. These are not pilots. Rao describes signing two double digit million dollar commitments during a 20 minute Uber ride to the studio.
    • The platform strategy is mostly horizontal. Anthropic will go vertical with offerings like Claude for Financial Services, Claude for Life Sciences, and Claude Security where it can demonstrate the model’s capabilities, but expects most application value to accrue to customers building on top.
    • Investors raised over $75 billion in equity since Rao joined, with another $50 billion in commitments tied to the Amazon and Google deals. Capital intensity is real, but the raises fund the upper end of the cone of uncertainty more than they fund current losses.
    • The Series E close coincided with the day the DeepSeek news broke, forcing investors to reassess their AI thesis in real time. Anthropic closed the round anyway.
    • Inside finance, Claude now produces statutory financial statements for every Anthropic legal entity, with a human checker. A library of more than 70 finance specific skills underpins workflows.
    • A custom Monthly Financial Review skill produces a 90 to 95 percent ready monthly close report, so leadership discussion shifts from reconciling numbers to debating implications.
    • An internal real time analytics platform called Anthrop Stats compresses weekly insight cycles from hours to about 30 minutes.
    • The biggest token user inside Anthropic’s finance team is the head of tax, focused on tax policy engines and workflow automation. The most senior people, not the youngest, are leading internal adoption.
    • Talent density beats talent mass. When Meta and others ran aggressive offer waves, Anthropic lost two people while peer labs lost dozens.
    • All seven Anthropic co founders remain at the company, as does most of the first 20 to 30 employees, which Rao credits to a collaborative, transparent, debate friendly culture and a real culture interview that can veto otherwise top tier candidates.
    • Dario Amodei holds an open all hands every two weeks, writes a short prepared document, and takes unscripted questions from anyone at the company.
    • AI safety investments in interpretability and alignment have a commercial side effect. Looking inside the model helps Anthropic build better models, and enterprises selling sensitive workloads want to trust the lab they hand customer data to.
    • Anthropic explicitly identifies as America first in its approach to model development, and engages closely with the US administration on capability releases such as Mythos.
    • The longer term product vision is the virtual collaborator: an agent with organizational context, access to the company’s tools, persistent memory, and the ability to work on ideas, not just tasks, over long horizons.
    • CoWork, Anthropic’s extension of the Claude Code paradigm into general knowledge work, is being adopted faster than Claude Code itself when indexed to the same point in its launch curve.
    • Anthropic’s product teams ship daily, with a fleet of agents working across the company on specific tasks. Everyone effectively becomes a manager of agents.
    • The dominant downside risks to Anthropic’s high end forecast are slower customer diffusion of model capability into real workflows, scaling laws flattening unexpectedly, and Anthropic losing its position at the frontier.
    • Rao is most excited about biotech and healthcare outcomes, especially the prospect that AI could push drug discovery and lab throughput up 10x or 100x, turning currently incurable diagnoses into treatable ones within a patient’s lifetime.

    Detailed Summary

    Compute as Lifeblood and the Cone of Uncertainty

    Rao opens with the claim that compute is the most important resource at Anthropic, and the most consequential decision class in the company. You cannot buy a gigawatt of compute next week. You have to anticipate demand a year or two in advance, and the cost of being wrong in either direction is high. Buy too much and the unit economics collapse. Buy too little and you cannot serve customers or stay at the frontier, which are described as the same failure mode. To navigate this, the team uses a cone of uncertainty rather than point estimates. Small differences in weekly growth compound into vastly different two year outcomes, and Anthropic tries to position itself toward the upper end of that cone while preserving optionality. Rao notes he has had to consciously break a lifetime of linear thinking and force himself into exponential models.

    Three Chip Platforms, One Orchestration Layer

    Anthropic uses Amazon’s Trainium, Google’s TPUs, and Nvidia’s GPUs fungibly. That was not free. Adopting TPUs at scale started around the third TPU generation, when outside observers thought it was a strange choice. Anthropic invested years into compilers and orchestration so workloads can flow across chips by generation and by job type. The team works deeply with Annapurna Labs at AWS to influence Trainium roadmaps because Anthropic stresses these chips harder than almost anyone. The result is what Rao believes is the most efficient utilization of compute across any frontier lab, with a dollar of compute going further inside Anthropic than anywhere else.

    Three Buckets and the Model Development Floor

    Compute gets allocated across model development, internal acceleration of employees, and customer serving. The conversations are collaborative rather than zero sum, but there is a hard floor on model development that the company refuses to cross even if it makes customer demand harder to serve in the short term. The thesis is simple. The returns to frontier intelligence are extremely high, especially in enterprise, so cutting model investment to chase near term revenue is a bad trade. Internal employee use is also explicitly protected. Rao notes that diverting that internal usage to external customers would unlock billions of additional revenue today, but the compounding benefit of accelerating researchers and engineers outweighs that.

    Intelligence Is Multi Dimensional

    Rao pushes back hard on the IQ framing of model progress. Benchmarks saturate quickly, and the real signal comes from how customers actually use the models. Anthropic looks at long horizon task completion, tool use, computer use, and time to result on agentic tasks. Two equally capable agents who differ only in speed produce dramatically different value, because the faster one compounds into more attempts and more outcomes. Frontier model leaps are also fuel efficient. The sedan to sports car analogy breaks down because each Opus generation, 4 to 4.5 to 4.6 to 4.7, delivers a step up in capability and a multiplier on per token efficiency.

    From 9 Billion to 30 Billion ARR in One Quarter

    The headline number for the quarter is a leap from about $9 billion of run rate revenue to over $30 billion, accomplished without onboarding a corresponding step up in compute, because new compute lands on ramps locked in 12 months prior. Rao attributes the leap to model capability gains, products that surface that intelligence in usable form factors, and an enterprise customer base that pulls more workloads onto Claude as each generation unlocks new use cases. Coding started the wave with Sonnet 3.5 and 3.6, and the same pattern is now playing out elsewhere in the economy.

    Recursive Self Improvement and Talent Density

    Over 90 percent of Anthropic’s code is now written by Claude Code, including most of Claude Code itself. Rao describes this as a structural reason to keep allocating internal compute to employees even when external demand is hungry. Recursive self improvement is not happening through models that need no humans. It is happening through researchers who set direction and use frontier models to compress months of work into days. Talent density beats talent mass. When Meta and other labs went after Anthropic researchers with very large packages, Anthropic lost two people while peer labs lost dozens.

    Procurement Strategy and the Layer Cake

    Compute lands as a layer cake. Last month Anthropic signed a 5 gigawatt TPU deal with Google and Broadcom starting in 2027, alongside an Amazon Trainium agreement for up to 5 gigawatts. The total is north of $100 billion in commitments. A new tie up with xAI’s Colossus facility in Memphis was announced just before the interview, intended for nearer term capacity to support consumer and prosumer growth. Anthropic evaluates near term and long term compute deals against the same set of variables: price, duration, location, chip type, and how efficiently the team can run it. The relationships are deeper than procurement. The hyperscalers are also distribution channels for the model.

    Platform First, Selective Vertical Bets

    Rao describes Anthropic as a platform first business, with most expected value accruing to customers building on the platform. The team will only go vertical when it can either demonstrate capabilities that are skating to where the puck is going, like Claude Code did before the models could fully support it, or when it wants to set a template for an industry vertical, as with Claude for Financial Services, Claude for Life Sciences, and Claude Security. He acknowledges that surprise capability jumps make customers anxious about the platform competing with them, and frames Anthropic’s mitigation as deeper partnerships, early access programs, and an emphasis on accelerating customer building rather than disintermediating it.

    Pricing, Jevons Paradox, and Return on Compute

    Pricing across Haiku, Sonnet, and Opus has been stable. The notable exception is Opus, which Anthropic deliberately repriced lower when launching Opus 4.5 because Opus class problems were being squeezed into Sonnet workloads. Efficiency gains made it possible to serve Opus profitably at the new level. The consumption response was a classic Jevons paradox, with usage rising far more than the price reduction would have predicted, and Opus 4.6 then slotted in at the same price with a capability bump. Margins are not framed as a per token markup. Compute is fungible across model development, internal acceleration, and customer serving, so Anthropic measures return on the entire compute envelope rather than software style variable cost per call.

    Fundraising, DeepSeek, and Capital Intensity

    Rao joined while Anthropic was closing its Series D, mid frontier model launch and during the FTX share liquidation. Investors initially questioned whether Anthropic needed a frontier model, whether AI safety and a real business could coexist, and why the sales team was so small. The Series E closed the same day the DeepSeek news broke, with markets violently re pricing AI in real time. Since Rao joined, Anthropic has raised over $75 billion, with another $50 billion tied to the Amazon and Google compute deals. The reason for the size of the raises is the cone of uncertainty, not current losses. Returns on compute today are described as robust.

    Mythos, Cyber Capability, and Phased Releases

    The Mythos release marks the first time Anthropic shipped a model under a deliberately phased rollout because of a specific capability spike. Cyber is the dimension that spiked. Where a prior model found 22 vulnerabilities in an open source codebase, Mythos found roughly 250. The defensive applications, automatically patching massive codebases, are genuinely valuable, but the offensive risk is real enough that Anthropic chose to release to a smaller group first and expand access over time. Rao positions this as a template for future capability spikes, not a permanent restriction. He also describes the relationship with the US administration as cooperative, including the Department of War interaction, with Anthropic supporting a regulatory framework that does not strangle innovation but takes responsibility seriously.

    Claude Inside Finance

    Anthropic’s finance team is one of the strongest internal case studies. Statutory financial statements for every legal entity are produced by Claude, with a human reviewer. A skill library of more than 70 finance specific skills underpins a Monthly Financial Review skill that drafts the monthly close at 90 to 95 percent ready, so leadership meetings shift from explaining the numbers to discussing what to do about them. An internal analytics platform called Anthrop Stats compresses weekly insight cycles from hours to 30 minutes. The biggest internal token user in finance is the head of tax, building policy engines, which Rao highlights as evidence that adoption is driven by the most senior people, not just younger engineers.

    Culture, Co Founders, and the Race to the Top

    Seven co founders should not, on paper, work as a leadership group. Rao argues it works because the culture was set early around collaboration, intellectual honesty, transparency, and humility. The culture interview is a real veto, not a checkbox. Dario Amodei runs an all hands every two weeks with a short written piece followed by unscripted questions, and decisions, once made, get clean alignment rather than residual politics. Anthropic frames its approach as a race to the top, where being a model for how to build the technology responsibly is itself a recruiting and retention advantage.

    The Virtual Collaborator and the Frontier Ahead

    The product vision Rao describes is the virtual collaborator. Not just a smarter chatbot, but an agent with organizational context, access to the company’s tools, memory, and the ability to work on ideas over long horizons. Coding was the first domain to feel this, but CoWork, Anthropic’s extension of the Claude Code pattern into general knowledge work, is being adopted faster than Claude Code was at the same age. Product development inside Anthropic already looks different. Teams ship daily, with fleets of agents working across the company, and individual humans increasingly act as managers of those fleets.

    Downside Risks and What Excites Him Most

    The three risks Rao names if asked to do a premortem on a softer year are slower customer diffusion of model capability into real workflows, scaling laws unexpectedly flattening, and Anthropic losing its frontier position to competitors. None of these are observed today, but he is unwilling to claim them with certainty. On the upside, he is most excited about biotech and healthcare. Lab throughput rising 10x or 100x, paired with AI assisted clinical workflows, could turn currently incurable diagnoses into treatable ones within a patient’s lifetime. That is the outcome he wants the technology to chase.

    Thoughts

    The most consequential structural point in this interview is the framing of compute as a single fungible resource pool measured by return on the entire envelope, not as a variable cost per inference call. That accounting shift, if you accept it, breaks most of the bear cases about AI lab unit economics. The bear argument almost always assumes that a token served to a customer is the only thing the chip did that day. Rao’s version is that the same fleet trains models in the morning, runs reinforcement learning at lunch, serves customers in the afternoon, and accelerates internal engineers in the evening. If even half of that is real, the right comparison is total compute spend versus total enterprise value created by the platform, and on that ratio Anthropic looks structurally strong rather than weak.

    The Jevons paradox on Opus pricing is the most actionable insight for anyone running an AI product. Most teams default to either chasing premium pricing on the newest model or undercutting to chase volume. Anthropic did something more disciplined: it left Sonnet and Haiku alone, dropped Opus when efficiency gains made it serveable, and watched aggregate usage rise faster than the price cut. The lesson is that frontier model pricing is not really a price problem. It is a capability access problem, and elasticity around the right tier is much higher than the standard SaaS playbook implies.

    The Mythos cyber jump deserves more attention than it has gotten. Going from 22 to 250 vulnerabilities found in the same codebase is the kind of capability discontinuity that genuinely changes the regulatory calculus. Anthropic is signaling that it can identify these discontinuities ahead of release and choose a deployment shape that respects them. Whether peer labs adopt similar discipline is the open question. Anthropic’s race to the top framing assumes they will be forced to. The competitive market may say otherwise.

    The hiring data point is the most underrated investor signal. Two departures while peer labs lost dozens, during the most aggressive talent war in tech history, is not a culture poster. It is a structural advantage that compounds every time another lab tries to buy its way to the frontier. Money can be matched. Conviction in the mission, transparent leadership, and a culture interview that can veto otherwise stellar candidates cannot. If you believe scaling laws hold, talent retention at this density is one of the few moats that actually scales with capital.

    Finally, the most interesting personal admission is that Krishna Rao, a finance leader trained at Blackstone and Cedar, is openly telling investors that linear thinking is the failure mode he had to break out of. The companies that pattern match this moment to prior technology waves are mispricing it, in both directions. The cone of uncertainty Anthropic uses internally is the right metaphor for everyone else too. If you are forecasting AI as if it is cloud in 2010, you are almost certainly wrong, and the magnitude of the error is much larger than it would be in any prior era.

    Watch the full conversation with Krishna Rao on Invest Like the Best here.

  • Bubbles, Parabolas and Speed Crashes: How AI Agents Are Ending Human Market Structure and Why This Is Not the Dot-Com Bubble

    The host opens this Saturday morning macro and AI markets video with a direct challenge to anyone calling the current move a bubble. The argument is that the market structure itself has changed, that AI agents now dominate trading and capital allocation, and that Charles Kindleberger’s Manias, Panics, and Crashes describes a world that no longer exists. The full hour-long conversation walks through earnings, PEG ratios, capex, the benchmark arbitrage trapping passive investors, the inflation regime shift, and where money is rotating now. Watch the original video here.

    TLDW

    AI is not a bubble in the Kindleberger sense because the market is no longer dominated by emotional human professionals. AI agents, retail risk-takers, and passive flows are reshaping price discovery while the spend is being funded by free cash flow from the most cash-rich companies in history, not bond-issuance manias like telecoms or oil. Earnings growth is 27 percent, semiconductor sales grew 88 percent year over year in March, OpenAI and Anthropic revenue is on near-vertical curves, Nvidia’s PE is at decade lows even as Cisco’s was 130 at the dot-com peak, and the PEG ratio for the S&P sits at 1.03 with one third of the host’s thematic basket under 1.0 while Microsoft, Amazon, Meta, Apple, and Alphabet all carry richer PEGs. The new regime brings speed crashes instead of multi-year recessions, persistent bottlenecks in power, chips, transportation, and chemicals, inflation pressure that pushes three-month bills below CPI for the first time since the inflation era, and a benchmark arbitrage forcing passive money to chase AI exposure. The host is selling two thirds of his Micron, rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum, and warning that tokenization launches scheduled for July 26 will be the next major regime change.

    Key Takeaways

    • The word bubble is being misapplied because the same people calling AI a bubble called QE, tariffs, oil, Bitcoin, and passive investing bubbles for fifteen years and were wrong every time.
    • Kindleberger’s Manias, Panics, and Crashes described a slow, linear, human-emotion-driven world. AI agents have no emotion, no memory of Druckenmiller’s 2000 top, and one goal: make money.
    • The simplest test for anyone bearish on AI is to ask how much they use artificial intelligence. If they have not used a tool like OpenClaw or similar agentic systems, they are still operating in the old market regime.
    • This buildout is funded by free cash flow and bond issuance at yields better than US Treasuries from companies with stronger balance sheets than the federal government, unlike the dot-com telecoms or 1970s oil majors.
    • The S&P 500 is up only 7 percent year to date. The bubble framing is being applied to a handful of names, not to broad indices that remain reasonably valued.
    • The agentic stage of AI started in late November and accelerated when OpenClaw went viral at the end of January. Token consumption is set to grow 15 to 50 times from the IQ stage.
    • Anthropic revenue is stair-stepping from 5 to 7 to 9 to 14 to 19 to 24 to 30 billion in annualized run rate, on pace to surpass Alphabet in revenue by mid-2028.
    • OpenAI’s backlog hit 1.3 to 1.4 trillion in the most recent earnings cycle and the company still does not have enough compute.
    • Dario Amodei told the world Anthropic was planning for 10 times growth per year. In Q1 they saw 80 times annualized growth, which is why compute is bottlenecked and Anthropic is renting from Amazon, Google, and Colossus.
    • S&P 500 earnings growth is 27.1 percent year over year. The only quarters that match are those coming out of recessions, and this is not a reopening trade.
    • 320 of 500 S&P companies have reported and the average earnings surprise is 20 percent. Forward estimates are up 25 percent year over year as analysts revise upward against the historical pattern.
    • Total semiconductor sales grew 88 percent year over year in March. Semis have moved in proportion to earnings, not in excess of them.
    • Cisco’s PE was 130 at the dot-com peak. Nvidia’s PE today is the lowest of the last decade because professionals cannot run concentrated positions in single names.
    • The Edward Yardeni PEG ratio for the S&P is 1.03. The hyperscalers are not cheap on PEG: Microsoft 1.4, Amazon 1.66, Meta 1.96, Apple 3, Alphabet near 5. Thirty of ninety-five names in the host’s thematic portfolio carry PEGs under 1.0.
    • Passive investing creates a benchmark arbitrage. Everyone long the S&P 500 through index funds is structurally underweight Intel, Nvidia, Micron, and every name actually going up. Pension funds and mutual funds are forced to chase AI exposure to keep up.
    • BlackRock’s Tony Kim at the Milken conference: compute and model layers added 8 trillion in market cap year to date while the service apps that make up two thirds of GDP lost 1.2 trillion. The benchmark arbitrage is already running.
    • Larry Fink predicted a futures market for computing power. Power plus chips is the oil of the intelligence economy.
    • Jensen Huang called this a 90 trillion dollar AI physical upgrade cycle. The one big beautiful bill bonus depreciation provision was designed to incentivize this capex magic.
    • The host is selling two thirds of his Micron position. The reasoning is the memory market started moving in September of last year, the DRAM ETF is the ninth most traded ETF with billion dollar daily volumes, and exhaustion indicators are flashing red.
    • Money from Micron is rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum. The view is that the energy and power side of the AI stack is lagging the semis and will catch up next.
    • Silver versus gold has not moved while Micron has gone parabolic. LME metals are breaking out. China is increasing gold purchases significantly month over month.
    • The expected CPI print of 3.7 percent will put three-month Treasury bills below CPI for the first time since the post-pandemic inflation era. That is when Bitcoin started its last major run.
    • Logistics Managers Index hit 69.9 in March, the fastest expansion since March 2022. Transportation prices are surging because there is no capacity. This typically only happens during tax cuts or post-COVID reopenings.
    • Payroll job creation in information, professional services, and financial activities is negative. AI is already replacing knowledge work. Job creation has shifted to mining, manufacturing, construction, trade, transportation, and utilities, which is structurally inflationary.
    • Whirlpool says appliance demand is at great financial crisis lows. The consumer PC and laptop market collapse is worse than 2008. AI is pulling capital and pricing power away from legacy consumer categories.
    • Mike Wilson’s data shows reacceleration across sectors, not just large cap tech. Small caps and median stocks are showing earnings growth too, just at smaller market caps.
    • Chevron’s CEO says global oil shortages are starting. Jeff Currie warns US storage tanks will run empty. Ships are still not transiting the Strait of Hormuz. Countries that learned this lesson will restock to higher inventory levels permanently.
    • The Renmac Bubble Watch threshold was crossed on a technical basis. The host considers technical exhaustion a stronger signal than narrative-driven bubble calls.
    • Goldman Sachs power demand reports, Guggenheim warnings on the power crunch, and BlackRock’s compute intensity research all triangulate on the same conclusion: capex needs are larger than current forecasts.
    • The thematic portfolio is up roughly 30 percent from March lows. Power, optical fiber, advanced packaging, chemicals, and rack-level infrastructure baskets are leading.
    • Sterling Infrastructure (STRL), Fluence batteries, ABB electrification, Hon Hai (Foxconn), Vistra, Eaton, and Soitec are highlighted as names lagging the megacaps but inside the same AI infrastructure trade.
    • John Roque at 22V Research is releasing weekly frozen rope charts, long-base breakouts across power, copper, grid equipment, utilities, natural gas, transportation, capital goods, and agriculture. They all map to the same AI plus inflation regime.
    • Bitcoin ETF outstanding shares hit new highs. BlackRock, Morgan Stanley, and Goldman are all running competitive products. Boomer and wealth manager allocation is accelerating into year end.
    • Tokenization rolls out July 26. Wall Street clearing has enlisted 50 firms. A16Z published their case in December 2024. The host considers this underweighted by most investors and is speaking on the topic at the II event in Fort Lauderdale.
    • Raoul Pal and Yoni Assia on the end of human trading: AI agents and crypto collide by moving finance from human speed to machine speed. Agents will trade, allocate, hedge, and shift capital through wallets and exchanges. Tokenization means ownership becomes programmable.
    • The new regime is bubbles, parabolas, and speed crashes. Corrections compress from years into months. The right strategy is to never go to cash, only to rebalance and slow down within the portfolio.
    • For traders, exhaustion indicators using 5-day and 14-day RSI plus DeMark signals identify potential speed crash setups. Intel and Micron are flashing red on those screens right now.

    Detailed Summary

    Why this is not Kindleberger’s world anymore

    The framing argument of the video is that Manias, Panics, and Crashes described a market dominated by human professionals operating with limited information and lagged feedback loops. When supply and demand fell out of sync, prices collapsed because nobody could see what was happening in real time. That world is gone. AI agents now manage a majority of professional fund flows. Information moves instantaneously. Retail investors trade differently than institutional pros, and the capital structure of the entire market has changed. The host argues that since the Great Financial Crisis, the combination of QE and exponential corporate growth produced the only companies in history worth 25 trillion dollars combined with no net debt. Their AI capex is funded by free cash flow and high-grade bonds, not panicked bond issuance like the dot-com telecoms or oil majors of the 1970s.

    The Druckenmiller anchor and why FOMO is the wrong lens

    The video reads the Stanley Druckenmiller story of buying six billion in tech at the 2000 top and losing three billion in six weeks. Every professional carries that scar. It has shaped a generation of money managers into seeing parabolic moves and immediately calling bubble. The host’s counter is that recession calls from wealthy professionals are themselves a form of hope. Cash-rich investors root for crashes because crashes give them entry points. If the bubble never breaks the way it broke in 2000, those investors stay locked out, and that is precisely what the AI regime is doing.

    Earnings, revenue, and the reality test

    The video walks through current numbers in detail. S&P 500 earnings growth is running 27.1 percent year over year, which only happens coming out of recessions. 320 companies have reported with an average 20 percent earnings surprise. Forward estimates were revised up 25 percent year over year, well above the historical pattern of starting-year estimates getting cut. Total semiconductor sales were up 88 percent year over year in March. Anthropic’s revenue trajectory is stair-stepping from 5 to 30 billion in annualized run rate on the back of Claude Opus 4.5, putting it on track to surpass Alphabet by mid-2028. OpenAI is sitting on a 1.3 to 1.4 trillion backlog and still cannot get enough compute. Dario Amodei told the public Anthropic planned for 10 times growth per year and saw 80 times in Q1.

    PE, PEG, and the valuation argument

    Cisco’s PE at the dot-com peak was 130. Nvidia, the indisputable lead dog of the AI buildout, currently has a PE at the lowest of its last decade. The S&P 500’s PE is roughly where it has been since the post-COVID money printing era, far below the dot-com peak. Edward Yardeni’s PEG ratio for the index sits at 1.03. The host built a PEG screen for his ninety-five name thematic portfolio. Thirty of those names trade at a PEG under 1.0. The hyperscalers everyone holds passively are the expensive ones: Microsoft 1.4, Amazon 1.66, Meta 1.96, Apple 3, Alphabet near 5. The capacity for forward PE compression sits in the names retail and active rotational money are buying, not in the index core.

    The benchmark arbitrage trap

    Most money is now in passive investing. By construction, an S&P 500 or MSCI World allocation is underweight the names that are actually rising. Pension funds, mutual funds, and any active manager benchmarked to those indices is forced to add AI exposure to keep pace. BlackRock’s Tony Kim made this point at Milken: 8 trillion in market cap has accrued to compute and model layers year to date, while service apps representing two thirds of GDP lost 1.2 trillion. The host calls this benchmark arbitrage and considers it the single most underappreciated driver of the current move.

    The 90 trillion dollar physical upgrade cycle

    Jensen Huang’s framing of a 90 trillion dollar AI upgrade includes autos, phones, computers, humanoids, robotics, and the military stack. The host considers this a global race between the US and China. The one big beautiful bill included bonus depreciation specifically to incentivize the capex push. Greg Brockman’s interview with Sequoia made the point that demand for intelligence is effectively unlimited, and that every company outside the hyperscalers, Morgan Stanley, Goldman, Eli Lilly, Merck, United Healthcare, needs their own data center compute or their margins will not keep up with competitors. In a capitalist system, that forces broad enterprise AI spending.

    Speed crashes replace recessions

    The new regime has corrections but they are fast. Since 2020 we have had multiple 20 percent corrections compressed into weeks instead of years. The host expects this pattern to continue for the next decade. Bottlenecks in power, chips, transportation, chemicals, and skilled labor will produce inflation spikes that trigger speed crashes, not traditional credit-cycle recessions. The Logistics Managers Index reading of 69.9 in March, with capacity contraction near record lows, signals exactly this kind of bottleneck environment. The host’s strategy in this regime is to never go to cash, only to rebalance and slow down within the portfolio.

    The inflation regime shift and the rotation out of Micron

    The expected CPI print of 3.7 percent will put three-month Treasury bills below CPI for the first time since the post-pandemic inflation era, restoring negative real yields. That was the condition under which Bitcoin first launched its major bull moves. The host has sold two thirds of his Micron position despite continued bullish conviction on the name, because the memory market is the most stretched on exhaustion indicators and the DRAM ETF is trading at unprecedented volume. The capital is rotating into Nvidia, Vistra, silver, Bitcoin, and Ethereum. Silver versus gold has not moved while semis went parabolic. LME metals are breaking out. China is increasing gold purchases. The energy and power side of the stack is the next leg up.

    AI is breaking the consumer and the labor market

    Whirlpool reports appliance demand at financial crisis lows. PCs and laptops are collapsing worse than 2008. Phones, autos, housing, all the categories Kindleberger’s framework was built around are under pressure because AI is pulling capital and pricing power into compute, power, and chemicals. Payroll job creation in information, professional services, and financial activities is negative as AI takes knowledge work. Job creation is rotating into mining, construction, manufacturing, trade, transportation, and utilities, which is structurally inflationary because those sectors require physical capacity and wages. That combination, wage inflation plus commodity inflation, makes it very difficult for the Fed to ease, even with Kevin Warsh likely taking over.

    Crypto, tokenization, and AI agents at machine speed

    The final section pivots to crypto. Bitcoin ETF outstanding shares hit new highs, BlackRock’s product remains dominant, and Morgan Stanley and Goldman have launched competing vehicles. Wealth managers and boomers are allocating. The Raoul Pal and Yoni Assia conversation on the end of human trading is the host’s headline reference: AI agents will trade, allocate, hedge, and shift capital at machine speed through programmable wallets and exchanges. Tokenization, scheduled for a major launch on July 26 with 50 Wall Street clearing firms onboarded, makes ownership programmable. A16Z laid out the case in December 2024. The host is speaking on tokenization at the II event in Fort Lauderdale May 13 through 15 and considers it the next regime-defining shift after agentic AI.

    Thoughts

    The strongest argument in this video is structural, not narrative. The shift from human professionals with anchored memories to AI agents and benchmark-driven passive flows is a real change in who sets prices. Whether or not you accept the host’s portfolio calls, the framing should make any investor pause before defaulting to dot-com pattern recognition. Cisco’s PE was 130 with no business model. Nvidia’s PE is at a decade low with a near monopoly on the picks and shovels of the largest capex cycle in industrial history. Those facts cannot both be true and produce the same outcome.

    The PEG framework is the cleanest test in the video. If you believe Nvidia, Micron, Intel, and the second-tier AI infrastructure names are bubbles, you are implicitly betting that earnings growth collapses. That bet was viable in 2000 because the companies driving the move had no earnings. It is much harder to bet against earnings growth when 320 companies have just printed a 20 percent average earnings beat and analysts are revising forward estimates up by 25 percent. The host’s argument is not that the prices are reasonable in absolute terms. It is that the bear case requires growth to fall off a cliff, and nothing in the order books, the capex commitments, or the compute backlog suggests that is imminent.

    The benchmark arbitrage point deserves more attention than it gets. If the majority of professional money is locked in passive structures that are by definition underweight the leading names, and if those managers are evaluated quarter to quarter against the benchmark they cannot match, the pressure to chase will compound. This is the opposite of the dot-com setup, where active managers were forced to add overpriced tech to keep up with the index. Here, the index itself is structurally underweight the trade, and the active managers chasing it are doing so against names with rational PEG ratios.

    The rotation thesis from Micron into power, silver, and crypto is more debatable. The energy and bottleneck story is real, but the timing of when the power trade catches up with the semi trade is the hard part. The host’s discipline of never going to cash and rebalancing through the cycle is a sensible response to a regime that produces speed crashes rather than slow drawdowns. The investors most hurt by this regime will not be the ones who are long the wrong names. They will be the ones who sit out waiting for an entry point that never comes.

    Tokenization is the most underappreciated thread in the video. If the July 26 rollout brings 50 clearing firms and real ownership programmability online, the second half of the year could produce a regime shift on top of the AI regime shift. AI agents transacting on tokenized assets at machine speed is the logical endpoint of the trends the host has been tracking, and it is the part of his framework that current market consensus has not yet priced.

    Watch the full conversation here.

  • Dario Amodei on the AGI Exponential: Anthropic’s High-Stakes Financial Model and the Future of Intelligence

    TL;DW (Too Long; Didn’t Watch)

    Anthropic CEO Dario Amodei joined Dwarkesh Patel for a high-stakes deep dive into the endgame of the AI exponential. Amodei predicts that by 2026 or 2027, we will reach a “country of geniuses in a data center”—AI systems capable of Nobel Prize-level intellectual work across all digital domains. While technical scaling remains remarkably smooth, Amodei warns that the real-world friction of economic diffusion and the ruinous financial risks of $100 billion training clusters are now the primary bottlenecks to total global transformation.


    Key Takeaways

    • The Big Blob Hypothesis: Intelligence is an emergent property of scaling compute, data, and broad distribution; specific algorithmic “cleverness” is often just a temporary workaround for lack of scale.
    • AGI is a 2026-2027 Event: Amodei is 90% certain we reach genius-level AGI by 2035, with a strong “hunch” that the technical threshold for a “country of geniuses” arrives in the next 12-24 months.
    • Software Engineering is the First Domino: Within 6-12 months, models will likely perform end-to-end software engineering tasks, shifting human engineers from “writers” to “editors” and strategic directors.
    • The $100 Billion Gamble: AI labs are entering a “Cournot equilibrium” where massive capital requirements create a high barrier to entry. Being off by just one year in revenue growth projections can lead to company-wide bankruptcy.
    • Economic Diffusion Lag: Even after AGI-level capabilities exist in the lab, real-world adoption (curing diseases, legal integration) will take years due to regulatory “jamming” and organizational change management.

    Detailed Summary: Scaling, Risk, and the Post-Labor Economy

    The Three Laws of Scaling

    Amodei revisits his foundational “Big Blob of Compute” hypothesis, asserting that intelligence scales predictably when compute and data are scaled in proportion—a process he likens to a chemical reaction. He notes a shift from pure pre-training scaling to a new regime of Reinforcement Learning (RL) and Test-Time Scaling. These allow models to “think” longer at inference time, unlocking reasoning capabilities that pre-training alone could not achieve. Crucially, these new scaling laws appear just as smooth and predictable as the ones that preceded them.

    The “Country of Geniuses” and the End of Code

    A recurring theme is the imminent automation of software engineering. Amodei predicts that AI will soon handle end-to-end SWE tasks, including setting technical direction and managing environments. He argues that because AI can ingest a million-line codebase into its context window in seconds, it bypasses the months of “on-the-job” learning required by human engineers. This “country of geniuses” will operate at 10-100x human speed, potentially compressing a century of biological and technical progress into a single decade—a concept he calls the “Compressed 21st Century.”

    Financial Models and Ruinous Risk

    The economics of building the first AGI are terrifying. Anthropic’s revenue has scaled 10x annually (zero to $10 billion in three years), but labs are trapped in a cycle of spending every dollar on the next, larger cluster. Amodei explains that building a $100 billion data center requires a 2-year lead time; if demand growth slows from 10x to 5x during that window, the lab collapses. This financial pressure forces a “soft takeoff” where labs must remain profitable on current models to fund the next leap.

    Governance and the Authoritarian Threat

    Amodei expresses deep concern over “offense-dominant” AI, where a single misaligned model could cause catastrophic damage. He advocates for “AI Constitutions”—teaching models principles like “honesty” and “harm avoidance” rather than rigid rules—to allow for better generalization. Geopolitically, he supports aggressive chip export controls, arguing that democratic nations must hold the “stronger hand” during the inevitable post-AI world order negotiations to prevent a global “totalitarian nightmare.”


    Final Thoughts: The Intelligence Overhang

    The most chilling takeaway from this interview is the concept of the Intelligence Overhang: the gap between what AI can do in a lab and what the economy is prepared to absorb. Amodei suggests that while the “silicon geniuses” will arrive shortly, our institutions—the FDA, the legal system, and corporate procurement—are “jammed.” We are heading into a world of radical “biological freedom” and the potential cure for most diseases, yet we may be stuck in a decade-long regulatory bottleneck while the “country of geniuses” sits idle in their data centers. The winner of the next era won’t just be the lab with the most FLOPs, but the society that can most rapidly retool its institutions to survive its own technological adolescence.

    For more insights, visit Anthropic or check out the full transcript at Dwarkesh Patel’s Podcast.

  • AI Industry Pioneers Advocate for Consideration of Potential Challenges Amid Rapid Technological Progress

    AI Industry Pioneers Advocate for Consideration of Potential Challenges Amid Rapid Technological Progress

    On Tuesday, a collective of industry frontrunners plans to express their concern about the potential implications of artificial intelligence technology, which they have a hand in developing. They suggest that it could potentially pose significant challenges to society, paralleling the severity of pandemics and nuclear conflicts.

    The anticipated statement from the Center for AI Safety, a nonprofit organization, will call for a global focus on minimizing potential challenges from AI. This aligns it with other significant societal issues, such as pandemics and nuclear war. Over 350 AI executives, researchers, and engineers have signed this open letter.

    Signatories include chief executives from leading AI companies such as OpenAI’s Sam Altman, Google DeepMind’s Demis Hassabis, and Anthropic’s Dario Amodei.

    In addition, Geoffrey Hinton and Yoshua Bengio, two Turing Award-winning researchers for their pioneering work on neural networks, have signed the statement, along with other esteemed researchers. Yann LeCun, the third Turing Award winner, who leads Meta’s AI research efforts, had not signed as of Tuesday.

    This statement arrives amidst escalating debates regarding the potential consequences of artificial intelligence. Innovations in large language models, as employed by ChatGPT and other chatbots, have sparked concerns about the misuse of AI in spreading misinformation or possibly disrupting numerous white-collar jobs.

    While the specifics are not always elaborated, some in the field argue that unmitigated AI developments could lead to societal-scale disruptions in the not-so-distant future.

    Interestingly, these concerns are echoed by many industry leaders, placing them in the unique position of suggesting tighter regulations on the very technology they are working to develop and advance.

    In an attempt to address these concerns, Altman, Hassabis, and Amodei recently engaged in a conversation with President Biden and Vice President Kamala Harris on the topic of AI regulation. Following this meeting, Altman emphasized the importance of government intervention to mitigate the potential challenges posed by advanced AI systems.

    In an interview, Dan Hendrycks, executive director of the Center for AI Safety, suggested that the open letter represented a public acknowledgment from some industry figures who previously only privately expressed their concerns about potential risks associated with AI technology development.

    While some critics argue that current AI technology is too nascent to pose a significant threat, others contend that the rapid progress of AI has already exceeded human performance in some areas. These proponents believe that the emergence of “artificial general intelligence,” or AGI, an AI capable of performing a wide variety of tasks at or beyond human-level performance, may not be too far off.

    In a recent blog post, Altman, along with two other OpenAI executives, proposed several strategies to manage powerful AI systems responsibly. They proposed increased cooperation among AI developers, further technical research into large language models, and the establishment of an international AI safety organization akin to the International Atomic Energy Agency.

    Furthermore, Altman has endorsed regulations requiring the developers of advanced AI models to obtain a government-issued license.

    Earlier this year, over 1,000 technologists and researchers signed another open letter advocating for a six-month halt on the development of the largest AI models. They cited fears about an unregulated rush to develop increasingly powerful digital minds.

    The new statement from the Center for AI Safety is brief, aiming to unite AI experts who share general concerns about powerful AI systems, regardless of their views on specific risks or prevention strategies.

    Geoffrey Hinton, a high-profile AI expert, recently left his position at Google to openly discuss potential AI implications. The statement has since been circulated and signed by some employees at major AI labs.

    The recent increased use of AI chatbots for entertainment, companionship, and productivity, combined with the rapid advancements in the underlying technology, has amplified the urgency of addressing these concerns.

    Altman emphasized this urgency in his Senate subcommittee testimony, saying, “We want to work with the government to prevent [potential challenges].”