PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Tag: AI infrastructure

  • Jensen Huang on Nvidia’s Supply Chain Moat, TPU Competition, China Export Controls, and Why Nvidia Will Not Become a Cloud (Dwarkesh Podcast Summary)

    TLDW (Too Long, Didn’t Watch)

    Jensen Huang sat down with Dwarkesh Patel for over 90 minutes covering Nvidia’s supply chain dominance, the TPU threat, why Nvidia will not become a hyperscaler, whether the US should sell AI chips to China, and why Nvidia does not pursue multiple chip architectures at once. Jensen framed Nvidia’s entire business as transforming “electrons into tokens” and argued that Nvidia’s real moat is not any single technology but the full stack ecosystem it has built over two decades. He was blunt about his regret over not investing in Anthropic and OpenAI earlier, passionate about keeping the American tech stack dominant worldwide, and dismissive of the idea that China’s chip industry can be meaningfully contained through export controls.

    Key Takeaways

    1. Nvidia’s moat is the ecosystem, not the chip. Jensen repeatedly emphasized that Nvidia’s competitive advantage comes from CUDA, its massive installed base, its deep partnerships across the entire supply chain, and the fact that it operates in every cloud. The moat is not a single product but an interlocking system that took 20+ years to build.

    2. Supply chain bottlenecks are temporary, energy bottlenecks are not. Jensen argued that CoWoS packaging, HBM memory, EUV capacity, and logic fabrication bottlenecks can all be resolved in two to three years with the right demand signal. The real constraint on AI scaling is energy policy, which takes far longer to fix.

    3. TPUs and ASICs are not an existential threat to Nvidia. Jensen was emphatic that no competitor has demonstrated better price-performance or performance-per-watt than Nvidia, and challenged TPU and Trainium to prove otherwise on public benchmarks like InferenceMAX and MLPerf. He described Anthropic as a “unique instance, not a trend” for TPU adoption.

    4. Jensen regrets not investing in Anthropic and OpenAI earlier. He admitted he did not deeply internalize how much capital AI labs needed and that traditional VC funding was not sufficient for companies at that scale. He described this as a clear miss, though he said Nvidia was not in a position to make multi-billion dollar investments at the time.

    5. Nvidia will not become a hyperscaler. Jensen’s philosophy is “do as much as needed, as little as possible.” Building cloud infrastructure is something other companies can do, so Nvidia supports neoclouds like CoreWeave, Nebius, and Nscale instead of competing with them. Nvidia invests in ecosystem partners rather than vertically integrating into cloud services.

    6. Jensen is strongly against US chip export controls on China. This was the longest and most heated segment of the interview. Jensen argued that China already has abundant compute, energy, and AI researchers, and that export controls have accelerated China’s domestic chip industry while causing the US to concede the world’s second-largest technology market. He compared the situation to how US telecom policy allowed Huawei to dominate global telecommunications.

    7. AI will cause software tool usage to skyrocket, not collapse. Jensen pushed back on the narrative that AI will commoditize software companies. He argued that agents will use existing tools at massive scale, causing the number of instances of products like Excel, Synopsys Design Compiler, and other enterprise tools to grow exponentially.

    8. Nvidia does not pick winners among AI labs. Jensen explained that Nvidia invests across multiple foundation model companies simultaneously and refuses to favor any single one. He cited his own company’s unlikely survival story as the reason for this humility: Nvidia’s original graphics architecture was “precisely wrong” and would have been counted out by anyone picking winners.

    9. Nvidia added Groq for premium token economics. Nvidia recently acquired Groq and is folding it into the CUDA ecosystem because the market is now segmenting into different token tiers. Some customers will pay premium prices for faster response times even at lower throughput, creating a new segment of the inference market.

    10. Without AI, Nvidia would still be very large. Jensen was clear that accelerated computing, not AI specifically, is the foundational mission of the company. Molecular dynamics, quantum chemistry, computational lithography, data processing, and physics simulation all benefit from GPU acceleration regardless of deep learning.

    Detailed Summary

    Nvidia’s Real Business: Electrons to Tokens

    Jensen opened the conversation by reframing Nvidia’s entire value proposition. When Dwarkesh suggested that Nvidia is fundamentally a software company that sends a GDS2 file to TSMC for manufacturing, Jensen pushed back hard. He described Nvidia’s job as transforming electrons into tokens, with everything in between representing an “incredible journey” of artistry, engineering, science, and invention. He said the transformation is far from deeply understood and the journey is far from over, making commoditization unlikely.

    Jensen described Nvidia as operating a philosophy of doing “as much as necessary and as little as possible.” Whatever Nvidia does not need to do itself, it partners with someone else and makes it part of the broader ecosystem. This is why Nvidia has what Jensen called probably the largest ecosystem of partners in the industry, spanning the full supply chain upstream and downstream, application developers, model makers, and all five layers of the AI stack.

    On the question of whether AI will commoditize software companies, Jensen offered a contrarian take. He argued that agents are going to use software tools at unprecedented scale, meaning the number of instances of products like Excel, Cadence design tools, and Synopsys compilers will skyrocket. Today the bottleneck is the number of human engineers. Tomorrow, those engineers will be supported by swarms of agents exploring design spaces and using the same tools humans use today. Jensen said the reason this has not happened yet is simply that the agents are not good enough at using tools. That will change.

    The Supply Chain Moat

    Dwarkesh pressed Jensen on Nvidia’s reported $100 billion (and potentially $250 billion) in purchase commitments with foundries, memory manufacturers, and packaging companies. The question was whether Nvidia’s real moat for the next few years is simply locking up scarce upstream components so that no competitor can get the memory and logic they need to build alternative accelerators.

    Jensen confirmed this is a significant advantage but framed it differently. He said Nvidia has made enormous explicit and implicit commitments upstream. The implicit commitments matter just as much: Jensen personally meets with CEOs across the supply chain to explain the scale of the coming AI industry, convince them to invest in capacity, and assure them that Nvidia’s downstream demand is large enough to justify that investment. Nvidia’s GTC conference serves this purpose too, bringing the entire ecosystem together so upstream suppliers can see downstream demand and vice versa.

    Jensen described a process of systematically “prefetching bottlenecks” years in advance. CoWoS advanced packaging was a major bottleneck two years ago, but Nvidia swarmed it with repeated doubling of capacity until TSMC recognized it as mainstream computing technology rather than a specialty product. More recently, Nvidia has invested in the silicon photonics ecosystem through partnerships with Lumentum and Coherent, invented new packaging technologies, licensed patents to keep the supply chain open, and even invested in new testing equipment like double-sided probing.

    When Dwarkesh asked about the ultimate physical bottlenecks, Jensen surprised him. The hardest bottleneck to solve is not CoWoS or HBM or EUV machines. It is plumbers and electricians needed to build data centers. Jensen used this as a launching point to criticize “doomers” who discourage people from pursuing careers in software engineering or radiology, arguing that scaring people out of these professions creates the real bottlenecks.

    On EUV and logic scaling specifically, Jensen was optimistic. He said no supply chain bottleneck lasts longer than two to three years. Once you can build one of something, you can build ten, and once you can build ten, you can build a million. The key is a clear demand signal. If TSMC is convinced of the demand, ASML will produce enough EUV machines. Meanwhile, Nvidia continues to improve computing efficiency by 10x to 50x per generation through architecture, algorithms, and system design.

    The TPU Question

    Dwarkesh pushed hard on whether Google’s TPUs represent a real threat, noting that two of the top three AI models (Claude and Gemini) were trained on TPUs. Jensen drew a sharp distinction between what Nvidia builds and what a TPU is. Nvidia builds accelerated computing, which serves molecular dynamics, quantum chromodynamics, data processing, fluid dynamics, particle physics, and AI. A TPU is a tensor processing unit optimized for matrix multiplies. Nvidia’s market reach is far greater than any TPU or ASIC can possibly have.

    Jensen emphasized programmability as Nvidia’s core architectural advantage. If you want to invent a new attention mechanism, build a hybrid SSM model, fuse diffusion and autoregressive techniques, or disaggregate computation in a novel way, you need a generally programmable architecture. The only way to achieve 10x or 100x performance leaps (versus the roughly 25% per year from Moore’s Law) is to fundamentally change the algorithm, and that requires the flexibility CUDA provides.

    On the specific question of whether hyperscalers with huge engineering teams can simply write their own kernels and bypass CUDA, Jensen acknowledged they do write custom kernels but argued that Nvidia’s engineers still routinely deliver 2x to 3x speedups when they optimize a partner’s stack. He described Nvidia’s GPUs as “F1 racers” that anyone can drive at 100 mph, but extracting peak performance requires deep architectural expertise. Nvidia uses AI itself to generate many of its optimized kernels.

    Jensen was particularly blunt about public benchmarks. He pointed to Dylan Patel’s InferenceMAX benchmark and said neither TPU nor Trainium has been willing to demonstrate their claimed performance advantages on it. He said Nvidia’s performance-per-TCO is the best in the world, “bar none,” and challenged anyone to prove otherwise.

    Regarding Anthropic’s multi-gigawatt deal with Broadcom and Google for TPUs, Jensen called it “a unique instance, not a trend.” He said without Anthropic, there would be essentially no TPU growth and no Trainium growth. He traced this back to his own mistake: when Anthropic and OpenAI needed multi-billion dollar investments from their compute suppliers to get off the ground, Nvidia was not in a position to provide that capital. Google and AWS were, and in return, Anthropic committed to using their compute.

    Nvidia’s Investment Strategy and Regrets

    Jensen was unusually candid about his regret over not investing in foundation model companies earlier. He said he did not deeply internalize how different AI labs were from typical startups. A traditional VC would never put $5 to $10 billion into a single AI lab, but that was exactly what companies like OpenAI and Anthropic needed. By the time Jensen understood this, Nvidia was not in a financial or cultural position to make those kinds of investments.

    Now, Nvidia has invested approximately $30 billion in OpenAI and $10 billion in Anthropic. Jensen said he is delighted to support both and considers their existence essential for the world. But he acknowledged that these investments came at much higher valuations than would have been possible years earlier.

    Jensen explained Nvidia’s broader investment philosophy: support everyone, do not pick winners. He invests in one foundation model company, he invests in all of them. This comes from hard-won humility. When Nvidia started, there were 60 3D graphics companies. Nvidia’s original architecture was “precisely wrong” and the company would have been at the top of most lists to fail. Jensen said he has enough humility from that experience to know that you cannot predict which AI company will ultimately succeed.

    Why Nvidia Will Not Become a Hyperscaler

    Dwarkesh pointed out that Nvidia has the cash to build and operate its own cloud infrastructure, bypassing the middleman ecosystem that converts CapEx into OpEx for AI labs. Jensen rejected this path based on his core operating philosophy.

    If Nvidia did not build its computing platform, NVLink, and the CUDA ecosystem, nobody else would have done it. He is “completely certain” of that. These are things Nvidia must do. But the world has lots of clouds. If Nvidia did not build a cloud, someone else would show up. So the answer is to support the ecosystem instead: invest in CoreWeave, Nscale, Nebius, and others to help them exist and scale, rather than competing with them.

    Jensen was clear that Nvidia is not trying to be in the financing business either. When OpenAI needed a $30 billion investment before its IPO, Nvidia stepped up because OpenAI needed it and Nvidia deeply believed in the company. But these are targeted ecosystem investments, not a strategic pivot into cloud services.

    On GPU allocation during shortages, Jensen pushed back on the narrative that Nvidia strategically “fractures” the market by giving allocations to smaller neoclouds. He said the process is straightforward: you forecast demand, you place a purchase order, and it is first in, first out. Nvidia never changes prices based on demand. Jensen said he prefers to be dependable and serve as the foundation of the industry rather than extracting maximum short-term value.

    The China Debate

    The longest and most heated section of the interview was Jensen’s case against US chip export controls on China. This was a genuine debate, with Dwarkesh pushing the national security argument and Jensen pushing back forcefully.

    Jensen’s core argument rested on several pillars. First, China already has abundant compute. They manufacture 60% or more of the world’s mainstream chips, have massive energy infrastructure (including empty data centers with full power), and employ roughly 50% of the world’s AI researchers. The threshold of compute needed to build models like Anthropic’s Mythos has already been reached and exceeded by China’s existing infrastructure.

    Second, export controls have backfired. They accelerated China’s domestic chip industry, forced their AI ecosystem to optimize for internal architectures instead of the American tech stack, and caused the United States to concede the second-largest technology market in the world. Jensen compared this directly to how US telecom policy allowed Huawei to dominate global telecommunications infrastructure.

    Third, Jensen argued that AI is a five-layer stack (energy, chips, computing platform, models, applications) and the US needs to win at every layer. Fixating on one layer (models) at the expense of another layer (chips) is counterproductive. If Chinese open source AI models end up optimized for non-American hardware and that stack gets exported to the global south, the Middle East, Africa, and Southeast Asia, the US will have lost something far more valuable than whatever marginal compute advantage the export controls provided.

    Dwarkesh countered with the Mythos example: Anthropic’s new model found thousands of high-severity zero-day vulnerabilities across every major operating system and browser, including one that had existed in OpenBSD for 27 years. If China had enough compute to train and deploy a model like Mythos at scale before the US could prepare, the cyber-offensive capabilities would be devastating.

    Jensen’s response was direct. Mythos was trained on “fairly mundane capacity” that is already abundantly available in China. The amount of compute is not the bottleneck for that kind of breakthrough. Great computer science is, and China has no shortage of brilliant AI researchers. He pointed to DeepSeek as evidence: most advances in AI come from algorithmic innovation, not raw hardware. If China’s researchers can achieve breakthroughs like DeepSeek with limited hardware, imagine what they could do with more.

    Jensen also argued for dialogue over confrontation. He said it is essential that American and Chinese AI researchers are talking to each other, and that both countries agree on what AI should not be used for. The idea that you can prevent AI risks by cutting off chip sales, when the real advances come from algorithms and computer science, reflects a fundamental misunderstanding of how AI progress works.

    The debate ended without resolution, but Jensen’s final point was sharp: “I’m not talking to somebody who woke up a loser. That loser attitude, that loser premise, makes no sense to me.”

    Why Not Multiple Chip Architectures?

    Near the end of the interview, Dwarkesh asked why Nvidia does not run multiple parallel chip projects with different architectures, like a Cerebras-style wafer-scale design or a Dojo-style huge package, or even one without CUDA.

    Jensen’s answer was simple: “We don’t have a better idea.” Nvidia simulates all of these alternative approaches in its internal simulators and they are provably worse. The company works on exactly the projects it wants to work on. If the workload were to change dramatically (not just the algorithms, but the actual market shape), Nvidia might add other accelerators.

    In fact, Nvidia recently did exactly this by acquiring Groq. The inference market is now segmenting into different tiers. Some customers will pay premium prices for extremely fast response times even if throughput is lower. This creates a new “high ASP token” segment that justifies a different point on the performance curve. But Jensen was clear: if he had more money, he would put it all behind Nvidia’s existing architecture, not diversify into alternatives.

    Nvidia Without AI

    Jensen closed by saying that even if the deep learning revolution had never happened, Nvidia would be “very, very large.” The premise of the company has always been that general-purpose computing cannot scale indefinitely and that domain-specific acceleration is the way forward. Molecular dynamics, seismic processing, image processing, computational lithography, quantum chemistry, and data processing all benefit from GPU acceleration regardless of AI. Jensen said the fundamental promise of accelerated computing has not changed “not even a little bit.”

    Thoughts

    This interview is one of the most revealing Jensen Huang conversations in years, partly because Dwarkesh actually pushes back instead of lobbing softballs. A few things stand out.

    The Anthropic regret is real and significant. Jensen is essentially admitting that Nvidia’s biggest strategic miss of the AI era was not understanding that foundation model companies needed supplier-level capital commitments, not VC funding. The fact that Google and AWS used compute investments to lock in Anthropic’s architecture choices has had downstream consequences that Nvidia is still working to unwind. When Jensen says Anthropic is “a unique instance, not a trend” for TPU adoption, he is simultaneously downplaying the threat and revealing exactly how seriously he takes it.

    The China debate is the highlight. Jensen’s argument is more nuanced than it first appears. He is not saying “sell China everything.” He is saying the current binary approach of near-total restriction has backfired by accelerating China’s domestic chip industry and pushing the Chinese AI ecosystem away from the American tech stack. His comparison to the US telecom industry losing global market share to Huawei is pointed and historically grounded. Whether you agree with his conclusion or not, the framing of AI as a five-layer stack where the US needs to compete at every layer is a useful mental model.

    The “electrons to tokens” framing is Jensen at his best. It is a simple metaphor that captures something genuinely complex about where value is created in the AI supply chain. And his insistence that the transformation is “far from deeply understood” is a subtle way of arguing that Nvidia’s competitive position will be durable because the problem space is not close to being solved.

    The Groq acquisition reveal is interesting for what it signals about the inference market. If Nvidia is creating a separate product tier for premium-priced, low-latency tokens, it suggests the company sees inference economics fragmenting significantly. This aligns with the broader trend of AI becoming an enterprise product where different customers have wildly different willingness to pay based on how they use tokens.

    Finally, Jensen’s refusal to diversify chip architectures is a bold bet. “We simulate it all in our simulator, provably worse” is an incredibly confident statement. History is full of companies that were right until they were not. But Nvidia’s track record of 50x generation-over-generation improvements through co-design across processors, fabric, libraries, and algorithms is hard to argue with. The question is whether the current paradigm of transformer-based models on GPU clusters represents a local or global optimum for AI compute.

  • The Genesis Mission: Inside the “Manhattan Project” for AI-Driven Science

    TL;DR

    On November 24, 2025, President Trump signed an Executive Order launching “The Genesis Mission.” This initiative aims to centralize federal data and high-performance computing under the Department of Energy to create a massive AI platform. Likened to the World War II Manhattan Project, its goal is to accelerate scientific discovery in critical fields like nuclear energy, biotechnology, and advanced manufacturing.

    Key Takeaways

    • The “Manhattan Project” of AI: The Administration frames this as a historic national effort comparable in urgency to the project that built the atomic bomb, aimed now at global technology dominance.
    • Department of Energy Leads: The Secretary of Energy will oversee the mission, leveraging National Labs and supercomputing infrastructure.
    • The “Platform”: A new “American Science and Security Platform” will be built to host AI agents, foundation models, and secure federal datasets.
    • Six Core Challenges: The mission initially focuses on advanced manufacturing, biotechnology, critical materials, nuclear energy, quantum information science, and semiconductors.
    • Data is the Fuel: The order prioritizes unlocking the “world’s largest collection” of federal scientific datasets to train these new AI models.

    Detailed Summary of the Executive Order

    The Executive Order, titled Launching the Genesis Mission, establishes a coordinated national effort to harness Artificial Intelligence for scientific breakthroughs. Here is how the directive breaks down:

    1. Purpose and Ambition

    The order asserts that America is currently in a race for global technology dominance in AI. To win this race, the Administration is launching the “Genesis Mission,” described as a dedicated effort to unleash a new age of AI-accelerated innovation. The explicit goal is to secure energy dominance, strengthen national security, and multiply the return on taxpayer investment in R&D.

    2. The American Science and Security Platform

    The core mechanism of this mission is the creation of the American Science and Security Platform. This infrastructure will provide:

    • Compute: Secure cloud-based AI environments and DOE national lab supercomputers.
    • AI Agents: Autonomous agents designed to test hypotheses, automate research workflows, and explore design spaces.
    • Data: Access to proprietary, federally curated, and open scientific datasets, as well as synthetic data generated by DOE resources.

    3. Timeline and Milestones

    The Secretary of Energy is on a tight schedule to operationalize this vision:

    • 90 Days: Identify all available federal computing and storage resources.
    • 120 Days: Select initial data/model assets and develop a cybersecurity plan for incorporating data from outside the federal government.
    • 270 Days: Demonstrate an “initial operating capability” of the Platform for at least one national challenge.

    4. Targeted Scientific Domains

    The mission is not open-ended; it focuses on specific high-impact areas. Within 60 days, the Secretary must submit a list of at least 20 challenges, spanning priority domains including Biotechnology, Nuclear Fission and Fusion, Quantum Information Science, and Semiconductors.

    5. Public-Private and International Collaboration

    While led by the DOE, the mission explicitly calls for bringing together “brilliant American scientists” from universities and pioneering businesses. The Secretary is tasked with developing standardized frameworks for IP ownership, licensing, and trade-secret protections to encourage private sector participation.


    Analysis and Thoughts

    “The Genesis Mission will… multiply the return on taxpayer investment into research and development.”

    The Data Sovereignty Play
    The most significant aspect of this order is the recognition of federal datasets as a strategic asset. By explicitly mentioning the “world’s largest collection of such datasets” developed over decades, the Administration is leveraging an asset that private companies cannot easily duplicate. This suggests a shift toward “Sovereign AI” where the government doesn’t just regulate AI, but builds the foundational models for science.

    Hardware over Software
    Placing this under the Department of Energy (DOE) rather than the National Science Foundation (NSF) or Commerce is a strategic signal. The DOE owns the National Labs (like Oak Ridge and Lawrence Livermore) and the world’s fastest supercomputers. This indicates the Administration views this as a heavy-infrastructure challenge—requiring massive energy and compute—rather than just a software problem.

    The “Manhattan Project” Framing
    Invoking the Manhattan Project sets an incredibly high bar. That project resulted in a singular, world-changing weapon. The Genesis Mission aims for a broader diffusion of “AI agents” to automate research. The success of this mission will depend heavily on the integration mentioned in Section 2—getting academic, private, and classified federal systems to talk to each other without compromising security.

    The Energy Component
    It is notable that nuclear fission and fusion are highlighted as specific challenges. AI is notoriously energy-hungry. By tasking the DOE with solving energy problems using AI, the mission creates a feedback loop: better AI designs better power plants, which power better AI.

  • Inside Microsoft’s AGI Masterplan: Satya Nadella Reveals the 50-Year Bet That Will Redefine Computing, Capital, and Control

    1) Fairwater 2 is live at unprecedented scale, with Fairwater 4 linking over a 1 Pb AI WAN

    Nadella walks through the new Fairwater 2 site and states Microsoft has targeted a 10x training capacity increase every 18 to 24 months relative to GPT-5’s compute. He also notes Fairwater 4 will connect on a one petabit network, enabling multi-site aggregation for frontier training, data generation, and inference.

    2) Microsoft’s MAI program, a parallel superintelligence effort alongside OpenAI

    Microsoft is standing up its own frontier lab and will “continue to drop” models in the open, with an omni-model on the roadmap and high-profile hires joining Mustafa Suleyman. This is a clear signal that Microsoft intends to compete at the top tier while still leveraging OpenAI models in products.

    3) Clarification on IP: Microsoft says it has full access to the GPT family’s IP

    Nadella says Microsoft has access to all of OpenAI’s model IP (consumer hardware excluded) and shared that the firms co-developed system-level designs for supercomputers. This resolves long-standing ambiguity about who holds rights to GPT-class systems.

    4) New exclusivity boundaries: OpenAI’s API is Azure-exclusive, SaaS can run elsewhere with limited exceptions

    The interview spells out that OpenAI’s platform API must run on Azure. ChatGPT as SaaS can be hosted elsewhere only under specific carve-outs, for example certain US government cases.

    5) Per-agent future for Microsoft’s business model

    Nadella describes a shift where companies provision Windows 365 style computers for autonomous agents. Licensing and provisioning evolve from per-user to per-user plus per-agent, with identity, security, storage, and observability provided as the substrate.

    6) The 2024–2025 capacity “pause” explained

    Nadella confirms Microsoft paused or dropped some leases in the second half of last year to avoid lock-in to a single accelerator generation, keep the fleet fungible across GB200, GB300, and future parts, and balance training with global serving to match monetization.

    7) Concrete scaling cadence disclosure

    The 10x training capacity target every 18 to 24 months is stated on the record while touring Fairwater 2. This implies the next frontier runs will be roughly an order of magnitude above GPT-5 compute.

    8) Multi-model, multi-supplier posture

    Microsoft will keep using OpenAI models in products for years, build MAI models in parallel, and integrate other frontier models where product quality or cost warrants it.

    Why these points matter

    • Industrial scale: Fairwater’s disclosed networking and capacity targets set a new bar for AI factories and imply rapid model scaling.
    • Strategic independence: MAI plus GPT IP access gives Microsoft a dual track that reduces single-partner risk.
    • Ecosystem control: Azure exclusivity for OpenAI’s API consolidates platform power at the infrastructure layer.
    • New revenue primitives: Per-agent provisioning reframes Microsoft’s core metrics and pricing.

    Pull quotes

      “We’ve tried to 10x the training capacity every 18 to 24 months.”

      “The API is Azure-exclusive. The SaaS business can run anywhere, with a few exceptions.”

      “We have access to the GPT family’s IP.”

    TL;DW

    • Microsoft is building a global network of AI super-datacenters (Fairwater 2 and beyond) designed for fast upgrade cycles and cross-region training at petabit scale.
    • Strategy spans three layers: infrastructure, models, and application scaffolding, so Microsoft creates value regardless of which model wins.
    • AI economics shift margins, so Microsoft blends subscriptions with metered consumption and focuses on tokens per dollar per watt.
    • Future includes autonomous agents that get provisioned like users with identity, security, storage, and observability.
    • Trust and sovereignty are central. Microsoft leans into compliant, sovereign cloud footprints to win globally.

    Detailed Summary

    1) Fairwater 2: AI Superfactory

    Microsoft’s Fairwater 2 is presented as the most powerful AI datacenter yet, packing hundreds of thousands of GB200 and GB300 accelerators, tied by a petabit AI WAN and designed to stitch training jobs across buildings and regions. The key lesson: keep the fleet fungible and avoid overbuilding for a single hardware generation as power density and cooling change with each wave like Vera Rubin and Rubin Ultra.

    2) The Three-Layer Strategy

    • Infrastructure: Azure’s hyperscale footprint, tuned for training, data generation, and inference, with strict flexibility across model architectures.
    • Models: Access to OpenAI’s GPT family for seven years plus Microsoft’s own MAI roadmap for text, image, and audio, moving toward an omni-model.
    • Application Scaffolding: Copilots and agent frameworks like GitHub’s Agent HQ and Mission Control that orchestrate many agents on real repos and workflows.

    This layered approach lets Microsoft compete whether the value accrues to models, tooling, or infrastructure.

    3) Business Models and Margins

    AI raises COGS relative to classic SaaS, so pricing blends entitlements with consumption tiers. GitHub Copilot helped catalyze a multibillion market in a year, even as rivals emerged. Microsoft aims to ride a market that is expanding 10x rather than clinging to legacy share. Efficiency focus: tokens per dollar per watt through software optimization as much as hardware.

    4) Copilot, GitHub, and Agent Control Planes

    GitHub becomes the control plane for multi-agent development. Agent HQ and Mission Control aim to let teams launch, steer, and observe multiple agents working in branches, with repo-native primitives for issues, actions, and reviews.

    5) Models vs Scaffolding

    Nadella argues model monopolies are checked by open source and substitution. Durable value sits in the scaffolding layer that brings context, data liquidity, compliance, and deep tool knowledge, exemplified by Excel Agent that understands formulas and artifacts beyond screen pixels.

    6) Rise of Autonomous Agents

    Two worlds emerge: human-in-the-loop Copilots and fully autonomous agents. Microsoft plans to provision agents with computers, identity, security, storage, and observability, evolving end-user software into an infrastructure business for agents as well as people.

    7) MAI: Microsoft’s In-House Frontier Effort

    Microsoft is assembling a top-tier lab led by Mustafa Suleyman and veterans from DeepMind and Google. Early MAI models show progress in multimodal arenas. The plan is to combine OpenAI access with independent research and product-optimized models for latency and cost.

    8) Capex and Industrial Transformation

    Capex has surged. Microsoft frames this era as capital intensive and knowledge intensive. Software scheduling, workload placement, and continual throughput improvements are essential to maximize returns on a fleet that upgrades every 18 to 24 months.

    9) The Lease Pause and Flexibility

    Microsoft paused some leases to avoid single-generation lock-in and to prevent over-reliance on a small number of mega-customers. The portfolio favors global diversity, regulatory alignment, balanced training and inference, and location choices that respect sovereignty and latency needs.

    10) Chips and Systems

    Custom silicon like Maia will scale in lockstep with Microsoft’s own models and OpenAI collaboration, while Nvidia remains central. The bar for any new accelerator is total fleet TCO, not just raw performance, and system design is co-evolved with model needs.

    11) Sovereign AI and Trust

    Nations want AI benefits with continuity and control. Microsoft’s approach combines sovereign cloud patterns, data residency, confidential computing, and compliance so countries can adopt leading AI while managing concentration risk. Nadella emphasizes trust in American technology and institutions as a decisive global advantage.


    Key Takeaways

    1. Build for flexibility: Datacenters, pricing, and software are optimized for fast evolution and multi-model support.
    2. Three-layer stack wins: Infrastructure, models, and scaffolding compound each other and hedge against shifts in where value accrues.
    3. Agents are the next platform: Provisioned like users with identity and observability, agents will demand a new kind of enterprise infrastructure.
    4. Efficiency is king: Tokens per dollar per watt drives margins more than any single chip choice.
    5. Trust and sovereignty matter: Compliance and credible guarantees are strategic differentiators in a bipolar world.
  • Composer: Building a Fast Frontier Model with Reinforcement Learning

    Composer represents Cursor’s most ambitious step yet toward a new generation of intelligent, high-speed coding agents. Built through deep reinforcement learning (RL) and large-scale infrastructure, Composer delivers frontier-level results at speeds up to four times faster than comparable models:contentReference[oaicite:0]{index=0}. It isn’t just another large language model; it’s an actively trained software engineering assistant optimized to think, plan, and code with precision — in real time.

    From Cheetah to Composer: The Evolution of Speed

    The origins of Composer go back to an experimental prototype called Cheetah, an agent Cursor developed to study how much faster coding models could get before hitting usability limits. Developers consistently preferred the speed and fluidity of an agent that responded instantly, keeping them “in flow.” Cheetah proved the concept, but it was Composer that matured it — integrating reinforcement learning and mixture-of-experts (MoE) architecture to achieve both speed and intelligence.

    Composer’s training goal was simple but demanding: make the model capable of solving real-world programming challenges in real codebases using actual developer tools. During RL, Composer was given tasks like editing files, running terminal commands, performing semantic searches, or refactoring code. Its objective wasn’t just to get the right answer — it was to work efficiently, using minimal steps, adhering to existing abstractions, and maintaining code quality:contentReference[oaicite:1]{index=1}.

    Training on Real Engineering Environments

    Rather than relying on synthetic datasets or static benchmarks, Cursor trained Composer within a dynamic software environment. Every RL episode simulated an authentic engineering workflow — debugging, writing unit tests, applying linter fixes, and performing large-scale refactors. Over time, Composer developed behaviors that mirror an experienced developer’s workflow. It learned when to open a file, when to search globally, and when to execute a command rather than speculate.

    Cursor’s evaluation framework, Cursor Bench, measures progress by realism rather than abstract metrics. It compiles actual agent requests from engineers and compares Composer’s solutions to human-curated optimal responses. This lets Cursor measure not just correctness, but also how well the model respects a team’s architecture, naming conventions, and software practices — metrics that matter in production environments.

    Reinforcement Learning as a Performance Engine

    Reinforcement learning is at the heart of Composer’s performance. Unlike supervised fine-tuning, which simply mimics examples, RL rewards Composer for producing high-quality, efficient, and contextually relevant work. It actively learns to choose the right tools, minimize unnecessary output, and exploit parallelism across tasks. The model was even rewarded for avoiding unsupported claims — pushing it to generate more verifiable and responsible code suggestions.

    As RL progressed, emergent behaviors appeared. Composer began autonomously running semantic searches to explore codebases, fixing linter errors, and even generating and executing tests to validate its own work. These self-taught habits transformed it from a passive text generator into an active agent capable of iterative reasoning.

    Infrastructure at Scale: Thousands of Sandboxed Agents

    Behind Composer’s intelligence is a massive engineering effort. Training large MoE models efficiently requires significant parallelization and precision management. Cursor’s infrastructure, built with PyTorch and Ray, powers asynchronous RL at scale. Their system supports thousands of simultaneous environments, each a sandboxed virtual workspace where Composer experiments safely with file edits, code execution, and search queries.

    To achieve this scale, the team integrated MXFP8 MoE kernels with expert and hybrid-sharded data parallelism. This setup allows distributed training across thousands of NVIDIA GPUs with minimal communication cost — effectively combining speed, scale, and precision. MXFP8 also enables faster inference without any need for post-training quantization, giving developers real-world performance gains instantly.

    Cursor’s infrastructure can spawn hundreds of thousands of concurrent sandboxed coding environments. This capability, adapted from their Background Agents system, was essential to unify RL experiments with production-grade conditions. It ensures that Composer’s training environment matches the complexity of real-world coding, creating a model genuinely optimized for developer workflows.

    The Cursor Bench and What “Frontier” Means

    Composer’s benchmark performance earned it a place in what Cursor calls the “Fast Frontier” class — models designed for efficient inference while maintaining top-tier quality. This group includes systems like Haiku 4.5 and Gemini Flash 2.5. While GPT-5 and Sonnet 4.5 remain the strongest overall, Composer outperforms nearly every open-weight model, including Qwen Coder and GLM 4.6:contentReference[oaicite:2]{index=2}. In tokens-per-second performance, Composer’s throughput is among the highest ever measured under the standardized Anthropic tokenizer.

    Built by Developers, for Developers

    Composer isn’t just research — it’s in daily use inside Cursor. Engineers rely on it for their own development, using it to edit code, manage large repositories, and explore unfamiliar projects. This internal dogfooding loop means Composer is constantly tested and improved in real production contexts. Its success is measured by one thing: whether it helps developers get more done, faster, and with fewer interruptions.

    Cursor’s goal isn’t to replace developers, but to enhance them — providing an assistant that acts as an extension of their workflow. By combining fast inference, contextual understanding, and reinforcement learning, Composer turns AI from a static completion tool into a real collaborator.

    Wrap Up

    Composer represents a milestone in AI-assisted software engineering. It demonstrates that reinforcement learning, when applied at scale with the right infrastructure and metrics, can produce agents that are not only faster but also more disciplined, efficient, and trustworthy. For developers, it’s a step toward a future where coding feels as seamless and interactive as conversation — powered by an agent that truly understands how to build software.

  • Why Every Nation Needs Its Own AI Strategy: Insights from Jensen Huang & Arthur Mensch

    In a world where artificial intelligence (AI) is reshaping economies, cultures, and security, the stakes for nations have never been higher. In a recent episode of The a16z Podcast, Jensen Huang, CEO of NVIDIA, and Arthur Mensch, co-founder and CEO of Mistral, unpack the urgent need for sovereign AI—national strategies that ensure countries control their digital futures. Drawing from their discussion, this article explores why every nation must prioritize AI, the economic and cultural implications, and practical steps to build a robust strategy.

    The Global Race for Sovereign AI

    The conversation kicks off with a powerful idea: AI isn’t just about computing—it’s about culture, economics, and sovereignty. Huang stresses that no one will prioritize a nation’s unique needs more than the nation itself. “Nobody’s going to care more about the Swedish culture… than Sweden,” he says, highlighting the risk of digital dependence on foreign powers. Mensch echoes this, framing AI as a tool nations must wield to avoid modern digital colonialization—where external entities dictate a country’s technological destiny.

    AI as a General-Purpose Technology

    Mensch positions AI as a transformative force, comparable to electricity or the internet, with applications spanning agriculture, healthcare, defense, and beyond. Yet Huang cautions against waiting for a universal solution from a single provider. “Intelligence is for everyone,” he asserts, urging nations to tailor AI to their languages, values, and priorities. Mistral’s M-Saaba model, optimized for Arabic, exemplifies this—outperforming larger models by focusing on linguistic and cultural specificity.

    Economic Implications: A Game-Changer for GDP

    The economic stakes are massive. Mensch predicts AI could boost GDP by double digits for countries that invest wisely, warning that laggards will see wealth drain to tech-forward neighbors. Huang draws a parallel to the electricity era: nations that built their own grids prospered, while others became reliant. For leaders, this means securing chips, data centers, and talent to capture AI’s economic potential—a must for both large and small nations.

    Cultural Infrastructure and Digital Workforce

    Huang introduces a compelling metaphor: AI as a “digital workforce” that nations must onboard, train, and guide, much like human employees. This workforce should embody local values and laws, something no outsider can fully replicate. Mensch adds that AI’s ability to produce content—text, images, voice—makes it a social construct, deeply tied to a nation’s identity. Without control, countries risk losing their cultural sovereignty to centralized models reflecting foreign biases.

    Open-Source vs. Closed AI: A Path to Independence

    Both Huang and Mensch advocate for open-source AI as a cornerstone of sovereignty. Mensch explains that models like Mistral Nemo, developed with NVIDIA, empower nations to deploy AI on their own infrastructure, free from closed-system dependency. Open-source also fuels innovation—Mistral’s releases spurred Meta and others to follow suit. Huang highlights its role in niche markets like healthcare and mining, plus its security edge: global scrutiny makes open models safer than opaque alternatives.

    Risks and Challenges of AI Adoption

    Leaders often worry about public backlash—will AI replace jobs? Mensch suggests countering this by upskilling citizens and showcasing practical benefits, like France’s AI-driven unemployment agency connecting workers to opportunities. Huang sees AI as “the greatest equalizer,” noting more people use ChatGPT than code in C++, shrinking the tech divide. Still, both acknowledge the initial hurdle: setting up AI systems is tough, though improving tools make it increasingly manageable.

    Building a National AI Strategy

    Huang and Mensch offer a blueprint for action:

    • Talent: Train a local workforce to customize AI systems.
    • Infrastructure: Secure chips from NVIDIA and software from partners like Mistral.
    • Customization: Adapt open-source models with local data and culture.
    • Vision: Prepare for agentic and physical AI breakthroughs in manufacturing and science.

    Huang predicts the next decade will bring AI that thinks, acts, and understands physics—revolutionizing industries vital to emerging markets, from energy to manufacturing.

    Why It’s Urgent

    The podcast ends with a clarion call: AI is “the most consequential technology of all time,” and nations must act now. Huang urges leaders to engage actively, not just admire from afar, while Mensch emphasizes education and partnerships to safeguard economic and cultural futures. For more, follow Jensen Huang (@nvidia) and Arthur Mensch (@arthurmensch) on X, or visit NVIDIA and Mistral’s websites.

  • How NVIDIA is Revolutionizing Computing with AI: Jensen Huang on AI Infrastructure, Digital Employees, and the Future of Data Centers

    NVIDIA CEO Jensen Huang discusses the company’s role in revolutionizing computing through AI, emphasizing decade-long investments in scalable, interconnected AI infrastructure, breakthroughs in efficiency, and the future of digital and embodied AI as transformative for industries globally.


    NVIDIA is transforming the landscape of computing, driving innovation at every level from data centers to digital employees. In a recent conversation with Jensen Huang, NVIDIA’s CEO, he offered a rare look at the strategic direction and long-term vision that has positioned NVIDIA as a leader in the AI revolution. Through decade-long infrastructure investments, NVIDIA is not just building hardware but creating “AI factories” that promise to impact industries globally.

    Decade-Long Investments in AI Infrastructure

    For NVIDIA, success has come from looking far into the future. Jensen Huang emphasized the company’s commitment to ten-year investments in scalable, efficient AI infrastructure. With an eye on exponential growth, NVIDIA has focused on creating solutions that can continue to meet demand as AI expands in complexity and scope. One of the cornerstones of this approach is NVLink technology, which enables GPUs to function as a unified supercomputer, allowing unprecedented scale for AI applications.

    This vision aligns with Huang’s goal of optimizing data centers for high-performance AI, making NVIDIA’s infrastructure not only capable of tackling today’s AI challenges but prepared for tomorrow’s even larger-scale demands.

    Outpacing Moore’s Law with Full-Stack Integration

    Huang highlighted how NVIDIA aims to surpass the limits of traditional computing, especially Moore’s Law, by focusing on a full-stack integration strategy. This strategy involves designing hardware and software as a cohesive unit, enabling a 240x reduction in AI computation costs while increasing efficiency. With this approach, NVIDIA has managed to achieve performance improvements that far exceed conventional expectations, driving both cost and energy usage down across its AI operations.

    The full-stack approach has enabled NVIDIA to continually upgrade its infrastructure and enhance performance, ensuring that each component of its architecture is optimized and aligned.

    The Evolution of Data Centers: From Storage to AI Factories

    One of NVIDIA’s groundbreaking shifts is the redefinition of data centers from traditional storage units to “AI factories” generating intelligence. Unlike conventional data centers focused on multi-tenant storage, NVIDIA’s new data centers produce “tokens” for AI models at an industrial scale. These tokens are used in applications across industries, from robotics to biotechnology. Huang believes that every industry will benefit from AI-generated intelligence, making this shift in data centers vital to global AI adoption.

    This AI-centric infrastructure is already making waves, as seen with NVIDIA’s 100,000-GPU supercluster built for X.AI. NVIDIA demonstrated its logistical prowess by setting up this supercluster rapidly, paving the way for similar large-scale projects in the future.

    The Role of AI in Science, Engineering, and Digital Employees

    NVIDIA’s infrastructure investments and technological advancements have far-reaching impacts, particularly in science and engineering. Huang shared that AI-driven methods are now integral to NVIDIA’s chip design process, allowing them to explore new design options and optimize faster than human engineers alone could. This innovation is just the beginning, as Huang envisions AI reshaping fields like biotechnology, materials science, and theoretical physics, creating opportunities for breakthroughs at a previously impossible scale.

    Beyond science, Huang foresees AI-driven digital employees as a major component of future workforces. AI employees could assist in roles like marketing, supply chain management, and chip design, allowing human workers to focus on higher-level tasks. This shift to digital labor marks a major milestone for AI and has the potential to redefine productivity and efficiency across industries.

    Embodied AI and Real-World Applications

    Huang believes that embodied AI—AI in physical form—will transform industries such as robotics and autonomous vehicles. Self-driving cars and robots equipped with AI will become more common, thanks to NVIDIA’s advancements in AI infrastructure. By training these AI models on NVIDIA’s systems, industries can integrate intelligent robots and vehicles without needing substantial changes to existing environments.

    This embodied AI will serve as a bridge between digital intelligence and the physical world, enabling a new generation of applications that go beyond the screen to interact directly with people and environments.

    Sustaining Innovation Through Compatibility and Software Longevity

    Huang stressed that compatibility and sustainability are central to NVIDIA’s long-term vision. NVIDIA’s CUDA platform has enabled the company to build a lasting ecosystem, allowing software created on earlier NVIDIA systems to operate seamlessly on newer ones. This commitment to software longevity means companies can rely on NVIDIA’s systems for years, making it a trusted partner for businesses that prioritize innovation without disruption.

    NVIDIA as the “AI Factory” of the Future

    As Huang puts it, NVIDIA has evolved beyond a hardware company and is now an “AI factory”—a company that produces intelligence as a commodity. Huang sees AI as a resource as valuable as energy or raw materials, with applications across nearly every industry. From providing AI-driven insights to enabling new forms of intelligence, NVIDIA’s technology is poised to transform global markets and create value on an industrial scale.

    Jensen Huang’s vision for NVIDIA is not just about staying ahead in the computing industry; it’s about redefining what computing means. NVIDIA’s investments in scalable infrastructure, software longevity, digital employees, and embodied AI represent a shift in how industries will function in the future. As Huang envisions, the company is no longer just producing chips or hardware but enabling an entire ecosystem of AI-driven innovation that will touch every aspect of modern life.