PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Tag: DeepSeek

  • Jensen Huang on Lex Fridman: NVIDIA’s CEO Reveals His Vision for the AI Revolution, Scaling Laws, and Why Intelligence Is Now a Commodity

    A deep breakdown of Lex Fridman Podcast #494 featuring Jensen Huang, CEO of NVIDIA, covering extreme co-design, the four AI scaling laws, CUDA’s origin story, the future of programming, AGI timelines, and what it takes to lead the world’s most valuable company.

    TLDW (Too Long, Didn’t Watch)

    Jensen Huang sat down with Lex Fridman for a sprawling two-and-a-half-hour conversation covering the full arc of NVIDIA’s evolution from a GPU gaming company to the engine of the AI revolution. Jensen explains how NVIDIA now thinks in terms of rack-scale and pod-scale computing rather than individual chips, breaks down his four AI scaling laws (pre-training, post-training, test time, and agentic), and reveals the near-existential bet the company made putting CUDA on GeForce. He shares his views on China’s tech ecosystem, his deep respect for TSMC, why he turned down the chance to become TSMC’s CEO, how Elon Musk’s systems engineering approach built Colossus in record time, and why he believes AGI already exists. He also discusses why the future of programming is really about “specification,” why intelligence is being commoditized while humanity is the true superpower, and how he manages the enormous pressure of leading a company that nations and economies depend on. His core message: do not let the democratization of intelligence cause you anxiety. Instead, let it inspire you.

    Key Takeaways

    1. NVIDIA No Longer Thinks in Chips. It Thinks in AI Factories.

    Jensen’s mental model of what NVIDIA builds has fundamentally changed. He no longer picks up a chip to represent a new product generation. Instead, his mental model is a gigawatt-scale AI factory with power generation, cooling systems, and thousands of engineers bringing it online. The unit of computing at NVIDIA has evolved from GPU to computer to cluster to AI factory. His next mental “click” is planetary-scale computing.

    2. Extreme Co-Design Is NVIDIA’s Secret Weapon

    The reason NVIDIA dominates is not just better GPUs. It is the extreme co-design of the entire stack: GPU, CPU, memory, networking, switching, power, cooling, storage, software, algorithms, and applications. Jensen explains that when you distribute workloads across tens of thousands of computers and want them to go a million times faster (not just 10,000 times), every single component becomes a bottleneck. This is a restatement of Amdahl’s Law at scale. NVIDIA’s organizational structure directly reflects this co-design philosophy. Jensen has 60+ direct reports, holds no one-on-ones, and runs every meeting as a collective problem-solving session where specialists across all domains are present and contribute.

    3. The Four AI Scaling Laws Are a Flywheel

    Jensen outlined four distinct scaling laws that form a continuous loop:

    Pre-training scaling: Larger models plus more data equals smarter AI. The industry panicked when people said data was running out, but synthetic data generation has removed that ceiling. Data is now limited by compute, not by human generation.

    Post-training scaling: Fine-tuning, reinforcement learning from human feedback, and curated data continue to scale AI capabilities beyond what pre-training alone achieves.

    Test-time scaling: Inference is not “easy” as many predicted. It is thinking, reasoning, planning, and search. It is far more compute-intensive than memorization and pattern matching. This is why inference chips cannot be commoditized the way many predicted.

    Agentic scaling: A single AI agent can spawn sub-agents, creating teams. This is like scaling a company by hiring more employees rather than trying to make one person faster. The experiences generated by agents feed back into pre-training, creating a flywheel.

    4. The CUDA Bet Nearly Killed NVIDIA

    Putting CUDA on GeForce was one of the most consequential technology decisions in modern history. It increased GPU costs by roughly 50%, which crushed the company’s gross margins at a time when NVIDIA was a 35% gross margin business. The company’s market cap dropped from around $7-8 billion to approximately $1.5 billion. But Jensen understood that install base defines a computing architecture, not elegance. He pointed to x86 as proof: a less-than-elegant architecture that defeated beautifully designed RISC alternatives because of its massive install base. CUDA on GeForce put a supercomputer in the hands of every researcher, every scientist, every student. It took a decade to recover, but that install base became the foundation of the deep learning revolution.

    5. NVIDIA’s Moat Is Trust, Velocity, and Install Base

    Jensen was direct about NVIDIA’s competitive advantage. The CUDA install base is the number one asset. Developers target CUDA first because it reaches hundreds of millions of computers, is in every cloud, every OEM, every country, every industry. NVIDIA ships a new architecture roughly every year. No company in history has built systems of this complexity at this cadence. And the trust that NVIDIA will maintain, improve, and optimize CUDA indefinitely is something developers can count on. If someone created “GUDA” or “TUDA” tomorrow, it would not matter. The install base, velocity of execution, ecosystem breadth, and earned trust create a compounding advantage that is nearly impossible to replicate.

    6. Jensen Believes AGI Is Already Here

    When asked about AGI timelines, Jensen said he believes AGI has been achieved. His reasoning is practical: an agentic system today could plausibly create a web service, achieve virality, and generate a billion dollars in revenue, even if temporarily. This is not meaningfully different from many internet-era companies that did the same thing with technology no more sophisticated than what current AI agents can produce. He does not believe 100,000 agents could build another NVIDIA, but he believes a single agent-driven viral product is within reach right now.

    7. The Future of Programming Is Specification, Not Syntax

    Jensen believes the number of programmers in the world will increase dramatically, not decrease. His reasoning: the definition of coding is expanding to include specification and architectural description in natural language. This expands the population of “coders” from roughly 30 million professional developers to potentially a billion people. Every carpenter, plumber, accountant, and farmer who can describe what they want a computer to build is now a coder. The artistry of the future is knowing where on the spectrum of specification to operate, from highly prescriptive to exploratory and open-ended.

    8. China Is the Fastest Innovating Country in the World

    Jensen gave a nuanced and detailed explanation of why China’s tech ecosystem is so formidable. About 50% of the world’s AI researchers are Chinese. China’s tech industry emerged during the mobile cloud era, so it was built on modern software from the start. The country’s provincial competition creates an insane internal competitive environment. And the cultural norm of knowledge-sharing through school and family networks means China effectively operates as an open-source ecosystem at all times. This is why Chinese companies contribute disproportionately to open source. Their engineers’ brothers, friends, and schoolmates work at competing companies, and sharing knowledge is the cultural default.

    9. The Power Grid Has Enormous Waste That AI Can Exploit

    Jensen proposed a pragmatic solution to the energy problem for AI data centers. Power grids are designed for worst-case conditions with margin, but 99% of the time they run at around 60% of peak capacity. That idle capacity is simply wasted. Jensen wants data centers to negotiate flexible contracts where they absorb excess power most of the time and gracefully degrade during rare peak demand periods. This requires three things: customers accepting that “six nines” uptime may not always be necessary, data centers that can dynamically shift workloads, and utilities that offer tiered power delivery contracts instead of all-or-nothing commitments.

    10. Jensen Turned Down the CEO Role at TSMC

    In 2013, TSMC founder Morris Chang offered Jensen the chance to become CEO of TSMC. Jensen confirmed the story is true and said he was deeply honored. But he had already envisioned what NVIDIA could become and felt it was his sole responsibility to make that vision happen. He sees the relationship with TSMC as one built on three decades of trust, hundreds of billions of dollars in business, and zero formal contracts.

    11. Elon Musk’s Systems Engineering Approach Is Instructive

    Jensen praised Elon Musk’s approach to building the Colossus supercomputer in Memphis in just four months. He highlighted several principles: Elon questions everything relentlessly, strips every process down to the minimum necessary, is physically present at the point of action, and his personal urgency creates urgency in every supplier. Jensen drew a parallel to NVIDIA’s own “speed of light” methodology, where every process is benchmarked against the physical limits of what is possible, not against historical baselines.

    12. Intelligence Is a Commodity. Humanity Is Not.

    Perhaps the most philosophical takeaway from the conversation: Jensen argued that intelligence is a functional, measurable thing that is being commoditized. He surrounded himself with 60 direct reports who are all “superhuman” in their respective domains, more educated and deeper in their specialties than he is. Yet he sits in the middle orchestrating all of them. This proves that intelligence alone does not determine success. Character, compassion, grit, determination, tolerance for embarrassment, and the ability to endure suffering are the real differentiators. Jensen wants the audience to understand that the word we should elevate is not intelligence but humanity.

    Detailed Summary

    From GPU Maker to AI Infrastructure Company

    The conversation opened with Jensen explaining NVIDIA’s evolution from chip-scale to rack-scale to pod-scale design. The Vera Rubin pod, announced at GTC, contains seven chip types, five purpose-built rack types, 40 racks, 1.2 quadrillion transistors, nearly 20,000 NVIDIA dies, over 1,100 Rubin GPUs, 60 exaflops of compute, and 10 petabytes per second of scale bandwidth. And that is just one pod. NVIDIA plans to produce roughly 200 of these pods per week.

    Jensen explained that extreme co-design is necessary because the problems AI must solve no longer fit inside a single computer. When you distribute a workload across 10,000 computers but want a million-fold speedup, everything becomes a bottleneck: computation, networking, switching, memory, power, cooling. This is fundamentally an Amdahl’s Law problem at planetary scale. If computation represents only 50% of the workload, speeding it up infinitely only doubles total throughput. Every layer must be co-optimized simultaneously.

    NVIDIA’s organizational structure is a direct reflection of this co-design philosophy. Jensen has more than 60 direct reports, almost all with deep engineering expertise. He does not do one-on-ones. Every meeting is a collective problem-solving session where the memory expert, the networking expert, the cooling expert, and the power delivery expert are all in the room together, attacking the same problem.

    The Strategic History of CUDA

    Jensen walked through the step-by-step journey from graphics accelerator to computing platform. The company invented a programmable pixel shader, then added IEEE-compatible FP32 to its shaders, then put C on top of that (called Cg), and eventually arrived at CUDA. The critical strategic decision was putting CUDA on GeForce, a consumer product.

    This was nearly an existential move. It increased GPU costs by roughly 50% and consumed all of the company’s gross profit at a time when NVIDIA was a 35% gross margin business. The market cap cratered from around $7-8 billion to approximately $1.5 billion. But Jensen understood a principle that many technologists overlook: install base defines a computing architecture. x86 survived not because it was elegant but because it was everywhere. CUDA on GeForce put a supercomputing capability in the hands of every gamer, every student, every researcher who built their own PC. When the deep learning revolution arrived, CUDA was already the foundation.

    How Jensen Leads and Makes Decisions

    Jensen described a leadership philosophy built on continuous reasoning in public. He does not make announcements in the traditional sense. Instead, he shapes the belief systems of his employees, board, partners, and the broader industry over months and years by reasoning through decisions step by step, using every new piece of external information as a brick in the foundation. By the time he formally announces a strategic direction, the reaction is not surprise but rather, “What took you so long?”

    He applies this same approach to his supply chain. He personally visits CEOs of DRAM companies, packaging companies, and infrastructure providers. He explains the dynamics of the industry, shares his vision of future demand, and helps them reason through why they should make multi-billion-dollar capital investments. Three years ago, he convinced DRAM CEOs that HBM memory would become mainstream for data centers, which sounded ridiculous at the time. Those companies had record years as a result.

    Jensen’s “speed of light” methodology is his framework for decision-making. Every process, every design, every cost is benchmarked against the physical limits of what is theoretically possible. He prefers this to continuous improvement, which he views as incrementalism. He would rather strip a 74-day process back to zero and ask, “If we built this from scratch today, how long would it take?” Often the answer is six days, and the remaining 68 days are filled with accumulated compromises that can be challenged individually.

    AI Scaling Laws and the Future of Compute

    Jensen broke down the four scaling laws in detail. The pre-training scaling law, which depends on model size and data volume, was thought to be hitting a wall when the industry worried about running out of high-quality human-generated data. Jensen argued this concern is misplaced. Synthetic data generation has effectively removed the ceiling, and the constraint is now compute, not data.

    Post-training continues to scale through fine-tuning and reinforcement learning. Test-time scaling was the most counterintuitive for the industry. Many predicted that inference would be “easy” and that inference chips would be small, cheap, and commoditized. Jensen saw this as fundamentally wrong. Inference is thinking: reasoning, planning, search, decomposing novel problems into solvable pieces. Thinking is much harder than reading, and test-time compute is intensely resource-hungry.

    Agentic scaling is the newest frontier. A single AI agent can spawn sub-agents, effectively multiplying intelligence the way a company scales by hiring. The experiences and data generated by agentic systems feed back into pre-training, creating a continuous improvement loop. Jensen described this as the reason NVIDIA designed the Vera Rubin rack architecture differently from the Grace Blackwell architecture. Grace Blackwell was optimized for running large language models. Vera Rubin is designed for agents, which need to access files, use tools, do research, and spin off sub-agents. NVIDIA anticipated this architectural shift two and a half years before tools like OpenClaw arrived.

    China, TSMC, and the Global Supply Chain

    Jensen provided a thoughtful analysis of China’s tech ecosystem. He identified several structural advantages: 50% of the world’s AI researchers are Chinese, the tech industry was born during the mobile cloud era (making it natively modern), provincial competition creates internal Darwinian pressure, and the culture of knowledge-sharing through school and family networks makes China effectively open-source by default.

    On TSMC, Jensen emphasized that the deepest misunderstanding about the company is that its technology is its only advantage. Their manufacturing orchestration system, which dynamically manages the shifting demands of hundreds of companies, is “completely miraculous.” Their culture uniquely balances bleeding-edge technology excellence with world-class customer service. And the trust that Jensen places in TSMC is extraordinary: three decades of partnership, hundreds of billions of dollars in business, and no formal contract.

    Jensen also discussed the AI supply chain more broadly. NVIDIA has roughly 200 suppliers contributing technology to each rack. Jensen personally manages these relationships, flying to supplier sites, explaining industry dynamics, and helping CEOs reason through multi-billion-dollar investment decisions. When asked if supply chain bottlenecks keep him up at night, he said no, because he has already communicated what NVIDIA needs, his partners have told him what they will deliver, and he believes them.

    The Energy Challenge and Space Computing

    On the energy front, Jensen proposed a practical approach to the power problem. Rather than waiting for new power generation, he wants to capture the enormous waste already present in the grid. Power infrastructure is designed for worst-case peak demand, but 99% of the time it runs far below capacity. AI data centers could absorb this excess capacity with flexible contracts that allow graceful degradation during rare peak periods.

    On space computing, NVIDIA already has GPUs in orbit for satellite imaging. Jensen acknowledged the cooling challenge (no conduction or convection in space, only radiation) but sees it as a future frontier worth cultivating. In the meantime, he is focused on the lower-hanging fruit of eliminating waste in the terrestrial power grid.

    On AGI, Jobs, and the Human Future

    Jensen stated directly that he believes AGI has been achieved, at least by the practical definition of an AI system capable of creating a billion-dollar company. He sees it as plausible that an agent could build a viral web service that briefly generates enormous revenue, just as many internet-era companies did with technology no more sophisticated than what current AI agents produce.

    On jobs, Jensen was both compassionate and clear-eyed. He told the story of radiology: computer vision became superhuman around 2019-2020, and the prediction was that radiologists would disappear. Instead, the number of radiologists grew because AI allowed them to study more scans, diagnose better, and serve more patients. The purpose of the job (diagnosing disease) did not change, even though the tools changed completely.

    He applied this principle broadly: the number of software engineers at NVIDIA will grow, not decline, because their purpose is solving problems, not writing lines of code. The number of programmers globally will grow because the definition of coding is expanding to include natural language specification, opening it up to potentially a billion people.

    His advice to anyone worried about their job is straightforward: go use AI now. Become expert in it. Every profession, from carpenter to pharmacist to lawyer, will be elevated by AI tools. The people who learn to use AI will be the ones who get hired, promoted, and empowered.

    Mortality, Succession, and Legacy

    The conversation closed with deeply personal reflections. Jensen said he really does not want to die. He sees the current moment as a “once in a humanity experience.” He does not believe in traditional succession planning. Instead, he believes the best succession strategy is to pass on knowledge continuously, every single day, in every meeting, as fast as possible. His hope is to die on the job, instantaneously, with no long period of suffering.

    He described a vision for a kind of digital continuity: sending a humanoid robot into space, continuously improving it in flight, and eventually uploading the consciousness derived from a lifetime of communications, decisions, and reasoning to catch up with it at the speed of light.

    On the emotional experience of leading NVIDIA, Jensen was candid about hitting psychological low points regularly. His coping mechanism is decomposition: break the problem into pieces, reason about what you can control, tell someone who can help, share the burden, and then deliberately forget what is behind you. He compared this to the mental discipline of great athletes who focus only on the next point.

    His final message was about the relationship between intelligence and humanity. Intelligence, he argued, is functional. It is being commoditized. Humanity, character, compassion, grit, tolerance for embarrassment, and the capacity for suffering are the true superpowers. The word society should elevate is not intelligence but humanity.

    Thoughts

    This is one of the most substantive CEO interviews of 2026. What makes it remarkable is not just the breadth of topics but the depth of reasoning Jensen demonstrates in real time. You can actually watch him think through problems on the spot, which is rare for someone at his level.

    A few things stand out. First, the CUDA origin story is one of the great strategic narratives in tech history. The decision to absorb a 50% cost increase on a consumer product, watching your market cap collapse by 80%, and holding the course for a decade because you understood the power of install base is the kind of conviction that separates generational companies from everyone else.

    Second, Jensen’s framing of the four scaling laws as a flywheel is the clearest articulation anyone has given of why AI compute demand will continue to accelerate. Most people understand pre-training. Fewer understand test-time scaling. Almost nobody is thinking about agentic scaling as a compute multiplier. Jensen has been thinking about it for years and already designed hardware for it before the software ecosystem caught up.

    Third, the discussion on jobs deserves attention. The radiology example is powerful because it is a completed experiment, not a prediction. The profession that was supposed to be eliminated first by AI instead grew. The mechanism is straightforward: when you automate the task, you expand the capacity of the purpose, and demand for the purpose increases. This does not mean there will be no pain or dislocation. Jensen acknowledged that explicitly. But the historical pattern is clear.

    Finally, the philosophical distinction between intelligence and humanity is the kind of framing that could genuinely help people navigate the anxiety of this moment. If you define your value by your intelligence alone, AI commoditization is terrifying. If you define your value by your character, your compassion, your tolerance for suffering, and your willingness to keep going when everything goes wrong, then AI is just the most powerful set of tools you have ever been given.

    Jensen Huang is 62 years old, has been running NVIDIA for 34 years, and shows no signs of slowing down. If anything, his conviction about the future is accelerating alongside his company’s growth.

    Watch the full episode: Lex Fridman Podcast #494 with Jensen Huang

  • Beyond the Bubble: Jensen Huang on the Future of AI, Robotics, and Global Tech Strategy in 2026

    In a wide-ranging discussion on the No Priors Podcast, NVIDIA Founder and CEO Jensen Huang reflects on the rapid evolution of artificial intelligence throughout 2025 and provides a strategic roadmap for 2026. From the debunking of the “AI Bubble” to the rise of physical robotics and the “ChatGPT moments” coming for digital biology, Huang offers a masterclass in how accelerated computing is reshaping the global economy.


    TL;DW (Too Long; Didn’t Watch)

    • The Core Shift: General-purpose computing (CPUs) has hit a wall; the world is moving permanently to accelerated computing.
    • The Jobs Narrative: AI automates tasks, not purposes. It is solving labor shortages in manufacturing and nursing rather than causing mass unemployment.
    • The 2026 Breakthrough: Digital biology and physical robotics are slated for their “ChatGPT moment” this year.
    • Geopolitics: A nuanced, constructive relationship with China is essential, and open source is the “innovation flywheel” that keeps the U.S. competitive.

    Key Takeaways

    • Scaling Laws & Reasoning: 2025 proved that scaling compute still translates directly to intelligence, specifically through massive improvements in reasoning, grounding, and the elimination of hallucinations.
    • The End of “God AI”: Huang dismisses the myth of a monolithic “God AI.” Instead, the future is a diverse ecosystem of specialized models for biology, physics, coding, and more.
    • Energy as Infrastructure: AI data centers are “AI Factories.” Without a massive expansion in energy (including natural gas and nuclear), the next industrial revolution cannot happen.
    • Tokenomics: The cost of AI inference dropped 100x in 2024 and could drop a billion times over the next decade, making intelligence a near-free commodity.
    • DeepSeek’s Impact: Open-source contributions from China, like DeepSeek, are significantly benefiting American startups and researchers, proving the value of a global open-source ecosystem.

    Detailed Summary

    The “Five-Layer Cake” of AI

    Huang explains AI not as a single app, but as a technology stack: EnergyChipsInfrastructureModelsApplications. He emphasizes that while the public focuses on chatbots, the real revolution is happening in “non-English” languages, such as the languages of proteins, chemicals, and physical movement.

    Task vs. Purpose: The Future of Labor

    Addressing the fear of job loss, Huang uses the “Radiologist Paradox.” While AI now powers nearly 100% of radiology applications, the number of radiologists has actually increased. Why? Because AI handles the task (scanning images), allowing the human to focus on the purpose (diagnosis and research). This same framework applies to software engineers: their purpose is solving problems, not just writing syntax.

    Robotics and Physical AI

    Huang is incredibly optimistic about robotics. He predicts a future where “everything that moves will be robotic.” By applying reasoning models to physical machines, we are moving from “digital rails” (pre-programmed paths) to autonomous agents that can navigate unknown environments. He foresees a trillion-dollar repair and maintenance industry emerging to support the billions of robots that will eventually inhabit our world.

    The “Bubble” Debate

    Is there an AI bubble? Huang argues “No.” He points to the desperate, unsatisfied demand for compute capacity across every industry. He notes that if chatbots disappeared tomorrow, NVIDIA would still thrive because the fundamental architecture of the world’s $100 trillion GDP is shifting from CPUs to GPUs to stay productive.


    Analysis & Thoughts

    Jensen Huang’s perspective is distinct because he views AI through the lens of industrial production. By calling data centers “factories” and tokens “output,” he strips away the “magic” of AI and reveals it as a standard industrial revolution—one that requires power, raw materials (data/chips), and specialized labor.

    His defense of Open Source is perhaps the most critical takeaway for policymakers. By arguing that open source prevents “suffocation” for startups and 100-year-old industrial companies, he positions transparency as a national security asset rather than a liability. As we head into 2026, the focus is clearly shifting from “Can the model talk?” to “Can the model build a protein or drive a truck?”

  • The BG2 Pod: A Deep Dive into Tech, Tariffs, and TikTok on Liberation Day

    In the latest episode of the BG2 Pod, hosted by tech luminaries Bill Gurley and Brad Gerstner, the duo tackled a whirlwind of topics that dominated headlines on April 3, 2025. Recorded just after President Trump’s “Liberation Day” tariff announcement, this bi-weekly open-source conversation offered a verbose, insightful exploration of market uncertainty, global trade dynamics, AI advancements, and corporate maneuvers. With their signature blend of wit, data-driven analysis, and insider perspectives, Gurley and Gerstner unpacked the implications of a rapidly shifting economic and technological landscape. Here’s a detailed breakdown of the episode’s key discussions.

    Liberation Day and the Tariff Shockwave

    The episode kicked off with a dissection of President Trump’s tariff announcement, dubbed “Liberation Day,” which sent shockwaves through global markets. Gerstner, who had recently spoken at a JP Morgan Tech conference, framed the tariffs as a doctrinal move by the Trump administration to level the trade playing field—a philosophy he’d predicted as early as February 2025. The initial market reaction was volatile: S&P and NASDAQ futures spiked 2.5% on a rumored 10% across-the-board tariff, only to plummet 600 basis points as details emerged, including a staggering 54% tariff on China (on top of an existing 20%) and 25% auto tariffs targeting Mexico, Canada, and Germany.

    Gerstner highlighted the political theater, noting Trump’s invite to UAW members and his claim that these tariffs flipped Michigan red. The administration also introduced a novel “reciprocal tariff” concept, factoring in non-tariff barriers like currency manipulation, which Gurley critiqued for its ambiguity. Exemptions for pharmaceuticals and semiconductors softened the blow, potentially landing the tariff haul closer to $600 billion—still a hefty leap from last year’s $77 billion. Yet, both hosts expressed skepticism about the economic fallout. Gurley, a free-trade advocate, warned of reduced efficiency and higher production costs, while Gerstner relayed CEOs’ fears of stalled hiring and canceled contracts, citing a European-Asian backlash already brewing.

    US vs. China: The Open-Source Arms Race

    Shifting gears, the duo explored the escalating rivalry between the US and China in open-source AI models. Gurley traced China’s decade-long embrace of open source to its strategic advantage—sidestepping IP theft accusations—and highlighted DeepSeek’s success, with over 1,500 forks on Hugging Face. He dismissed claims of forced open-sourcing, arguing it aligns with China’s entrepreneurial ethos. Meanwhile, Gerstner flagged Washington’s unease, hinting at potential restrictions on Chinese models like DeepSeek to prevent a “Huawei Belt and Road” scenario in AI.

    On the US front, OpenAI’s announcement of a forthcoming open-weight model stole the spotlight. Sam Altman’s tease of a “powerful” release, free of Meta-style usage restrictions, sparked excitement. Gurley praised its defensive potential—leveling the playing field akin to Google’s Kubernetes move—while Gerstner tied it to OpenAI’s consumer-product focus, predicting it would bolster ChatGPT’s dominance. The hosts agreed this could counter China’s open-source momentum, though global competition remains fierce.

    OpenAI’s Mega Funding and Coreweave’s IPO

    The conversation turned to OpenAI’s staggering $40 billion funding round, led by SoftBank, valuing the company at $260 billion pre-money. Gerstner, an investor, justified the 20x revenue multiple (versus Anthropic’s 50x and X.AI’s 80x) by emphasizing ChatGPT’s market leadership—20 million paid subscribers, 500 million weekly users—and explosive demand, exemplified by a million sign-ups in an hour. Despite a projected $5-7 billion loss, he drew parallels to Uber’s turnaround, expressing confidence in future unit economics via advertising and tiered pricing.

    Coreweave’s IPO, meanwhile, weathered a “Category 5 hurricane” of market turmoil. Priced at $40, it dipped to $37 before rebounding to $60 on news of a Google-Nvidia deal. Gerstner and Gurley, shareholders, lauded its role in powering AI labs like OpenAI, though they debated GPU depreciation—Gurley favoring a shorter schedule, Gerstner citing seven-year lifecycles for older models like Nvidia’s V100s. The IPO’s success, they argued, could signal a thawing of the public markets.

    TikTok’s Tangled Future

    The episode closed with rumors of a TikTok US deal, set against the April 5 deadline and looming 54% China tariffs. Gerstner, a ByteDance shareholder since 2015, outlined a potential structure: a new entity, TikTok US, with ByteDance at 19.5%, US investors retaining stakes, and new players like Amazon and Oracle injecting fresh capital. Valued potentially low due to Trump’s leverage, the deal hinges on licensing ByteDance’s algorithm while ensuring US data control. Gurley questioned ByteDance’s shift from resistance to cooperation, which Gerstner attributed to preserving global value—90% of ByteDance’s worth lies outside TikTok US. Both saw it as a win for Trump and US investors, though China’s approval remains uncertain amid tariff tensions.

    Broader Implications and Takeaways

    Throughout, Gurley and Gerstner emphasized uncertainty’s chilling effect on markets and innovation. From tariffs disrupting capex to AI’s open-source race reshaping tech supremacy, the episode painted a world in flux. Yet, they struck an optimistic note: fear breeds buying opportunities, and Trump’s dealmaking instincts might temper the tariff storm, especially with China. As Gurley cheered his Gators and Gerstner eyed Stargate’s compute buildout, the BG2 Pod delivered a masterclass in navigating chaos with clarity.

  • How AI is Revolutionizing Writing: Insights from Tyler Cowen and David Perell

    TLDW/TLDR

    Tyler Cowen, an economist and writer, shares practical ways AI transforms writing and research in a conversation with David Perell. He uses AI daily as a “secondary literature” tool to enhance reading and podcast prep, predicts fewer books due to AI’s rapid evolution, and emphasizes the enduring value of authentic, human-centric writing like memoirs and personal narratives.

    Detailed Summary of Video

    In a 68-minute YouTube conversation uploaded on March 5, 2025, economist Tyler Cowen joins writer David Perell to explore AI’s impact on writing and research. Cowen details his daily AI use—replacing stacks of books with large language models (LLMs) like o1 Pro, Claude, and DeepSeek for podcast prep and leisure reading, such as Shakespeare and Wuthering Heights. He highlights AI’s ability to provide context quickly, reducing hallucinations in top models by over tenfold in the past year (as of February 2025).

    The discussion shifts to writing: Cowen avoids AI for drafting to preserve his unique voice, though he uses it for legal background or critiquing drafts (e.g., spotting obnoxious tones). He predicts fewer books as AI outpaces long-form publishing cycles, favoring high-frequency formats like blogs or Substack. However, he believes “truly human” works—memoirs, biographies, and personal experience-based books—will persist, as readers crave authenticity over AI-generated content.

    Cowen also sees AI decentralizing into a “Republic of Science,” with models self-correcting and collaborating, though this remains speculative. For education, he integrates AI into his PhD classes, replacing textbooks with subscriptions to premium models. He warns academia lags in adapting, predicting AI will outstrip researchers in paper production within two years. Perell shares his use of AI for Bible study, praising its cross-referencing but noting experts still excel at pinpointing core insights.

    Practical tips emerge: use top-tier models (o1 Pro, Claude, DeepSeek), craft detailed prompts, and leverage AI for travel or data visualization. Cowen also plans an AI-written biography by “open-sourcing” his life via blog posts, showcasing AI’s potential to compile personal histories.

    Article Itself

    How AI is Revolutionizing Writing: Insights from Tyler Cowen and David Perell

    Artificial Intelligence is no longer a distant sci-fi dream—it’s a tool reshaping how we write, research, and think. In a recent YouTube conversation, economist Tyler Cowen and writer David Perell unpack the practical implications of AI for writers, offering a roadmap for navigating this seismic shift. Recorded on March 5, 2025, their discussion blends hands-on advice with bold predictions, grounded in Cowen’s daily AI use and Perell’s curiosity about its creative potential.

    Cowen, a prolific author and podcaster, doesn’t just theorize about AI—he lives it. He’s swapped towering stacks of secondary literature for LLMs like o1 Pro, Claude, and DeepSeek. Preparing for a podcast on medieval kings Richard II and Henry V, he once ordered 20-30 books; now, he interrogates AI for context, cutting prep time and boosting quality. “It’s more fun,” he says, describing how he queries AI about Shakespearean puzzles or Wuthering Heights chapters, treating it as a conversational guide. Hallucinations? Not a dealbreaker—top models have slashed errors dramatically since 2024, and as an interviewer, he prioritizes context over perfect accuracy.

    For writing, Cowen draws a line: AI informs, but doesn’t draft. His voice—cryptic, layered, parable-like—remains his own. “I don’t want the AI messing with that,” he insists, rejecting its smoothing tendencies. Yet he’s not above using it tactically—checking legal backgrounds for columns or flagging obnoxious tones in drafts (a tip from Agnes Callard). Perell nods, noting AI’s knack for softening managerial critiques, though Cowen prefers his weirdness intact.

    The future of writing, Cowen predicts, is bifurcated. Books, with their slow cycles, face obsolescence—why write a four-year predictive tome when AI evolves monthly? He’s shifted to “ultra high-frequency” outputs like blogs and Substack, tackling AI’s rapid pace. Yet “truly human” writing—memoirs, biographies, personal narratives—will endure. Readers, he bets, want authenticity over AI’s polished slop. His next book, Mentors, leans into this, drawing on lived experience AI can’t replicate.

    Perell, an up-and-coming writer, feels the tension. AI’s prowess deflates his hard-earned skills, yet he’s excited to master it. He uses it to study the Bible, marveling at its cross-referencing, though it lacks the human knack for distilling core truths. Both agree: AI’s edge lies in specifics—detailed prompts yield gold, vague ones yield “mid” mush. Cowen’s tip? Imagine prompting an alien, not a human—literal, clear, context-rich.

    Educationally, Cowen’s ahead of the curve. His PhD students ditch textbooks for AI subscriptions, weaving it into papers to maximize quality. He laments academia’s inertia—AI could outpace researchers in two years, yet few adapt. Perell’s takeaway? Use the best models. “You’re hopeless without o1 Pro,” Cowen warns, highlighting the gap between free and cutting-edge tools.

    Beyond writing, AI’s horizon dazzles. Cowen envisions a decentralized “Republic of Science,” where models self-correct and collaborate, mirroring human progress. Large context windows (Gemini’s 2 million tokens, soon 10-20 million) will decode regulatory codes and historical archives, birthing jobs in data conversion. Inside companies, he suspects AI firms lead secretly, turbocharging their own models.

    Practically, Cowen’s stack—o1 Pro for queries, Claude for thoughtful prose, DeepSeek for wild creativity, Perplexity for citations—offers a playbook. He even plans an AI-crafted biography, “open-sourcing” his life via blog posts about childhood in Fall River or his dog, Spinosa. It’s low-cost immortality, a nod to AI’s archival power.

    For writers, the message is clear: adapt or fade. AI won’t just change writing—it’ll redefine what it means to create. Human quirks, stories, and secrets will shine amid the deluge of AI content. As Cowen puts it, “The truly human books will stand out all the more.” The revolution’s here—time to wield it.

  • Global Madness Unleashed: Tariffs, AI, and the Tech Titans Reshaping Our Future

    As the calendar turns to March 21, 2025, the world economy stands at a crossroads, buffeted by market volatility, looming trade policies, and rapid technological shifts. In the latest episode of the BG2 Pod, aired March 20, venture capitalists Bill Gurley and Brad Gerstner dissect these currents with precision, offering a window into the forces shaping global markets. From the uncertainty surrounding April 2 tariff announcements to Google’s $32 billion acquisition of Wiz, Nvidia’s bold claims at GTC, and the accelerating AI race, their discussion—spanning nearly two hours—lays bare the high stakes. Gurley, sporting a Florida Gators cap in a nod to March Madness, and Gerstner, fresh from Nvidia’s developer conference, frame a narrative of cautious optimism amid palpable risks.

    A Golden Age of Uncertainty

    Gerstner opens with a stark assessment: the global economy is traversing a “golden age of uncertainty,” a period marked by political, economic, and technological flux. Since early February, the NASDAQ has shed 10%, with some Mag 7 constituents—Apple, Amazon, and others—down 20-30%. The Federal Reserve’s latest median dot plot, released just before the podcast, underscores the gloom: GDP forecasts for 2025 have been cut from 2.1% to 1.7%, unemployment is projected to rise from 4.3% to 4.4%, and inflation is expected to edge up from 2.5% to 2.7%. Consumer confidence is fraying, evidenced by a sharp drop in TSA passenger growth and softening demand reported by Delta, United, and Frontier Airlines—a leading indicator of discretionary spending cuts.

    Yet the picture is not uniformly bleak. Gerstner cites Bank of America’s Brian Moynihan, who notes that consumer spending rose 6% year-over-year, reaching $1.5 trillion quarterly, buoyed by a shift from travel to local consumption. Conversations with hedge fund managers reveal a tactical retreat—exposures are at their lowest quartile—but a belief persists that the second half of 2025 could rebound. The Atlanta Fed’s GDP tracker has turned south, but Gerstner sees this as a release of pent-up uncertainty rather than an inevitable slide into recession. “It can become a self-fulfilling prophecy,” he cautions, pointing to CEOs pausing major decisions until the tariff landscape clarifies.

    Tariffs: Reciprocity or Ruin?

    The specter of April 2 looms large, when the Trump administration is set to unveil sectoral tariffs targeting the “terrible 15” countries—a list likely encompassing European and Asian nations with perceived trade imbalances. Gerstner aligns with the administration’s vision, articulated by Vice President JD Vance in a recent speech at an American Dynamism event. Vance argued that globalism’s twin conceits—America monopolizing high-value work while outsourcing low-value tasks, and reliance on cheap foreign labor—have hollowed out the middle class and stifled innovation. China’s ascent, from manufacturing to designing superior cars (BYD) and batteries (CATL), and now running AI inference on Huawei’s Ascend 910 chips, exemplifies this shift. Treasury Secretary Scott Bessent frames it as an “American detox,” a deliberate short-term hit for long-term industrial revival.

    Gurley demurs, championing comparative advantage. “Water runs downhill,” he asserts, questioning whether Americans will assemble $40 microwaves when China commands 35% of the global auto market with superior products. He doubts tariffs will reclaim jobs—automation might onshore production, but employment gains are illusory. A jump in tariff revenues from $65 billion to $1 trillion, he warns, could tip the economy into recession, a risk the U.S. is ill-prepared to absorb. Europe’s reaction adds complexity: *The Economist*’s Zanny Minton Beddoes reports growing frustration among EU leaders, hinting at a pivot toward China if tensions escalate. Gerstner counters that the goal is fairness, not protectionism—tariffs could rise modestly to $150 billion if reciprocal concessions materialize—though he concedes the administration’s bellicose tone risks misfiring.

    The Biden-era “diffusion rule,” restricting chip exports to 50 countries, emerges as a flashpoint. Gurley calls it “unilaterally disarming America in the race to AI,” arguing it hands Huawei a strategic edge—potentially a “Belt and Road” for AI—while hobbling U.S. firms’ access to allies like India and the UAE. Gerstner suggests conditional tariffs, delayed two years, to incentivize onshoring (e.g., TSMC’s $100 billion Arizona R&D fab) without choking the AI race. The stakes are existential: a misstep could cede technological primacy to China.

    Google’s $32 Billion Wiz Bet Signals M&A Revival

    Amid this turbulence, Google’s $32 billion all-cash acquisition of Wiz, a cloud security firm founded in 2020, signals a thaw in mergers and acquisitions. With projected 2025 revenues of $1 billion, Wiz commands a 30x forward revenue multiple—steep against Google’s 5x—adding just 2% to its $45 billion cloud business. Gerstner hails it as a bellwether: “The M&A market is back.” Gurley concurs, noting Google’s strategic pivot. Barred by EU regulators from bolstering search or AI, and trailing AWS’s developer-friendly platform and Microsoft’s enterprise heft, Google sees security as a differentiator in the fragmented cloud race.

    The deal’s scale—$32 billion in five years—underscores Silicon Valley’s capacity for rapid value creation, with Index Ventures and Sequoia Capital notching another win. Gerstner reflects on Altimeter’s misstep with Lacework, a rival that faltered on product-market fit, highlighting the razor-thin margins of venture success. Regulatory hurdles loom: while new FTC chair Matthew Ferguson pledges swift action—“go to court or get out of the way”—differing sharply from Lina Khan’s inertia, Europe’s penchant for thwarting U.S. deals could complicate closure, slated for 2026 with a $3.2 billion breakup fee at risk. Success here could unleash “animal spirits” in M&A and IPOs, with CoreWeave and Cerebras rumored next.

    Nvidia’s GTC: A $1 Trillion AI Gambit

    At Nvidia’s GTC in San Jose, CEO Jensen Huang—clad in a leather jacket evoking Steve Jobs—addressed 18,000 attendees, doubling down on AI’s explosive growth. He projects a $1 trillion annual market for AI data centers by 2028, up from $500 billion, driven by new workloads and the overhaul of x86 infrastructure with accelerated computing. Blackwell, 40x more capable than Hopper, powers robotics (a $5 billion run rate) to synthetic biology. Yet Nvidia’s stock hovers at $115, 20x next year’s earnings—below Costco’s 50x—reflecting investor skittishness over demand sustainability and competition from DeepSeek and custom ASICs.

    Huang dismisses DeepSeek R1’s “cheap intelligence” narrative, insisting compute needs are 100x what was estimated a year ago. Coding agents, set to dominate software development by year-end per Zuckerberg and Musk, fuel this surge. Gurley questions the hype—inference, not pre-training, now drives scaling, and Huang’s “chief revenue destroyer” claim (Blackwell obsoleting Hopper) risks alienating customers on six-year depreciation cycles. Gerstner sees brilliance in Nvidia’s execution—35,000 employees, a top-tier supply chain, and a four-generation roadmap—but both flag government action as the wildcard. Tariffs and export controls could bolster Huawei, though Huang shrugs off near-term impacts.

    AI’s Consumer Frontier: OpenAI’s Lead, Margin Mysteries

    In consumer AI, OpenAI’s ChatGPT reigns with 400 million weekly users, supply-constrained despite new data centers in Texas. Gerstner calls it a “winner-take-most” market—DeepSeek briefly hit #2 in app downloads but faded, Grok lingers at #65, Gemini at #55. “You need to be 10x better to dent this inertia,” he says, predicting a Q2 product blitz. Gurley agrees the lead looks unassailable, though Meta and Apple’s silence hints at brewing counterattacks.

    Gurley’s “negative gross margin AI theory” probes deeper: many AI firms, like Anthropic via AWS, face slim margins due to high acquisition and serving costs, unlike OpenAI’s direct model. With VC billions fueling negative margins—pricing for share, not profit—and compute costs plummeting, unit economics are opaque. Gerstner contrasts this with Google’s near-zero marginal costs, suggesting only direct-to-consumer AI giants can sustain the capex. OpenAI leads, but Meta, Amazon, and Elon Musk’s xAI, with deep pockets, remain wildcards.

    The Next 90 Days: Pivot or Peril?

    The next 90 days will define 2025. April 2 tariffs could spark a trade war or a fairer field; tax cuts and deregulation promise growth, but AI’s fate hinges on export policies. Gerstner’s optimistic—Nvidia at 20x earnings and M&A’s resurgence signal resilience—but Gurley warns of overreach. A trillion-dollar tariff wall or a Huawei-led AI surge could upend it all. As Gurley puts it, “We’ll turn over a lot of cards soon.” The world watches, and the outcome remains perilously uncertain.

  • The AI Revolution Unveiled: Jonathan Ross on Groq, NVIDIA, and the Future of Inference


    TL;DR

    Jonathan Ross, Groq’s CEO, predicts inference will eclipse training in AI’s future, with Groq’s Language Processing Units (LPUs) outpacing NVIDIA’s GPUs in cost and efficiency. He envisions synthetic data breaking scaling limits, a $1.5 billion Saudi revenue deal fueling Groq’s growth, and AI unlocking human potential through prompt engineering, though he warns of an overabundance trap.

    Detailed Summary

    In a captivating 20VC episode with Harry Stebbings, Jonathan Ross, the mastermind behind Groq and Google’s original Tensor Processing Unit (TPU), outlines a transformative vision for AI. Ross asserts that inference—deploying AI models in real-world scenarios—will soon overshadow training, challenging NVIDIA’s GPU stronghold. Groq’s LPUs, engineered for affordable, high-volume inference, deliver over five times the cost efficiency and three times the energy savings of NVIDIA’s training-focused GPUs by avoiding external memory like HBM. He champions synthetic data from advanced models as a breakthrough, dismantling scaling law barriers and redirecting focus to compute, data, and algorithmic bottlenecks.

    Groq’s explosive growth—from 640 chips in early 2024 to over 40,000 by year-end, aiming for 2 million in 2025—is propelled by a $1.5 billion Saudi revenue deal, not a funding round. Partners like Aramco fund the capital expenditure, sharing profits after a set return, liberating Groq from financial limits. Ross targets NVIDIA’s 40% inference revenue as a weak spot, cautions against a data center investment bubble driven by hyperscaler exaggeration, and foresees AI value concentrating among giants via a power law—yet Groq plans to join them by addressing unmet demands. Reflecting on Groq’s near-failure, salvaged by “Grok Bonds,” he dreams of AI enhancing human agency, potentially empowering 1.4 billion Africans through prompt engineering, while urging vigilance against settling for “good enough” in an abundant future.

    The Big Questions Raised—and Answered

    Ross’s insights provoke profound metaphorical questions about AI’s trajectory and humanity’s role. Here’s what the discussion implicitly asks, paired with his responses:

    • What happens when creation becomes so easy it redefines who gets to create?
      • Answer: Ross champions prompt engineering as a revolutionary force, turning speech into a tool that could unleash 1.4 billion African entrepreneurs. By making creation as simple as talking, AI could shift power from tech gatekeepers to the masses, sparking a global wave of innovation.
    • Can an underdog outrun a titan in a scale-driven game?
      • Answer: Groq can outpace NVIDIA, Ross asserts, by targeting inference—a massive, underserved market—rather than battling over training. With no HBM bottlenecks and a scalable Saudi-backed model, Groq’s agility could topple NVIDIA’s inference share, proving size isn’t everything.
    • What’s the human cost when machines replace our effort?
      • Answer: Ross likens LPUs to tireless employees, predicting a shift from labor to compute-driven economics. Yet, he warns of “financial diabetes”—a loss of drive in an AI-abundant world—urging us to preserve agency lest we become passive consumers of convenience.
    • Is the AI gold rush a promise or a pipe dream?
      • Answer: It’s both. Ross foresees billions wasted on overhyped data centers and “AI t-shirts,” but insists the total value created will outstrip losses. The winners, like Groq, will solve real problems, not chase fleeting trends.
    • How do we keep innovation’s spirit alive amid efficiency’s rise?
      • Answer: By prioritizing human agency and delegation—Ross’s “anti-founder mode”—over micromanagement, he says. Groq’s 25 million token-per-second coin aligns teams to innovate, not just optimize, ensuring efficiency amplifies creativity.
    • What’s the price of chasing a future that might not materialize?
      • Answer: Seven years of struggle taught Ross the emotional and financial toll is steep—Groq nearly died—but strategic bets (like inference) pay off when the wave hits. Resilience turns risk into reward.
    • Will AI’s pursuit drown us in wasted ambition?
      • Answer: Partially, yes—Ross cites VC’s “Keynesian Beauty Contest,” where cash floods copycats. But hyperscalers and problem-solvers like Groq will rise above the noise, turning ambition into tangible progress.
    • Can abundance liberate us without trapping us in ease?
      • Answer: Ross fears AI could erode striving, drawing from his boom-bust childhood. Prompt engineering offers liberation—empowering billions—but only if outliers reject “good enough” and push for excellence.

    Jonathan Ross’s vision is a clarion call: AI’s future isn’t just about faster chips or bigger models—it’s about who wields the tools and how they shape us. Groq’s battle with NVIDIA isn’t merely corporate; it’s a referendum on whether innovation can stay human-centric in an age of machine abundance. As Ross puts it, “Your job is to get positioned for the wave”—and he’s riding it, challenging us to paddle alongside or risk being left ashore.

  • The DeepSeek Revolution: Financial Markets in TurmoilA Sputnik Moment for AI and Finance

    The DeepSeek Revolution: Financial Markets in TurmoilA Sputnik Moment for AI and Finance

    On January 27, 2025, the financial markets experienced significant upheaval following the release of DeepSeek’s latest AI model, R1. This event has been likened to a modern “Sputnik moment,” highlighting its profound impact on the global economic and technological landscape.

    Market Turmoil: A Seismic Shift

    The unveiling of DeepSeek R1 led to a sharp decline in major technology stocks, particularly those heavily invested in AI development. Nvidia, a leading AI chip manufacturer, saw its shares tumble by approximately 11.5%, signaling a potential loss exceeding $340 billion in market value if the trend persists. This downturn reflects a broader market reassessment of the AI sector’s financial foundations, especially concerning the substantial investments in high-cost AI infrastructure.

    The ripple effects were felt globally, with tech indices such as the Nasdaq 100 and Europe’s Stoxx 600 technology sub-index facing a combined market capitalization reduction projected at $1.2 trillion. The cryptocurrency market was not immune, as AI-related tokens experienced a 13.3% decline, with notable losses in assets like Near Protocol and Internet Computer (ICP).

    DeepSeek R1: A Paradigm Shift in AI

    DeepSeek’s R1 model has been lauded for its advanced reasoning capabilities, reportedly surpassing established Western models like OpenAI’s o1. Remarkably, R1 was developed at a fraction of the cost, challenging the prevailing notion that only vast financial resources can produce cutting-edge AI. This achievement has prompted a reevaluation of the economic viability of current AI investments and highlighted the rapid technological advancements emerging from China.

    The emergence of R1 has also intensified discussions regarding the effectiveness of U.S. export controls aimed at limiting China’s technological progress. By achieving competitive AI capabilities with less advanced hardware, DeepSeek underscores the potential limitations and unintended consequences of such sanctions, suggesting a need for a strategic reassessment in global tech policy.

    Broader Implications: Economic and Geopolitical Considerations

    The market’s reaction to DeepSeek’s R1 extends beyond immediate financial losses, indicating deeper shifts in economic power, technological leadership, and geopolitical influence. China’s rapid advancement in AI capabilities signifies a pivotal moment in the global race for technological dominance, potentially leading to a reallocation of capital from Western institutions to Chinese entities and reshaping global investment trends.

    Furthermore, this development reaffirms the critical importance of computational resources, such as GPUs, in the AI race. The narrative that more efficient use of computing power can lead to models exhibiting human-like intelligence positions computational capacity not merely as a tool but as a cornerstone of this new technological era.

    DeepSeek’s Strategic Approach: Efficiency and Accessibility

    DeepSeek’s strategy emphasizes efficiency and accessibility. The R1 model was developed using a pure reinforcement learning approach, a departure from traditional methods that often rely on supervised learning. This method allowed the model to develop reasoning capabilities autonomously, without initial reliance on human-annotated datasets.

    In terms of cost, DeepSeek’s R1 model offers a significantly more affordable option compared to its competitors. For instance, where OpenAI’s o1 costs $15 per million input tokens and $60 per million output tokens, DeepSeek’s R1 costs $0.55 per million input tokens and $2.19 per million output tokens. This cost-effectiveness makes advanced AI technology more accessible to a broader audience, including developers, businesses, and educational institutions.

    Global Reception and Future Outlook

    The global reception to DeepSeek’s R1 has been mixed. While some industry leaders have praised the model’s efficiency and performance, others have expressed skepticism regarding its rapid development and the potential implications for data security and ethical considerations.

    Looking ahead, DeepSeek plans to continue refining its models and expanding its offerings. The company aims to democratize AI by making advanced models accessible to a wider audience, challenging the current market leaders, and potentially reshaping the future landscape of artificial intelligence.

    Wrap Up

    DeepSeek’s R1 model has not merely entered the market; it has redefined it, challenging established players, prompting a reevaluation of investment strategies, and potentially ushering in a new era where AI capabilities are more evenly distributed globally. As we navigate this juncture, the pertinent question is not solely who will lead in AI but how this technology will shape our future across all facets of human endeavor. Welcome to 2025, where the landscape has shifted, and the race is on.