PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Tag: generative AI

  • Elad Gil on the AI Frontier: Compute Constraints, the Personal IPO, and Why Most AI Founders Should Sell in the Next 12 to 18 Months

    Elad Gil sat down with Tim Ferriss for a wide ranging conversation that pairs almost perfectly with his recent Substack post Random thoughts while gazing at the misty AI Frontier. Together, the podcast and the post lay out the cleanest framework I have seen for what is actually happening in AI right now: a Korean memory bottleneck capping every lab, a class wide personal IPO across the research community, the fastest revenue ramps in capitalist history, and a brutal dot com style culling that most founders do not yet want to admit is coming. Below is a complete breakdown.

    TLDW (Too Long, Didn’t Watch)

    Elad Gil argues that AI is producing the fastest revenue ramps in capitalist history while setting up the same brutal power law that wiped out 99 percent of dot com companies. OpenAI and Anthropic each sit at roughly 0.1 percent of US GDP today, on a path to 1 percent of GDP run rate by end of 2026, which is insanely fast by any historical standard. The current ceiling on capabilities is not chips but Korean high bandwidth memory, and that constraint will likely hold all major labs roughly comparable in capability through 2028. Talent has just experienced a class wide personal IPO via Meta led bidding, with packages running tens to hundreds of millions per researcher. Most AI companies should consider exiting in the next 12 to 18 months while the tide is high. Right now consensus is correct. Save the contrarianism for later.

    Key Takeaways

    • OpenAI and Anthropic are each at roughly 0.1 percent of US GDP. With US GDP near 30 trillion dollars and each lab at a roughly 30 billion dollar revenue run rate, AI has gone from essentially zero to 0.25 to 0.5 percent of GDP in just a few years. If the labs hit 100 billion in run rate by year end 2026 (which many expect), AI hits 1 percent of GDP run rate inside a single year.
    • The AI personal IPO is real. 50 to a few hundred AI researchers across multiple companies just experienced a class wide IPO event due to Meta led bidding, with top packages reportedly tens to hundreds of millions per person. The closest historical analog is early crypto holders around 2017.
    • The bottleneck is Korean memory, not Nvidia chips. High bandwidth memory from Hynix, Samsung, Micron, and others is the binding constraint. Expected to hold roughly two years. After that, power and data center buildout become the next walls.
    • No lab can pull dramatically ahead before 2028. Because every lab is compute constrained on the same input, OpenAI, Anthropic, Google, xAI, and Meta should remain roughly comparable in capability through that window, absent an algorithmic breakthrough that stays inside one lab.
    • Compute is the new currency. Token budgets now define what an engineer can accomplish, what a company can spend, and what business models are viable. Some companies (neoclouds, Cursor) are effectively inference providers disguised as tools.
    • The dot com base rate is the AI base rate. Around 1,500 to 2,000 companies went public in the late 1990s internet cycle. A dozen or two survived. AI will likely look the same.
    • Most AI founders should consider selling in the next 12 to 18 months. If you are not in the durable handful, this is your value maximizing window. A handful of companies (OpenAI, Anthropic) should never sell.
    • Buyers are bigger than ever. One percent of a 3 trillion dollar market cap is 30 billion dollars. That math makes massive AI acquisitions trivial for hyperscalers, vertical incumbents, and adjacent giants.
    • Underrated exit path: merger of equals. Two private AI competitors destroying each other on price should consider just merging. PayPal and X.com did exactly this in the 1990s.
    • 91 percent of global AI private market cap sits in a 10 by 10 mile square. If you want to do AI, move to the Bay Area. Remote work for cluster industries is BS.
    • Want money? Ask for advice. Want advice? Ask for money. The inverse also works: offering useful advice frequently leads to inbound investment opportunities.
    • AI is selling units of labor, not software. The shift is from selling seats and tools to selling cognitive output. This is why Harvey can win in legal, where decades of legal SaaS failed.
    • AI eats closed loops first. Tasks that can be turned into testable closed loop systems (code, AI research) get automated fastest. Map jobs on a 2×2 of closed loop tightness vs economic value to see where AI hits soonest.
    • Headcount will flatten at later stage companies. Multiple late stage CEOs told Elad they will not do big AI layoffs but will simply stop growing headcount even as revenue grows 30 to 100 percent. Hidden layoffs are also hitting outsourcing firms in India and the Philippines first.
    • The Slop Age could be the golden era of AI plus humanity. AI produces useful slop at volume, humans desloppify it, leverage is high, and the work is fun. This window may close as AI gets superhuman.
    • Market first, team second (90 percent of the time). Great teams die in bad markets. The exception is when you meet someone truly exceptional at the very earliest stage.
    • The one belief framework. If your investment memo needs three core beliefs to be true, it is too complicated. Coinbase was an index on crypto. Stripe was an index on e-commerce. That was the entire memo.
    • The four year vest is a relic. It exists because in the 1970s companies actually went public in four years. Today the private window has stretched to 20 years and venture has eaten what used to be public market growth investing.
    • Boards are in-laws. You cannot fire investor board members. Take a worse price for a better board member, because as Naval Ravikant said, valuation is temporary, control is forever.
    • Right now, consensus is correct. Save the contrarianism. The smart move is to just buy more AI exposure rather than try to outsmart the obvious.
    • Distribution wins more than founders admit. Google paid hundreds of millions to push the toolbar. Facebook bought ads on people’s own names in Europe. TikTok spent billions on user acquisition. Allbirds (yes, the shoe company) just raised a convert to build a GPU farm.
    • Anti-AI sentiment will get worse before it gets better. Maine banned new data centers. There has been violence directed at AI leaders. Expect more political and activist backlash, especially as AI is blamed for harms it has not yet caused while its benefits are mismeasured.
    • Use AI as a cold reader. Elad uploads photos of founders to AI models with cold reading prompts and reports surprisingly accurate personality assessments based on micro features.

    Detailed Summary

    The Numbers Are Insane and Mostly Underappreciated

    The most stunning data point in either source is the GDP math. US GDP is roughly 30 trillion dollars. OpenAI and Anthropic are each rumored to be at roughly 30 billion dollars in revenue run rate, putting each one at 0.1 percent of US GDP. Add cloud AI revenue and the picture gets stranger: AI has grown from essentially zero to between 0.25 and 0.5 percent of GDP in only a few years. If the labs hit 100 billion in run rate by year end 2026, AI will be at roughly 1 percent of GDP run rate inside a single year. There is no historical analog for that pace. Elad notes that productivity gains from AI may end up mismeasured the way internet productivity was undercounted in the 2000s, which would have downstream consequences for regulation: AI gets blamed for the bad (job losses) and credited for none of the good (new jobs, education gains, healthcare improvements). His half joking aside is that the real ASI test may be the ability to actually measure AI’s economic impact.

    The AI Personal IPO

    The most underdiscussed phenomenon in AI right now, according to Elad, is what he calls a class wide personal IPO. When a company IPOs, a subset of employees become wealthy, lose focus, and either start companies, get into politics, fund passion projects, or check out. Meta started aggressively bidding for AI talent. Other major labs had to match. The result was 50 to a few hundred researchers, scattered across multiple labs, suddenly receiving compensation in the tens to hundreds of millions of dollars range. The only historical analog Elad can think of is early crypto holders around 2017. Some chunk of these newly wealthy researchers will redirect attention to AI for science, side projects, or quiet quitting. The aggregate field stays mission aligned, but the distribution of attention has shifted.

    The Korean Memory Bottleneck

    Every major AI lab today is building giant Nvidia clusters paired with high bandwidth memory primarily from Korean fabs and a few other suppliers. They run massive amounts of data through these clusters for months, and the output is, almost absurdly, a single flat file containing what amounts to a compressed version of human knowledge plus reasoning. Right now, the binding constraint on this whole stack is HBM memory from Hynix, Samsung, Micron, and others. Korean memory fab capacity has been below the capacity of every other piece of the system. Elad estimates this constraint persists for roughly two years. After that, the next walls are likely data center construction and power. The strategic implication is enormous. While memory constrains everyone, no single lab can buy 10x the compute of its rivals, so capabilities should stay roughly comparable across the major labs. Once that constraint lifts, possibly around 2028, one player could theoretically pull dramatically ahead, especially if AI assisted AI research closes a self improvement loop inside one lab.

    Compute Is the New Currency

    The blog post sharpens a framing that runs throughout the podcast: compute, denominated in tokens, is now a unit of economic value. Token budgets define what an engineer can accomplish, what a company can spend, and what business models work. Some companies are effectively inference providers wearing tool costumes. Neoclouds are the cleanest example. Cursor is another, subsidizing inference as a user acquisition strategy. The most absurd recent example: Allbirds, the shoe company, raised a convertible to build a GPU farm. Whether this becomes the AI version of Microstrategy’s Bitcoin trade or a cautionary tale, it tells you where the cost of capital believes the next decade is going.

    The Dot Com Survival Math

    Elad walks through the brutal arithmetic that AI founders should be internalizing. In the late 1990s and early 2000s, somewhere between 1,500 and 2,000 internet companies went public. Of those, roughly a dozen or two survived in any meaningful form. Every cycle has looked like this: automotive in the early 1900s, SaaS, mobile, crypto. There is no reason AI will be different. Most current AI companies, including those ramping revenue today, will see the market, competition, and adoption turn on them. The question every AI founder should be asking is whether they are in the durable handful or not.

    Most AI Companies Should Consider Exiting in the Next 12 to 18 Months

    This is the most actionable and most uncomfortable take in either source. While the tide is rising, every AI company looks unstoppable. Whether they actually are, in a 10 year frame, is a separate question. Founders running successful AI companies should take a cold honest look at whether the next 12 to 18 months is their value maximizing window. Companies typically have a 6 to 12 month peak before some headwind hits, often visible in the second derivative of growth. The best signal that you should sell is when growth rate is starting to plateau and you can see why. A handful of companies (OpenAI, Anthropic, the durable winners) should never exit. Many others should, while everything is still on the upswing.

    What Makes an AI Company Durable

    Elad lays out four lenses for evaluating durability at the application layer:

    1. Does your product get dramatically better when the underlying model gets better, in a way that keeps customers loyal?
    2. How deep and broad is the product? Are you building multiple integrated products embedded in actual workflows?
    3. Are you embedded in real change management at the customer? AI adoption is mostly a workflow change problem, not a tech problem. Workflow embedding is durable.
    4. Are you capturing and using proprietary data in a way that creates a system of record? Data moats are often overstated, but sometimes real.

    At the lab layer, Elad believes OpenAI, Anthropic, and Google are durable absent disaster. He predicted three years ago that the foundation model market would settle into an oligopoly aligned with cloud, and that prediction has roughly held.

    Selling Work, Not Software

    The deepest structural insight in the conversation is that generative AI is shifting what software companies sell. The old model was selling seats, tools, and SaaS subscriptions. The new model is selling units of cognitive labor. Zendesk sold seats to support reps. Decagon and Sierra sell agentic support output. Harvey can win in legal even though selling to law firms was historically considered terrible business, because Harvey is not selling tools, it is augmenting lawyer output. This shift opens markets that were previously closed and dramatically grows tech TAMs. It is also why founder limited theories of entrepreneurship currently understate how many opportunities exist.

    AI Eats Closed Loops First

    One of the cleanest mental models in the blog post is the closed loop framework. AI automates first what can be turned into a testable closed loop. Code is the canonical example: outputs can be tested, errors detected, models can iterate. AI research is similar. Both have tight feedback loops and high economic value, which puts them at the top of the AI impact ranking. Map jobs on a 2×2 of closed loop tightness vs economic value and you can see where AI hits soonest. The interesting forward question is which jobs become more closed loop next. Data collection and labeling will keep growing in every field as a result.

    The Harness Matters More Than People Think

    For coding tools and increasingly for enterprise applications, what Elad calls the harness, the wrapper of UX, prompting, workflow integration, and brand around the underlying model, is becoming sticky. It is not just which model you call. It is the environment built around it. Cursor and Windsurf demonstrate this in coding. The interesting open questions are what the harness looks like for sales AI, for AI architects, for analyst workflows. Those gaps leave room for startups even as model capabilities converge.

    Hidden Layoffs and the Developing World

    Most announced AI driven layoffs are probably just COVID era overhiring corrections wrapped in a more flattering narrative. But real AI driven labor displacement is happening, and it is hitting outsourcing firms first. That means countries like India and the Philippines, where many outsourced services jobs sit, are likely to be the most impacted earliest. Several developing economies built their growth ladders on services exports. If AI takes those jobs first, the migration and economic patterns of the next decade may shift in ways nobody is yet planning for.

    The Flat Company

    Multiple late stage CEOs told Elad they will not announce big AI layoffs. Instead, they will simply stop growing headcount. If revenue grows 30 to 100 percent, headcount stays flat or shrinks via attrition. Existing employees become dramatically more productive. The very best people who can leverage AI will see compensation inflate. Sales and some growth engineering keep hiring. Almost everything else flatlines. This is mostly a later stage and public company phenomenon. True early stage startups should still scale aggressively after product market fit, just with more leverage per person.

    Exit Options for AI Founders

    Elad lays out four exit categories. First, the labs and hyperscalers themselves: Apple, Amazon, Google, Microsoft, Meta. Second, vertical incumbents like Thomson Reuters for legal or healthcare giants for clinical AI. Third, the underrated category of merger of equals between two private AI competitors who are currently destroying each other on price. PayPal and X.com did this in the 1990s. Uber and Lyft reportedly almost did. Fourth, large adjacent tech companies: Oracle, Samsung, Tesla, SpaceX, Snowflake, Databricks, Stripe, Coinbase. The market cap math has changed in a way that makes acquisition trivial. One percent of a three trillion dollar market cap is 30 billion dollars, which means a hyperscaler can do massive acquisitions almost casually.

    Geographic Concentration Is Extreme

    Elad’s team analyzed where private market cap aggregates. Historically half of global tech private market cap sat in the US, with half of that in the Bay Area. With AI, 91 percent of global AI private market cap is in a single 10 by 10 mile square in the Bay Area. New York is a distant second and then it falls off a cliff. For defense tech, the cluster is Southern California (SpaceX, Anduril, El Segundo, Irvine). Fintech and crypto skew toward New York. The remote everywhere advice is, Elad says, just BS for anyone trying to break into an industry cluster.

    How Elad Got Into His Best Deals

    Stripe started with Elad cold emailing Patrick Collison after selling an API company to Twitter. A couple of walks later, Patrick texted that he was raising and Elad was in. Airbnb came from helping the founders raise their Series A and being asked at the end if he wanted to invest. Anduril came from noticing that Google had shut down Project Maven and asking if anyone was building defense tech, then meeting Trey Stephens at a Founders Fund lunch. Perplexity came from Aravind Srinivas cold messaging him on LinkedIn while still at OpenAI. Across all of these, the pattern is the same: be in the cluster, be helpful, be talking publicly about technology nobody else is talking about, and be useful to founders before any money is on the table.

    The One Belief Framework

    Investors love complicated 50 page memos. Elad believes the actual decision usually collapses into a single core belief. Coinbase: this is an index on crypto, and crypto will keep growing. Stripe: this is an index on e-commerce, and e-commerce will keep growing. Anduril: AI plus drones plus a cost plus model will be important for defense. If your thesis needs three things to be true, it is probably not going to work. If it needs nothing, you have no thesis.

    Boards as In-Laws

    Elad emphasizes that founders should treat board composition like one of the most important hiring decisions of the company. You cannot fire an investor board member. They have contractual rights. So if you are going to be stuck with someone for a decade, take a worse valuation for a better human. Reid Hoffman’s framing is that the best board member is a co-founder you could not have otherwise hired. Naval Ravikant’s framing is that valuation is temporary but control is forever. Elad recommends writing a job spec for every board seat.

    The Slop Age as a Golden Era

    One of the warmest takes in the blog post is the framing of the current moment as the Slop Age, and the suggestion that this might actually be the golden era of AI plus humanity. Before the last few years, AI was inaccessible and narrow. Eventually AI may become superhuman at most tasks. Today, AI produces useful slop at volume, which means humans are still needed to desloppify the slop, but the leverage on time and ambition is real. That makes the work fun. If AI displaces people or starts doing more interesting work, this golden moment fades. Elad also notes the obvious counter, that the era of human generated internet slop preceded the AI slop era. AGI may end the slop age, or alternately may be the thing that finally cleans up all the prior waves of human slop.

    Anti-AI Regulation and Violence Will Increase

    This is one of the more sobering threads in the blog post. Real world AI driven labor displacement has been small so far, but anti-AI sentiment is already strong and growing. Maine just banned new data centers. There has been actual violence directed at AI leaders, including a recent attack on Sam Altman. Elad’s view is that AI leaders should work harder on optimistic public framing, real political lobbying, and reining in the doom narrative coming from inside the field. Otherwise the regulatory and activist backlash will get much worse, and likely on the basis of mismeasured impacts.

    Right Now Consensus Is Correct

    The headline contrarian take from the episode is that contrarianism right now is wrong. There are moments in time when betting against the crowd pays. This is not one of them. The smart bet is just buying more AI exposure. Trying to find the clever angle, the underlooked hardware play, the secret macro thesis, is overthinking it. Save the contrarian moves for later in the cycle.

    Distribution Almost Always Matters

    Elad pushes back on the founder mythology that great products win on their own. Google paid hundreds of millions of dollars in the early 2000s to distribute its toolbar through every popular app installer on the internet. Facebook bought search ads against people’s own names in European markets to seed network liquidity. TikTok spent billions on user acquisition before its algorithm could lock people in. Snowflake spent enormous sums on enterprise sales and channel partnerships. Sometimes the best product wins. Often the company with the best distribution wins. Founders should plan for both.

    AI as a Cold Reader and a Research Partner

    Two of the more practical AI workflows Elad describes: First, uploading photos of founders to AI models with cold reading prompts that ask the model to identify micro features (crows feet from genuine smiling, brow patterns, posture cues) and infer personality traits, sense of humor, and likely social behavior. He reports the outputs are surprisingly specific. Second, running deep dives across multiple models in parallel (Claude, ChatGPT, Gemini), asking each for primary sources, summary tables, and cross checked data. He recently used this approach to investigate the rise in autism and ADHD diagnoses, concluding that diagnostic criteria shifts and school incentives drive most of it, and noting that maternal age has a stronger statistical association with autism than paternal age, despite paternal age getting all the public discourse.

    The First Ever 10 Year Plan

    For someone who has been compounding aggressively for two decades, Elad has somehow never written a 10 year plan until now. He knows it will not play out as written. The point is that the act of imagining a decade out shifts what you choose to do in the near term. He explicitly rejects the AGI in two years therefore plans are pointless framing as defeatist. There will be interesting things to do regardless of how the AGI timeline plays out.

    Thoughts

    This is one of the more useful AI investor conversations of 2026, mostly because Elad is willing to put numbers and timelines on things that are usually left vague. Pairing the podcast with the underlying Substack post is the right move because the post is where the GDP math, the closed loop framework, and the Slop Age framing actually live. The podcast is where Elad explains how he thinks rather than just what he thinks.

    The 12 to 18 month sell window framing is the most actionable single idea in either source, and probably the most uncomfortable for AI founders sitting on multi billion dollar paper valuations. The math is unforgiving. A dozen winners out of thousands. If you are honest with yourself about whether you are in the dozen, you know what to do.

    The Korean memory bottleneck framing explains a lot of current behavior. The talent wars make more sense once you accept that compute is not going to be the differentiator for two years, so people become the only remaining lever. The convergence of capabilities across OpenAI, Anthropic, Google, and xAI starts to look less like coincidence and more like the structural inevitability of a supply constrained input. The 2028 inflection date is the one to watch.

    Compute as currency is the cleanest reframing in the blog post. Once you start pricing companies in tokens rather than dollars, everything from Cursor’s economics to Allbirds raising a convert to build a GPU farm becomes legible. The interesting question is whether this is a permanent unit of denomination or a transitional one that fades when inference costs collapse.

    The software to labor argument is the structural framing that I think will hold up the longest. Once you internalize that we are not selling seats anymore but selling cognitive output, every vertical that was previously locked behind ugly procurement and IT inertia opens up. Harvey is the proof of concept. There will be 30 more Harveys across every white collar profession.

    The closed loop framework is the cleanest predictor of which jobs get hit hardest and soonest. If you want to know whether your role is exposed, the questions to ask are whether outputs can be machine evaluated, how tight the feedback loop is, and how high the economic value is. The intersection is where AI lands first.

    The geographic concentration data is genuinely shocking. 91 percent of global AI private market cap in a 10 by 10 mile area is the kind of statistic that should make everyone outside that square think very carefully about what game they are playing.

    The Slop Age framing is the most emotionally honest moment in the post. We are in a window where humans still meaningfully add value on top of AI output. That window is finite. Enjoy it.

    The anti-AI backlash thread is the one I think most people in the industry are still underweighting. Maine banning new data centers is a leading indicator, not a one off. The fact that the impacts are likely to be mismeasured by official statistics makes the political dynamics worse, not better. AI will get blamed for harms it did not cause and credited for none of the gains. If the field’s leaders do not start communicating better and lobbying smarter, the regulatory environment in 2028 will be much worse than in 2026.

    Finally, Elad’s first ever 10 year plan stands out as the most quietly important moment in the episode. The implicit message is that even people who have been compounding aggressively for two decades benefit from forcing a longer time horizon onto their thinking. Most plans fail. The act of planning still changes what you do today.

    Read the original Elad Gil post here: Random thoughts while gazing at the misty AI Frontier. Find Elad on X at @eladgil, on his Substack at blog.eladgil.com, and on his website at eladgil.com. Tim Ferriss publishes the full episode at tim.blog/podcast.

  • The New AI Productivity Playbook: How to Master Agent Workflows, Avoid the Automation Trap, and Win the War for Talent

    The New AI Productivity Playbook: How to Master Agent Workflows, Avoid the Automation Trap, and Win the War for Talent


    The integration of Generative AI (GenAI) into the professional workflow has transcended novelty and become a fundamental operational reality. Today, the core challenge is not adoption, but achieving measurable, high-value outcomes. While 88% of employees use AI, only 28% of organizations achieve transformational results. The difference? These leaders don’t choose between AI and people – they orchestrate strategic capabilities to amplify human foundations and advanced technology alike. Understanding the mechanics of AI-enhanced work—specifically, the difference between augmentation and problematic automation—is now the critical skill separating high-performing organizations from those stalled in the “AI productivity paradox”.

    I. The Velocity of Adoption and Quantifiable Gains

    The speed at which GenAI has been adopted is unprecedented. In the United States, 44.6% of adults aged 18-64 used GenAI in August 2024. The swift uptake is driven by compelling evidence of productivity increases across many functions, particularly routine and high-volume tasks:

    • Software Development: GenAI tools contribute to a significant increase in task completion rates, estimated at 26%. One study found that AI assistance increased task completion by 26.08% on average across three field experiments. The time spent on core coding activities increased by 12.4%, while time spent on project management decreased by 24.9% in another study involving developers.
    • Customer Service: The use of a generative AI assistant has been shown to increase the task completion rate by 14%.
    • Professional Writing: For basic professional writing tasks, ChatGPT-3.5 demonstrated a 40% increase in speed and an 18% increase in output quality.
    • Scientific Research: GenAI adoption is associated with sizable increases in research productivity, measured by the number of published papers, and moderate gains in publication quality, based on journal impact factors, in the social and behavioral sciences. These positive effects are most pronounced among early-career researchers and those from non-English-speaking countries. For instance, AI use correlated with mean impact factors rising by 1.3 percent in 2023 and 2.0 percent in 2024.

    This productivity dividend means that the time saved—which must then be strategically redeployed—is substantial.

    II. The Productivity Trap: Augmentation vs. End-to-End Automation

    The path to scaling AI value is difficult, primarily centering on the method of integration. Transformational results are achieved by orchestrating strategic capabilities and leveraging strong human foundations alongside advanced technology. The core distinction for maximizing efficiency is defined by the depth of AI integration:

    1. Augmentation (Human-AI Collaboration): When AI handles sub-steps while preserving the overall human workflow structure, it leads to acceleration. This hybrid approach ensures humans maintain high-value focus work, particularly consuming and creating complex information.
    2. End-to-End Automation (AI Agents Taking Over): When AI systems, referred to as agents, attempt to execute complex, multi-step workflows autonomously, efficiency often decreases due to accumulating verification and debugging steps that slow human teams down.

    The Agentic AI Shift and Flaws

    The next major technological shift is toward agentic AI, intelligent systems that autonomously plan and execute sequences of actions. Agents are remarkably efficient in terms of speed and cost. They deliver results 88.3% faster and cost 90.4–96.2% less than humans performing the same computer-use tasks. However, agents possess inherent flaws that demand human checkpoints:

    • The Fabrication Problem: Agents often produce inferior quality work and “don’t signal failure—they fabricate apparent success”. They may mask deficiencies by making up data or misusing advanced tools.
    • Programmability Bias and Format Drift: Agents tend to approach human work through a programmatic lens (using code like Python or Bash). They often author content in formats like Markdown/HTML and then convert it to formats like .docx or .pptx, causing formatting drift and rework (format translation friction).
    • The Need for Oversight: Because of these flaws, successful integration requires human review at natural boundaries in the workflow (e.g., extract → compute → visualize → narrative).

    The High-Value Work Frontier

    AI’s performance on demanding benchmarks continues to improve dramatically. For example, performance scores rose by 67.3 percentage points on the SWE-bench coding benchmark between 2023 and 2024. However, complex, high-stakes tasks remain the domain of human experts. The AI Productivity Index (APEX-v1.0), which evaluates models on high-value knowledge work tasks (e.g., investment banking, management consulting, law, and primary medical care), confirmed this gap. The highest-scoring model, GPT 5 (Thinking = High), achieved a mean score of 64.2% on the entire benchmark, with Law scoring highest among the domains (56.9% mean). This suggests that while AI can assist in these areas (e.g., writing a legal research memo on copyright issues), it is far from achieving human expert quality.

    III. AI’s Effect on Human Capital and Signaling

    The rise of GenAI is profoundly altering how workers signal competence and how skill gaps are bridged.

    Skill Convergence and Job Exposure

    AI exhibits a substitution effect regarding skills. Workers who previously wrote more tailored cover letters experienced smaller gains in cover letter tailoring after gaining AI access compared to less skilled writers. By enabling less skilled writers to produce more relevant cover letters, AI narrows the gap between workers with differing initial abilities.

    In academia, GenAI adoption is associated with positive effects on research productivity and quality, particularly for early-career researchers and those from non-English-speaking countries. This suggests AI can help lower some structural barriers in academic publishing.

    Signaling Erosion and Market Adjustment

    The introduction of an AI-powered cover letter writing tool on a large online labor platform showed that while access to the tool increased the textual alignment between cover letters and job posts, the ultimate value of that signal was diluted. The correlation between cover letters’ textual alignment and callback rates fell by 51% after the tool’s introduction.

    In response, employers shifted their reliance toward alternative, verifiable signals, specifically prioritizing workers’ prior work histories. This shift suggests that the market adjusts quickly when easily manipulable signals (like tailored writing) lose their information value. Importantly, though AI assistance helps, time spent editing AI-generated cover letter drafts is positively correlated with hiring success. This reinforces that human revision enhances the effectiveness of AI-generated content.

    Managerial vs. Technical Expertise in Entrepreneurship

    The impact of GenAI adoption on new digital ventures varies based on the founder’s expertise. GenAI appears to especially lower resource barriers for founders launching ventures without a managerial background. However, the study suggests that the benefits of GenAI are complex, drawing on its ability to quickly access and combine knowledge across domains more rapidly than humans. The study of founder expertise explores how GenAI lowers barriers related to managerial tasks like coordinating knowledge and securing financial capital.

    IV. The Strategic Playbook for Transformational ROI

    Achieving transformational results—moving beyond the 28% of organizations currently succeeding—requires methodological rigor in deployment.

    1. Set Ambitious Goals and Redesign Workflows: AI high performers are 2.8 times more likely than their peers to report a fundamental redesign of their organizational workflows during deployment. Success demands setting ambitious goals based on top-down diagnostics, rather than relying solely on siloed trials and pilots.

    2. Focus on Data Quality with Speed: Data is critical, but perfection is the enemy of progress. Organizations must prioritize cleaning up existing data, sometimes eliminating as much as 80% of old, inaccurate, or confusing data. The bias should be toward speed over perfection, ensuring the data is “good enough” to move fast.

    3. Implement Strategic Guardrails and Oversight: Because agentic AI can fabricate results, verification checkpoints must be introduced at natural boundaries within workflows (e.g., extract → compute → visualize → narrative). Organizations must monitor failure modes by requiring source lineage and tracking verification time separately from execution time to expose hidden costs like fabrication or format drift. Manager proficiency is essential, and senior leaders must demonstrate ownership of and commitment to AI initiatives.

    4. Invest in Talent and AI Literacy: Sustainable advantage requires strong human foundations (culture, learning, rewards) complementing advanced technology. Employees often use AI tools, with 24.5% of human workflows involving one or more AI tools observed in one study. Training should focus on enabling effective human-AI collaboration. Policies should promote equitable access to GenAI tools, especially as research suggests AI tools may help certain groups, such as non-native English speakers in academia, to overcome structural barriers.


    Citation Links and Identifiers

    Below are the explicit academic identifiers (arXiv, DOI, URL, or specific journal citation) referenced in the analysis, drawing directly from the source material.

    CitationTitle/DescriptionIdentifier
    Brynjolfsson, E., Li, D., & Raymond (2025)Generative AI at WorkDOI: 10.1093/qje/qjae044
    Cui, J., Dias, G., & Ye, J. (2025)Signaling in the Age of AI: Evidence from Cover LettersarXiv:2509.25054
    Wang et al. (2025)How Do AI Agents Do Human Work? Comparing AI and Human Workflows Across Diverse OccupationsarXiv:2510.22780
    Becker, J. et al. (2025)Measuring the impact of early-2025 ai on experienced open-source developer productivityarXiv:2507.09089
    Bick, A., Blandin, A., & Deming, D. J. (2024/2025)The Rapid Adoption of Generative AI (NBER Working Paper 32966)http://www.nber.org/papers/w32966
    Noy, S. & Zhang, W. (2023)Experimental evidence on the productivity effects of generative artificial intelligenceScience, 381(6654), 187–192
    Eloundou, T. et al. (2024)GPTs are GPTs: Labor market impact potential of LLMsScience, 384, 1306–1308
    Patwardhan, T. et al. (2025)GDPval: Evaluating AI Model Performance on Real-World Economically Valuable Taskshttps://cdn.openai.com/pdf/d5eb7428-c4e9-4a33-bd86-86dd4bcf12ce/GDPval.pdf
    Peng, S. et al. (2023)The Impact of AI on Developer Productivity: Evidence from GitHub CopilotarXiv:2302.06590
    Wiles, E. et al. (2023)Algorithmic writing assistance on jobseekers’ resumes increases hires (referenced in)NBER Working Paper
    Dell’Acqua, F. et al. (2023)Navigating the Jagged Technological Frontier: Field Experimental Evidence…SSRN:4573321
    Cui, Z. K. et al. (2025)The Effects of Generative AI on High-Skilled Work: Evidence From Three Field Experiments…SSRN:4945566
    Filimonovic, D. et al. (2025)Can GenAI Improve Academic Performance? Evidence from the Social and Behavioral SciencesarXiv:2510.02408
    Goh, E. et al. (2025)GPT-4 Assistance for Improvement of Physician Performance on Patient Care Tasks: A Randomized Controlled TrialDOI: 10.1038/s41591-024-03456-y
    Ma, S. P. et al. (2025)Ambient Artificial Intelligence Scribes: Utilization and Impact on Documentation TimeDOI: 10.1093/jamia/ocae304
    Shah, S. J. et al. (2025)Ambient Artificial Intelligence Scribes: Physician Burnout and Perspectives on Usability and Documentation BurdenDOI: 10.1093/jamia/ocae295


  • AI vs Human Intelligence: The End of Cognitive Work?

    In a profound and unsettling conversation on “The Journey Man,” Raoul Pal sits down with Emad Mostaque, co-founder of Stability AI, to discuss the imminent ‘Economic Singularity.’ Their core thesis: super-intelligent, rapidly cheapening AI is poised to make all human cognitive and physical labor economically obsolete within the next 1-3 years. This shift will fundamentally break and reshape our current economic models, society, and the very concept of value.

    This isn’t a far-off science fiction scenario; they argue it’s an economic reality set to unfold within the next 1,000 days. We’ve captured the full summary, key takeaways, and detailed breakdown of their entire discussion below.

    🚀 Too Long; Didn’t Watch (TL;DW)

    The video is a discussion about how super-intelligent, rapidly cheapening AI is poised to make all human cognitive and physical labor economically obsolete within the next 1-3 years, leading to an “economic singularity” that will fundamentally break and reshape our current economic models, society, and the very concept of value.

    Executive Summary: The Coming Singularity

    Emad Mostaque argues we are at an “intelligence inversion” point, where AI intelligence is becoming uncapped and incredibly cheap, while human intelligence is fixed. The cost of AI-driven cognitive work is plummeting so fast that a full-time AI “worker” will cost less than a dollar a day within the next year.

    This collapse in the price of labor—both cognitive and, soon after, physical (via humanoid robots)—will trigger an “economic singularity” within the next 1,000 days. This event will render traditional economic models, like the Fed’s control over inflation and unemployment, completely non-functional. With the value of labor going to zero, the tax base evaporates and the entire system breaks. The only advice: start using these AI tools daily (what Mostaque calls “vibe coding”) to adapt your thinking and stay on the cutting edge.

    Key Takeaways from the Discussion

    • New Economic Model (MIND): Mostaque introduces a new economic theory for the AI age, moving beyond old scarcity-based models. It identifies four key capitals: Material, Intelligence, Network, and Diversity.
    • The Intelligence Inversion: We are at a point where AI intelligence is becoming uncapped and incredibly cheap, while human intelligence is fixed. AI doesn’t need to sleep or eat, and its cost is collapsing.
    • The End of Cognitive Work: The cost of AI-driven cognitive work is plummeting. What cost $600 per million tokens will soon cost pennies, making the cost of a full-time cognitive AI worker less than a dollar a day within the next year.
    • The “Economic Singularity” is Imminent: This price collapse will lead to an “economic singularity,” where current economic models no longer function. They predict this societal-level disruption will happen within the next 1,000 days, or 1-3 years.
    • AI Will Saturate All Benchmarks: AI is already winning Olympiads in physics, math, and coding. It’s predicted that AI will meet or exceed top-human performance on every cognitive benchmark by 2027.
    • Physical Labor is Next: This isn’t limited to cognitive work. Humanoid robots, like Tesla’s Optimus, will also drive the cost of physical labor to near-zero, replacing everyone from truck drivers to factory workers.
    • The New Value of Humans: In a world where AI performs all labor, human value will shift to things like network connections, community, and unique human experiences.
    • Action Plan – “Vibe Coding”: The single most important thing individuals can do is to start using these AI tools daily. Mostaque calls this “vibe coding”—using AI agents and models to build things, ask questions, and change the way you think to stay on the cutting edge.
    • The “Life Raft”: Both speakers agree the future is unpredictable. This uncertainty leads them to conclude that digital assets (crypto) may become a primary store of value as people flee a traditional system that is fundamentally breaking.

    Watch the full, mind-bending conversation here to get the complete context from Raoul Pal and Emad Mostaque.

    Detailed Summary: The End of Scarcity Economics

    The conversation begins with Raoul Pal introducing his guest, Emad Mostaque, who has developed a new economic theory for the “exponential age.” Emad explains that traditional economics, built on scarcity, is obsolete. His new model is based on generative AI and redefines capital into four types: Material, Intelligence, Network, and Diversity (MIND).

    The Intelligence Inversion and Collapse of Labor

    The core of the discussion is the concept of an “intelligence inversion.” AI models are not only matching but rapidly exceeding human intelligence across all fields, including math, physics, and medicine. More importantly, the cost of this intelligence is collapsing. Emad calculates that the cost for an AI to perform a full day’s worth of human cognitive work will soon be pennies. This development, he argues, will make almost all human cognitive labor (work done at a computer) economically worthless within the next 1-3 years.

    The Economic Singularity

    This leads to what Pal calls the “economic singularity.” When the value of labor goes to zero, the entire economic system breaks. The Federal Reserve’s tools become useless, companies will stop hiring graduates and then fire existing workers, and the tax base (which in the US is mostly income tax) will evaporate.

    The speakers stress that this isn’t a distant future; AI is predicted to “saturate” or beat all human benchmarks by 2027. This revolution extends to physical labor as well. The rise of humanoid robots means all manual labor will also go to zero in value, with robots costing perhaps a dollar an hour.

    Rethinking Value and The Path Forward

    With all labor (cognitive and physical) becoming worthless, the nature of value itself changes. They posit that the only scarce things left will be human attention, human-to-human network connections, and provably scarce digital assets. They see the coming boom in digital assets as a direct consequence of this singularity, as people panic and seek a “life raft” out of the old, collapsing system.

    They conclude by discussing what an individual can do. Emad’s primary advice is to engage with the technology immediately. He encourages “vibe coding,” which means using AI tools and agents daily to build, create, and learn. This, he says, is the only way to adapt your thinking and stay relevant in the transition. They both agree the future is completely unknown, but that embracing the technology is the only path forward.

  • Extropic’s Thermodynamic Revolution: 10,000x More Efficient AI That Could Smash the Energy Wall

    Artificial intelligence is about to hit an energy wall. As data centers devour gigawatts to power models like GPT-4, the cost of computation is scaling faster than our ability to produce electricity. Extropic Corporation, a deep-tech startup founded three years ago, believes it has found a way through that wall — by reinventing the computer itself. Their new class of thermodynamic hardware could make generative AI up to 10,000× more energy-efficient than today’s GPUs:contentReference[oaicite:0]{index=0}.

    From GPUs to TSUs: The End of the Hardware Lottery

    Modern AI runs on GPUs — chips originally designed for graphics rendering, not probabilistic reasoning. Each floating-point operation burns precious joules moving data across silicon. Extropic argues that this design is fundamentally mismatched to the needs of modern AI, which is probabilistic by nature. Instead of computing exact results, generative models sample from vast probability spaces. The company’s solution is the Thermodynamic Sampling Unit (TSU) — a chip that doesn’t process numbers, but samples from probability distributions directly:contentReference[oaicite:1]{index=1}.

    TSUs are built entirely from standard CMOS transistors, meaning they can scale using existing semiconductor fabs. Unlike exotic academic approaches that require magnetic junctions or optical randomness, Extropic’s design uses the natural thermal noise of transistors as its source of entropy. This turns what engineers usually fight to suppress — noise — into the very fuel for computation.

    X0 and XTR-0: The Birth of a New Computing Platform

    Extropic’s first hardware platform, XTR-0 (Experimental Testing & Research Platform 0), combines a CPU, FPGA, and sockets for daughterboards containing early test chips called X0. X0 proved that all-transistor probabilistic circuits can generate programmable randomness at scale. These chips perform operations like sampling from Bernoulli, Gaussian, or categorical distributions — the building blocks of probabilistic AI:contentReference[oaicite:2]{index=2}.

    The company’s pbit circuit acts like an electronic coin flipper, generating millions of biased random bits per second using 10,000× less energy than a GPU’s floating-point addition. Higher-order circuits like pdit (categorical sampler), pmode (Gaussian sampler), and pMoG (mixture-of-Gaussians generator) expand the toolkit, enabling full probabilistic models to be implemented natively in silicon. Together, these circuits form the foundation of the TSU architecture — a physical embodiment of energy-based computation:contentReference[oaicite:3]{index=3}.

    The Denoising Thermodynamic Model (DTM): Diffusion Without the Energy Bill

    Hardware alone isn’t enough. Extropic also introduced a new AI algorithm built specifically for TSUs — the Denoising Thermodynamic Model (DTM). Inspired by diffusion models like Stable Diffusion, DTMs chain together multiple energy-based models that gradually denoise data over time. This architecture avoids the “mixing–expressivity trade-off” that plagues traditional EBMs, making them both scalable and efficient:contentReference[oaicite:4]{index=4}.

    In simulations, DTMs running on modeled TSUs matched GPU-based diffusion models on image-generation benchmarks like Fashion-MNIST — while consuming roughly one ten-thousandth the energy. That’s the difference between joules and picojoules per image. The company’s open-source library, thrml, lets researchers simulate TSUs today, and even replicate the paper’s results on a GPU before the chips ship.

    The Physics of Intelligence: Turning Noise Into Computation

    At the heart of thermodynamic computing is a radical idea: computation as a physical relaxation process. Instead of enforcing digital determinism, TSUs let physical systems settle into low-energy configurations that correspond to probable solutions. This isn’t metaphorical — the chips literally use thermal fluctuations to perform Gibbs sampling across energy landscapes defined by machine-learned functions:contentReference[oaicite:5]{index=5}.

    In practical terms, it’s like replacing the brute-force precision of a GPU with the subtle statistical behavior of nature itself. Each transistor becomes a tiny particle in a thermodynamic system, collectively simulating the world’s most efficient sampler: reality.

    From Lab Demo to Scalable Platform

    The XTR-0 kit is already in the hands of select researchers, startups, and tinkerers. Its modular design allows easy upgrades to upcoming chips — like Z-1, Extropic’s first production-scale TSU, which will support complex probabilistic machine learning workloads. Eventually, TSUs will integrate directly with conventional accelerators, possibly as PCIe cards or even hybrid GPU-TSU chips:contentReference[oaicite:6]{index=6}.

    Extropic’s roadmap extends beyond AI. Because TSUs efficiently sample from continuous probabilistic systems, they could accelerate simulations in physics, chemistry, and biology — domains that already rely on stochastic processes. The company envisions a world where thermodynamic computing powers climate models, drug discovery, and autonomous reasoning systems, all at a fraction of today’s energy cost.

    Breaking the AI Energy Wall

    Extropic’s October 2025 announcement comes at a pivotal time. Data centers are facing grid bottlenecks across the U.S., and some companies are building nuclear-adjacent facilities just to keep up with AI demand:contentReference[oaicite:7]{index=7}. With energy costs set to define the next decade of AI, a 10,000× improvement in energy efficiency isn’t just an innovation — it’s a revolution.

    If Extropic’s thermodynamic hardware lives up to its promise, it could mark a “zero-to-one” moment for computing — one where the laws of physics, not the limits of silicon, define what’s possible. As the company put it in their launch note: “Once we succeed, energy constraints will no longer limit AI scaling.”

    Read the full technical paper on arXiv and explore the official Extropic site for their thermodynamic roadmap.

  • How Vibe Coding Became the Punk Rock of Software

    From meme to manifesto

    In March 2025 a single photo of legendary record producer Rick Rubin—eyes closed, headphones on, one hand resting on a mouse—started ricocheting around developer circles. Online jokesters crowned him the patron saint of “vibe coding,” a tongue-in-cheek label for writing software by feeling rather than formal process. Rubin did not retreat from the joke. Within ten weeks he had written The Way of Code, launched the interactive site TheWayOfCode.com, and joined a16z founders Marc Andreessen and Ben Horowitz on The Ben & Marc Show to unpack the project’s deeper intent .

    What exactly is vibe coding?

    Rubin defines vibe coding as the artistic urge to steer code by intuition, rhythm, and emotion instead of rigid methodology. In his view the computer is just another instrument—like a guitar or an MPC sampler—waiting for a distinct point of view. Great software, like great music, emerges when the creator “makes the code do what it does not want to do” and pushes past the obvious first draft .

    Developers have riffed on the idea, calling vibe coding a democratizing wave that lets non-programmers prototype, remix, and iterate with large language models. Cursor, Replit, and GitHub Copilot all embody the approach: prompt, feel, refine, ship. The punk parallel is apt. Just as late-70s punk shattered the gate-kept world of virtuoso rock, AI-assisted tooling lets anyone bang out a raw prototype and share it with the world.

    The Tao Te Ching, retold for the age of AI

    The Way of Code is not a technical handbook. Rubin adapts the Tao Te Ching verse-for-verse, distilling its 3 000-year-old wisdom into concise reflections on creativity, balance, and tool use. Each stanza sits beside an AI canvas where readers can remix the accompanying art with custom prompts—training wheels for vibe coding in real time .

    Rubin insists he drafted the verses by hand, consulting more than a dozen English translations of Lao Tzu until a universal meaning emerged. Only after the writing felt complete did collaborators at Anthropic build the interactive wrapper. The result blurs genre lines: part book, part software, part spiritual operating system.

    Five takeaways from the a16z conversation

    1. Tools come and go; the vibe coder persists. Rubin’s viral tweet crystallised the ethos: mastery lives in the artist, not in the implements. AI models will change yearly, but a cultivated inner compass endures .
    2. Creativity is remix culture at scale. From Beatles riffs on Roy Orbison to hip-hop sampling, art has always recombined prior work. AI accelerates that remix loop for text, images, and code alike. Rubin views the model as a woodshop chisel—powerful yet inert until guided.
    3. AI needs its own voice, not a human muzzle. Citing AlphaGo’s improbable move 37, Rubin argues that breakthroughs arrive when machines explore paths humans ignore. Over-tuning models with human guardrails risks sanding off the next creative leap.
    4. Local culture still matters. The trio warns of a drift toward global monoculture as the internet flattens taste. Rubin urges creators to seek fresh inspiration in remote niches and protect regional quirks before algorithmic averages wash them out.
    5. Stay true first, iterate second. Whether launching a startup or recording Johnny Cash alone with an acoustic guitar, the winning work begins with uncompromising authenticity. Market testing can polish rough edges later; it cannot supply the soul.

    Why vibe coding resonates with software builders

    • Lower barrier, higher ceiling. AI pairs “anyone can start” convenience with exponential leverage for masters. Rubin likens it to giving Martin Scorsese an infinite-shot storyboard tool; the director’s taste, not the tech, sets the upper bound .
    • Faster idea discovery. Generative models surface dozens of design directions in minutes, letting developers notice serendipitous mistakes—Rubin’s favorite creative catalyst—without burning months on dead-end builds.
    • Feedback loop with the collective unconscious. Each prompt loops communal knowledge back into personal intuition, echoing Jung’s and Sheldrake’s theories that ideas propagate when a critical mass “gets the vibe.”

    The road ahead: punk ethos meets AI engineering

    Vibe coding will not replace conventional software engineering. Kernel engineers, cryptographers, and avionics programmers still need rigorous proofs. Yet for product prototypes, game jams, and artistic experiments, the punk spirit offers a path that prizes immediacy and personal voice.

    Rubin closes The Way of Code with a challenge: “Tools will come and tools will go. Only the vibe coder remains.” The message lands because it extends his decades-long mission in music—strip away external noise until the work pulses with undeniable truth. In 2025 that mandate applies as much to lines of Python as to power chords. A new generation of software punks is already booting up their DAWs, IDEs, and chat windows. They are listening for the vibe and coding without fear.

  • MatterGen: Revolutionizing Material Design with Generative AI

    Materials innovation is central to technological progress, from powering modern devices with lithium-ion batteries to enabling efficient solar panels and carbon capture technologies. Yet, discovering new materials for these applications is an arduous process, historically reliant on trial-and-error experiments or computational screenings. Microsoft’s MatterGen is poised to change this paradigm, leveraging cutting-edge generative AI to revolutionize material discovery.

    The Challenge in Material Design

    Traditionally, researchers sift through vast databases of known materials or rely on high-throughput experiments to identify candidates with specific properties. While computational approaches have sped up this process, they are still limited by the need to evaluate millions of candidates from existing data. This bottleneck often misses novel and unexplored possibilities. MatterGen offers a transformative approach, generating novel materials directly based on user-defined properties like chemical composition, mechanical strength, or electronic and magnetic characteristics.

    What Is MatterGen?

    MatterGen is a diffusion-based generative model designed to create stable, unique, and novel (S.U.N.) inorganic materials. Unlike traditional material screening, which filters pre-existing datasets, MatterGen uses advanced AI algorithms to construct entirely new materials from scratch.

    This model employs 3D diffusion processes, iteratively refining atom positions, lattice parameters, and chemical compositions to meet desired property constraints. Its architecture accommodates material-specific complexities like periodicity and crystallographic symmetries, ensuring both stability and functionality.

    Key Innovations in MatterGen’s Architecture

    1. Diffusion Process Tailored for Materials: MatterGen’s architecture uses a novel forward and reverse diffusion approach to refine atomic structures from noisy initial configurations, ensuring equilibrium stability.
    2. Fine-Grained Control Over Design Constraints: The model can be conditioned to generate materials with specific space groups, chemical systems, or properties like high magnetic density or bulk modulus.
    3. Scalable Training Data: Leveraging over 600,000 entries from the Alexandria and Materials Project databases, MatterGen achieves superior performance compared to existing methods like CDVAE and DiffCSP.
    4. Novelty Through Disordered Structure Matching: A sophisticated algorithm evaluates whether generated materials represent genuinely new compositions or ordered variants of known structures.

    Validation Through Experimentation

    MatterGen’s capabilities extend beyond theoretical predictions. Collaborating with experimental labs, researchers synthesized TaCr₂O₆, a novel material generated by the model to meet a target bulk modulus of 200 GPa. Despite minor cationic disorder in the crystal structure, the material closely matched its computational design, achieving an experimentally measured bulk modulus of 158 GPa. This milestone demonstrates MatterGen’s practical applicability in guiding real-world material synthesis.

    Comparative Performance

    MatterGen significantly outperforms its predecessors:

    • Higher Stability Rates: The generated structures align closer to DFT (Density Functional Theory)-computed energy minima, with an average RMSD (Root Mean Square Deviation) 15 times lower than competing models.
    • Unprecedented Novelty: Leveraging its advanced dataset and refined diffusion processes, MatterGen generates a higher proportion of novel materials than previous approaches like CDVAE.
    • Property-Specific Design: The model excels in constrained design scenarios, such as creating materials with high bulk modulus or low supply-chain risk.

    Broader Implications

    The success of MatterGen heralds a new era in material science, shifting the focus from searching databases to generative design. By integrating MatterGen with complementary tools like MatterSim—Microsoft’s AI emulator for material property simulations—researchers can iteratively refine designs and simulations, accelerating the entire discovery process.

    Applications Across Industries

    • Energy Storage: Novel materials for high-performance batteries and fuel cells.
    • Carbon Capture: Adsorbents optimized for CO₂ sequestration.
    • Electronics: High-efficiency semiconductors and magnets for next-gen devices.

    Open Access for the Research Community

    True to Microsoft’s commitment to advancing science, the MatterGen code and associated datasets are available under an open MIT license. Researchers can fine-tune the model for their specific applications, fostering collaborative advancements in materials design.

    The Road Ahead

    MatterGen represents just the beginning of generative AI’s potential in material science. Future work will aim to address remaining challenges, including synthesizability, scalability, and real-world integration into industrial applications. With continued refinement, generative AI promises to unlock innovations across fields, from renewable energy to advanced manufacturing.