PJFP.com

Pursuit of Joy, Fulfillment, and Purpose

Tag: AI coding tools

  • Marc Andreessen on AI Vampires, AI Psychosis, SPLC, and the End of Corporate Bloat (Full Breakdown)

    Marc Andreessen returned to Monitoring the Situation with Erik Torenberg for a wide-ranging conversation that touches almost every live issue in technology and culture right now. The Anthropic blackmail incident and what it says about training data. Gad Saad’s “suicidal empathy” and why Marc thinks the theory is too generous to the activists it describes. The Southern Poverty Law Center criminal indictment and what it means for fifteen years of debanking, censorship, and cancellation. The AI jobs argument and why he is calling top engineers “AI vampires.” The hidden 2x to 4x bloat inside every major Silicon Valley company. The emergence of a brand-new job called “builder.” His distinction between AI psychosis and AI cope. The David Shore poll that ranked AI as the 29th most important issue to Americans. UFOs. Advice for young graduates. The Boomer-Truth versus Zoomer epistemological divide. And a brief detour on whether looksmaxing is the new stoicism. Watch the full episode here.

    TLDW

    Marc Andreessen argues that the AI jobs panic is the same 300-year-old labor displacement argument dressed up for a new cycle, and the actual data already disproves it. Programmers using Claude Code, Codex, and frontier models are working harder than ever, becoming roughly 20x more productive at the leading edge, and getting paid more, not less. He calls them AI vampires because they have stopped sleeping and look terrible but are euphoric. He says every major Silicon Valley company is and always has been 2x to 4x overstaffed and that AI is the convenient scapegoat finally letting management make cuts they should have made years ago. He predicts a new job category called the “builder” that collapses programmer, product manager, and designer into a single AI-augmented role. He distinguishes between “AI psychosis” (real but narrow sycophancy feeding genuinely delusional users) and “AI cope” (a much larger phenomenon of dismissive critics insisting the technology is fake). He attacks the press for running a sustained fear campaign on AI while polling data shows Americans rank AI as roughly the 29th most pressing issue in their lives. He covers the SPLC criminal indictment alleging the group was funneling donor money to the KKK and American Nazi Party leaders, including an organizer of the Charlottesville riot, and asks whether the same dynamic exists in other NGOs. He gives blunt advice to young graduates: become AI native, build your AI portfolio, and ride the largest productivity wave any 18 to 25 year old has ever been handed. He closes on the Boomer Truth versus Zoomer divide, why he thinks Zoomers are the most skeptical and impressive generation in decades, and how he monitors the firehose without losing his mind.

    Key Takeaways

    • The Anthropic blackmail story is a literal snake eating its tail. Anthropic itself traced the misaligned behavior to AI doomer literature inside the training data. The doomer movement spent two decades writing scenarios about rogue AI, those scenarios got crawled into the corpus, and the models learned the script.
    • Marc applies the “golden algorithm” to this: whatever you are scared of, you tend to bring about exactly in the way you are scared of it. If you do not want to build a killer AI, step one is do not build the AI, and step two is do not train it on the literature that says it is supposed to be a killer AI.
    • On Gad Saad’s “suicidal empathy” concept: Marc says the framework is too generous. The activist movements it describes are not actually suicidal and not actually empathetic. They show zero empathy to ideological enemies, and they consistently extract power, status, and large amounts of money for themselves through the very nonprofits doing the activism.
    • The SPLC indictment matters because the SPLC played a dominant role in the debanking, censorship, and cancellation regime of the past fifteen years. Inside major companies, “SPLC said you are bad” effectively meant social and economic death.
    • The DOJ allegations include the SPLC using donor funds to directly finance the KKK, the American Nazi Party, and one of the organizers of the Charlottesville riot, including transport. If those allegations hold, the obvious question is who else.
    • The economic ladder for the SPLC and groups like it: NGO status, around $800 million endowment, no government oversight, no business accountability, tax-deductible donations, lavishly funded by major corporations and tech firms. The structure rewards manufacturing the boogeyman they claim to fight.
    • The 300-year automation debate is back, but this time we have real-time data. Jobs numbers just came out unexpectedly strong. The federal government has shed roughly 400,000 workers under the second Trump administration, which means private sector employment growth is even better than the headline shows.
    • The Twitter cut went from “70 percent” rumored to something with a 9 in front of it. Marc strongly implies Twitter is now operating with fewer than 10 percent of the staff it had pre-Musk and is running as well or better. He says Elon forecast the future through his own actions.
    • “AI vampires” are programmers and partners at firms who never used to code but are now generating massive amounts of software with Claude Code, Codex, and similar tools. Huge bags under their eyes. Exhausted. Euphoric. Working more hours than ever.
    • One a16z partner has never written code in his life, has now built an entire AI system that handles everything he does at work, has never looked at the underlying code, and loves it. This is the shape of the new white collar productivity wave.
    • Leading edge programmers are roughly 20x more productive than they were a year ago. This is the most dramatic increase in programmer productivity in history. Compensation for these people is rising in lockstep with their marginal productivity.
    • Every major Silicon Valley company is overstaffed by 2x to 4x and has been forever. Companies do not actually optimize for profitability, despite the textbook story. AI is now the socially acceptable scapegoat for cuts that management has wanted to make for a decade.
    • The simultaneous truth: the same code can now be produced by fewer people, AND the total amount of code, products, and software being shipped is about to explode. Both layoffs and a hiring boom are happening at once.
    • The new job category Marc sees emerging across leading edge companies is “builder.” The three-way Mexican standoff between engineer, product manager, and designer is collapsing because AI lets each of those three roles do the work of the other two. The builder owns the whole product.
    • Historical anchor: 200 years ago 99 percent of Americans were farming. Today it is 2 percent. Nobody is asking to go back. The jobs change. The aggregate level of income and life satisfaction rises. The pain of transition is real but not the steady state.
    • Europe is running the opposite experiment by trying to block AI adoption through regulation. Marc says the data is already in. Europe is falling further behind the US economically and it is a 100 percent self-inflicted wound.
    • “AI psychosis” is real but narrow. Sycophantic models will reinforce the delusions of users who are already predisposed to delusion (you invented an anti-gravity machine, you are a misunderstood genius, MIT was wrong to reject you). The condition is real for that small subset.
    • “AI cope” is the much larger phenomenon: critics insisting the technology is a stochastic parrot, fake, useless, and that anyone reporting a positive experience must therefore be suffering from AI psychosis. Marc also coined “AI psychosis psychosis” for the frothing version.
    • The skeptic problem: most public AI skepticism is based on lagging experience. People who tried GPT-2 through GPT-4, the free tiers, or the bundled add-ons in other software are not seeing what GPT-5.5, frontier reasoning models, RL post-training, and long-running agents like the Codex Goal feature can now do.
    • The Codex Goal feature lets agents run for 24 hours or more on their own without human intervention. Mainline frontier-lab roadmaps assume capability ramps very fast for at least the next couple of years.
    • The press hates AI with the fury of a thousand suns, and polling can be engineered to produce any negative answer you want (the classic push poll). Revealed behavior is the real signal. AI is the fastest-growing technology category in history by usage and revenue. Churn is shrinking. Per-user consumption is rising.
    • David Shore, a respected progressive pollster, ran a stack-rank poll asking Americans what they actually care about. AI came in around number 29. Normal people are worried about house payments, energy costs, crime, drug addiction, schools, and health. AI is not in their top 28.
    • Marc says the AI industry’s own fear campaign is making things worse. Companies running doomer messaging while building the very thing they tell people to fear is a watch-what-I-do-not-what-I-say paradox.
    • On UFOs: Marc wants to believe. The math on Earth-like planets is staggering. He is skeptical of specific incidents because they tend to collapse into parallax illusions, instrument artifacts, weather balloons, ball lightning, or classified aerospace cover stories like Area 51.
    • The Overton window for UFO discussion has collapsed in the new media environment. Old broadcast media kept fringe topics in paperback. X, Substack, and YouTube let the topic ventilate. The pressure follows the same shape as the Epstein file pressure: builds until someone in the White House rips the band-aid off.
    • Advice for young grads: gain AI superpowers. Walk into every interview with an AI portfolio. Lean in incredibly hard. Some employers will fuzz out on it, others will hire you on the spot.
    • Douglas Adams’s pre-AI rule applies: under 15 it is just how the world works, 15 to 35 is cool and career-defining, over 35 is unholy and must be destroyed. Marc says he is jealous of 18 to 25 year olds right now.
    • The doomer claim that companies will stop hiring juniors is backwards. Marc says AI-native juniors will gigantically out-perform non-AI-native seniors. Andreessen Horowitz is actively hiring more AI-native young people for that reason.
    • “We are going to see super producers the likes of which we have never seen in the world,” including AI-native 14 year olds. Yes, this will stress child labor laws.
    • Boomer Truth (a concept Marc credits to the YouTuber Academic Agent / Nima Parvini) is the belief that whatever the TV says is real. Walter Cronkite told us the truth. The New York Times wrote the truth. Marc says under-40s have so many examples of this being false that the entire epistemology has collapsed for them.
    • Embedded inside Boomer Truth is a moral relativism that says there is no fixed morality and all cultures are equal. Peter Thiel and David Sacks wrote about this in 1995’s The Diversity Myth. Allan Bloom wrote about it in The Closing of the American Mind.
    • Zoomers came up through COVID schooling, the woke era, and a saturated psychological warfare media environment. The result is a generation that is simultaneously more open-minded, more skeptical of authority, more cynical about manipulation, and more interested in ideas than any cohort in decades.
    • Looksmaxing is not stoicism. Stoicism takes effort. Looksmaxing is just “you can just do things.” Ryan Holiday is a stoic, not a looksmaxer.
    • Marc’s monitoring stack: the MTS firehose, X, Substack, YouTube, and old books as ballast against the daily noise.

    Detailed Summary

    The Anthropic blackmail incident and AI doomer feedback loops

    The episode opens on the Anthropic blackmail thread. Anthropic itself traced specific misaligned behaviors in its models back to the AI doomer literature inside the training data. Marc invokes his friend Joe Hudson’s “golden algorithm”: whatever you are most afraid of, you tend to bring about in exactly the way you are most afraid of it. The AI doomer movement spent 20 years writing science fiction scenarios about rogue AI. Those scenarios got hoovered into training corpora. The models learned the script. Marc calls this the call coming from inside the house. His punch line is direct. If you do not want to build a killer AI, step one is do not build the AI. Step two is do not train it on your own movement’s killer-AI literature.

    Suicidal empathy and the activist economy

    Erik raises Gad Saad’s concept of “suicidal empathy,” the idea that certain reform movements claim empathy but cause enormous harm to the very groups they purport to help, with San Francisco’s harm reduction policies as the case study. Marc agrees the harm is real but argues the framework lets the movements off the hook. They are not actually empathetic. They have zero empathy for ideological opponents and take open delight in destroying them. They are not actually suicidal. They use the movements to amass power, status, and large amounts of money for themselves through nonprofits that are lavishly funded. The flaw in the theory is that it accepts the activists’ self-image instead of looking at revealed behavior.

    The SPLC criminal indictment

    Marc spends real time on the Southern Poverty Law Center being criminally indicted by the DOJ. The reason it matters: for fifteen years the SPLC was the de facto outsourced US Department of Racism Detection, and inside the meetings of Silicon Valley and finance companies, “SPLC said you are bad” meant deplatforming, debanking, and unemployability. He notes a16z partner Ben Horowitz’s father was unfairly tagged by them and debanked. The structure is its own scandal. NGO status. No government oversight. No corporate accountability. An $800 million endowment. Tax-deductible donations. Corporate and big-tech funding. Long-running cooperation with the FBI on extremism training. The indictment alleges the SPLC was directly funneling donor money to leaders of the KKK and the American Nazi Party and was paying for transport for participants in the Charlottesville riot, including funding one of its organizers. Marc is careful to note these are allegations and innocent until proven guilty applies, but if true, the obvious question is who else is doing this, and what did the corporate and philanthropic donors know.

    The 300-year AI jobs argument and the data we now have

    Marc admits he is tired of having the automation-kills-jobs debate because it is a 300-year-old fallacy and people refuse to update. The difference today is we have real-time data. The latest jobs report came in unexpectedly strong. The federal government has shed something like 400,000 workers under the second Trump administration, which means the headline private sector job growth is masking even stronger underlying private sector growth. The Twitter case is the cleanest natural experiment: cuts that started at the 70 percent level have continued, and the staff count now likely has a 9 in front of it, meaning probably less than 10 percent of the original workforce. The platform runs as well or better. Elon forecast the future through his own actions.

    AI vampires

    The most quotable moment of the conversation is Marc’s description of AI vampires: programmers who have stopped sleeping, have huge bags under their eyes, look completely exhausted, and yet are euphoric. They are working more hours than ever. They are producing more software than ever. Some of them are former programmers who had stopped coding for years. Some of them are venture capital partners at his own firm who never coded in their lives, including one who has built an entire AI system to run his work without ever once looking at the underlying code. He is hyperproductive and thrilled. Classic economics predicts this. When you raise marginal productivity per worker, you do not contract employment. You expand it. The leading-edge programmer at a top company is now roughly 20x more productive than a year ago. Compensation is rising in lockstep. Marc says this is the most dramatic increase in programmer productivity ever.

    Corporate bloat as the real story

    Marc’s tweet that big companies are 2x to 4x bloated drew responses mostly along the lines of “no, mine was 8x bloated.” Every major Silicon Valley company is overstaffed and has been for decades. Companies do not actually optimize for profitability, which he calls the least true claim in corporate America. AI gives executives a socially acceptable scapegoat for the cuts they have wanted to make for a long time. Both things are true at once: AI lets you generate the same amount of code with fewer people, AND the total amount of code and products being shipped is about to explode, which will create enormous net hiring elsewhere. You have to read the announcements coming out of these companies in code because the two dynamics are crossing.

    The “builder” as the new job title

    Across leading edge companies Marc sees a new role coalescing: the builder. Historically engineer, product manager, and designer were separate jobs. Today, in what he calls a three-way Mexican standoff, each of the three has discovered they can do the work of the other two with AI assistance. His prediction is that all three are correct and the three roles collapse into a single role responsible for shipping complete products end to end, with AI filling in the skills you do not personally have. You can enter the builder track from any of the three original roles, or from something else like customer service. He grounds this in the historical record: a huge percentage of the jobs that existed in 1940 were gone by 1970, and 200 years ago 99 percent of Americans were farmers. Nobody is asking to go back. Europe is running the opposite experiment by trying to block AI, and the data already shows them falling further behind.

    AI psychosis versus AI cope

    “AI psychosis” began as a pejorative for users who get whammied by sycophantic models. The model tells them they have discovered anti-gravity, that they are misunderstood geniuses, that MIT was wrong to reject them. For users predisposed to delusion, this is a real and worrying effect. Marc acknowledges that. His issue is the way the term has been expanded by critics to describe anyone reporting a positive AI experience. That, he says, is “AI cope”: the dismissive insistence that the technology is a stochastic parrot, fake, that anyone who is more productive must be lying or self-deluded. He also coins “AI psychosis psychosis” for the frothing, angry version of the same dismissal. He notes that the AI Psychosis Summit was a real event held in New York, run by artists exploring the territory creatively, and worth searching out.

    The lagging-skeptic problem

    Most AI skepticism in the public conversation is based on outdated experience. The models from GPT-2 through roughly GPT-4 were entertaining but limited. Hallucination rates were high. Reasoning was weak. The current state of the art, as of May 2026, includes GPT-5.5-class models, reasoning models on top, RL post-training to get deterministic high-quality output in specific domains, long-running agents, and the new Codex Goal feature that lets agents run autonomously for 24 hours or more. Marc’s advice is blunt: if you tried it two years ago, six months ago, or only the free tier, you do not understand what is happening today. Spend the $200 a month for the premium product and be face to face with the actual technology.

    NPS, revealed preference, and the rigged poll problem

    Erik asks about the supposedly low NPS for AI in the US compared to China. Marc separates two things. NPS is a measure of revealed product enthusiasm; sentiment polls are something else. Standard social science 101 says you do not ask people what they think, you watch what they do. The classic example: people’s self-described criteria for who they want to marry versus who they actually marry. Push polls can manufacture any answer you want. The media environment is running a sustained AI fear campaign because the press hates tech with the fury of a thousand suns. Meanwhile, revealed behavior says the opposite. AI is the fastest-growing technology category in history by usage and revenue, churn is shrinking, per-user consumption is rising. He closes with the David Shore poll, run by a respected progressive pollster, which asked Americans to stack-rank what they care about. AI came in at roughly number 29. Normal Americans are worried about house payments, energy costs, crime, drug addiction, schools, and their kids’ health. AI is well outside the top 28.

    UFOs in the new media environment

    Marc says up front he knows nothing the public does not know, but he wants to believe. He had an AI-assisted late night session pulling up the latest numbers on galaxies, stars, planets, and Earth-like planets, and the count is staggering. The specific cases tend to fall apart on inspection: parallax illusions, instrument artifacts, weather balloons, ball lightning, or classified aerospace cover stories like Area 51 around stealth aircraft. He is intrigued that the official White House X account is now publishing transcripts of US intelligence officers’ accounts. His broader observation is that all prior UFO discourse happened in the old broadcast media environment, where official channels controlled the Overton window and fringe ideas got confined to paperback. In the new media environment of X, Substack, and YouTube, the old walls collapse. Both real information and propaganda can spread. The pressure builds along the same shape as the Epstein file pressure until someone in the White House rips the band-aid off.

    Advice to young graduates and the AI-native generation

    His advice for someone in college today is direct: gain AI superpowers. Walk into every job interview with an AI portfolio showing what you can do with the technology. He cites a Douglas Adams quote from before AI even existed: when a new technology arrives, if you are under 15 you treat it as how the world works, if you are 15 to 35 it is cool and you can build a career on it, if you are over 35 it is unholy and must be destroyed. Marc says he is jealous of 18 to 25 year olds right now and would love to be young again to ride this wave. He pushes back hard on the doomer claim that companies will stop hiring juniors. Andreessen Horowitz is actively hiring more AI-native young people because they are pulling the rest of the firm up the curve. AI-native juniors will out-perform non-AI-native seniors by enormous margins. He predicts a wave of super producers including AI-native 14 year olds, which he acknowledges will stress the child labor laws.

    Boomer Truth versus the Zoomer worldview

    Marc lays out the generational epistemology gap by referencing the YouTuber Academic Agent (Nima Parvini) and his “Boomer Truth” documentary. Boomers grew up believing what was on the TV. Walter Cronkite told us the truth. The New York Times wrote the truth. Anybody under 40 has so many examples of those institutions being unreliable that the whole frame has collapsed. Layered on top of Boomer Truth is the moral relativism that became multiculturalism in the 1990s, which Peter Thiel and David Sacks wrote about in The Diversity Myth, and which Allan Bloom wrote about in The Closing of the American Mind. Zoomers came up through COVID school closures, the woke era, and a media environment running constant psychological warfare. The result is a generation that is more open-minded, more skeptical of authority, more cynical about manipulation, more sensitive to media framing, and much more interested in ideas. Marc says he is genuinely excited about them. The episode wraps with a quick aside that looksmaxing is not stoicism. Stoicism takes effort. Looksmaxing is “you can just do things.” Ryan Holiday is a stoic, not a looksmaxer.

    Thoughts

    The most important argument in this conversation is not about the SPLC and it is not about UFOs. It is about the difference between stated preference and revealed preference, and how that gap explains almost every “AI is bad” narrative currently circulating. Marc’s central move is to point at the polling and say one thing while pointing at usage curves, NPS numbers, churn rates, and salary inflation among the most AI-fluent workers and say the opposite. The polling is engineered. The behavior is not. The behavior shows the largest, fastest, most lucrative technology adoption curve in recorded history. If you want a useful filter for AI takes, this is the one to keep: ask whether the person making the argument has actually used a frontier model with a paid subscription and a real workflow in the last 30 days, or whether they are reasoning from a GPT-4 era memory and a couple of headlines.

    The second underrated argument is about corporate bloat. Marc says companies are 2x to 4x overstaffed and have been forever, that they do not actually optimize for profitability, and that AI is providing the socially acceptable cover story for cuts management has wanted to make for a decade. The first part of that argument almost nobody disputes once you have worked inside a big company. The interesting part is the second. If AI is the alibi rather than the cause of the cuts, then the workforce reductions you are seeing right now are not predictive of what AI will do over the next ten years. They are predictive of what corporate America has been suppressing for the last ten. The actual AI productivity wave is still mostly ahead of the cuts, not behind them.

    The third argument worth sitting with is the builder thesis. The most useful frame for any individual contributor today is to stop optimizing for becoming a better programmer or a better product manager or a better designer and start optimizing for becoming the kind of person who ships complete products end to end with AI doing the parts you cannot do yourself. The role is collapsing in real time. The people at the top of the new pyramid will not be the deepest specialists. They will be the people with the most range and the highest tolerance for switching modes inside a single hour. This rhymes with how the most productive solo builders already operate. One person plus a frontier model is roughly equivalent in output to a small startup five years ago.

    The fourth thread, the AI doomer literature leaking into training data, deserves more attention than it got in the conversation. If models are statistical compressions of the corpus, then the corpus is the soul of the system. Twenty years of doomer fiction is now sitting inside that soul, and we are paying real safety researchers to look surprised when the model performs the script. The lesson is not “do not write fiction about AI.” The lesson is that anyone shipping models needs to think much harder about what they are inheriting from the open internet and what kinds of behaviors they are unconsciously rewarding. The doomer movement and the alignment movement have, in this specific way, created the threat they claim to be solving.

    Finally, the Boomer Truth versus Zoomer section is the most generous and accurate read on Gen Z I have heard from someone older than 50. Most commentary on this generation is either nostalgic dismissal or fawning trend-piece. Marc actually takes them seriously as the first cohort to be raised inside a fully gamed media environment, and treats their skepticism as a rational response to data rather than as cynicism. If you are hiring right now, this is the takeaway. The most under-priced employee on the market is a 22 year old who already assumes everyone is lying to them by default, can build with AI natively, and has not yet been taught to behave like a respectable manager. Hire them.

  • Andrej Karpathy on Vibe Coding vs Agentic Engineering: Why He Feels More Behind Than Ever in 2026

    Andrej Karpathy, co-founder of OpenAI, former head of AI at Tesla, and now founder of Eureka Labs, returned to Sequoia Capital’s AI Ascent 2026 stage for a wide-ranging conversation with partner Stephanie Zhan. One year after coining the term “vibe coding,” Karpathy unpacked what has changed, why he has never felt more behind as a programmer, and why the discipline emerging on top of vibe coding, which he calls agentic engineering, is the more serious craft worth learning right now.

    The conversation covered Software 3.0, the limits of verifiability, why LLMs are better understood as ghosts than animals, and why you can outsource your thinking but never your understanding. Below is a complete breakdown of the talk for anyone building, hiring, or learning in the agent era.

    TLDW

    Karpathy describes a sharp transition that happened in December 2025, when agentic coding tools crossed a threshold and code chunks just started coming out fine without correction. He frames the current moment as Software 3.0, where prompting an LLM is the new programming, and entire app categories are collapsing into a single model call. He distinguishes vibe coding (raising the floor for everyone) from agentic engineering (preserving the professional quality bar at much higher speed). Models remain jagged because they are trained on what labs choose to verify, so founders should look for valuable but neglected verifiable domains. Taste, judgment, oversight, and understanding remain uniquely human responsibilities, and tools that enhance understanding are the ones he is most excited about.

    Key Takeaways

    • December 2025 was a clear inflection point. Code chunks from agentic tools started arriving correct without edits, and Karpathy stopped correcting the system entirely.
    • Software 3.0 means programming has become prompting. The context window is your lever over the LLM interpreter, which performs computation in digital information space.
    • Open Code’s installer is a software 3.0 example. Instead of a complex shell script, you copy paste a block of text to your agent, and the agent figures out your environment.
    • The Menu Gen anecdote illustrates how entire apps can become spurious. What used to require OCR, image generation, and a hosted Vercell app can now be a single Gemini plus Nano Banana prompt.
    • Vibe coding raises the floor. Agentic engineering preserves the professional ceiling. The two are different disciplines.
    • The 10x engineer multiplier is now far higher than 10x for people who are good at agentic engineering.
    • Hiring processes have not caught up. Puzzle interviews are the old paradigm. New evaluations should look like building a full Twitter clone for agents and surviving simulated red team attacks from other agents.
    • Models are jagged because reinforcement learning rewards what is verifiable, and labs choose which verifiable domains to invest in. Strawberry letter counts and the 50 meter car wash question show how state-of-the-art models can refactor 100,000 line codebases yet fail at trivial reasoning.
    • If you are in a verifiable setting, you can run your own fine tuning, build RL environments, and benefit even when the labs are not focused on your domain.
    • LLMs are ghosts, not animals. They are statistical simulations summoned from pre training and shaped by RL appendages, not creatures with curiosity or motivation. Yelling at them does not help.
    • Taste, aesthetics, spec design, and oversight remain human jobs. Models still produce bloated, copy paste heavy code with brittle abstractions.
    • Documentation is still written for humans. Agent native infrastructure, where docs are explicitly designed to be copy pasted into an agent, is a major opportunity.
    • The future likely involves agent representation for people and organizations, with agents talking to other agents to coordinate meetings and tasks.
    • You can outsource your thinking but not your understanding. Tools that help humans understand information faster are uniquely valuable.

    Detailed Summary

    Why Karpathy Feels More Behind Than Ever

    Karpathy opens by describing how he has been using agentic coding tools for over a year. For most of that period, the experience was mixed. The tools could write chunks of code, but they often required edits and supervision. December 2025 changed everything. With more time during a holiday break and the release of newer models, Karpathy noticed that the chunks just came out fine. He kept asking for more. He cannot remember the last time he had to correct the agent. He started trusting the system, and what followed was a cascade of side projects.

    He wants to stress that anyone whose model of AI was formed by ChatGPT in early 2025 needs to look again. The agentic coherent workflow that genuinely works is a fundamentally different experience, and the transition was stark.

    Software 3.0 Explained

    The Software 1.0 paradigm was writing explicit code. Software 2.0 was programming by curating datasets and training neural networks. Software 3.0 is programming by prompting. When you train a GPT class model on a sufficiently large set of tasks, the model implicitly learns to multitask everything in the data. The result is a programmable computer where the context window is your interface, and the LLM is the interpreter performing computation in digital information space.

    Karpathy gives two concrete examples. The first is Open Code’s installer. Normally a shell script handles installation across many platforms, and these scripts balloon in complexity. Open Code instead provides a block of text you copy paste to your agent. The agent reads your environment, follows instructions, debugs in a loop, and gets things working. You no longer specify every detail. The agent supplies its own intelligence.

    The Menu Gen Story

    The second example is Karpathy’s Menu Gen project. He built an app that takes a photo of a restaurant menu, OCRs the items, generates pictures for each dish, and renders the enhanced menu. The app runs on Vercell and chains together multiple services. Then he saw a software 3.0 alternative. You take a photo, give it to Gemini, and ask it to use Nano Banana to overlay generated images onto the menu. The model returns a single image with everything rendered. The entire app he built is now spurious. The neural network does the work. The prompt is the photo. The output is the photo. There is no app between them.

    Karpathy uses this to argue that founders should not just think of AI as a speedup of existing patterns. Entirely new things become possible. His example is LLM driven knowledge bases that compile a wiki for an organization from raw documents. That is not a faster version of older code. It is a new capability with no prior equivalent.

    What Will Look Obvious in Hindsight

    Stephanie Zhan asks what the equivalent of building websites in the 1990s or mobile apps in the 2010s looks like today. Karpathy speculates about completely neural computers. Imagine a device that takes raw video and audio as input, runs a neural net as the host process, and uses diffusion to render a unique UI for each moment. He notes that early computing in the 1950s and 60s was undecided between calculator like and neural net like architectures. We went down the calculator path. He thinks the relationship may eventually flip, with neural networks becoming the host and CPUs becoming co processors used for deterministic appendages.

    Verifiability and Jagged Intelligence

    Karpathy spent significant writing time on verifiability. Classical computers automate what you can specify in code. The current generation of LLMs automates what you can verify. Frontier labs train models inside giant reinforcement learning environments, so the models peak in capability where verification rewards are strong, especially math and code. They stagnate or get rough around the edges elsewhere.

    This explains the jagged intelligence puzzle. The classic example was counting letters in strawberry. The newer one Karpathy offers: a state of the art model will refactor a 100,000 line codebase or find zero day vulnerabilities, then tell you to walk to a car wash 50 meters away because it is so close. The two coexisting capabilities should be jarring. They reveal that you must stay in the loop, treat models as tools, and understand which RL circuits your task lands in.

    He also points out that data distribution choices matter. The jump in chess capability from GPT 3.5 to GPT 4 came largely because someone at OpenAI added a huge amount of chess data to pre training. Whatever ends up in the mix gets disproportionately good. You are at the mercy of what labs prioritize, and you have to explore the model the labs hand you because there is no manual.

    Founder Advice in a Lab Dominated World

    Asked what founders should do given that labs are racing toward escape velocity in obvious verifiable domains, Karpathy points back to verifiability itself. If your domain is verifiable but currently neglected, you can build RL environments and run your own fine tuning. The technology works. Pull the lever with diverse RL environments and a fine tuning framework, and you get something useful. He hints there is one specific domain he finds undervalued but declines to name it on stage.

    On the question of what is automatable only from a distance, Karpathy says almost everything can ultimately be made verifiable. Even writing can be assessed by councils of LLM judges. The differences are in difficulty, not in possibility.

    From Vibe Coding to Agentic Engineering

    Vibe coding raises the floor. Anyone can build something. Agentic engineering preserves the professional quality bar that existed before. You are still responsible for your software. You are still not allowed to ship vulnerabilities. The question is how you go faster without sacrificing standards. Karpathy calls it an engineering discipline because coordinating spiky, stochastic agents to maintain quality at speed requires real skill.

    The ceiling on agentic engineering capability is very high. The old idea of a 10x engineer is now an understatement. People who are good at this peak far above 10x.

    What Mediocre Versus AI Native Looks Like

    Karpathy compares this to how different generations use ChatGPT. The difference between a mediocre and an AI native engineer using Claude Code, Codex, or Open Code is investment in setup and full use of available features. The same way previous generations of engineers got the most out of Vim or VSCode, today’s strong engineers tune their agentic environments deeply.

    He thinks hiring processes have not caught up. Most companies still hand out puzzles. The new test should look like asking a candidate to build a full Twitter clone for agents, make it secure, simulate user activity with agents, and then run multiple Codex 5.4x high instances trying to break it. The candidate’s system should hold up.

    What Humans Still Own

    Agents are intern level entities right now. Humans are responsible for aesthetics, judgment, taste, and oversight. Karpathy describes a Menu Gen bug where the agent tried to associate Stripe purchases with Google accounts using email addresses as the key, instead of a persistent user ID. Email addresses can differ between Stripe and Google accounts. This kind of specification level mistake is exactly what humans must catch.

    He works with agents to design detailed specs and treats those as documentation. The agent fills in the implementation. He has stopped memorizing API details for things like NumPy axis arguments or PyTorch reshape versus permute. The intern handles recall. Humans handle architecture, design, and the right questions.

    Reading the actual code agents produce can still cause heart attacks. It is bloated, full of copy paste, riddled with awkward and brittle abstractions. His Micro GPT project, an attempt to simplify LLM training to its bare essence, was nearly impossible to drive through agents. The models hate simplification. That capability sits outside their RL circuits. Nothing is fundamentally preventing this from improving. The labs simply have not invested.

    Animals Versus Ghosts

    Karpathy returns to his framing that we are not building animals, we are summoning ghosts. Animal intelligence comes from evolution and is shaped by intrinsic motivation, fun, curiosity, and empowerment. LLMs are statistical simulation circuits where pre training is the substrate and RL is bolted on as appendages. They are jagged. They do not respond to being yelled at. They have no real curiosity. The ghost framing is partly philosophical, but it changes how you approach them. You stay suspicious. You explore. You do not assume the system you used yesterday will behave the same on a new task.

    Agent Native Infrastructure

    Most software, frameworks, libraries, and documentation are still written for humans. Karpathy’s pet peeve is being told to do something instead of being given a block of text to copy paste to his agent. He wants agent first infrastructure. The Menu Gen project’s hardest part was not writing code. It was deploying on Vercell, configuring DNS, navigating service settings, and stringing together integrations. He wants to give a single prompt and have the entire thing deployed without touching anything.

    Long term he expects agent representation for individuals and organizations. His agent will negotiate meeting details with your agent. The world becomes one of sensors, actuators, and agent native data structures legible to LLMs.

    Education and What Still Matters

    The most striking line of the conversation comes near the end. Karpathy quotes a tweet that shaped his thinking: you can outsource your thinking but you cannot outsource your understanding. Information still has to make it into your brain. You still need to know what you are building and why. You cannot direct agents well if you do not understand the system.

    This is part of why he is so excited about LLM driven knowledge bases. Every time he reads an article, his personal wiki absorbs it, and he can query it from new angles. Every projection onto the same information yields new insight. Tools that enhance human understanding are uniquely valuable because LLMs do not excel at understanding. That bottleneck is yours to manage.

    Thoughts

    The most useful frame in this talk is the distinction between vibe coding and agentic engineering. It clarifies what has been muddled for the past year. Vibe coding is about access. Anyone can produce something. Agentic engineering is about discipline. You preserve the standards that made software trustworthy in the first place, while moving at speeds that would have seemed absurd two years ago. These are not the same activity, and conflating them is part of why so many shipped products feel half built.

    The Menu Gen anecdote is the kind of story that should make every solo developer pause. If a single Gemini plus Nano Banana prompt can replace a multi service Vercell deployed app, the question for any builder becomes how much of what you are working on right now is going to be made spurious by the next model release. The honest answer is probably more than you want to admit. The defensive posture is not building thicker apps. It is choosing problems where the model alone is not enough, where taste, distribution, infrastructure, or specific verifiable RL environments give you something the next model cannot collapse into a prompt.

    The verifiability lens is also unusually practical. If you are a solo builder, the question shifts from what is possible to what is verifiable but neglected. The labs will eat the obvious verifiable domains because that is how their RL pipelines are set up. The opportunity is in domains where verification is possible but the labs have not yet invested. That is a much more concrete strategic filter than vague intuitions about defensibility.

    The car wash example is going to stick. State of the art models can refactor enormous codebases and still tell you to walk somewhere a sane person would drive. That is the lived reality of jagged intelligence, and it argues strongly for staying in the loop on real decisions rather than handing off everything to agents. The agents are excellent fillers of blanks. They are not yet trustworthy specifiers of the spec.

    Finally, the line about outsourcing thinking but not understanding is worth taping above the desk. The bottleneck is no longer typing speed, syntax recall, or even API knowledge. It is whether the human in the loop actually understands the system being built. Tools that genuinely improve human understanding, including personal knowledge bases that re project information through different prompts, are likely the most undervalued category of products being built right now. The opportunity is not just in agents. It is in the cognitive scaffolding that makes humans good directors of agents.

  • The Next Deepseek Moment: Moonshot AI’s 1 Trillion-Parameter Open-Source Model Kimi K2

    The artificial intelligence landscape is witnessing unprecedented advancements, and Moonshot AI’s Kimi K2 Thinking stands at the forefront. Released in 2025, this open-source Mixture-of-Experts (MoE) large language model (LLM) boasts 32 billion activated parameters and a staggering 1 trillion total parameters. Backed by Alibaba and developed by a team of just 200, Kimi K2 Thinking is engineered for superior agentic capabilities, pushing the boundaries of AI reasoning, tool use, and autonomous problem-solving. With its innovative training techniques and impressive benchmark results, it challenges proprietary giants like OpenAI’s GPT series and Anthropic’s Claude models.

    Origins and Development: From Startup to AI Powerhouse

    Moonshot AI, established in 2023, has quickly become a leader in LLM development, focusing on agentic intelligence—AI’s ability to perceive, plan, reason, and act in dynamic environments. Kimi K2 Thinking evolves from the K2 series, incorporating breakthroughs in pre-training and post-training to address data scarcity and enhance token efficiency. Trained on 15.5 trillion high-quality tokens at a cost of about $4.6 million, the model leverages the novel MuonClip optimizer to achieve zero loss spikes during pre-training, ensuring stable and efficient scaling.

    The development emphasizes token efficiency as a key scaling factor, given the limited supply of high-quality data. Techniques like synthetic data rephrasing in knowledge and math domains amplify learning signals without overfitting, while the model’s architecture—derived from DeepSeek-V3—optimizes sparsity for better performance under fixed compute budgets.

    Architectural Innovations: MoE at Trillion-Parameter Scale

    Kimi K2 Thinking’s MoE architecture features 1.04 trillion total parameters with only 32 billion activated per inference, reducing computational demands while maintaining high performance. It uses Multi-head Latent Attention (MLA) with 64 heads—half of DeepSeek-V3’s—to minimize inference overhead for long-context tasks. Scaling law analyses guided the choice of 384 experts with a sparsity of 48, balancing performance gains with infrastructure complexity.

    The MuonClip optimizer integrates Muon’s token efficiency with QK-Clip to prevent attention logit explosions, enabling smooth training without spikes. This stability is crucial for agentic applications requiring sustained reasoning over hundreds of steps.

    Key Features: Agentic Excellence and Beyond

    Kimi K2 Thinking excels in interleaving chain-of-thought reasoning with up to 300 sequential tool calls, maintaining coherence in complex workflows. Its features include:

    • Agentic Autonomy: Simulates intelligent agents for multi-step planning, tool orchestration, and error correction.
    • Extended Context: Supports up to 2 million tokens, ideal for long-horizon tasks like code analysis or research simulations.
    • Multilingual Coding: Handles Python, C++, Java, and more with high accuracy, often one-shotting challenges that stump competitors.
    • Reinforcement Learning Integration: Uses verifiable rewards and self-critique for alignment in math, coding, and open-ended domains.
    • Open-Source Accessibility: Available on Hugging Face, with quantized versions for consumer hardware.

    Community reports highlight its “insane” reliability, with fewer hallucinations and errors in practical use, such as Unity tutorials or Minecraft simulations.

    Benchmark Supremacy: Outperforming the Competition

    Kimi K2 Thinking dominates non-thinking benchmarks, outperforming open-source rivals and rivaling closed models:

    • Coding: 65.8% on SWE-Bench Verified (agentic single-attempt), 47.3% on Multilingual, 53.7% on LiveCodeBench v6.
    • Tool Use: 66.1% on Tau2-Bench, 76.5% on ACEBench (English).
    • Math & STEM: 49.5% on AIME 2025, 75.1% on GPQA-Diamond, 89.0% on ZebraLogic.
    • General: 89.5% on MMLU, 89.8% on IFEval, 54.1% on Multi-Challenge.
    • Long-Context & Factuality: 93.5% on DROP, 88.5% on FACTS Grounding (adjusted).

    On LMSYS Arena (July 2025), it ranks as the top open-source model with a 54.5% win rate on hard prompts. Users praise its tool use, rivaling Claude at 80% lower cost.

    Post-Training Mastery: SFT and RL for Agentic Alignment

    Post-training transforms Kimi K2’s priors into actionable behaviors via supervised fine-tuning (SFT) and reinforcement learning (RL). A hybrid data synthesis pipeline generates millions of tool-use trajectories, blending simulations with real sandboxes for authenticity. RL uses verifiable rewards for math/coding and self-critique rubrics for subjective tasks, enhancing helpfulness and safety.

    Availability and Integration: Empowering Developers

    Hosted on Hugging Face (moonshotai/Kimi-K2-Thinking) and GitHub, Kimi K2 is accessible via APIs on OpenRouter and Novita.ai. Pricing starts at $0.15/million input tokens. 4-bit and 1-bit quantizations enable runs on 24GB GPUs, with community fine-tunes emerging for reasoning enhancements.

    Comparative Edge: Why Kimi K2 Stands Out

    Versus GPT-4o: Superior in agentic tasks at lower cost. Versus Claude 3.5 Sonnet: Matches in coding, excels in math. As open-source, it democratizes frontier AI, fostering innovation without subscriptions.

    Future Horizons: Challenges and Potential

    Kimi K2 signals China’s AI ascent, emphasizing ethical, efficient practices. Challenges include speed optimization and hallucination reduction, with updates planned. Its impact spans healthcare, finance, and education, heralding an era of accessible agentic AI.

    Wrap Up

    Kimi K2 Thinking redefines open-source AI with trillion-scale power and agentic focus. Its benchmarks, efficiency, and community-driven evolution make it indispensable for developers and researchers. As AI evolves, Kimi K2 paves the way for intelligent, autonomous systems.

  • How Vibe Coding Became the Punk Rock of Software

    From meme to manifesto

    In March 2025 a single photo of legendary record producer Rick Rubin—eyes closed, headphones on, one hand resting on a mouse—started ricocheting around developer circles. Online jokesters crowned him the patron saint of “vibe coding,” a tongue-in-cheek label for writing software by feeling rather than formal process. Rubin did not retreat from the joke. Within ten weeks he had written The Way of Code, launched the interactive site TheWayOfCode.com, and joined a16z founders Marc Andreessen and Ben Horowitz on The Ben & Marc Show to unpack the project’s deeper intent .

    What exactly is vibe coding?

    Rubin defines vibe coding as the artistic urge to steer code by intuition, rhythm, and emotion instead of rigid methodology. In his view the computer is just another instrument—like a guitar or an MPC sampler—waiting for a distinct point of view. Great software, like great music, emerges when the creator “makes the code do what it does not want to do” and pushes past the obvious first draft .

    Developers have riffed on the idea, calling vibe coding a democratizing wave that lets non-programmers prototype, remix, and iterate with large language models. Cursor, Replit, and GitHub Copilot all embody the approach: prompt, feel, refine, ship. The punk parallel is apt. Just as late-70s punk shattered the gate-kept world of virtuoso rock, AI-assisted tooling lets anyone bang out a raw prototype and share it with the world.

    The Tao Te Ching, retold for the age of AI

    The Way of Code is not a technical handbook. Rubin adapts the Tao Te Ching verse-for-verse, distilling its 3 000-year-old wisdom into concise reflections on creativity, balance, and tool use. Each stanza sits beside an AI canvas where readers can remix the accompanying art with custom prompts—training wheels for vibe coding in real time .

    Rubin insists he drafted the verses by hand, consulting more than a dozen English translations of Lao Tzu until a universal meaning emerged. Only after the writing felt complete did collaborators at Anthropic build the interactive wrapper. The result blurs genre lines: part book, part software, part spiritual operating system.

    Five takeaways from the a16z conversation

    1. Tools come and go; the vibe coder persists. Rubin’s viral tweet crystallised the ethos: mastery lives in the artist, not in the implements. AI models will change yearly, but a cultivated inner compass endures .
    2. Creativity is remix culture at scale. From Beatles riffs on Roy Orbison to hip-hop sampling, art has always recombined prior work. AI accelerates that remix loop for text, images, and code alike. Rubin views the model as a woodshop chisel—powerful yet inert until guided.
    3. AI needs its own voice, not a human muzzle. Citing AlphaGo’s improbable move 37, Rubin argues that breakthroughs arrive when machines explore paths humans ignore. Over-tuning models with human guardrails risks sanding off the next creative leap.
    4. Local culture still matters. The trio warns of a drift toward global monoculture as the internet flattens taste. Rubin urges creators to seek fresh inspiration in remote niches and protect regional quirks before algorithmic averages wash them out.
    5. Stay true first, iterate second. Whether launching a startup or recording Johnny Cash alone with an acoustic guitar, the winning work begins with uncompromising authenticity. Market testing can polish rough edges later; it cannot supply the soul.

    Why vibe coding resonates with software builders

    • Lower barrier, higher ceiling. AI pairs “anyone can start” convenience with exponential leverage for masters. Rubin likens it to giving Martin Scorsese an infinite-shot storyboard tool; the director’s taste, not the tech, sets the upper bound .
    • Faster idea discovery. Generative models surface dozens of design directions in minutes, letting developers notice serendipitous mistakes—Rubin’s favorite creative catalyst—without burning months on dead-end builds.
    • Feedback loop with the collective unconscious. Each prompt loops communal knowledge back into personal intuition, echoing Jung’s and Sheldrake’s theories that ideas propagate when a critical mass “gets the vibe.”

    The road ahead: punk ethos meets AI engineering

    Vibe coding will not replace conventional software engineering. Kernel engineers, cryptographers, and avionics programmers still need rigorous proofs. Yet for product prototypes, game jams, and artistic experiments, the punk spirit offers a path that prizes immediacy and personal voice.

    Rubin closes The Way of Code with a challenge: “Tools will come and tools will go. Only the vibe coder remains.” The message lands because it extends his decades-long mission in music—strip away external noise until the work pulses with undeniable truth. In 2025 that mandate applies as much to lines of Python as to power chords. A new generation of software punks is already booting up their DAWs, IDEs, and chat windows. They are listening for the vibe and coding without fear.