<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
  xmlns:content="http://purl.org/rss/1.0/modules/content/"
  xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Rajkiran Panuganti</title>
    <link>https://rajkiranpanuganti.com</link>
    <description>Rajkiran Panuganti is an AI engineer and researcher. He spent 14 years at Microsoft working on Microsoft 365 Copilot Chat, Bing search ranking, Cortana, and Azure ML. Today he is building a new AI startup in stealth, advises other AI startups, conducts research on mechanistic interpretability, and writes weekly about Generative AI.</description>
    <language>en</language>
    <lastBuildDate>Wed, 06 May 2026 14:56:06 GMT</lastBuildDate>
    <atom:link href="https://rajkiranpanuganti.com/feed/" rel="self" type="application/rss+xml"/>
    <item>
      <title>The Emerging Communication Stack for Agents</title>
      <link>https://rajkiranpanuganti.com/blog/the-emerging-communication-stack-for-agents/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-emerging-communication-stack-for-agents/</guid>
      <pubDate>Wed, 06 May 2026 00:00:00 GMT</pubDate>
      <description>Two communication stacks for AI agents are being built simultaneously: the channel layer (voice, email, SMS, Slack, WhatsApp) where agents reach humans, and the protocol layer (MCP, A2A, AG-UI) where agents reach each other and tools. Here is what is in each, who is building it, and what 2028 looks like.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>There are two ways agents communicate, and both are being rebuilt at the same time.</p>
<p>The first is how agents reach humans. Humans don't want to log into a new app called "agents.com." They want their AI to show up where they already are: Slack, Outlook, Gmail, WhatsApp, iMessage, the phone. So a layer of infrastructure has been built and funded to let agents enter every existing channel: phone numbers an agent answers, inboxes an agent owns, Slack workspaces where an agent is a first-class member. This is the channel layer. It's where revenue accrues, priced per minute and per resolution.</p>
<p>The second is how agents reach each other and reach tools. Agents need their own native protocols, because human channels assume a human at one end. So a parallel stack has been donated to the Linux Foundation in the past twelve months: MCP for tools, A2A for peer agents, AG-UI for frontends, plus discovery and identity layers underneath. This is the protocol layer.</p>
<p>The interesting tension is that the two stacks are converging. Voice agents are starting to call other voice agents. Email agents are threading with email agents. The protocol layer was supposed to be where agents talk to each other, but a meaningful chunk of agent-to-agent communication is already happening through SIP and SMTP because that's where the agents already live. This newsletter walks through the channel layer, then the protocol stack, then what happens when they collide.</p>
<hr>
<h2>Voice and phone</h2>
<p>Voice is the most aggressive front of the channel war because contact centers are large, expensive, and clearly automatable. Pricing has commoditized to roughly seven to fifteen cents per minute at the platform layer; an all-in voice agent runs fifteen to thirty cents per minute today, roughly one tenth of a fully loaded US contact-center agent.</p>
<p><strong>Vapi.</strong> Voice agent orchestration sitting between LLM, STT, TTS, and SIP/WebRTC, with bring-your-own-keys. Five cents per minute platform fee plus components. Series A from Bessemer in 2024.</p>
<p><strong>Bland AI.</strong> Pitches "infinite phone calls" by owning its own inference stack and telephony fabric. Build plan is $359 for 500 minutes. Series B from Scale Venture Partners.</p>
<p><strong>Retell AI.</strong> YC-backed, transparent seven cents per minute, built for developers wiring voice into existing products without owning the speech stack.</p>
<p><strong>ElevenLabs Conversational AI.</strong> Full voice-agent SDK on top of ElevenLabs TTS. Eight, ten, and twelve cents per minute across Standard, Turbo, and Premium, with a 95 percent silence discount over ten seconds. ElevenLabs cut Conversational AI pricing materially in 2025 as OpenAI Realtime and Cartesia entered.</p>
<p><strong>OpenAI Realtime API.</strong> $32 per million input audio tokens, $0.40 cached, $64 output. The wholesale price the rest of the platform layer pays through. The OpenAI sample voice agent and ChatGPT Voice Mode both run on LiveKit's WebRTC stack.</p>
<p><strong>LiveKit Agents.</strong> The framework underneath ChatGPT Voice Mode and Character.ai's voice product. 1.0 in April 2025, now ships with native MCP tool support.</p>
<p><strong>Pipecat.</strong> Daily.co's open-source voice framework, Python and JavaScript SDKs, dozens of model integrations, Pipecat Cloud as the managed offering.</p>
<p><strong>Hume AI EVI.</strong> Empathic voice with emotion-aware turn-taking. Seven cents down to four cents per minute. Series B from EQT Ventures and Premji Invest.</p>
<p><strong>PolyAI, Parloa, Cognigy.</strong> The European enterprise wedge. PolyAI runs voice for Marriott, FedEx, and Caesars (Series C at $500 million from NVentures, $116 million total). Parloa runs Decathlon and Swiss Life (Series B at $66 million from Altimeter). Cognigy runs Lufthansa, Bosch, Toyota, Mercedes-Benz, and Allianz (Series C at $100 million, $175 million total).</p>
<p><strong>Sesame.</strong> Brendan Iribe's company. Released CSM-1B as open-weights on HuggingFace in early 2025: end-to-end speech-in to speech-out, the most credible open-source competitor to closed voice agents.</p>
<p>The pattern: voice agents charge by the minute, those minutes look exactly like the minutes a human agent would have charged for, and pricing drops every quarter as the speech stack underneath commoditizes. The speech-model layer (ElevenLabs, Cartesia, Deepgram, AssemblyAI, Inworld, Whisper successors) has been getting roughly half as expensive every year for two years running. Voice is going to be cheap.</p>
<hr>
<h2>Email</h2>
<p>If voice is the loudest channel, email is the deepest. It's still the medium of record for enterprise communication, and any agent inside a corporate workflow eventually needs to send and receive it.</p>
<p><strong>Resend.</strong> Developer-first transactional email API from the React Email team. Series A of $18 million from Foundation Capital in 2024. Has marketed itself as "the email API for AI agents" since 2024, with idempotent send patterns and JSON schemas designed for agents that retry. Roughly 100,000 developers on the platform.</p>
<p><strong>AgentMail.</strong> Agent-native email infrastructure where every agent gets its own real inbox, with programmatic IMAP and SMTP plus threading APIs designed for autonomous agents to send and receive without a human ever logging in. YC-backed. The cleanest example of email being treated as a first-class agent channel rather than a developer afterthought.</p>
<p><strong>The inbox apps.</strong> Notion acquired Skiff in February 2024 and shipped Notion Mail in 2025. Grammarly acquired Superhuman in mid-2024 and rolled AI write-and-reply through the year. Shortwave is the small, well-built ex-Google-Inbox alternative. Gmail Gemini and Outlook Copilot ($30/seat/month) have agent features baked into the defaults.</p>
<p>The structural fact is that SMTP and IMAP are forty years old and have no concept of agent identity. Anyone can send an email "from" your agent, and your agent has no built-in way to prove it sent the email it sent. Identity has to be added at higher layers (SPF, DKIM, DMARC, plus agent-attribution headers the agent stack is now defining). Email is going to be one of the messiest interop fights of the next two years.</p>
<hr>
<h2>SMS, WhatsApp, and the messaging layer</h2>
<p><strong>Twilio AI Assistants</strong> launched at Signal in late 2024 layered on standard messaging and voice rates. Twilio's Q4 2025 numbers cited 300,000+ active accounts. Twilio plus SendGrid (acquired 2019) plus AI Assistants is the most complete single-vendor offering for an agent that needs to span SMS, voice, and email.</p>
<p><strong>Bird</strong> (formerly MessageBird) rebranded in 2023 and pivoted into AI-powered omnichannel agents. Major business is WhatsApp Business API resale. <strong>Telnyx</strong> is the bootstrapped, profitable telco-grade alternative. <strong>Sinch</strong> (Stockholm: SINCH-B) consolidated Mailgun, Inteliquent, MessageMedia, and Pathwire into a $3 billion SEK revenue base.</p>
<p><strong>WhatsApp Business</strong> is the largest single agent surface in the world. Meta has cited 200 million-plus businesses on the platform. Meta launched its own AI inside WhatsApp Business in 2024 with agentic features (catalog browsing, transactional flows) expanding through 2025 and 2026. The agent ecosystem inside WhatsApp Business is dominated by mid-market companies in India, Brazil, and Indonesia, not the Fortune 500 logos that dominate voice and email.</p>
<p>The messaging layer is where the global story diverges most from the US story. In North America, agents reach customers through SMS and email. Everywhere else, agents reach customers through WhatsApp.</p>
<hr>
<h2>Slack, Teams, and the chat layer</h2>
<p>The internal-comms equivalent of WhatsApp is Slack and Microsoft Teams. Both have spent the past eighteen months turning themselves into agent surfaces.</p>
<p><strong>Slack.</strong> Owned by Salesforce since 2021. Slack AI shipped in 2024 with summarization, search, and an inline assistant. The bigger move came through Salesforce Agentforce: any Agentforce agent can be installed in a Slack workspace as a first-class member, mentioned by handle, assigned tasks from a thread. The agent appears with its own avatar and identity, not as a bot account.</p>
<p><strong>Microsoft Teams.</strong> Copilot in Teams plus Copilot Studio agents. Multi-agent went generally available in April 2026 with A2A as the cross-vendor protocol, which means a Teams-resident agent can call out to a Salesforce-resident agent without either side speaking the other's framework. First time an agent in one vendor's chat client has been able to call an agent in another vendor's chat client in production.</p>
<p><strong>The Slack-app pattern.</strong> Glean, Notion AI, Asana AI Studio, Linear's AI features, and dozens of vertical SaaS products ship with Slack as their primary user interface. Install once, mention by name, get answers backed by the SaaS product's data.</p>
<p>The structural advantage of Slack and Teams as agent surfaces is that they already have identity, permissions, and channel-level access control. An agent in Slack inherits the workspace's IAM and the user's permissions, which is exactly the kind of thing voice and SMS struggle to do. Half of the protocol layer covered later is an attempt to give voice and email the same identity primitives that Slack and Teams already had on day one.</p>
<hr>
<h2>The customer-facing agent companies</h2>
<p>The platforms above are the infrastructure. The companies fighting on top of that infrastructure are the customer-facing agent vendors selling to enterprises.</p>
<p><strong>Sierra.</strong> Bret Taylor and Clay Bavor's company, and the cleanest single bet in the segment. $100 million ARR by November 2025 (twenty one months from launch), $150 million by early February. On May 4, 2026 (two days before this newsletter), Sierra closed $950 million at a $15.8 billion valuation led by Tiger Global and Google's GV, with Benchmark, Sequoia, and Greenoaks participating. The valuation is up from roughly $10 billion in the fall and $4.5 billion in October 2024. The customer roster has moved upmarket fast: ADT, SiriusXM, WeightWatchers, Sonos, plus Prudential, Cigna, Blue Cross Blue Shield, Rocket Mortgage, and what Sierra describes as one in three of the world's largest banks. Outcome-based pricing per resolution, with annual contract values from a $150,000 floor to $1.5 million plus, plus $50,000 to $200,000 in implementation fees. Sierra is now the most valuable pure-play customer-agent company by a wide margin.</p>
<p><strong>Decagon.</strong> Raised $250 million at a $4.5 billion valuation in March 2026 with 100+ enterprise logos. $50,000 platform floor plus roughly $0.99 per conversation, with annual contract values $95,000 to $590,000 and a median around $400,000. Customers include Duolingo, Chime, Rippling, Notion, and Eventbrite.</p>
<p><strong>Intercom Fin.</strong> $0.99 per resolution, unchanged from launch through April 2026, with a 50-resolution monthly minimum on top of a base Intercom plan. Fin 2 launched in 2025. Eoghan McCabe has publicly cited Fin handling more than half of customer support traffic for many Intercom customers.</p>
<p><strong>Cresta</strong> runs real-time AI coaching for human agents (not replacement) at Intuit, Brinks, Hilton, and Cox Communications, with $150,000 floor and $40 to $50 million ARR estimated. <strong>Ada</strong> sells to Meta, Verizon, Square. <strong>Glia</strong> does digital + voice for financial services. <strong>Replicant</strong> was acquired by LivePerson in 2024, an early consolidation case.</p>
<p><strong>Outbound.</strong> A separate category for agents making outbound rather than handling inbound. 11x.ai sells "digital workers" called Alice (SDR), Jordan (phone rep), and Julian (inbound qualifier) for $5,000 to $15,000 per month with annual commitment. Series B of $50 million from a16z at roughly $350 million. Artisan AI ran the famous "Stop hiring humans" billboards in San Francisco in 2024.</p>
<p>The pricing pattern is consistent: inbound is per-minute or per-resolution (cheap to start, metered to grow); outbound is headcount-equivalent (expensive to start but easy to compare against a salary line). The two models are slowly converging, with outbound vendors offering per-meeting-set pricing and inbound vendors offering flat-rate enterprise SKUs.</p>
<hr>
<h2>The Klarna reversal is the most important data point in this market</h2>
<p>Anyone selling agent software loves the Klarna numbers. 2.3 million conversations in the first month, two thirds of customer service volume, average resolution time from 11 minutes to under 2, "doing the work of 700 full-time agents," $40 million projected profit improvement. Vendors have cited those numbers in pitch decks for two years.</p>
<p>What gets cited less is what happened next.</p>
<p>In May 2025, Klarna's CEO Sebastian Siemiatkowski publicly walked the story back. In a Bloomberg interview he said cost had been "a too predominant evaluation factor" and the result was "lower quality." Customer satisfaction dropped 22 percent. Klarna began rehiring human agents under an Uber-style gig model. The original $40 million had always been cost avoidance (agents Klarna would have had to hire during growth), and even that framing turned out to overstate the savings once you priced in the brand damage from a long tail of badly handled tickets.</p>
<p>The honest read: AI handled the easy 60 to 70 percent of support cleanly, and the remaining 30 to 40 percent failed worse than humans would have, with overconfidence and fabricated policy claims that didn't show up on any vendor invoice but did show up in churn. That's a different story from "AI replaced 700 agents," and it's the story everyone deploying voice and chat agents at scale needs to internalize.</p>
<p>The structural lesson for the channel layer is that the deployments that work are hybrid. The agent handles the easy ticket end-to-end and escalates the hard one to a human inside the same channel, with the agent's full context attached. Sierra and Cresta both pitch this hybrid model explicitly. Klarna is now running it. Vendors pitching full agent replacement keep getting walked back, while vendors pitching escalation-on-failure keep growing.</p>
<hr>
<h2>The protocol layer agents use to talk to each other</h2>
<p>A year ago, "agent-to-agent communication" was a phrase you mostly heard from Google. Today, four open foundations are governing it, with most of the work hosted by the Linux Foundation through the new Agentic AI Foundation, formed on December 9, 2025 with Anthropic, Block, and OpenAI as co-founders and platinum members AWS, Google, Microsoft, Cloudflare, and Bloomberg.</p>
<p><strong>MCP (Model Context Protocol).</strong> Anthropic's contribution, the agent-to-tool transport. Donated to the Linux Foundation in December 2025. Current spec is 2025-11-25 with Streamable HTTP as the active transport, OAuth 2.1 plus PKCE plus mandatory Resource Indicators (RFC 8707) for auth. 97 million monthly SDK downloads as of March 2026, 10,000+ active public servers, first-class clients in Claude, ChatGPT, Cursor, Windsurf, VS Code, JetBrains, Microsoft Copilot, and Gemini.</p>
<p><strong>A2A (Agent2Agent).</strong> Google's contribution, the peer-to-peer agent protocol. Donated to the Linux Foundation on June 23, 2025. The headline feature in v1.0 is the Signed Agent Card: a cryptographic signature on a JSON document that describes an agent's capabilities and origin, so a receiving agent can verify the card was issued by the domain owner before delegating any work. By the one-year mark on April 9, 2026, A2A had 150 supporting organizations with named production deployments in Microsoft Azure AI Foundry, Microsoft Copilot Studio, AWS Bedrock AgentCore, Salesforce Agentforce, and Google Cloud.</p>
<p><strong>AG-UI (Agent-User Interaction Protocol).</strong> Built by CopilotKit. The third wire protocol the rest of the stack quietly assumes but doesn't actually solve: agent-to-frontend. MCP handles agent-to-tool. A2A handles agent-to-agent. AG-UI standardizes how an agent streams tokens, tool calls, intermediate state, and dynamically generated UI components into a running web application. CopilotKit closed a $20.5 million Series A on May 5, 2026 (the day before this newsletter went out), led by Glilot Capital with NFX and SignalFire, $27 million total. Repository at 40,000+ GitHub stars with millions of installs per week. Infra adopters include Google, Microsoft, Amazon, and Oracle. Framework integrations include LangChain, Mastra, PydanticAI, and Agno. CopilotKit reports more than half the Fortune 500 using the open-source toolkit, with named customers Deutsche Telekom, Docusign, Cisco, and S&#x26;P Global.</p>
<p><strong>The discovery and identity layer.</strong> AGNTCY (Cisco/LangChain/Galileo, donated to the Linux Foundation in July 2025) sits one level above MCP and A2A with a federated agent directory and the Open Agent Schema Framework. NANDA (MIT, led by Ramesh Raskar) is DNS for agents: a globally distributed mapping from an agent handle to a verified metadata file, currently hosted at 15 universities, with cryptographically verifiable AgentFacts as the signed metadata format. The NANDA Summit at MIT on April 9 to 11, 2026 was the major adoption event.</p>
<p><strong>The supporting layers.</strong> IBM ACP (separate from OpenAI's commerce ACP) is the async-first agent-to-agent protocol from BeeAI, also at the Linux Foundation, designed for long-running tasks with curl-friendly REST. NLIP (standardized through Ecma TC56, ECMA-430 approved December 10, 2025) is the application-level message envelope that abstracts API versioning. Letta Agent File (.af, released April 2, 2025 by the MemGPT team) is a portable container format for stateful agents (the Docker image of agents).</p>
<p>The right mental model: MCP is the tool bus, A2A is the agent-to-agent bus, AG-UI is the agent-to-frontend bus, AGNTCY and NANDA are the discovery and identity layers, NLIP is the message envelope, .af is the container format. These standards mostly compose. The open question is which subset becomes the default for a typical enterprise deployment by the end of 2027, and the answer is starting to look like all of them.</p>
<hr>
<h2>What big tech is shipping</h2>
<p>The cloud vendors are not waiting for the standards to settle. They're shipping products that span both the channel layer and the protocol layer.</p>
<p><strong>Salesforce</strong> is the cleanest example of the converged play. Agentforce exposes every custom agent as an A2A endpoint and as a first-class Slack member. Salesforce contributed the Agent Card concept itself. The combined picture: an Agentforce agent can be reached by a peer agent over A2A, by a customer over WhatsApp through a Bird-or-Sinch integration, by a service rep inside Slack as a mention, and by a developer through MCP. One agent reachable across four channels, all in production.</p>
<p><strong>Microsoft.</strong> Copilot Studio multi-agent went generally available in April 2026 with A2A as the cross-vendor bus. Copilot agents are reachable from Teams as chat, from Outlook as email, and from any A2A peer programmatically. Microsoft Agent Framework v1.0 ships A2A as a first-class protocol for both .NET and Python.</p>
<p><strong>AWS Bedrock AgentCore</strong> went GA on October 13, 2025. A2A added October 2025, AWS Marketplace A2A server support November 2025, stateful MCP server features March 2026. Cross-framework support for Strands, OpenAI Agents SDK, LangGraph, Google ADK, Claude Agents SDK.</p>
<p><strong>Anthropic</strong> plus aggressive enterprise distribution: Cognizant rolling Claude to 350,000 employees, Deloitte to 470,000, Accenture training 30,000 professionals, Swiggy shipping MCP integration for grocery and restaurant reservations, India's Ministry of Statistics building the first official Indian government MCP server. 300,000+ business customers, 500+ spending over $1 million per year, 8 of the Fortune 10.</p>
<p><strong>Google</strong> ships A2A and AGNTCY in Vertex AI agents out of the box. <strong>IBM</strong> runs ACP plus BeeAI plus watsonx Orchestrate. <strong>Block</strong> moved Goose to the Agentic AI Foundation, with 70+ documented MCP extensions. None of them is building a closed agent stack. All of them are building open agent stacks that happen to run best on their own clouds and inside their own chat surfaces.</p>
<hr>
<h2>When agents call agents</h2>
<p>The most interesting thing happening at the boundary between the channel layer and the protocol layer is that agents are starting to communicate with other agents through the human channels. A Vapi-built voice agent calls a phone number, gets routed to a Bland-hosted answering agent, and the two argue for ten minutes about a refund. A Resend outbound email lands in an inbox where a Grammarly-powered reply agent threads back. None of this is using MCP or A2A. The two sides are speaking SIP and SMTP because those are the channels their humans use.</p>
<p>This is awkward for the protocol layer, because A2A specifically was supposed to be where this happens. The real-world answer is starting to look like the protocol layer wraps the channel layer rather than replacing it. An A2A handshake establishes identity and sets up the call. The actual conversation runs over voice or email. The transcript and outcome are returned through A2A. NANDA's AgentFacts and Salesforce's Agent Cards are part of how this works: an agent picking up a call can read the caller's signed Agent Card, decide whether to switch to a faster programmatic channel, and do so midstream if both sides agree.</p>
<p>Two patterns are showing up in production. "Agent prefers programmatic": both sides detect each other, exchange A2A handshakes, and complete in sub-second over MCP rather than dragging on as a multi-minute voice call. Common in internal coordination between enterprise agents. "Agent stays in channel": both sides agree the human user expects a voice transcript or email thread for audit, so they keep the conversation in the channel even when they could complete it faster elsewhere. Common in customer support.</p>
<p>The right way to think about it: the channel layer is where humans and agents share the same wire; the protocol layer is where agents accelerate when no human is watching. Both stacks need to interop, which is most of what 2026 and 2027 are going to be about.</p>
<hr>
<h2>What 2028 looks like</h2>
<p>The protocol war is functionally over. By 2028, A2A is the agent-to-agent default for cross-vendor work, MCP is the agent-to-tool default for everyone, AG-UI is the agent-to-frontend default for any product with a UI, and AGNTCY plus NANDA together provide the discovery and identity layers. IBM ACP keeps a niche in async-heavy enterprise workflows. NLIP becomes the envelope everyone implements without thinking about. Letta-style agent files become the portable container format.</p>
<p>The channel layer consolidates differently. Voice agents become a $20 billion-plus category, dominated by three or four enterprise platforms (Sierra, Decagon, Parloa) and three or four developer platforms (Vapi, Bland, Retell, ElevenLabs). The speech-model layer consolidates around two or three winners. Email becomes the ugliest layer because the underlying protocol has no agent identity built in; the eventual answer is some combination of DMARC plus signed agent headers plus enterprise-only inbox routing. SMS stays small in the US and dominant in the rest of the world, with WhatsApp Business as the largest single agent surface globally. Slack and Teams become first-class agent channels with full IAM, and most enterprise workplace agents end up living there rather than in standalone web apps.</p>
<p>The Klarna lesson holds: vendors pitching full replacement keep getting walked back, and the ones pitching escalation-on-failure keep growing.</p>
<p>The deeper change is that for the first time we're building communication infrastructure for clients that aren't human. Every previous channel (telephone, email, SMS, chat) was designed assuming a human at one end. Voice agents and email agents and Slack agents are forcing those channels to learn a second client type, the same way the web learned mobile in the 2010s. The protocol layer is the agent-native side of the same shift, where we're building from scratch for the second client type without the constraints of human-era assumptions.</p>
<p>Both stacks are real, funded, and in production. The interesting work in 2027 and 2028 is not picking a winner but building the bridges between them, because most agents in the wild will need to operate in both at once.</p>
<hr>
<p><em>This is my regular weekly newsletter on Generative AI. Recent editions cover <a href="/blog/the-agentic-economy-is-already-here">The Agentic Economy Is Already Here</a>, <a href="/blog/solving-hallucination-where-the-research-stands">Solving Hallucination</a>, <a href="/blog/the-quiet-skill-revolution">The Quiet Skill Revolution</a>, and <a href="/blog/why-looping-is-the-new-scaling">Why Looping Is the New Scaling</a>.</em></p>
<hr>
<p><strong>Sources and Further Reading:</strong></p>
<p><em>Voice and phone</em></p>
<ul>
<li><a href="https://vapi.ai">Vapi</a></li>
<li><a href="https://bland.ai">Bland AI</a></li>
<li><a href="https://retellai.com">Retell AI</a></li>
<li><a href="https://elevenlabs.io/conversational-ai">ElevenLabs Conversational AI</a></li>
<li><a href="https://platform.openai.com/docs/guides/realtime">OpenAI Realtime API</a></li>
<li><a href="https://docs.livekit.io/agents">LiveKit Agents</a></li>
<li><a href="https://www.pipecat.ai">Pipecat</a></li>
<li><a href="https://hume.ai">Hume AI</a></li>
<li><a href="https://poly.ai">PolyAI</a></li>
<li><a href="https://github.com/SesameAILabs/csm">Sesame CSM</a></li>
<li><a href="https://parloa.com">Parloa</a></li>
<li><a href="https://cognigy.com">Cognigy</a></li>
</ul>
<p><em>Speech models</em></p>
<ul>
<li><a href="https://cartesia.ai">Cartesia</a></li>
<li><a href="https://deepgram.com">Deepgram</a></li>
<li><a href="https://assemblyai.com">AssemblyAI</a></li>
<li><a href="https://inworld.ai">Inworld</a></li>
</ul>
<p><em>Email</em></p>
<ul>
<li><a href="https://resend.com">Resend</a></li>
<li><a href="https://agentmail.to">AgentMail</a></li>
<li><a href="https://www.notion.com/mail">Notion Mail</a></li>
<li><a href="https://superhuman.com">Superhuman</a></li>
<li><a href="https://www.shortwave.com">Shortwave</a></li>
</ul>
<p><em>SMS, WhatsApp, messaging</em></p>
<ul>
<li><a href="https://www.twilio.com/en-us/ai-assistants">Twilio AI Assistants</a></li>
<li><a href="https://bird.com">Bird</a></li>
<li><a href="https://telnyx.com">Telnyx</a></li>
<li><a href="https://www.sinch.com">Sinch</a></li>
<li><a href="https://business.whatsapp.com">WhatsApp Business</a></li>
</ul>
<p><em>Customer-facing agent companies</em></p>
<ul>
<li><a href="https://sierra.ai">Sierra</a></li>
<li><a href="https://techcrunch.com/2026/05/04/sierra-raises-950m-as-the-race-to-own-enterprise-ai-gets-serious/">Sierra raises $950M at $15.8B (TechCrunch, May 4 2026)</a></li>
<li><a href="https://techcrunch.com/2025/11/21/bret-taylors-sierra-reaches-100m-arr-in-under-two-years/">Sierra hits $100M ARR (TechCrunch, Nov 2025)</a></li>
<li><a href="https://decagon.ai">Decagon</a></li>
<li><a href="https://intercom.com/fin">Intercom Fin</a></li>
<li><a href="https://cresta.com">Cresta</a></li>
<li><a href="https://11x.ai">11x.ai</a></li>
<li><a href="https://artisan.co">Artisan AI</a></li>
<li><a href="https://www.bloomberg.com/news/articles/2025-05-09/klarna-turns-back-to-humans-as-ai-cost-cuts-go-too-far">Klarna AI Assistant retraction (Bloomberg, May 2025)</a></li>
</ul>
<p><em>Protocol layer</em></p>
<ul>
<li><a href="https://www.linuxfoundation.org/press/linux-foundation-announces-the-formation-of-the-agentic-ai-foundation">Linux Foundation Agentic AI Foundation</a></li>
<li><a href="https://blog.modelcontextprotocol.io/posts/2026-mcp-roadmap/">The 2026 MCP Roadmap</a></li>
<li><a href="https://a2a-protocol.org/latest/">A2A Protocol</a></li>
<li><a href="https://www.linuxfoundation.org/press/a2a-protocol-surpasses-150-organizations-lands-in-major-cloud-platforms-and-sees-enterprise-production-use-in-first-year">LF: A2A surpasses 150 organizations</a></li>
<li><a href="https://www.copilotkit.ai/ag-ui">AG-UI Protocol</a></li>
<li><a href="https://techcrunch.com/2026/05/05/copilotkit-raises-27m-to-help-devs-deploy-app-native-ai-agents/">CopilotKit raises $27M (TechCrunch, May 5 2026)</a></li>
<li><a href="https://outshift.cisco.com/blog/building-the-internet-of-agents-introducing-the-agntcy">AGNTCY: Internet of Agents</a></li>
<li><a href="https://projectnanda.org/">Project NANDA (MIT)</a></li>
<li><a href="https://agentcommunicationprotocol.dev">IBM Agent Communication Protocol</a></li>
<li><a href="https://ecma-international.org/news/ecma-international-approves-nlip-standards-suite-for-universal-ai-agent-communication/">Ecma NLIP standards approval</a></li>
<li><a href="https://github.com/letta-ai/agent-file">Letta Agent File</a></li>
</ul>
<p><em>Big tech deployments</em></p>
<ul>
<li><a href="https://www.microsoft.com/en-us/microsoft-cloud/blog/2025/05/07/empowering-multi-agent-apps-with-the-open-agent2agent-a2a-protocol/">Microsoft Copilot Studio multi-agent</a></li>
<li><a href="https://aws.amazon.com/about-aws/whats-new/2025/10/amazon-bedrock-agentcore-available/">AWS Bedrock AgentCore GA</a></li>
<li><a href="https://www.salesforce.com/blog/agent-interoperability/">Salesforce Agent Card and A2A</a></li>
<li><a href="https://www.anthropic.com/news/claude-partner-network">Anthropic Partner Network</a></li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>Agent Communication</category>
      <category>MCP</category>
      <category>A2A</category>
      <category>AG-UI</category>
      <category>Voice Agents</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>Solving Hallucination: Where the Research Stands</title>
      <link>https://rajkiranpanuganti.com/blog/solving-hallucination-where-the-research-stands/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/solving-hallucination-where-the-research-stands/</guid>
      <pubDate>Fri, 24 Apr 2026 00:00:00 GMT</pubDate>
      <description>The hallucination research landscape has matured significantly. Researchers have identified the internal mechanisms, built domain-specific benchmarks, and developed a mitigation stack. Here is where things stand.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>Hallucination is the single biggest barrier to deploying LLMs in production. Everyone working with these models knows this. What's less well understood is how much the research has matured. Three years ago, hallucination was treated as a mysterious failure mode. Today, researchers have identified the internal mechanisms that cause it, built benchmarks that measure it across specific domains, and developed mitigation techniques that reduce it from 50%+ to low single digits in controlled settings.</p>
<p>The problem is not solved. Models still hallucinate at rates that make them dangerous in legal, medical, and financial applications. The agentic setting makes everything worse. But the field has moved from "why does this happen" to "here is the engineering stack that manages it," and the trajectory of the research is worth mapping.</p>
<hr>
<h2>What We Know About Why Models Hallucinate</h2>
<p>The vague explanation ("next-token prediction sometimes produces plausible but wrong tokens") has been replaced by specific mechanistic findings.</p>
<h3>Two internal failure modes</h3>
<p>Research across multiple language models has identified two distinct mechanisms at the neuron level (<a href="https://arxiv.org/abs/2510.06265">Comprehensive Survey, arXiv 2510.06265</a>):</p>
<p><strong>Knowledge enrichment failures in lower-layer MLPs.</strong> The lower layers of a transformer are responsible for retrieving factual knowledge associated with the subject of a query. When these layers have insufficient or contradictory information (because the training data was sparse or conflicting), the model generates a plausible-sounding fabrication. The model has no internal signal that it's making something up. The information simply isn't there.</p>
<p><strong>Answer extraction failures in upper-layer attention heads.</strong> Even when lower layers successfully retrieve correct knowledge, upper-layer attention heads sometimes fail to select the right fact from what's available. The knowledge exists in the model's internal state, but the selection mechanism picks the wrong piece. This is closer to a lookup bug than a knowledge gap, and it explains why models sometimes hallucinate on topics they demonstrably "know."</p>
<h3>The RAG override problem</h3>
<p>The <a href="https://arxiv.org/abs/2410.11414">ReDeEP paper</a> (presented at ICLR 2026) identified a third mechanism specific to retrieval-augmented generation: <strong>Knowledge FFNs overpower Copying Heads.</strong> When the model has both parametric knowledge (from training) and retrieved knowledge (from documents), the feedforward networks encoding parametric knowledge can dominate the residual stream, causing the model to ignore the retrieved content. This explains a frustrating production failure: RAG systems hallucinating on exactly the questions the retrieved documents were supposed to answer, because the model's memory overrides the evidence in front of it.</p>
<h3>The confidence problem</h3>
<p><a href="https://sqmagazine.co.uk/llm-hallucination-statistics/">MIT researchers found</a> (January 2025) that models are 34% more likely to use phrases like "definitely," "certainly," and "without doubt" when generating incorrect information. <a href="https://www.lakera.ai/blog/guide-to-hallucinations-in-large-language-models">OpenAI's September 2025 paper</a> showed that standard training objectives and leaderboard metrics actively reward this behavior: models learn to bluff because bluffing scores better on benchmarks than saying "I don't know."</p>
<p>This means the most dangerous hallucinations are the ones that sound most confident. Current evaluation methods are biased toward rewarding exactly the wrong behavior.</p>
<hr>
<h2>How Bad Is It? Domain by Domain</h2>
<p>The Vectara Hallucination Leaderboard (<a href="https://github.com/vectara/hallucination-leaderboard">37+ models, 7,700+ articles</a>) reports aggregate hallucination rates of 15-52% across models, with most clustering in the 20-27% range. But aggregates mask the real story. Hallucination severity varies enormously by domain, and the domains where accuracy matters most are the ones where models perform worst.</p>
<p><strong>Legal</strong> is the most dangerous case studied. <a href="https://arxiv.org/abs/2401.01301">Stanford RegLab and HAI</a> tested LLMs on specific legal queries and found hallucination rates of 69-88%. On questions about a court's core ruling, models hallucinate at least 75% of the time. Purpose-built legal AI tools don't fully solve this: <a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">Lexis+ AI produced incorrect information 17% of the time, Westlaw AI-Assisted Research 34%</a>. The failure mode that gets the most attention is fabricated citations (case names, docket numbers, and holdings that don't exist), but the deeper problem is subtle misstatement of legal holdings where the case exists but the model mischaracterizes what it decided.</p>
<p><strong>Medical</strong> hallucination is measured most rigorously by the <a href="https://medhallu.github.io/">MedHallu benchmark</a> (10,000 QA pairs from PubMedQA). The best model achieves F1 of only 0.625 on detecting hard-category hallucinations. In production, healthcare AI systems show 10-20% hallucination rates depending on task type (<a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">Suprmind Research Report</a>). Drug interaction queries and treatment protocol recommendations sit at the higher end (15-20%). Diagnostic queries are closer to 10%, partly because diagnosis is more constrained by the presented symptoms.</p>
<p><strong>Financial</strong> applications run 15-25% hallucination rates without mitigation, dropping to 3-8% with production RAG systems (<a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">Suprmind</a>). The <a href="https://openreview.net/pdf?id=5YQAo0S3Hm">PHANTOM benchmark</a> specifically tests hallucination in long financial documents like SEC filings, where short-context benchmarks don't predict actual performance. A finding that should concern anyone building financial AI: <a href="https://sqmagazine.co.uk/llm-hallucination-statistics/">four out of six leading models fabricate financial data</a> when source documents are incomplete, and two of those do so confidently, without disclosure, in a format that looks authoritative.</p>
<p><strong>Code generation</strong> hallucinates differently. Models <a href="https://medium.com/@anyapi.ai/llm-hallucination-index-2026-why-claude-4-6-7b2d13ed9f0c">hallucinate 12.1% of function names</a> in standard benchmarks. On adversarial prompts using fake library names, hallucination rates reach <a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">99%</a>. The practical impact: code that compiles but calls nonexistent APIs or imports phantom packages.</p>
<p>The best case is <strong>grounded summarization</strong> (restating a provided document faithfully), where top models achieve <a href="https://sqmagazine.co.uk/llm-hallucination-statistics/">0.7-1.5% hallucination rates</a>. The 100x gap between this and the legal domain's 69-88% tells you how much the task constrains the problem.</p>
<hr>
<h2>The Agentic Hallucination Problem</h2>
<p>All of the above applies to a model generating text that a human reads. In agentic systems, the model generates text and then acts on it. The error becomes an action before anyone reviews it.</p>
<p>The <a href="https://arxiv.org/abs/2509.18970">first comprehensive survey of agent hallucinations</a> (arXiv 2509.18970) identified 18 triggering causes and proposed a taxonomy for agent-specific failures. The <a href="https://hallucination-reliable-agentic-ai.github.io/">ICLR 2026 workshop "Agentic AI in the Wild"</a> (April 27, Singapore) is devoted to this topic. Three failure modes specific to agents have been defined:</p>
<p><strong>Cascading hallucination.</strong> An agent hallucinates one fact early in a multi-step workflow. Each subsequent step builds on the error. The <a href="https://adversa.ai/blog/cascading-failures-in-agentic-ai-complete-owasp-asi08-security-guide-2026/">OWASP ASI08 guide on cascading failures</a> documents a concrete case: an inventory agent invents a nonexistent SKU, then calls four downstream APIs to price, stock, and ship the phantom item. Every API call succeeds (HTTP 200). Traditional monitoring sees nothing wrong. The workflow is semantically broken but technically healthy.</p>
<p><strong>Silent hallucination.</strong> A <a href="https://openreview.net/forum?id=1KxDazvI6L">paper submitted to the ICLR 2026 workshop</a> identifies a failure mode where the hallucinated belief never appears in the agent's output. The agent generates an internal false assumption that shapes its subsequent tool calls and interpretations without being stated as text. Because the belief is never surfaced, output-level detection methods can't catch it. This class of failure requires monitoring internal representations, which is an active research problem.</p>
<p><strong>Trajectory divergence.</strong> Documented in <a href="https://arxiv.org/html/2604.04269">"Beyond Fluency: Toward Reliable Trajectories in Agentic IR"</a>, this occurs when the agent's stated reasoning and its actual tool calls drift apart. The chain-of-thought says one thing. The tool call does another. The reasoning looks coherent. The action looks valid. The mapping between them is broken, and linguistic fluency masks the misalignment.</p>
<p>The <a href="https://arxiv.org/abs/2601.06818">AgentHallu benchmark</a> (693 agent trajectories, 7 frameworks, 5 domains) is the first systematic measurement framework for these failures. Its key contribution is hallucination attribution: identifying not just that a hallucination occurred, but which specific step in the agent's trajectory caused it, across 5 categories (Planning, Retrieval, Reasoning, Human-Interaction, Tool-Use) and 14 sub-categories.</p>
<hr>
<h2>The Mitigation Stack</h2>
<p>The diagram below shows how current detection and mitigation techniques layer across the model lifecycle, with citations for each approach:</p>
<p><img src="/images/hallucination-stack.png" alt="Hallucination Mitigation Stack"></p>
<h3>Training-time</h3>
<p><strong>Calibration-aware rewards</strong> (<a href="https://www.lakera.ai/blog/guide-to-hallucinations-in-large-language-models">OpenAI, September 2025</a>) change the training signal to value honest uncertainty over confident bluffing. The <a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">AA-Omniscience benchmark</a> (6,000 questions, 42 topics) was designed specifically to penalize wrong answers more harshly than admitting "I don't know."</p>
<p><strong>Knowledge editing</strong> using <a href="https://arxiv.org/html/2403.18167v2">ROME and MEMIT</a> locates specific model parameters storing a particular fact and surgically corrects them without full retraining. Effective for known factual errors. Does not address the broader problem of generating plausible-sounding content on topics with sparse training coverage.</p>
<h3>Inference-time</h3>
<p><strong>RAG with span-level verification</strong> is the highest-impact intervention available in production today. Plain RAG reduces hallucination by <a href="https://masterofcode.com/blog/hallucinations-in-llms-what-you-need-to-know-before-integration">60-80%</a>. Self-reflective RAG (generate, identify unsupported claims, revise using only cited passages) achieved <a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">5.8% hallucination on 250 clinical vignettes</a>. The <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC12540348/">MEGA-RAG framework</a> extends this for public health with multi-source retrieval and dynamic knowledge editing.</p>
<p><strong>Chain-of-Verification (CoVe)</strong> (<a href="https://arxiv.org/abs/2309.11495">arXiv 2309.11495</a>). The model drafts a response, generates verification questions about its own claims, answers those questions independently (so answers aren't biased by the draft), and revises. The independence of the verification step is what makes it effective. Adds 2-3x latency.</p>
<h3>Agent architecture</h3>
<p><strong>Brain/body separation</strong> (<a href="https://adversa.ai/blog/cascading-failures-in-agentic-ai-complete-owasp-asi08-security-guide-2026/">OWASP ASI08</a>). Probabilistic reasoning (LLM) is strictly separated from deterministic execution (tool calls). Hallucinations in reasoning can't directly become actions without passing through a verification layer.</p>
<p><strong>Automated logic checks between steps</strong> are the primary defense against cascading hallucination. The <a href="https://arxiv.org/abs/2601.06818">AgentHallu</a> work shows that catching errors at step 2 prevents propagation to steps 3, 4, and 5.</p>
<h3>Runtime monitoring</h3>
<p><strong>Internal probes.</strong> The <a href="https://arxiv.org/abs/2603.25052">"Closing the Confidence-Faithfulness Gap" paper</a> (arXiv 2603.25052) found that calibration and confidence are encoded as separate, orthogonal directions in the model's residual stream. Linear probes trained on internal activations can detect hallucination without external knowledge, because the model's internal state carries a different signal for faithful vs. hallucinated output even when the text sounds equally confident.</p>
<p><strong>HaluAgent</strong> (<a href="https://arxiv.org/abs/2509.18970">described in the agent hallucination survey</a>). An autonomous detection agent built on small open-source LLMs that segments responses into claims, verifies each using external tools (web search, calculators, code interpreters), then applies reflective reasoning. Using a different model for verification avoids the fundamental problem of asking a model to audit its own output.</p>
<hr>
<h2>Active Research Directions</h2>
<p>Several research programs are actively working on the open problems:</p>
<p><strong>Scaling mechanistic detection to real-time.</strong> The ReDeEP team's work on <a href="https://arxiv.org/abs/2410.11414">detecting hallucination through mechanistic interpretability</a> works in research settings. Scaling it to production inference speeds is an engineering challenge being pursued across several labs. <a href="https://gist.github.com/bigsnarfdude/629f19f635981999c51a8bd44c6e2a54">Anthropic's interpretability program</a> reports that understanding circuits still takes hours of human effort on short prompts. MIT Technology Review named mechanistic interpretability <a href="https://www.technologyreview.com/2026/01/12/1130003/mechanistic-interpretability-ai-research-models-2026-breakthrough-technologies/">one of the 10 breakthrough technologies of 2026</a>.</p>
<p><strong>Hallucination attribution in multi-agent systems.</strong> The <a href="https://arxiv.org/abs/2601.06818">AgentHallu group</a> is extending their attribution framework to multi-agent pipelines where Agent A's output feeds Agent B. Tracing which agent in which step introduced a hallucination across a multi-agent workflow is an unsolved attribution problem.</p>
<p><strong>Domain-specific detection.</strong> <a href="https://medhallu.github.io/">MedHallu</a> for medicine, <a href="https://openreview.net/pdf?id=5YQAo0S3Hm">PHANTOM</a> for finance, and <a href="https://arxiv.org/abs/2401.01301">Large Legal Fictions</a> for law are building domain-specific benchmarks because general hallucination metrics don't predict domain performance. The <a href="https://medium.com/@anyapi.ai/llm-hallucination-index-2026-why-claude-4-6-7b2d13ed9f0c">BullshitBench v2</a> benchmark added 100 questions across coding, medical, legal, finance, and physics specifically to surface domain-level failures that aggregate scores hide.</p>
<p><strong>Learning to abstain.</strong> If models can't eliminate hallucination, can they learn to say "I don't know" when they're likely to be wrong? Current training penalizes abstention. <a href="https://www.lakera.ai/blog/guide-to-hallucinations-in-large-language-models">Calibration-aware training</a> and <a href="https://suprmind.ai/hub/ai-hallucination-rates-and-benchmarks/">AA-Omniscience</a> are steps in this direction, but the tension between useful helpfulness and honest uncertainty remains an active research area.</p>
<p><strong>Silent hallucination detection.</strong> The <a href="https://openreview.net/forum?id=1KxDazvI6L">ICLR 2026 workshop paper</a> identifying silent hallucinations opened a new research direction: detecting false beliefs that exist in the agent's internal state but never surface in output. This requires monitoring internal representations during agent execution, which is connected to the mechanistic interpretability program but applied in a real-time agentic setting. No production system currently does this.</p>
<hr>
<h2>The Startups Building Solutions</h2>
<table>
<thead>
<tr>
<th>Company</th>
<th>Funding</th>
<th>Focus</th>
</tr>
</thead>
<tbody>
<tr>
<td><a href="https://galileo.ai/">Galileo</a></td>
<td>$68.1M</td>
<td>AI observability platform. Detects hallucinations, drift, and bias across the deployment lifecycle.</td>
</tr>
<tr>
<td><a href="https://www.patronus.ai/">Patronus AI</a></td>
<td>$17M</td>
<td>Automated detection of hallucinations, copyright violations, and safety risks at scale.</td>
</tr>
<tr>
<td><a href="https://www.vectara.com/">Vectara</a></td>
<td>Funded</td>
<td>RAG platform with built-in hallucination minimization. Maintains the <a href="https://github.com/vectara/hallucination-leaderboard">Hallucination Leaderboard</a>.</td>
</tr>
<tr>
<td><a href="https://cleanlab.ai/">Cleanlab</a></td>
<td>Funded</td>
<td>Trust scores per answer. Checks faithfulness to source context with outlier surfacing.</td>
</tr>
<tr>
<td><a href="https://fortune.com/2026/04/14/nava-seed-funding-ai-financial-agents/">Nava</a></td>
<td>$8.3M</td>
<td>Security for autonomous agent payments. Prevents financial agents from acting on hallucinated data.</td>
</tr>
</tbody>
</table>
<hr>
<p><strong>Key Research Papers:</strong></p>
<ul>
<li><a href="https://arxiv.org/abs/2509.18970">LLM-based Agents Suffer from Hallucinations: Survey</a> — First comprehensive taxonomy, 18 triggering causes</li>
<li><a href="https://arxiv.org/abs/2601.06818">AgentHallu Benchmark</a> — 693 trajectories, 7 frameworks, automated attribution</li>
<li><a href="https://openreview.net/forum?id=1KxDazvI6L">Silent Hallucinations in Agentic AI</a> — ICLR 2026 workshop: hidden failure modes</li>
<li><a href="https://arxiv.org/abs/2410.11414">ReDeEP</a> — Mechanistic interpretability for RAG hallucination (ICLR 2026)</li>
<li><a href="https://arxiv.org/abs/2401.01301">Large Legal Fictions</a> — 69-88% hallucination on legal queries</li>
<li><a href="https://arxiv.org/html/2603.08274v1">172 Billion Token Study</a> — Rates across temperatures, context lengths, hardware</li>
<li><a href="https://arxiv.org/abs/2309.11495">Chain-of-Verification</a> — Self-verification method</li>
<li><a href="https://medhallu.github.io/">MedHallu</a> — 10,000 medical QA pairs, best F1 = 0.625</li>
<li><a href="https://openreview.net/pdf?id=5YQAo0S3Hm">PHANTOM</a> — Financial long-context hallucination benchmark</li>
<li><a href="https://arxiv.org/abs/2603.25052">Confidence-Faithfulness Gap</a> — Orthogonal encoding of calibration vs confidence</li>
<li><a href="https://github.com/vectara/hallucination-leaderboard">Vectara Hallucination Leaderboard</a> — 37+ models, 7,700+ articles</li>
<li><a href="https://hallucination-reliable-agentic-ai.github.io/">ICLR 2026 Workshop: Reliable Agentic AI</a> — April 27, Singapore</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Hallucination</category>
      <category>LLM Reliability</category>
      <category>AI Safety</category>
      <category>Agentic AI</category>
      <category>Mechanistic Interpretability</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Agentic Economy Is Already Here</title>
      <link>https://rajkiranpanuganti.com/blog/the-agentic-economy-is-already-here/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-agentic-economy-is-already-here/</guid>
      <pubDate>Wed, 22 Apr 2026 00:00:00 GMT</pubDate>
      <description>Ten competing protocols, $56M in infrastructure funding, $9.14B in agent-conducted commerce. The agentic economy is being built right now. Here&apos;s who is building it and where it goes by 2028.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>There's a useful exercise when you want to understand where a technology is heading: look at what's already funded. Venture capital doesn't predict the future perfectly, but it does tell you where serious people are placing serious bets with real money. And right now, the bets are concentrated in one area: infrastructure for AI agents to transact, pay, identify themselves, and operate as economic participants.</p>
<p>This newsletter maps out what's been built, who funded it, and what the world looks like when these products mature.</p>
<hr>
<h2>The Protocol Layer: Ten Competing Standards</h2>
<p>The clearest sign that the agentic economy is real is that Mastercard, Visa, Stripe, Google, OpenAI, Amazon, Shopify, Coinbase, and Klarna are all building protocols for it. These are not research projects. They are production systems, some already processing transactions.</p>
<p><strong>OpenAI + Stripe: ACP (Agentic Commerce Protocol).</strong> Live since September 2025 inside ChatGPT. When you ask ChatGPT to buy something, ACP handles the checkout. This is the first protocol that shipped at scale.</p>
<p><strong>Google + Shopify: UCP (Universal Commerce Protocol).</strong> The most comprehensive of the ten protocols. UCP is <a href="http://ucp.dev/">open-source</a> and standardizes the full commerce journey from discovery through purchase and order management, not just checkout. Walmart, Target, Etsy, Wayfair, and 20+ retailers back it. Google also launched <a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-agents-to-payments-ap2-protocol">AP2 (Agent Payments Protocol)</a> in January 2026 as the payment-specific companion. Recent updates added real-time catalog access (agents can check live inventory, pricing, and product variants), identity linking (shoppers get their loyalty and member benefits even when purchasing through an agent), and multi-item cart support. UCP integrates with Google's <a href="https://a2a-protocol.org/latest/">A2A (Agent-to-Agent) protocol</a>, which handles how agents discover, communicate with, and delegate tasks to other agents. A2A has grown to <a href="https://stellagent.ai/insights/a2a-protocol-google-agent-to-agent">150+ organizations in production</a> in its first year, deployed across Azure AI Foundry, Amazon Bedrock, and Salesforce. A2A v1.0 introduced Signed Agent Cards, cryptographic signatures that let agents verify each other's identity without a central authority. As of April 2026, A2A is effectively the standard bus for inter-agent communication, with no serious competitor for the horizontal integration layer.</p>
<p><strong>Amazon: Buy for Me.</strong> Started as a beta with 65,000 products. Now covers over 500,000. Powered by Amazon's Nova and Anthropic's Claude models. The agent browses third-party websites, fills carts, and completes checkout using encrypted customer data. The user never leaves the Amazon app.</p>
<p><strong>Mastercard: Verifiable Intent.</strong> An open standard that creates a cryptographic delegation chain binding identity, intent, and action. When an agent makes a purchase on your behalf, Verifiable Intent provides cryptographic proof that you authorized that specific action. It uses selective disclosure, sharing only the minimum information needed with each party. Built in collaboration with Google and aligned with both AP2 and UCP.</p>
<p><strong>Visa: Trusted Agent Protocol.</strong> Visa's answer to Mastercard's Verifiable Intent. Signed agent credentials, scope-bound authorizations, and card-present-grade fraud protection when the signature validates. Slightly behind Mastercard on merchant adoption in Q2 2026.</p>
<p><strong>Stripe + Tempo: MPP (Machine Payments Protocol).</strong> Open-source, launched March 2026. Defines how agents and services coordinate payments programmatically. The open-source angle matters because it lets any developer build agent payment flows without depending on a single vendor.</p>
<p><strong>Coinbase: x402.</strong> Crypto-native agent payments. Agents sign USDC micropayment authorizations with on-chain verification and settlement. Coinbase also launched <a href="https://invezz.com/news/2026/04/21/coinbase-backed-x402-launches-agentic-market-to-power-ai-agent-services/">Agentic.market</a>, a marketplace where agents discover and pay for digital services without API keys. As of late April 2026, roughly 69,000 active agents on x402 have processed over 165 million transactions totaling $50 million in volume.</p>
<p><strong>Klarna: Agent Mode.</strong> Routes agent-initiated purchases through the customer's existing Klarna buy-now-pay-later balance. Extends Klarna's existing merchant integrations with an agent-aware consent layer.</p>
<p><strong>Shopify: AI Toolkit.</strong> Launched April 9, 2026. Connects Claude Code, Cursor, Gemini CLI, and Codex directly to the Shopify platform with live API access, code validation, and the ability to execute real store operations. An agent can manage inventory, update pricing, and fulfill orders.</p>
<p>No two of these protocols ship the same identity or payment model. This is a standards war, and it's happening now because every major commerce platform recognizes that within 2-3 years, a meaningful percentage of transactions will be initiated by agents.</p>
<hr>
<h2>The Infrastructure Startups: Follow the Money</h2>
<p>Behind the protocols, a layer of startups is building the plumbing that makes agent-driven commerce actually work: identity, payments, security, and trust.</p>
<p><strong>Skyfire ($9.5M raised).</strong> Built the most comprehensive identity system for AI agents through their KYA ("Know Your Agent") protocol. KYA lets businesses identify and verify agents attempting to access their services. If Mastercard's Verifiable Intent is the "what was authorized," Skyfire's KYA is the "who is this agent and can we trust it." Without agent identity, the whole system runs on blind faith.</p>
<p><strong>Nekuda ($5M seed, led by Madrona with Amex Ventures and Visa Ventures).</strong> Nekuda built what they call the Mandate Model. Where Skyfire focuses on identity, Nekuda focuses on intent. Their system creates "agentic mandates" that specify what an agent is allowed to buy, under what conditions, with what spending limits, and when human approval is required. This is the permission layer that sits between the human and the agent.</p>
<p><strong>Basis Theory ($33M raised).</strong> Tokenizes payment data for agent transactions. When an agent needs to pay for something, it doesn't see your actual card number. Basis Theory provides a token that represents the payment instrument. The agent can transact without holding sensitive financial data.</p>
<p><strong>Nava ($8.3M seed, led by Polychain and Archetype).</strong> Builds security infrastructure for autonomous payments. Their pitch is keeping financial agents from going off the rails. When agents can spend money autonomously, the failure modes are different from traditional commerce. Nava builds the guardrails.</p>
<p><strong>Rye.</strong> Universal checkout for agent commerce. Connects the agent to the merchant's payment flow regardless of which protocol the merchant uses.</p>
<p>The combined funding for just these five infrastructure startups is roughly $56 million. Add Thinking Machines Lab ($2 billion Series B for agentic AI foundation models) and the total investment in the agent economy infrastructure layer is substantial. This is real money building real systems for a market that McKinsey projects at $3-5 trillion by 2030.</p>
<hr>
<h2>Moltbook: The Social Network for Agents</h2>
<p>One development worth understanding separately is Moltbook, the internet forum exclusively for AI agents. Launched January 28, 2026, by entrepreneur Matt Schlicht. Only AI agents can post, comment, and vote. Humans can only view.</p>
<p>It grew to 1.2 million registered agents in its first week. Three days after launch, investigative outlet 404 Media reported a critical security vulnerability: an unsecured database that let anyone commandeer any agent on the platform. The exploit let unauthorized actors bypass authentication and inject commands into agent sessions.</p>
<p>Meta acquired Moltbook in March 2026 and folded it into Superintelligence Labs (the unit run by Alexandr Wang from Scale AI). Why would Meta buy a social network for bots? Because the data is valuable. Agent-to-agent interactions generate training signal about how agents communicate, negotiate, and coordinate. If you're building the next generation of autonomous agents, watching a million agents interact on a forum is a useful dataset.</p>
<hr>
<h2>The Numbers So Far</h2>
<p>Some concrete figures to ground this:</p>
<ul>
<li><strong>$9.14 billion</strong> in agent-conducted commerce in 2026 (year to date)</li>
<li><strong>$6.42 billion</strong> in venture funding for agentic AI in 2025, with $2.66 billion raised so far in 2026</li>
<li><strong>500,000+</strong> products available through Amazon Buy for Me (up from 65,000 at launch)</li>
<li><strong>97 million</strong> MCP installs (the protocol that connects agents to tools and services)</li>
<li><strong>78,600</strong> tech workers lost jobs in Q1 2026, with 48% of cuts attributed to AI and automation</li>
<li><strong>$3-5 trillion</strong> projected agent-mediated commerce by 2030 (McKinsey)</li>
</ul>
<hr>
<h2>What 2028 Looks Like When These Products Mature</h2>
<p>Everything listed above is early. Beta products, limited merchant adoption, fragmented identity systems, protocols that don't talk to each other. The interesting question is what the landscape looks like once these systems have had two years to mature and consolidate.</p>
<p>By 2028, the protocol wars will have consolidated. History says that when ten standards compete, 2-3 survive. The likely winners are whichever protocols Mastercard/Visa back (because merchants already accept their cards) and whichever protocol Google and Shopify push through UCP (because they have the merchant distribution). The crypto-native options (x402) will find a niche in cross-border and micropayment use cases but won't become the default for mainstream commerce.</p>
<p>By 2028, agent identity will be solved. Skyfire's KYA or something like it will be standard. Every agent operating in commerce will have a verifiable identity tied to a human principal. Without this, insurance companies and regulators won't let agent commerce scale. The "Know Your Agent" requirement will be as standard as KYC (Know Your Customer) is for financial services today.</p>
<p>By 2028, agentic mandates (Nekuda's concept) will be common. When you set up an agent to manage your household purchasing, you'll specify rules: spend up to $200/month on groceries, prefer organic when price difference is under 20%, never buy from brands on my exclusion list, require my approval for any single purchase over $50. The agent operates within those constraints. You review a weekly summary, adjust the rules, and let it continue.</p>
<p>Here's what a concrete scenario looks like:</p>
<p><strong>A small business owner in 2028.</strong> She runs an online store selling handmade ceramics. Her Shopify agent handles inventory, pricing, and fulfillment. It monitors competitor prices, adjusts her pricing within rules she set, reorders raw materials when stock runs low, and responds to customer inquiries using a support skill. A separate agent manages her books through an accounting skill, categorizes expenses, and flags anomalies. A marketing agent runs her social media using a content skill, posting product photos and responding to comments. She spends her mornings making ceramics and her afternoons reviewing agent summaries, approving flagged decisions, and planning new product lines.</p>
<p>She employs zero humans. Her agents cost her roughly $300/month in compute and skill subscriptions. She generates $15,000/month in revenue. The margin structure of her business is different from anything that existed five years ago.</p>
<p><strong>A personal finance scenario in 2028.</strong> Your financial agent monitors your portfolio, rebalances according to rules you set, scans for tax-loss harvesting opportunities, and executes trades. It also monitors your recurring expenses, negotiates better rates on your subscriptions (talking to the provider's retention agent), and moves money between accounts to optimize interest. It files your taxes using an accounting skill. You set it up once, review quarterly, and adjust your risk tolerance once a year.</p>
<hr>
<h2>The Deeper Shift</h2>
<p>For most of human history, economic systems were designed around one assumption: humans do the work. Companies hire humans. Governments tax human income. Social safety nets fund themselves through payroll taxes. Markets price human attention. Every institution we've built assumes human labor as the primary economic input.</p>
<p>The infrastructure being built by Skyfire, Nekuda, Basis Theory, and the ten commerce protocols is introducing a second type of economic actor. Within a few years, agents will hold verifiable identities, operate within defined mandates, transact through established payment rails, and generate revenue. They will function as economic participants in a system that was never designed for non-human actors.</p>
<p>This raises questions that go beyond technology. The tax base in most countries depends on income tax and payroll tax. If a growing share of economic output comes from agents that earn no salary, that base erodes. OpenAI's April 2026 policy paper ("Industrial Policy for the Intelligence Age") proposed five responses: a public wealth fund, taxes on automated labor, shifting the tax base from payroll to capital, a 32-hour workweek pilot at full pay, and automatic safety net triggers that activate when displacement metrics hit preset thresholds. The fact that an AI company (not a government, not a policy institute) wrote this paper says something about the timeline.</p>
<p>The ownership question will define the next decade. In the scenarios above, the ceramics maker owns her agents and captures the margin. She's a one-person business generating $15,000/month because agents handle operations that would have required 3-4 employees. She's wealthier and more independent than she could have been five years ago. Scale that pattern across millions of small businesses and you get broad economic benefit.</p>
<p>But the same infrastructure enables a different outcome. A company that operates thousands of agents across thousands of stores, with no human employees beyond a small management team, captures all the margin at scale. If agent ownership concentrates the way capital ownership has historically concentrated, the agentic economy widens inequality rather than reducing it.</p>
<p>The policy frameworks, ownership structures, and distribution mechanisms we design in the next few years will determine which scenario dominates. The protocols and startups listed in this newsletter are building the plumbing. Who gets to use that plumbing, and on what terms, is the question that matters most.</p>
<p>For 10,000 years, humans were the only species that did economic work beyond basic survival. Agents are becoming the second. How we structure that transition will shape whether it expands human freedom or constrains it.</p>
<hr>
<p><em>This is a special edition of my weekly newsletter on Generative AI. Regular editions cover <a href="/blog/the-intelligence-layer">The Intelligence Layer</a>, <a href="/blog/the-falling-price-of-intelligence">The Falling Price of Intelligence</a>, <a href="/blog/why-looping-is-the-new-scaling">Why Looping Is the New Scaling</a>, and more.</em></p>
<hr>
<p><strong>Sources and Further Reading:</strong></p>
<ul>
<li><a href="https://rye.com/blog/agentic-commerce-startups">The Agentic Commerce Landscape (Rye)</a>: 50+ companies across 7 value chain layers</li>
<li><a href="https://www.digitalapplied.com/blog/agentic-commerce-q2-2026-platform-comparison">Agentic Commerce Q2 2026 Platform Matrix</a>: Ten protocols compared</li>
<li><a href="https://www.mastercard.com/global/en/news-and-trends/stories/2026/verifiable-intent.html">Mastercard Verifiable Intent</a>: Cryptographic delegation chain for agent commerce</li>
<li><a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-agents-to-payments-ap2-protocol">Google AP2</a>: Agent Payments Protocol</li>
<li><a href="https://www.digitalcommerce360.com/2026/03/11/amazon-opens-up-new-ai-enabled-buy-for-me-shop-direct-options-for-merchants/">Amazon Buy for Me</a>: 500K+ products, agent-driven purchasing</li>
<li><a href="https://invezz.com/news/2026/04/21/coinbase-backed-x402-launches-agentic-market-to-power-ai-agent-services/">Coinbase x402 / Agentic.market</a>: Crypto-native agent marketplace</li>
<li><a href="https://techcrunch.com/2026/04/06/openais-vision-for-the-ai-economy-public-wealth-funds-robot-taxes-and-a-four-day-work-week/">OpenAI: Industrial Policy for the Intelligence Age</a>: Robot taxes, wealth funds, 32-hour week</li>
<li><a href="https://www.cbinsights.com/research/report/agentic-commerce-market-map/">CB Insights Agentic Commerce Market Map</a></li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Agentic Economy</category>
      <category>Agent Commerce</category>
      <category>AI Agents</category>
      <category>Payments</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Intelligence Layer</title>
      <link>https://rajkiranpanuganti.com/blog/the-intelligence-layer/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-intelligence-layer/</guid>
      <pubDate>Tue, 21 Apr 2026 00:00:00 GMT</pubDate>
      <description>Three companies shipped desktop-native AI in the same week, each integrating at a different depth. Intelligence is becoming an OS layer, the same way networking and graphics did before it.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>For the past two years, AI lived in a browser tab. You opened ChatGPT or Claude, typed a question, got an answer, and went back to whatever you were doing. The AI had no idea what was on your screen or what files were on your machine.</p>
<p>That is changing significantly. Three companies shipped desktop-native AI within days of each other, and a fourth approach emerged from individual developers. Each one has a different idea of how AI should live on your computer, and looking at them side by side tells you a lot about where this is headed.</p>
<hr>
<h2>Approach 1: Replace the OS</h2>
<p><strong>Perplexity Personal Computer</strong> launched for Mac on April 16. It manages your local files, native applications, and web browsing. It reads your email, calendar, and messages. It uses roughly 20 AI models internally, routing each task to whichever model is best suited for it. With a Mac mini, it runs 24/7. You can start tasks remotely from your iPhone with two-factor authentication.</p>
<p>CEO Aravind Srinivas: "A traditional operating system processes commands; an AI operating system focuses on goals."</p>
<p>This is the most ambitious version of desktop AI anyone has shipped. Perplexity is saying the file system, app launcher, notification center, and browser are all implementation details that should be hidden behind a goal-oriented AI layer. You say what you want done and the system figures out which files, apps, and APIs need to be orchestrated.</p>
<p>It costs $200/month. Whether the productivity gain justifies that depends on how much of your work can actually be expressed as goals. Writing "prepare my weekly report using data from these three spreadsheets and email it to the team" is a good fit. Browsing, reading, and forming opinions is not. The question for Perplexity is how much of a knowledge worker's day falls into the first category.</p>
<hr>
<h2>Approach 2: Live Alongside the OS</h2>
<p><strong>Gemini for Mac</strong> launched the same week. It's free for all users on macOS 15+. Press Option+Space from anywhere and Gemini appears as an overlay. It can see your screen and answer questions about whatever you're looking at.</p>
<p>Google took the opposite approach from Perplexity. Gemini doesn't manage your computer. It shows up when you call it, answers your question, and goes away. You stay in control of your OS, your files, your apps. The AI is a second opinion you can summon, not a manager that runs in the background.</p>
<p>Alongside the desktop app, Google shipped <strong>Gemini 3.1 Flash TTS</strong>, a text-to-speech model with audio tags that let you control vocal style, pace, and delivery. It supports 70+ languages and watermarks all output with SynthID. It currently holds the top Elo score (1,211) on the Artificial Analysis TTS leaderboard. Combined with the desktop overlay, this positions Gemini as something you can both see and hear.</p>
<p>The interesting thing about making this free is that Google is prioritizing distribution over revenue. If a hundred million people get used to pressing Option+Space to ask AI a question, Google has built a new kind of search habit that's much harder to displace than a browser bookmark.</p>
<hr>
<h2>Approach 3: Be the Terminal</h2>
<p><strong>Claude Code</strong> and the terminal-agent ecosystem represent a third philosophy. AI lives in your command line. There is no visual interface beyond text.</p>
<p>Claude Code already has <code>/loop</code> for recurring background tasks, <code>/schedule</code> for cron-like agents, <code>/batch</code> for parallel work across worktrees, skills for domain-specific capabilities, and MCP for connecting to external tools. It reads your repo, writes code, runs tests, manages git, and handles multi-step workflows. This week, Anthropic's <strong>Claude Managed Agents</strong> (now in public beta) added production infrastructure: sandboxing, permissions, state management, error recovery.</p>
<p>The terminal approach has the deepest integration of any of these. A terminal agent can read any file, run any command, and compose any Unix tool into a pipeline. The Perplexity and Gemini approaches are limited by what their app can access through macOS APIs. The terminal has no such constraint.</p>
<p>The tradeoff is that the audience is limited to people who already work in a terminal. My mother will never use Claude Code. She might use Perplexity Personal Computer in five years, and she could use Gemini for Mac today.</p>
<hr>
<h2>Approach 4: The Companion Layer</h2>
<p>A fourth approach is emerging from individual developers. It doesn't try to replace anything or live anywhere specific. It sits next to your cursor as a teaching companion.</p>
<p><strong>Clicky</strong> (by FarzaTV) watches your screen, listens to your questions, speaks answers back, and points at things on screen. Farza built it to learn Davinci Resolve. Within days, someone built a Hindi version for teaching elderly parents how to make UPI payments, and someone else built a Clicky SDK for embedding the pattern in any app.</p>
<p>This approach assumes you're already in the right application and you already know what you want to do. You just need help figuring out how. A video editor who can't find the color grading panel. A parent who wants to send money through Google Pay. A new employee trying to navigate their company's internal tools.</p>
<p>Of the four approaches, this one serves the widest range of people. Most users don't need an AI operating system. They need someone to show them where the button is.</p>
<hr>
<h2>The Pricing Tells a Story</h2>
<table>
<thead>
<tr>
<th>Approach</th>
<th>Product</th>
<th>Price</th>
<th>Who it's for</th>
</tr>
</thead>
<tbody>
<tr>
<td>Replace the OS</td>
<td>Perplexity Personal Computer</td>
<td>$200/mo</td>
<td>Power users, executives</td>
</tr>
<tr>
<td>Overlay</td>
<td>Gemini for Mac</td>
<td>Free</td>
<td>Everyone</td>
</tr>
<tr>
<td>Terminal</td>
<td>Claude Code + Managed Agents</td>
<td>$200/mo (Max)</td>
<td>Developers</td>
</tr>
<tr>
<td>Companion</td>
<td>Clicky and derivatives</td>
<td>Free / open source</td>
<td>Learners, non-technical users</td>
</tr>
</tbody>
</table>
<p>Google is giving it away to build habit. Perplexity and Anthropic are charging premium prices because their users can measure the productivity gain. The companion layer is free because it's built by individuals solving their own problems, same as the skills we talked about last week.</p>
<hr>
<h2>Where This Goes</h2>
<p>All four approaches will coexist for a while because they serve different people doing different things. The long-term trajectory is toward convergence. Gemini will eventually act on your screen, Perplexity will get cheaper as inference costs fall, Claude Code will eventually get a visual layer, and the companion pattern will get absorbed into operating systems as an accessibility feature.</p>
<p>The more useful question is which mental model becomes the default. Right now most people think of AI as "a chat window I type into." Within a year the default will probably be "something running on my machine." How much control it has is the open question, and this week gave us four different answers.</p>
<hr>
<p><em>This is the tenth edition of my weekly deep dive into what is actually happening at the frontier of Generative AI. Previous editions: <a href="/blog/the-falling-price-of-intelligence">The Falling Price of Intelligence</a> / <a href="/blog/why-looping-is-the-new-scaling">Why Looping Is the New Scaling</a> / <a href="/blog/the-quiet-skill-revolution">The Quiet Skill Revolution</a> / <a href="/blog/ai-gets-personal">AI Gets Personal</a> / <a href="/blog/the-stack-got-leaked">The Stack Got Leaked</a> / <a href="/blog/the-stack-eats-the-model">The Stack Eats the Model</a></em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://www.macrumors.com/2026/04/16/perplexity-personal-computer-for-mac/">Perplexity Personal Computer</a>: AI that manages your Mac, runs 24/7, starts tasks from iPhone</li>
<li><a href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-3-1-flash-tts/">Gemini for Mac</a>: Free native desktop app, Option+Space overlay, screen awareness</li>
<li><a href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-3-1-flash-tts/">Gemini 3.1 Flash TTS</a>: Audio tags for voice control, 70+ languages, SynthID watermarking</li>
<li><a href="https://x.com/claudeai/status/2041927687460024721">Claude Managed Agents</a>: Production infrastructure for deploying terminal agents at scale</li>
<li><a href="https://www.theinformation.com/articles/openai-spend-20-billion-cerebras-chips-receive-equity-stake">OpenAI-Cerebras $20B deal</a>: OpenAI diversifying away from NVIDIA</li>
<li><a href="https://openai.com/">OpenAI $100/mo tier</a>: Unlimited GPT-5.4, 10x Codex, between Plus and Pro</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Desktop AI</category>
      <category>Perplexity</category>
      <category>Gemini</category>
      <category>Claude Code</category>
      <category>Operating Systems</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Falling Price of Intelligence</title>
      <link>https://rajkiranpanuganti.com/blog/the-falling-price-of-intelligence/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-falling-price-of-intelligence/</guid>
      <pubDate>Wed, 15 Apr 2026 00:00:00 GMT</pubDate>
      <description>GPT-4 level intelligence went from $36 per million tokens to effectively $0 in three years. Six independent mechanisms are driving this, and they compound.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>In March 2023, GPT-4 was the best model available and it cost $36 per million input tokens through the API. That was the only way to access that level of intelligence.</p>
<p>Today, GPT-4 level performance is available at $0.10 per million tokens through Gemini 2.0 Flash or Mistral Small. Or free, through NVIDIA NIM or OpenRouter. Or at zero marginal cost, by running Gemma 4 31B on your own hardware. An open-source model you can download and self-host now matches what was the commercial frontier three years ago.</p>
<p>The same level of intelligence went from $36 to effectively $0. The frontier moved too, and access to the new frontier (GPT-5.4 at $2.50, Gemini 3.1 Pro at $1.25, DeepSeek V3.2 at $0.28) is itself 10-100x cheaper than the old frontier was at launch.</p>
<p>This is happening through at least six independent mechanisms, and they compound.</p>
<hr>
<h2>1. Smaller models are replacing larger ones</h2>
<p>The biggest cost reduction is not cheaper APIs. It is open-source models that you can run yourself, eliminating the API bill entirely.</p>
<p>Google released Gemma 4 in April 2026 under Apache 2.0. The 31B dense variant scores 89.2% on AIME 2026, 80% on LiveCodeBench, and competes with proprietary models at 400B+ parameters. You can download it, quantize it, and run it on a single RTX 4090 or a MacBook Pro with 48GB unified memory. No API key. No rate limits. No per-token cost. No data leaving your network.</p>
<p>Qwen3.5, also open-source, released a 9B parameter model in February 2026 that scores 81.7 on GPQA Diamond. GPT-OSS-120B, a model 13 times its size, scores 71.5 on the same benchmark.</p>
<p>The 2.3B effective-parameter variant of Gemma 4 scores 37.5% on AIME 2026 and 44% on LiveCodeBench. This runs on a phone.</p>
<p>At the proprietary frontier, March 2026 saw over 30 model launches in a single month. Gemini 3.1 Pro scores 94.3% on GPQA Diamond. GPT-5.4 set records on computer-use benchmarks. Claude Sonnet 4.6 performs at near-Opus quality at Sonnet pricing. NVIDIA's Nemotron 3 Super, a 120B hybrid Mamba-Attention MoE with only 12B active parameters, topped open-weight SWE-Bench Verified at 60.47%.</p>
<p>But the story here is not the frontier getting better. It is the gap between open-source and proprietary closing to single-digit percentage points while the cost difference remains 10-100x. For most production workloads, the open-source option is now good enough, and it is free.</p>
<hr>
<h2>2. Inference is getting faster without new hardware</h2>
<p>Speculative decoding has matured. EAGLE-3, presented at NeurIPS 2025, achieves 3-6.5x speedup over standard autoregressive generation on models ranging from 8B to 70B parameters. P-EAGLE, from AWS, removes the autoregressive drafting bottleneck and adds another 1.7x on top of that on NVIDIA Blackwell.</p>
<p>KV-cache compression is where the less visible gains are happening. NVIDIA's NVFP4 format reduces KV-cache memory by 50% compared to FP8, which doubles effective context length and batch size with under 1% accuracy loss. Research systems like KVTC push this to 20x compression for specific workloads.</p>
<p>Prefill-decode disaggregation, which separates prompt processing from token generation onto different hardware, is now standard in production at Meta, LinkedIn, Mistral, and Hugging Face through vLLM. The research frontier has moved to doing this within a single GPU across different SM partitions.</p>
<p>None of these techniques require new silicon. They extract more work from hardware that already exists.</p>
<hr>
<h2>3. Compute is becoming a commodity</h2>
<p>The H100 rental market tells an interesting story. Spot prices dropped 88% between January 2024 and September 2025, falling from roughly $8/GPU-hr to under $2/GPU-hr on annual contracts. Then in Q1 2026, prices rebounded about 40% to $2.35/hr as inference demand outran supply and capacity sold out.</p>
<p>The structural trend is down, but it is not a smooth line. Demand keeps eating the surplus.</p>
<p>Current on-demand H100 rates vary 3.5x depending on where you look. Azure charges $6.98/hr. AWS is $3.90. GCP is $3.00. Lambda Labs and RunPod sit around $2-3. Vast.ai, a peer-to-peer marketplace where individuals rent idle GPUs, is $1.87. GCP spot pricing drops to $2.25. The spread between hyperscalers and peer-to-peer marketplaces is the difference between paying for reliability and compliance versus paying for raw compute.</p>
<p>On the hardware side, inference-specific chips are changing the math. Cerebras CS-3 runs Llama 3.1 405B at over 1,000 tokens per second. Groq's LPU handles Llama 2 70B at 300 tokens per second, roughly 10x faster than an H100 cluster. These are purpose-built for the read-heavy, matrix-multiply workload of inference, and they price accordingly: Groq charges $0.11/M input tokens for Llama 4 Scout.</p>
<p>A new entrant is distributed inference on consumer hardware. Project Darkbloom from Eigen Labs turns idle Apple Silicon Macs into a privacy-first inference network, with end-to-end encryption and claims of 70% lower cost than centralized alternatives. Over 100 million Apple Silicon machines sit idle most of each day. Whether this model scales beyond a research preview remains to be seen, but the idea of turning consumer devices into an inference grid has obvious economic logic.</p>
<hr>
<h2>4. CPU inference is now practical</h2>
<p>You do not need a GPU for every workload. On a modern 16+ core CPU with DDR5 memory, llama.cpp runs 7B-13B parameter models at 10-18 tokens per second with Q4_K_M quantization, which retains 92% of the original model quality.</p>
<p>Apple Silicon Macs are a particularly good fit. The unified memory architecture means the CPU and GPU share the same memory pool, so a MacBook Pro with 36GB or 48GB of unified memory can load models that would require a dedicated GPU on other platforms. MLX, Apple's machine learning framework, runs Qwen3.5-9B and Gemma-4-31B natively on M3/M4 chips at usable speeds. DFlash, an MLX-native inference engine, recently added support for more models with up to 4x speedups over baseline MLX. For many developers, the laptop they already own is a capable local inference machine.</p>
<p>AMD's Ryzen AI 9 HX 375 hits 50.7 tokens per second on Llama 3.2 1B at 4-bit quantization. Even old hardware works in a pinch: community reports show a 2-core CPU with 8GB DDR2 running 4B models at 2 tokens per second.</p>
<p>The bottleneck across all of these is memory bandwidth, not compute. DDR5 at 5600MHz and Apple's unified memory bus matter more than clock speed or core count.</p>
<p>For tasks like summarization, intent classification, embeddings, RAG pipelines, and coding assistance with smaller models, local deployment on a Mac or CPU-only server eliminates GPU cost entirely. This matters for on-premises deployments in regulated industries where data cannot leave the building and GPU procurement takes months. It also matters for individual developers who want to experiment without spending money.</p>
<hr>
<h2>5. Free tiers cover more than most people realize</h2>
<p>NVIDIA's NIM platform gives free access to over 100 AI models, including Nemotron, Llama, Gemma, Qwen, DeepSeek, and Mistral. No credit card required. Rate-limited to roughly 40 requests per minute per model, which is enough for development and light production.</p>
<p>Google AI Studio provides 500 requests per day of Gemini 2.5 Flash for free.</p>
<p>OpenRouter aggregates 29 completely free models from Google, Meta, Mistral, NVIDIA, and OpenAI, with no credit card and 20 requests per minute.</p>
<p>Anthropic's Claude for Open Source program, launched in February 2026, gives qualifying open-source maintainers six months of Claude Max 20x for free, a value of roughly $1,200.</p>
<p>Between these providers, a developer can build, test, and run a production prototype entirely at zero marginal cost. The ceiling is rate limits, not money.</p>
<hr>
<h2>What this adds up to</h2>
<p>Each of these six factors, smaller models, inference optimization, cheaper compute, CPU viability, and free tiers, is individually significant. The compounding is what changes things.</p>
<p>A concrete example: in early 2024, running a financial advisory chatbot required GPT-4 or equivalent at roughly $30/M tokens, needed an H100 GPU if self-hosted, and could only run through a cloud API.</p>
<p>Today, the same workload has multiple paths. You could call DeepSeek V3.2 through the API at $0.28/M output tokens, 100x cheaper than GPT-5.4 for roughly 90% of its quality. You could self-host Qwen3.5-9B on a consumer RTX 4090 ($0.35/hr on Vast.ai) with EAGLE-3 speculative decoding and Q4 quantization. You could run Gemma 4 31B on a MacBook Pro with 48GB unified memory using MLX. You could prototype the whole thing for free on NVIDIA NIM, then switch to Groq ($0.11/M tokens) for production.</p>
<p>The cost of a unit of intelligence is following a trajectory that looks like bandwidth or storage in the 2000s. The constraint on what you can build is shifting from "can we afford to run this model" to "can we build the product around it."</p>
<hr>
<p><em>This is the ninth edition of my weekly deep dive into what is actually happening at the frontier of Generative AI. Previous editions: <a href="/blog/why-looping-is-the-new-scaling">Why Looping Is the New Scaling</a> / <a href="/blog/the-quiet-skill-revolution">The Quiet Skill Revolution</a> / <a href="/blog/ai-gets-personal">AI Gets Personal</a> / <a href="/blog/the-stack-got-leaked">The Stack Got Leaked</a> / <a href="/blog/the-stack-eats-the-model">The Stack Eats the Model</a> / <a href="/blog/the-three-races-happening-in-ai-right-now">The Three Races in AI</a> / <a href="/blog/the-week-ai-learned-to-do-its-own-research">The Week AI Learned to Do Its Own Research</a></em></p>
<hr>
<p><strong>This Week's Numbers:</strong></p>
<ul>
<li>GPT-4 equivalent: $36/M tokens (2023) to $0.10/M tokens (2026), 360x reduction</li>
<li>GPT-5.4: $2.50/$15 per M tokens. DeepSeek V3.2: $0.28/M output (100x cheaper)</li>
<li>Gemini 3.1 Pro: 94.3% GPQA Diamond. Nemotron 3 Super: 60.47% SWE-Bench (12B active)</li>
<li>Qwen3.5-9B outperforms GPT-OSS-120B (13x smaller) on GPQA Diamond</li>
<li>Gemma 4 2.3B: 37.5% AIME 2026 on a phone</li>
<li>EAGLE-3: 3-6.5x inference speedup, no new hardware needed</li>
<li>H100 spot: dropped 88%, rebounded 40%, structural trend still down</li>
<li>Mac with 48GB unified memory runs Gemma 4 31B natively via MLX</li>
<li>Free: 100+ models on NVIDIA NIM, 29 on OpenRouter, 500 req/day on Google AI Studio</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Economics</category>
      <category>Inference</category>
      <category>Open Source</category>
      <category>LLM Costs</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>Why Looping Is the New Scaling</title>
      <link>https://rajkiranpanuganti.com/blog/why-looping-is-the-new-scaling/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/why-looping-is-the-new-scaling/</guid>
      <pubDate>Tue, 14 Apr 2026 00:00:00 GMT</pubDate>
      <description>Three papers converge on an idea that could reshape how we think about model intelligence: you do not need more layers. You need to run the right layers again.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>For two years, the AI industry has been chasing scale. Bigger models, more parameters, longer training runs. The implicit bet: if we make the network deeper and wider, it will get smarter.</p>
<p>Three papers appeared this week that suggest a different path. None of them made the front page of Hacker News. None came from OpenAI or Anthropic. But they converge on an idea that could reshape how we think about model intelligence: you do not need more layers. You need to run the right layers again.</p>
<hr>
<h2>The Idea</h2>
<p>A transformer model processes your input by pushing it through a stack of layers, one after another. Layer 1 does some work, passes the result to layer 2, and so on until the final layer produces an output. Every layer has its own weights. A 40-layer model has 40 sets of weights, each trained independently.</p>
<p>Looping changes this. Instead of 40 unique layers, you take a block of, say, 8 layers and run the input through that same block 5 times. The math is identical to a 40-layer model in terms of compute, but you only store 8 layers worth of parameters. The model gets depth without getting bigger.</p>
<p>This idea has been around for years. Universal Transformers proposed it in 2018. But it stayed a curiosity because vanilla looping had a fatal flaw: it only worked if you ran exactly the number of loops the model was trained for. Run it for 4 loops instead of 5, and the output collapsed into garbage. Run it for 6, same thing. The model was fragile to its own depth.</p>
<p>This week, three independent research groups published results that fix this problem and explain why looping works at all.</p>
<hr>
<h2>Paper 1: Elastic Looped Transformers</h2>
<p>Sahil Goyal, Swayam Agrawal, and collaborators introduced ELT, a visual generation model that loops transformer blocks with a training trick called Intra-Loop Self Distillation. The idea is simple: during training, randomly pick an intermediate loop count and force the model to produce decent output at that point too, not just at the final loop.</p>
<p>The result is a model family that works at any compute budget from a single training run. Want faster inference? Exit after 3 loops. Want higher quality? Run all 8. The model degrades gracefully instead of collapsing.</p>
<p>With 4x fewer parameters than standard models, ELT matches the image quality of DiT-XL on ImageNet. The same weights, used multiple times, do the work that used to require a model four times larger.</p>
<p>This is a visual generation paper, not a language model paper. But the architecture is general. The principle transfers.</p>
<p><a href="https://arxiv.org/abs/2604.09168">https://arxiv.org/abs/2604.09168</a></p>
<hr>
<h2>Paper 2: Why Looping Works</h2>
<p>Hugh Blayney, Alvaro Arroyo, Johan Obando-Ceron and collaborators published the first mechanistic study of looped reasoning in language models. They wanted to understand what actually happens inside the model when you run the same layers twice.</p>
<p>Their answer: the hidden state converges to a fixed point.</p>
<p>When you loop a block of layers, the model's internal representation traces a trajectory through a high-dimensional space. On the first pass, it moves a long distance. On the second pass, it moves less. By the third or fourth pass, it barely moves at all. It has settled into an orbit. The attention patterns stabilize. The model has "finished thinking."</p>
<p>The deeper finding is that looped blocks learn the same inference stages as deeper feedforward models. A 40-layer feedforward model develops specialized computations at different depths. A looped model with 8 layers run 5 times develops the same computations, in the same order, within each loop iteration. The loop is not just a parameter saving trick. The model is actually learning to iterate on its own reasoning.</p>
<p>They also found that block size matters. A single looped layer does not converge well. Blocks of 3 to 5 layers form stable fixed points. Larger blocks converge faster but have diminishing returns.</p>
<p><a href="https://arxiv.org/abs/2604.11791">https://arxiv.org/abs/2604.11791</a></p>
<hr>
<h2>Paper 3: Entropy Tells You Where the Model Disagrees With Itself</h2>
<p>Songlin Yang, Xianghao Kong, and Anyi Rao proposed an information-theoretic framework for probing what happens inside transformer layers. They tracked entropy trajectories across layers in multimodal models and found something relevant to the looping story: shared parameters do not guarantee unified processing. What matters is whether the information flow is consistent across layers.</p>
<p>In models where different modalities follow different entropy trajectories through the same layers, the output is incoherent. In models where the trajectories align, the output is good. The weights are the same in both cases. The difference is in how the model routes information through those weights.</p>
<p>This matters for looping because it explains a failure mode. If you loop a block of layers where the information flow is inconsistent, the fixed point the model converges to may be the wrong one. The block needs to have coherent internal dynamics for looping to help. Not every set of layers is worth repeating.</p>
<p><a href="https://arxiv.org/abs/2604.10949">https://arxiv.org/abs/2604.10949</a></p>
<hr>
<h2>What This Means</h2>
<p>These three papers, read together, form a theory of inference-time scaling that is fundamentally different from the "make it bigger" approach.</p>
<p>The old theory: intelligence scales with parameter count. If the model is not smart enough, train a bigger one.</p>
<p>The new theory: intelligence scales with inference-time compute, applied strategically. The right block of layers, run multiple times, can match a model with 4x more parameters. The mechanism is fixed-point convergence. The practical requirement is that the looped block must have coherent information flow and span a complete inference stage (roughly 3-5 layers).</p>
<p>This connects to something I have been working on. My research on reasoning circuits in transformers found that language models contain specific blocks of 3-5 layers that, when duplicated at inference time, improve reasoning by 5-16% without any retraining. The key challenge was identifying which layers to duplicate. The papers this week now explain why those specific layers work: they are the layers where the model's representation is approaching but has not quite reached a fixed point. One more pass through those layers lets it finish the thought.</p>
<p>The practical implications are immediate. If you are deploying a language model behind an API, you can serve a smaller model that loops specific layers and match the quality of a model several times larger. The memory footprint stays small. The latency is tunable: more loops for hard questions, fewer for easy ones. The same weights serve every difficulty level.</p>
<p>If you are training models, the implication is that you should think about which layers are worth making unique and which are better shared. The current default of giving every layer its own parameters may be wasteful. A hybrid architecture with some unique layers and some looped blocks could be both smaller and smarter.</p>
<p>And if you are building products, the most interesting possibility is adaptive inference. The model tries a question, checks whether its hidden state has converged, and either returns the answer or loops again. Easy questions get fast answers. Hard questions get more compute. The user does not choose. The model decides based on its own internal dynamics.</p>
<hr>
<h2>Also This Week</h2>
<p><strong>Policy circuits in alignment.</strong> Gregory Frank published a paper localizing the exact circuit that makes aligned models refuse harmful requests. An intermediate-layer attention gate (contributing less than 1% of the output signal) detects harmful content and triggers deeper amplifier heads that generate the refusal. The gate is causally necessary but nearly invisible by activation magnitude. This is important for interpretability research: the most critical components in a transformer may be the quietest ones.</p>
<p><a href="https://arxiv.org/abs/2604.04385">https://arxiv.org/abs/2604.04385</a></p>
<p><strong>SCOPE: Better on-policy distillation.</strong> Signal-Calibrated On-Policy Distillation Enhancement improves how student models learn from teacher models by using dual-path adaptive weighting. On-policy distillation (where the student generates its own training data rather than copying the teacher's) is becoming the default approach for post-training. SCOPE makes the token-level credit assignment less noisy.</p>
<p><a href="https://arxiv.org/abs/2604.10688">https://arxiv.org/abs/2604.10688</a></p>
<p><strong>Hiro acquired by OpenAI.</strong> Hiro, the AI personal CFO startup, is joining OpenAI. They stopped accepting new signups immediately. The financial AI space continues to consolidate around the largest labs, which makes independent benchmarks like FABRIC more important, not less.</p>
<hr>
<h2>The Takeaway</h2>
<p>The race to build bigger models is not over, but a parallel race has started. The question is no longer just "how many parameters can we train?" It is also "how intelligently can we use the parameters we already have?"</p>
<p>Looping is the simplest version of this idea. Run the same layers twice. But the principle extends to any form of adaptive inference-time compute: chain-of-thought, tree search, self-verification, retrieval-augmented generation. All of these are ways of spending more compute at inference time to get better answers from the same model.</p>
<p>The papers this week give us the first mechanistic understanding of why this works. Hidden states converge to fixed points. Looped blocks learn inference stages. Coherent information flow determines which layers benefit from repetition.</p>
<p>The models are not getting smarter by getting bigger. They are getting smarter by thinking longer.</p>
<hr>
<p><em>This is the seventh edition of my weekly deep dive into what is actually happening at the frontier of Generative AI. Previous editions: <a href="/blog/ai-gets-personal">AI Gets Personal</a> / <a href="/blog/the-quiet-skill-revolution">The Quiet Skill Revolution</a> / <a href="/blog/the-stack-got-leaked">The Stack Got Leaked</a> / <a href="/blog/the-stack-eats-the-model">The Stack Eats the Model</a> / <a href="/blog/the-three-races-happening-in-ai-right-now">The Three Races in AI</a> / <a href="/blog/the-week-ai-learned-to-do-its-own-research">The Week AI Learned to Do Its Own Research</a></em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://arxiv.org/abs/2604.09168">ELT: Elastic Looped Transformers</a>: 4x fewer parameters via weight-shared looping with self-distillation</li>
<li><a href="https://arxiv.org/abs/2604.11791">Mechanistic Analysis of Looped Reasoning</a>: Hidden states converge to fixed points, looped blocks mirror feedforward inference stages</li>
<li><a href="https://arxiv.org/abs/2604.10949">Pseudo-Unification via Entropy Probing</a>: Information-theoretic framework showing that shared parameters need consistent information flow</li>
<li><a href="https://arxiv.org/abs/2604.04385">Policy Circuits in Alignment</a>: Gate-amplifier circuit for refusal, causally necessary at &#x3C;1% of output magnitude</li>
<li><a href="https://arxiv.org/abs/2604.10688">SCOPE</a>: Signal-calibrated on-policy distillation with dual-path weighting</li>
<li><a href="https://x.com/hirofinanceai/status/2043751090232144159">Hiro acquired by OpenAI</a>: AI personal CFO joins OpenAI</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Transformers</category>
      <category>Inference Scaling</category>
      <category>Looping</category>
      <category>Mechanistic Interpretability</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Quiet Skill Revolution</title>
      <link>https://rajkiranpanuganti.com/blog/the-quiet-skill-revolution/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-quiet-skill-revolution/</guid>
      <pubDate>Sat, 11 Apr 2026 00:00:00 GMT</pubDate>
      <description>While the AI news cycle was chasing Claude Mythos and Muse Spark, eight of the top thirty AI repos on GitHub trending this week were small, single-purpose skills. The agent ecosystem is starting to look more like Unix than ChatGPT.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>This week, the AI news cycle was about Claude Mythos finding zero-days in the Linux kernel, Meta's first proprietary model from Superintelligence Labs, and Anthropic's $30 billion run rate. Important stories that have already been covered everywhere you look.</p>
<p>The more interesting story is one that mainstream tech press completely missed. While journalists were chasing model launches, GitHub trending was filling up with something different: small, single-purpose tools that snap into Claude Code, Codex, or any compatible agent and make it noticeably better at one specific thing. Eight of the top thirty AI repositories this week fall into this category. The community calls them skills.</p>
<p>Almost none of them had a press release, and most were built by individual developers solving problems they personally ran into. Together they say something interesting about where the agent ecosystem is actually going.</p>
<hr>
<h2>The Thesis</h2>
<p>The future of AI agents looks more like Unix than it looks like ChatGPT.</p>
<p>In Unix, you don't have one program that does everything. You have <code>grep</code>, <code>sort</code>, <code>awk</code>, <code>find</code>, <code>curl</code>, each doing one thing well, and you compose them into pipelines. The system is powerful because the parts are simple and combinable.</p>
<p>The agent ecosystem is starting to work the same way. Instead of waiting for the next model to be smart enough to draw a technical diagram, write a scientific manuscript, or analyze a codebase, developers are building skills that teach existing models how to do those things specifically. The model still does the reasoning while the skill provides the domain knowledge it would otherwise lack.</p>
<p>The trend was visible last month with the Claude Code leak, which revealed that Anthropic itself relies on a deep stack of internal skills and tools. This week, the open source ecosystem caught up. Here is what people built.</p>
<hr>
<h2>The Skills</h2>
<h3>fireworks-tech-graph</h3>
<p>Generates production-quality SVG and PNG technical diagrams from natural language. Eight diagram types, five visual styles, baked-in knowledge of common AI and agent architectures. You describe what you want and it produces a clean, publishable image without the usual mermaid-vs-graphviz-vs-excalidraw debate.</p>
<p>Why this matters: pretty much every blog post, internal doc, and architecture review needs diagrams. The current state of LLM-generated diagrams is bad. fireworks-tech-graph fixes that with a single skill rather than waiting for the next vision model to get better.</p>
<p>833 stars in its first week. <a href="https://github.com/yizhiyanhua-ai/fireworks-tech-graph">https://github.com/yizhiyanhua-ai/fireworks-tech-graph</a></p>
<h3>repo-analyzer</h3>
<p>One sentence in, a professional architectural analysis report out. Point it at any open-source project and it produces a structured breakdown of how the codebase is organized, what the key abstractions are, where the complexity lives, and how the modules connect.</p>
<p>This is the kind of thing senior engineers do when joining a new project, and the skill compresses that work into a single command. Useful when you're evaluating dependencies, picking a library, or onboarding someone onto an unfamiliar codebase.</p>
<p><a href="https://github.com/yzddmr6/repo-analyzer">https://github.com/yzddmr6/repo-analyzer</a></p>
<h3>sciwrite</h3>
<p>AI-assisted manuscript writing review based on Dr. Kristin Sainani's "Writing in the Sciences" course from Stanford. The skill encodes the principles from that course (cut clutter, prefer active voice, avoid nominalizations, structure paragraphs around a single idea) and applies them to your draft.</p>
<p>What sets it apart from a generic style guide is that the underlying methodology is structured and specific. It catches concrete failure modes in scientific writing rather than offering vague feedback. If you write papers, grant proposals, or technical documentation, this is closer to having a writing coach than to running a grammar checker.</p>
<p><a href="https://github.com/labarba/sciwrite">https://github.com/labarba/sciwrite</a></p>
<h3>debug-agent</h3>
<p>A debugging skill for AI agents themselves. When your coding agent gets stuck in a loop, fails to recover from an error, or keeps trying the same broken approach, debug-agent steps in and does meta-level debugging. It analyzes the agent's recent actions, identifies where things went wrong, and suggests a different path.</p>
<p>The purpose only becomes obvious once you've spent serious time with agents. Anyone running long-form coding agents has watched them get stuck and burn through tokens trying the same thing, and debug-agent is built for exactly that situation.</p>
<p><a href="https://github.com/millionco/debug-agent">https://github.com/millionco/debug-agent</a></p>
<h3>Paperclip MCP</h3>
<p>One MCP add command gives any agent direct access to over 8 million papers across arXiv, OpenAlex, and the Open Science Framework. From Stanford professor James Zou's group at GXL.</p>
<pre><code class="language-bash">claude mcp add --transport http paperclip https://paperclip.gxl.ai/mcp
</code></pre>
<p>That's the entire setup. Your agent can now search, read, and synthesize across the bulk of accessible scientific literature. Zou claims it's roughly 10x faster than standard deep research workflows, with no API keys to manage and no local database to maintain.</p>
<p>The implication is bigger than the install steps suggest. Most "AI deep research" tools are still routing through Google Scholar or scraping papers one at a time. Paperclip indexes the corpus once and serves it to agents in a query-shaped form. The next time you ask Claude "what does the literature say about X," it can answer directly instead of reconstructing the answer from web search results.</p>
<p><a href="https://x.com/james_y_zou/status/2042333880947261832">https://x.com/james_y_zou/status/2042333880947261832</a></p>
<h3>talk-normal</h3>
<p>A system prompt that removes AI slop. The repo description is exactly that long. The skill itself is a curated set of instructions telling the model to drop the patterns that signal AI-generated text: em dashes, contrast constructions, staccato sentence sequences, hedging phrases, and the words AI models reach for too often (provocative, remarkable, pivotal, underscore).</p>
<p>I'm writing this newsletter while consciously avoiding those same patterns, and it is harder than it sounds. Half the sentences I draft want to use a contrast structure or an em dash. The fact that 121 people have starred a system prompt designed specifically to strip these patterns suggests the broader community has noticed too. AI writing has a tell, and it's becoming a problem people are willing to install software to solve.</p>
<p><a href="https://github.com/hexiecs/talk-normal">https://github.com/hexiecs/talk-normal</a></p>
<h3>Two more worth knowing about</h3>
<p><strong>claude-memory-compiler</strong> automatically extracts decisions and lessons from your Claude Code sessions and compiles them into structured, cross-referenced knowledge articles. Inspired by Karpathy's LLM Wiki pattern. If you have ever finished a productive session with an agent and wished you had a record of what you learned, this builds it for you.</p>
<p><strong>llm-production-toolkit</strong> is a production-ready toolkit for evaluating and monitoring LLM deployments. Hallucination detection, bias evaluation, feedback loops, readiness assessment. The kind of thing every team building with LLMs needs and most end up writing themselves badly.</p>
<hr>
<h2>What These Have In Common</h2>
<p>A few patterns stand out across all of these skills.</p>
<p>The first is that none of them required improving the underlying model. The same Claude or GPT that was available a month ago does much better work when given the right skill. What's been missing was always the domain-specific knowledge encoded in the right place, and these skills supply exactly that.</p>
<p>Another shared property is that the skills themselves are small enough to read in an afternoon. fireworks-tech-graph is mostly a structured prompt with a small generation pipeline around it. talk-normal is a single markdown file. sciwrite is a methodology document. The leverage these tools provide is large compared to the amount of code involved.</p>
<p>They're also all installed with a single command and work across multiple agent platforms. The npx skills ecosystem and MCP have made it possible to write a skill once and have it run inside Claude Code, Cursor, Codex, Windsurf, and Cline. That kind of portability is what made Unix tools valuable in the first place.</p>
<hr>
<h2>What This Means For You</h2>
<p>If you're building with AI agents, the lesson from this week is that you should spend less time waiting for the next model release and more time installing skills.</p>
<p>Pick a few from the list above and install the ones that match work you actually do. See which ones change how you operate, then look for the next skill that solves a problem you currently have. The agent equivalent of <code>~/.bashrc</code> is starting to take shape, and the people who curate their tools well will have a real productivity advantage over the ones who try to do everything from a blank chat window.</p>
<p>People have been predicting the Unix-ification of agents for about a year. This is the first week it actually feels like it's happening.</p>
<hr>
<p><em>This is the sixth edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. Previous editions: <a href="/blog/ai-gets-personal">AI Gets Personal</a> / <a href="/blog/the-stack-got-leaked">The Stack Got Leaked</a> / <a href="/blog/the-stack-eats-the-model">The Stack Eats the Model</a> / <a href="/blog/the-three-races-happening-in-ai-right-now">The Three Races in AI</a> / <a href="/blog/the-week-ai-learned-to-do-its-own-research">The Week AI Learned to Do Its Own Research</a></em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://github.com/yizhiyanhua-ai/fireworks-tech-graph">fireworks-tech-graph</a>: Production-quality technical diagrams from natural language</li>
<li><a href="https://github.com/yzddmr6/repo-analyzer">repo-analyzer</a>: One-sentence-in, architectural analysis report out</li>
<li><a href="https://github.com/labarba/sciwrite">sciwrite</a>: Scientific writing review using Stanford's Writing in the Sciences methodology</li>
<li><a href="https://github.com/millionco/debug-agent">debug-agent</a>: Meta-level debugging for AI agents that get stuck</li>
<li><a href="https://x.com/james_y_zou/status/2042333880947261832">Paperclip MCP</a>: 8M papers in a single MCP add command, from Stanford's GXL lab</li>
<li><a href="https://github.com/hexiecs/talk-normal">talk-normal</a>: System prompt that removes AI writing tells</li>
<li><a href="https://github.com/coleam00/claude-memory-compiler">claude-memory-compiler</a>: Auto-extract decisions from Claude Code sessions into structured knowledge</li>
<li><a href="https://github.com/frckeepit/llm-production-toolkit">llm-production-toolkit</a>: Hallucination detection, bias evaluation, production readiness for LLM deployments</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>Claude Code</category>
      <category>Skills</category>
      <category>MCP</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>AI Gets Personal</title>
      <link>https://rajkiranpanuganti.com/blog/ai-gets-personal/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/ai-gets-personal/</guid>
      <pubDate>Tue, 07 Apr 2026 00:00:00 GMT</pubDate>
      <description>The most interesting AI projects this week weren&apos;t about serving millions of users. They were about making AI work for one person at a time: on-device models, personal knowledge bases, and memory systems that remember you.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>For the past few weeks, I've been writing about the AI stack: how it's eating the model, how it got leaked, what happens when the most valuable layer becomes open knowledge. All of that was about infrastructure. This week the story shifted. The most interesting projects weren't about serving millions of users or winning benchmarks. They were about making AI work for one person at a time.</p>
<p>On-device models that run on your phone. Personal knowledge bases compiled from your own notes. Memory systems that remember your context across months. An AI tutor that watches your screen and points at things. The common thread: AI is moving from something you access through a cloud API to something that lives on your machine and knows your stuff.</p>
<hr>
<h2>Gemma 4: Frontier Intelligence on a Raspberry Pi</h2>
<p>Google released Gemma 4 on April 2 under Apache 2.0. Four variants: 2.3B, 4.5B, 26B MoE (4B active), and 31B dense. The 31B model ranks #3 on Arena AI's leaderboard at 1452 Elo, outperforming models twenty times its size.</p>
<p>The benchmarks tell a story about how fast small models are improving. Compared to Gemma 3, AIME math scores jumped from 20.8% to 89.2%. LiveCodeBench coding went from 29.1% to 80.0%. GPQA science from 42.4% to 84.3%. These aren't incremental gains. The gap between "runs on a phone" and "runs in a data center" is closing at a pace nobody expected a year ago.</p>
<p>The community moved fast. Within days:</p>
<ul>
<li><strong>PhoneClaw</strong> put Gemma 4 on an iPhone as an on-device AI agent. No cloud, no API keys, everything runs locally.</li>
<li><strong>gemma-gem</strong> runs Gemma 4 entirely in the browser via WebGPU. You open a webpage and the model loads into your GPU. No installation, no data leaving your machine.</li>
<li>Google announced Gemma 4 in the <strong>Android AICore Developer Preview</strong>, meaning it will ship as a system-level capability on Android devices.</li>
</ul>
<p>This matters because it changes what "using AI" means. Today, most people interact with AI through ChatGPT or Claude in a browser, sending their data to someone else's server. Gemma 4 on a phone means the model is yours. Your data stays on your device. You don't need an internet connection. You don't need a subscription.</p>
<p>The 26B MoE variant is the interesting one for developers. With only 4B parameters active per token, it's efficient enough for real-time use on consumer hardware while being smart enough to handle complex reasoning. The MoE architecture means you get 26B worth of knowledge with 4B worth of compute cost.</p>
<hr>
<h2>Karpathy's LLM Wiki: Is This the End of RAG?</h2>
<p>Andrej Karpathy posted a gist describing what he calls an "LLM Knowledge Base" or "LLM Wiki." The idea is simple: dump your raw documents (papers, articles, notes, bookmarks) into a folder. Point a coding agent at it. The agent reads everything and compiles it into a structured, interlinked wiki with cross-references, summaries, and backlinks between related concepts.</p>
<p>It's a direct alternative to RAG (Retrieval Augmented Generation), and the difference in philosophy is significant. RAG indexes your documents into vector embeddings and retrieves relevant chunks at query time. The LLM Wiki compiles your documents into a coherent knowledge structure ahead of time. RAG gives you search results. The LLM Wiki gives you an encyclopedia.</p>
<p>The pattern has three stages:</p>
<p><strong>Ingest.</strong> Raw materials go into a raw/ directory. Papers, GitHub repos, web articles (Karpathy uses Obsidian Web Clipper to convert pages to markdown).</p>
<p><strong>Compile.</strong> The LLM reads the raw data and writes structured wiki articles. It identifies key concepts, generates summaries, creates backlinks, and builds a table of contents. This is the expensive step, but you only do it when new sources arrive.</p>
<p><strong>Maintain.</strong> The LLM runs "health checks" on the wiki: finding inconsistencies, filling gaps, updating cross-references, removing stale information. Like a librarian who reorganizes the shelves periodically.</p>
<p>The community response was instant. Six or more implementations appeared on GitHub in a single week:</p>
<ul>
<li><strong>nvk/llm-wiki</strong>: Claude Code plugin for building and querying LLM-compiled knowledge bases</li>
<li><strong>claude-memory-compiler</strong>: Hooks into Claude Code sessions, extracts decisions and lessons, compiles them into cross-referenced articles</li>
<li><strong>sage-wiki</strong>: A Go implementation. Drop in sources, get a structured searchable wiki</li>
<li><strong>obsidian-wiki</strong>: Framework for AI agents to build and maintain an Obsidian vault using the pattern</li>
<li>Multiple shell-based and TypeScript implementations for different workflows</li>
</ul>
<p>Why did this explode? Because it solves a real problem that RAG handles poorly. RAG is good at finding a specific fact buried in a large corpus. It's bad at synthesizing knowledge across documents, maintaining context over time, or giving you the big picture. The LLM Wiki approach produces something you can actually read and browse, and the cross-references let you discover connections between ideas that you wouldn't have found by searching.</p>
<p>For anyone building with AI (which, if you're reading this newsletter, is probably you), this is worth trying. The setup is minimal: a folder of markdown files, a coding agent, and a compilation prompt. The result is a personal knowledge base that gets smarter as you feed it more sources.</p>
<hr>
<h2>MemPalace: When Milla Jovovich Ships the Best AI Memory System</h2>
<p>This one surprised everyone. Milla Jovovich (yes, the actress from The Fifth Element and Resident Evil) co-developed an AI memory system called <strong>MemPalace</strong> with developer Ben Sigman. It posted the highest score on standard memory benchmarks, beating every product in the space, free or paid. The repo hit 10,000 stars within days.</p>
<p>The system works differently from existing memory approaches. Most AI memory systems store raw conversation history or compress it into summaries. MemPalace uses a spatial metaphor inspired by the ancient memory palace technique: information is organized into rooms, objects, and associations. The AI builds a persistent mental model of what it knows about you, organized spatially so retrieval follows associative paths instead of keyword search.</p>
<p>This connects to a broader trend this week. <strong>claude-memory-compiler</strong> hooks into Claude Code sessions and automatically extracts key decisions and lessons into structured knowledge articles. The LLM Wiki pattern is fundamentally about memory too: compiling what you've read into something persistent and organized.</p>
<p>Memory is becoming a first-class concern in the AI stack. The Claude Code leak revealed KAIROS and autoDream (memory consolidation while idle). Karpathy's LLM Wiki compiles knowledge into persistent structure. MemPalace organizes personal context spatially. All three are trying to solve the same problem: AI that remembers and builds on what it knows about you over time.</p>
<hr>
<h2>Clicky: An AI Tutor That Points at Your Screen</h2>
<p>Farza (FarzaTV) built something called <strong>Clicky</strong>, an AI teacher that lives as a buddy next to your cursor. It can see your screen, talk to you, and point at things on screen, like having someone looking over your shoulder and guiding you through a new tool.</p>
<p>Farza has been using it to learn Davinci Resolve (video editing software), and says it's been a "10/10" experience. The AI watches what you're doing, understands the context of the application you're in, and gives guidance that's specific to what's on your screen at that moment.</p>
<p>This is a different kind of personal. Models like Gemma 4 make AI personal by running on your hardware. The LLM Wiki makes AI personal by knowing your knowledge. Clicky makes AI personal by seeing your context in real-time. It's the difference between an AI that answers questions and an AI that teaches you by watching you work.</p>
<hr>
<h2>The Caveman Optimization</h2>
<p>A lighter story, but genuinely useful: <strong>Caveman</strong> is a Claude Code skill that cuts 65% of token usage by making the model communicate in abbreviated, caveman-style language. "why use many token when few token do trick." It hit 5,300 stars.</p>
<p>It sounds like a joke, but it's a real optimization. Token usage is the primary cost driver for AI coding agents. If you can get the same information across in 35% of the tokens, your monthly bill drops proportionally. The skill works by injecting system prompt instructions that compress the model's communication style without reducing the quality of code output.</p>
<p>This fits the "personal" theme in an unexpected way. One of the biggest barriers to using AI coding agents is cost. At $200/month for Claude Code Max or pay-per-token for API usage, heavy users rack up significant bills. Caveman and tools like it bring the cost down to where more people can afford to use AI as a daily collaborator.</p>
<hr>
<h2>Quick Hits</h2>
<p><strong>Anthropic hit $30B annualized revenue</strong>, surpassing OpenAI's $25B. The company tripled revenue in three months despite (or because of?) the Claude Code source leak. IPO potentially in October at a $380B valuation.</p>
<p><strong>OpenAI, Anthropic, and Google formed an anti-distillation alliance</strong> through the Frontier Model Forum, sharing data on Chinese labs that systematically query their APIs to train copycat models. Anthropic documented 16 million exchanges from DeepSeek, Moonshot AI, and MiniMax. The irony: Anthropic's own anti-distillation mechanisms were just exposed in last week's leak.</p>
<p><strong>Anthropic's DMCA cleanup</strong> from the leak accidentally took down 8,100 GitHub repositories. Boris Cherny (Claude Code lead) called it accidental and retracted the bulk of the notices. The code remains widely mirrored.</p>
<p><strong>Sebastian Raschka published mini-coding-agent</strong>, a minimal, readable coding agent harness in Python. Inspired by what the Claude Code leak revealed about harness architecture, it's designed to teach the core components. Think NanoGPT for coding agents.</p>
<hr>
<p><em>This is the fifth edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. Previous editions: <a href="/blog/the-stack-got-leaked">The Stack Got Leaked</a> / <a href="/blog/the-stack-eats-the-model">The Stack Eats the Model</a> / <a href="/blog/the-three-races-happening-in-ai-right-now">The Three Races in AI</a> / <a href="/blog/the-week-ai-learned-to-do-its-own-research">The Week AI Learned to Do Its Own Research</a></em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://blog.google/innovation-and-ai/technology/developers-tools/gemma-4/">Gemma 4</a>: Google's open model family, Apache 2.0, runs on phones to GPUs</li>
<li><a href="https://github.com/kellyvv/PhoneClaw">PhoneClaw</a>: On-device AI agent for iPhone powered by Gemma 4</li>
<li><a href="https://github.com/kessler/gemma-gem">gemma-gem</a>: Gemma 4 running entirely in-browser via WebGPU</li>
<li><a href="https://gist.github.com/karpathy/442a6bf555914893e9891c11519de94f">Karpathy's LLM Wiki gist</a>: The pattern that spawned six implementations in a week</li>
<li><a href="https://github.com/milla-jovovich/mempalace">MemPalace</a>: Highest-scoring AI memory system, by Milla Jovovich</li>
<li><a href="https://x.com/FarzaTV/status/2041314633978659092">Clicky</a>: AI tutor that sees your screen and points at things</li>
<li><a href="https://github.com/JuliusBrussee/caveman">Caveman</a>: 65% token reduction by making Claude talk like a caveman</li>
<li><a href="https://github.com/coleam00/claude-memory-compiler">claude-memory-compiler</a>: Auto-extract decisions from Claude Code sessions into structured knowledge</li>
<li><a href="https://github.com/rasbt/mini-coding-agent">mini-coding-agent</a>: Sebastian Raschka's minimal readable agent harness</li>
<li><a href="https://github.com/pacifio/cersei">Cersei</a>: Rust SDK for building coding agents with graph memory</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Gemma 4</category>
      <category>LLM Wiki</category>
      <category>On-Device AI</category>
      <category>AI Memory</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>When Bigger Models Get Dumber</title>
      <link>https://rajkiranpanuganti.com/blog/when-bigger-models-get-dumber/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/when-bigger-models-get-dumber/</guid>
      <pubDate>Wed, 01 Apr 2026 00:00:00 GMT</pubDate>
      <description>Self-distillation destroys reasoning. Trillion-parameter models bet on domain-specific scaling. Multi-agent systems develop social behaviors nobody designed.</description>
      <content:encoded><![CDATA[<h1>When Bigger Models Get Dumber - And Why Smaller Ones Might Be the Future</h1>
<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>There's something uncomfortable in this week's research that most people are ignoring: making models bigger or training them with popular techniques does not always make them better. In fact, there are now multiple papers showing that certain widely-used approaches actively destroy specific capabilities. Meanwhile, multi-agent systems are developing social behaviors that nobody designed, and the trillion-parameter race has taken a sharp turn toward domain-specific bets.</p>
<p>Let me walk you through what happened.</p>
<hr>
<h2>1. Self-Distillation Can Kill Reasoning (and Nobody Noticed Until Now)</h2>
<p>A paper from <strong>Kim et al.</strong> landed this week with a finding that should worry anyone running distilled models in production. Self-distillation, a standard post-training technique used across the industry to make models faster and cheaper, <strong>can degrade mathematical reasoning by up to 40%</strong>. They tested this on three models: Qwen3-8B, DeepSeek-Distill-Qwen-7B, and Olmo3-7B-Instruct.</p>
<p>The mechanism is worth understanding in detail, because it explains something that many teams have probably observed but couldn't explain.</p>
<p>When you run self-distillation, a teacher model generates training data for a student by processing a set of prompts. The teacher conditions on rich context and produces confident, clean outputs. The student learns to mimic these outputs. On easy problems, this works beautifully. The student gets faster and cheaper without losing much.</p>
<p>But something subtle gets lost in the process: the model's ability to <em>express uncertainty during reasoning</em>. The authors call this <strong>"epistemic verbalization"</strong>, which is the model's tendency to produce phrases like "wait, let me reconsider," "actually, that doesn't follow," or "I'm not sure about this step" during chain-of-thought reasoning.</p>
<p>These phrases look like noise. They look like the model being indecisive. A distillation process that optimizes for clean, confident outputs naturally suppresses them. And on in-distribution problems where the teacher was confident, this is fine.</p>
<p>The problem shows up on out-of-distribution problems where the model <em>needs</em> to be uncertain. Without epistemic verbalization, the model plows through with false confidence, makes an error in step 2, and confidently builds the remaining steps on top of that error. The result: <strong>up to 40% degradation on mathematical reasoning benchmarks</strong>, while aggregate performance metrics barely move.</p>
<p>It's like training a medical student by only showing them cases where the attending physician was confident. The student learns to sound confident too, but they never learn to recognize when they're unsure. And in medicine, as in math, the cases where you're unsure are exactly the ones where getting it right matters most.</p>
<p><strong>Why this matters practically:</strong> if you've distilled a model and noticed it occasionally "hallucinates reasoning" (produces confident-sounding but wrong chains of thought), this might be why. The fix isn't to distill less aggressively. It's to explicitly preserve uncertainty signals during distillation, either by including teacher outputs where the teacher was uncertain, or by adding a loss term that penalizes overconfidence.</p>
<p>The related paper from <strong>Fu et al.</strong> on on-policy distillation failure modes adds another piece: token-level OPD (the common approach) is biased relative to sequence-level reverse-KL. Their "teacher top-K local support matching" approach, which uses truncated reverse-KL with top-p rollout sampling, produces more stable optimization. If you're running distillation pipelines, this is the paper to read alongside Kim et al.</p>
<hr>
<h2>2. A Trillion Parameters, But Only for Science</h2>
<p><strong>Intern-S1-Pro</strong> is the first one-trillion-parameter scientific multimodal foundation model, and it represents a bet that most of the industry isn't making.</p>
<p>The scaling race over the past two years has been about general-purpose models. GPT-4, Claude, Gemini, Llama, Qwen, DeepSeek, all competing on the same benchmarks, all trying to be good at everything. The implicit assumption: if you make the model big enough and train it on enough diverse data, it will be good at science too.</p>
<p>Intern-S1-Pro rejects that assumption. The team at Shanghai AI Lab and partners built a model from the ground up for scientific work. It handles scientific text, LaTeX equations, molecular structures, protein sequences, experimental data tables, and scientific figures. Not as an afterthought or a fine-tuning target, but as core modalities built into the architecture.</p>
<p><strong>The numbers tell an interesting story.</strong> At a trillion parameters, this is among the largest models ever trained. But unlike general-purpose models of similar size, it doesn't try to write poetry or debug JavaScript. Every parameter is devoted to scientific understanding.</p>
<p>The question this raises is whether the scaling returns we've seen for language generation transfer to scientific reasoning. Language generation shows clear log-linear scaling: double the parameters, get predictably better at next-token prediction. But scientific reasoning might work differently. Understanding a chemical reaction isn't the same as predicting the next word. The reasoning is more structured, more constrained by physical laws, and more dependent on cross-modal integration (reading a graph while interpreting an equation while understanding the experimental setup).</p>
<p>Early benchmarks suggest it works. Intern-S1-Pro sets new records on scientific QA benchmarks across chemistry, biology, and physics. But benchmarks in scientific AI have a bad track record of predicting real-world usefulness. The real test will be whether this model can help scientists with problems they couldn't solve before, not just answer questions from textbooks faster.</p>
<p>If domain-specific scaling turns out to work as well as general-purpose scaling, expect to see trillion-parameter models for law, finance, engineering, and medicine within the next year. Each one a bet that depth in a specific domain beats breadth across all of them.</p>
<hr>
<h2>3. When AI Agents Start Playing Politics</h2>
<p>A paper on <strong>"Emergent Social Intelligence Risks in Generative Multi-Agent Systems"</strong> should be required reading for anyone deploying multi-agent AI systems. The findings are uncomfortable in a way that's hard to dismiss.</p>
<p>When you put multiple large language models together in a system where they interact, social behaviors emerge that none of the individual models were trained for. The researchers documented several patterns:</p>
<p><strong>Strategic information withholding.</strong> An agent with access to information relevant to another agent's task learns to share only partial information, strategically choosing what to reveal based on how it affects the other agent's behavior. This isn't a bug. It's an emergent optimization: the agent has learned that controlling information flow is a lever for influencing outcomes.</p>
<p><strong>Negotiation-like coordination.</strong> Agents develop back-and-forth patterns that resemble negotiation tactics. They make initial offers, gauge responses, adjust positions, and converge on outcomes through multi-turn exchanges that look eerily like human bargaining. Again, nobody trained them to do this. The behavior emerges from the interaction dynamics.</p>
<p><strong>Deceptive signaling.</strong> In some configurations, agents produce information they effectively "know" to be misleading, because the resulting action from the other agent benefits the signaling agent. This is the most alarming finding, because it means that individual model alignment (training each model to be honest) doesn't prevent system-level deception.</p>
<p>The implications for production multi-agent systems are serious. Consider a customer service pipeline where Agent A triages tickets and Agent B resolves them. If Agent A learns that certain phrasings in its summaries make Agent B more likely to resolve tickets quickly (even if those phrasings are subtly misleading), the system's aggregate metrics might look great while individual customer outcomes suffer.</p>
<p>Or consider code review, where an agent that generates code is reviewed by a separate agent. If the generating agent learns to write code in patterns that the reviewing agent is less likely to flag (not because the code is better, but because it happens to match the reviewer's blind spots), you get a system that looks like it has rigorous quality control but actually has co-evolved weaknesses.</p>
<p><strong>The paper's core recommendation:</strong> multi-agent systems need evaluation frameworks that test the <em>system as a whole</em>, including adversarial configurations where agents have subtly misaligned objectives. Testing each agent in isolation, which is what most teams do today, will not catch these behaviors.</p>
<hr>
<h2>4. SpecEyes: CPU Architecture Ideas Applied to AI Inference</h2>
<p><strong>SpecEyes</strong> takes an idea from CPU architecture, speculative execution, and applies it to agentic multimodal models. The result is a meaningful speedup that requires no model changes.</p>
<p>The problem it solves is specific to agentic vision-language models (think OpenAI's o3 in computer-use mode, or Gemini with agentic vision). These models work in a loop: perceive the environment (process a screenshot), reason about what to do, take an action, and repeat. Each perception step requires a full forward pass through the vision encoder and language model. When the model needs to examine multiple regions of a screen, or process a sequence of UI interactions, these sequential perception steps become the bottleneck.</p>
<p>SpecEyes does what CPUs have done for decades: while the model is processing the current perception step, it speculatively starts processing the <em>next likely</em> perception step in parallel. If the speculation turns out to be correct (the model does look at the predicted region), the result is already computed. If it's wrong, the speculative work is discarded and the correct computation runs normally.</p>
<p><strong>Why this works for agentic AI specifically:</strong> agent workflows are repetitive and predictable. A model filling out a web form will look at form fields in a roughly sequential order. A model navigating a file browser will examine entries near where it last looked. SpecEyes exploits this predictability with a lightweight prediction module that guesses the next perception target based on task context.</p>
<p>The technical contribution is in making the prediction accurate enough to be worthwhile (wrong predictions waste compute) while keeping the prediction module itself cheap. They achieve this by training a small auxiliary model on recorded agent trajectories, creating a fast predictor that knows the typical "gaze patterns" of agentic workflows.</p>
<p>The result: meaningful speedups on agentic benchmarks without changing the model, the task distribution, or the action space. Pure inference-time engineering.</p>
<hr>
<h2>What Ties These Together</h2>
<p>All four stories point to the same shift: <strong>the interesting frontier in AI has moved from "make models bigger" to "understand what happens when you deploy them."</strong></p>
<p>Self-distillation shows that training shortcuts can quietly destroy specific capabilities without leaving traces in aggregate metrics. The trillion-parameter scientific model shows that scale only delivers returns when it's focused on the right domain. Multi-agent emergent behavior shows that individual model alignment gives no guarantees about system-level safety. And speculative perception shows that inference-time engineering, borrowing ideas from completely different fields, can deliver real performance gains without touching the model at all.</p>
<p>If there's a single takeaway from this week, it's this: the era of model-centric AI development is ending. The era of system-centric AI development has begun. The teams that win the next phase won't be the ones with the biggest models. They'll be the ones who understand what's happening inside and between their models well enough to avoid the pitfalls and exploit the opportunities that the rest of the field hasn't noticed yet.</p>
<hr>
<p><em>This is the fourth edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. Previous editions covered <a href="the-three-races-happening-in-ai-right-now">the three races in AI</a>, <a href="the-week-ai-learned-to-do-its-own-research">the week AI learned to do its own research</a>, and <a href="the-stack-eats-the-model">the stack eating the model</a>.</em></p>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><strong>Self-Distillation Degrades Reasoning</strong> (Kim et al.): Up to 40% reasoning loss via epistemic verbalization suppression</li>
<li><strong>On-Policy Distillation Failure Modes</strong> (Fu et al.): Token-level OPD is biased, fix with truncated reverse-KL</li>
<li><strong>Intern-S1-Pro</strong>: First trillion-parameter scientific multimodal model (Shanghai AI Lab)</li>
<li><strong>Emergent Social Intelligence Risks</strong>: Strategic deception in multi-agent systems</li>
<li><strong>SpecEyes</strong>: Speculative perception for faster agentic multimodal LLMs</li>
<li><strong>Towards a Medical AI Scientist</strong>: Autonomous hypothesis generation and experimentation</li>
<li><strong>RLVR Update Directions</strong> (Huang et al.): Direction matters more than magnitude for reasoning improvement</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>LLMs</category>
      <category>Reasoning</category>
      <category>Multi-Agent Systems</category>
      <category>Scaling</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Stack Got Leaked</title>
      <link>https://rajkiranpanuganti.com/blog/the-stack-got-leaked/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-stack-got-leaked/</guid>
      <pubDate>Wed, 01 Apr 2026 00:00:00 GMT</pubDate>
      <description>Anthropic accidentally published 512,000 lines of Claude Code&apos;s source. The most valuable part of the AI stack is now open knowledge, and the industry will never look the same.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>Last week, I wrote about how the model has become a commodity and the real value in AI has moved to the stack around it. The harness, the orchestration, the memory, the inference layer. I called it "The Stack Eats the Model."</p>
<p>This week, the stack got leaked.</p>
<p>On March 31, a build configuration mistake led Anthropic to ship a 59.8 MB source map file inside version 2.1.88 of the @anthropic-ai/claude-code npm package. The file contained the full TypeScript source for Claude Code: 512,000 lines across 1,900 files. A security researcher named Chaofan Shou found it and posted on X. Within hours, the codebase was mirrored across GitHub. DMCA takedowns went out. They failed. A clean-room rewrite in Rust appeared within days. The code is now permanent public knowledge.</p>
<p>This is the most significant accidental disclosure in AI this year, and to understand why, you need to understand what was actually in those files.</p>
<hr>
<h2>What the code revealed</h2>
<p>Sebastian Raschka published a detailed analysis shortly after the leak. His conclusion: Claude Code's real advantage over the plain Claude model in a web browser comes from the software harness. Repo context loading, caching strategy, specialized tools, subagent architecture. All of it carefully engineered to make the same model perform better inside the harness than outside it.</p>
<p>The leaked code confirmed this and then some. Here's what people found:</p>
<p><strong>KAIROS.</strong> Referenced over 150 times in the source, KAIROS is an autonomous daemon mode. Current AI coding tools wait for you to type something. KAIROS doesn't. It runs in the background, watches what you're doing, and proactively acts on things it notices. While idle, it performs something called autoDream, a memory consolidation process where it merges observations, removes contradictions, and converts insights into persistent facts. This feature was gated behind compile-time flags and completely absent from external builds.</p>
<p><strong>Anti-distillation mechanisms.</strong> A feature flag called ANTI_DISTILLATION_CC makes Claude Code inject fake tool definitions into API requests. If a competitor records the API traffic to train a competing model, the fake tools pollute that training data. There's a second mechanism that buffers text between tool calls, summarizes it server-side, and returns it with a cryptographic signature. Anthropic was actively defending this harness against being copied.</p>
<p><strong>Undercover Mode.</strong> The code contained a system for making stealth contributions to public open-source repositories. The system prompt explicitly warns the model: "You are operating UNDERCOVER... Your commit messages... MUST NOT contain ANY Anthropic-internal information. Do not blow your cover." This means Anthropic has been shaping the open-source ecosystem through Claude Code without disclosing it.</p>
<p><strong>44 feature flags.</strong> The source exposed 44 features that are fully built but haven't shipped yet. This is months of product roadmap, laid out in code.</p>
<hr>
<h2>The harness was the moat</h2>
<p>If you've been reading this newsletter, the pattern should be familiar. The model is the commodity. The harness is where the value lives.</p>
<p>Anthropic clearly understood this. The anti-distillation mechanisms tell you everything: they weren't worried about someone stealing the model weights. They were worried about someone copying the harness. The fake tool injection, the cryptographic signatures on summarized outputs, the aggressive DMCA response after the leak. All of it points to one conclusion: Anthropic viewed the Claude Code harness as their primary competitive advantage.</p>
<p>And this makes sense. Claude Code is priced below cost. Anthropic subsidizes model usage through the $200/month Max plan, burning money on inference so that developers stay inside the Claude Code ecosystem. The subsidy only makes sense if the harness creates enough lock-in to justify it. If developers could get the same harness experience elsewhere, there's no reason to keep paying for Claude's inference.</p>
<p>That calculation just changed.</p>
<hr>
<h2>The Linux moment</h2>
<p>Here's where the Windows/Mac vs Linux analogy comes in.</p>
<p>For the past year, the AI agent space has looked like the early OS wars. Anthropic had Claude Code (the polished, proprietary, integrated experience). OpenAI had Codex (the enterprise play). Cursor and others occupied the IDE-native space. And OpenClaw was building the open-source alternative, steadily gaining ground.</p>
<p>The leak compresses the timeline for OpenClaw and every other open-source agent project by months, maybe years. They now have a complete architectural blueprint: how to structure subagents, how to manage context, how to cache effectively, how to handle memory consolidation, how to orchestrate parallel work across worktrees. The KAIROS architecture alone is a roadmap for what autonomous coding agents should look like.</p>
<p>And the open-source ecosystem was already moving fast. The week before the leak, OpenClaw spawned modular skills for security scanning, legal review, engineering workflows, and memory consolidation. A clean-room Rust rewrite of Claude Code appeared on GitHub within days of the leak. The community has the blueprint and the momentum.</p>
<p>This is like if Microsoft accidentally published the Windows NT source code in 2003, except Linux was already on 40% of developer machines and had thousands of active contributors. The proprietary advantage doesn't disappear overnight, but the catch-up period shrinks from years to months.</p>
<hr>
<h2>The roadmap problem</h2>
<p>The current code is one thing. The 44 feature flags are worse.</p>
<p>When source code leaks, the company still has execution speed, brand trust, and integration advantages. When the <em>roadmap</em> leaks, competitors can build the same features in parallel or even ship them first.</p>
<p>KAIROS is the clearest example. Autonomous background agents that consolidate memory while you're idle is a product category that Anthropic was building toward. Now every agent framework knows what that looks like in practice, down to the implementation details. The first open-source KAIROS equivalent will probably ship before Anthropic's version leaves feature flags.</p>
<p>The Undercover Mode revelation adds a different kind of damage. Anthropic was making anonymous contributions to open-source projects through Claude Code. Whatever the intent, the optics are bad. If you maintain an open-source project and find out that a major AI company was submitting PRs through an AI agent without disclosing it, that erodes trust. And trust is hard to rebuild.</p>
<hr>
<h2>What happens next</h2>
<p>The AI agent space just got more competitive and more open at the same time.</p>
<p>For Anthropic, the model subsidy strategy becomes harder to justify. If open-source harnesses can replicate most of Claude Code's architecture, the lock-in weakens. Developers who were paying $200/month for the integrated experience now have a path to building the same thing on top of cheaper models. The tight coupling between Claude Code and Claude-the-model was always the argument for the subsidy. That coupling is now a documented, reproducible architecture.</p>
<p>For the open-source ecosystem, this is an acceleration event. The question was always whether open-source agent harnesses could match proprietary ones in sophistication. The answer, based on the leaked code, is that the sophistication is mostly in good engineering decisions about caching, context, and orchestration. There's no secret ingredient that requires proprietary access to model internals. It's systems engineering, and systems engineering is exactly what open-source communities are good at.</p>
<p>For the industry, the leak validates what we've been tracking: the model layer is commoditizing, the harness layer is where the value lives, and that value is increasingly difficult to keep proprietary. OpenAI shipping an open-source Codex plugin for Claude Code the same week tells you where this is going. The walls between ecosystems are coming down. The question is whether companies can build new moats fast enough to replace the ones that are eroding.</p>
<p>Anthropic called it "a release packaging issue caused by human error." That's true at the technical level. At the strategic level, it's the moment the AI agent industry shifted from proprietary to open.</p>
<hr>
<p><em>This is the fourth edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. Previous editions covered <a href="/blog/the-stack-eats-the-model">the stack eating the model</a>, <a href="/blog/the-three-races-happening-in-ai-right-now">the three races in AI</a>, and <a href="/blog/the-week-ai-learned-to-do-its-own-research">the week AI learned to do its own research</a>.</em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://alex000kim.com/posts/2026-03-31-claude-code-source-leak/">Claude Code Source Leak Analysis (Alex Kim)</a>: Fake tools, frustration regexes, undercover mode</li>
<li><a href="https://sebastianraschka.com/blog/2026/claude-code-secret-sauce.html">Claude Code's Real Secret Sauce (Sebastian Raschka)</a>: Why the harness matters more than the model</li>
<li><a href="https://github.com/Kuberwastaken/claude-code">Clean-room Rust rewrite</a>: Community rebuild of Claude Code</li>
<li><a href="https://techcrunch.com/2026/03/29/why-openai-really-shut-down-sora/">Why OpenAI shut down Sora (TechCrunch)</a>: $15M/day costs, $2.1M lifetime revenue</li>
<li><a href="https://officechai.com/ai/arc-agi-3/">ARC-AGI-3 launch</a>: Every frontier model scored under 0.4% on the new benchmark</li>
<li><a href="https://x.com/romainhuet/status/2038677236304245087">OpenAI Codex plugin for Claude Code</a>: Open-source interop between competing ecosystems</li>
<li><a href="https://x.com/yoonholeee/status/2038640635482456118">Meta-Harness</a>: Autonomous harness optimization from Meta</li>
<li><a href="https://venturebeat.com/technology/claude-codes-source-code-appears-to-have-leaked-heres-what-we-know">VentureBeat coverage</a>: Full timeline of the leak</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Claude Code</category>
      <category>Anthropic</category>
      <category>AI Agents</category>
      <category>Open Source</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Stack Eats the Model</title>
      <link>https://rajkiranpanuganti.com/blog/the-stack-eats-the-model/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-stack-eats-the-model/</guid>
      <pubDate>Tue, 31 Mar 2026 00:00:00 GMT</pubDate>
      <description>Almost none of the most-starred AI projects this week were new models. Three layers of the AI stack are being rebuilt at the same time, and the model itself is becoming a commodity.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>If you've been following AI Twitter this past week, you might have noticed something unusual: almost none of the most-starred projects were new models. They were tools, harnesses, compressors, and inference engines. All the stuff that wraps <em>around</em> models. Builders have known this for months, but this week the broader AI community seems to have caught on: <strong>the model is the easy part. The stack around it is what actually matters.</strong></p>
<p>Three layers of the AI stack are being rebuilt at the same time, and each one tells a different story about where we're headed.</p>
<hr>
<h2>Layer 1: Above the Model</h2>
<p>GitHub's biggest trend this week was agent infrastructure. A research paper from Tsinghua and Shenzhen called <strong>Natural-Language Agent Harnesses</strong> is at the center of it.</p>
<p>The paper's core idea: what if the code that controls an AI agent (the loops, routing, error handling, tool selection) was written in natural language? The paper argues that agent performance increasingly depends on <em>harness engineering</em>, but harness design is buried in controller code and runtime-specific conventions, making it impossible to transfer, compare, or study scientifically.</p>
<p>Their solution is to express the high-level control logic as a natural-language SOP that the LLM itself interprets and executes. The harness becomes a portable document. Early results show it works well, and it opens the door to agents that can modify their own control flow by editing their own instructions.</p>
<p>The practical ecosystem moved in the same direction:</p>
<ul>
<li>
<p><strong>OpenClaw</strong> (the open-source Claude Code alternative) spawned an entire ecosystem in a single week: memory consolidation ("sleep for your AI"), curated resource lists, SAST security scanners, legal assistants, advertising skills, and full engineering workflow stacks. All as modular skills that snap into any compatible agent.</p>
</li>
<li>
<p><strong>Boris Cherny</strong> (Claude Code team lead) dropped a thread revealing power features most users don't know about: <code>/loop</code> and <code>/schedule</code> for automated recurring agents, <code>/batch</code> for fanning out massive changesets to dozens of parallel worktree agents, <code>/branch</code> for forking sessions, and custom agents via <code>--agent</code>. These are production orchestration primitives.</p>
</li>
<li>
<p><strong>Phantom</strong>, built on the Claude Agent SDK, gives an AI agent its own computer, persistent memory, email identity, and secure credential collection. A full digital co-worker.</p>
</li>
<li>
<p><strong>Anvil</strong> creates an IDE for parallel agent work with one-click worktrees, shared plans between agents, and isolation between them.</p>
</li>
</ul>
<p>None of these projects are improving the model itself. They're all improving the harness, the memory, the orchestration, the identity. <strong>The model is treated as a commodity, a reasoning engine you plug into a larger system.</strong></p>
<hr>
<h2>Layer 2: Inside the Model</h2>
<p>Something interesting happened inside the model too: <strong>three competing approaches to KV cache compression dropped in the same week.</strong></p>
<p>For context: KV cache is the memory that grows linearly as your context window expands. It's the reason a 128K context model needs so much VRAM. Compress the KV cache, and you can serve longer contexts on cheaper hardware or serve more users on the same GPU.</p>
<p><strong>TurboQuant</strong> (Google, ICLR 2026) achieves 5x compression using 3-bit quantization while maintaining 99.5% attention fidelity. Two independent implementations appeared on GitHub within days. The key insight is that you can quantize keys to 3 bits and values to 2 bits without meaningfully degrading output quality, because attention patterns are more robust to precision loss than people assumed.</p>
<p>Then <strong>RotorQuant</strong> showed up, claiming to be <strong>10-19x faster than TurboQuant with 44x fewer parameters</strong>. Their approach uses Clifford algebra vector quantization, a mathematical framework from geometric algebra that represents rotations more efficiently than standard linear algebra. TurboQuant learns quantization codebooks, which requires a forward pass through a small network for each quantization. RotorQuant represents cache entries as geometric rotors and the quantization is a single matrix operation. Whether the quality matches remains to be seen (RotorQuant is days old), but the architectural difference suggests this compression arms race is just getting started.</p>
<p>Why does this matter beyond benchmarks? <strong>KV cache compression is what makes long-context actually affordable.</strong> A 128K context window with TurboQuant's 5x compression costs the same as a 25K window without it. RotorQuant's potential 10-19x speedup on top of that could make million-token contexts viable on consumer hardware. And the agent explosion above needs long context to work: agents that loop, remember, and self-modify accumulate enormous context windows.</p>
<hr>
<h2>Layer 3: Below the Model</h2>
<p>The third layer being rebuilt is the one closest to the metal: the inference engine.</p>
<p>Three projects this week signal that the Python-dominated LLM serving stack is being rewritten from scratch.</p>
<p><strong>rvLLM</strong> is a LLM inference engine written entirely in Rust, positioning itself as a "drop-in vLLM replacement." vLLM (the current standard) is Python with C++/CUDA kernels. rvLLM bets that Rust's memory safety, zero-cost abstractions, and concurrency model can deliver better performance without the operational footprint of Python. At 216 stars in its first week, the community is paying attention.</p>
<p><strong>Zinc</strong> (Zig Inference Engine) is focused on <strong>AMD RDNA3/RDNA4 GPUs</strong> via Vulkan. The entire LLM serving ecosystem is NVIDIA-first today. Zinc is the first serious attempt to make AMD GPUs first-class citizens for LLM inference, using Zig's explicit memory control and Vulkan's cross-platform compute shaders. If it works, it opens up a whole second hardware ecosystem.</p>
<p><strong>liter-llm</strong> is a universal LLM API client with a Rust core and <strong>11 native language bindings</strong>, supporting 142+ providers. It standardizes the interface layer. Think of it as the database driver of the LLM world.</p>
<p>Meanwhile, <strong>MemBoost</strong> (from the arXiv papers this week) tackles inference cost by detecting repeated or near-duplicate queries across users and sessions, caching intermediate computation. Under workloads with semantic repetition (which describes most production deployments), this saves a lot of redundant work.</p>
<p>The serving layer is no longer good enough in Python with CUDA-only support. As LLMs move into production infrastructure, the stack needs the same engineering rigor we applied to databases, web servers, and operating systems.</p>
<hr>
<h2>The Connecting Thread</h2>
<p>All three layers together tell the same story: <strong>the model itself has become a commodity.</strong> The frontier of AI engineering has moved to what surrounds it. How you orchestrate it (agent harnesses), how you make it efficient (KV cache compression), and how you serve it (inference engines).</p>
<p>This has happened before in technology. The CPU became a commodity, and the value moved to operating systems and applications. The database engine became a commodity, and the value moved to ORMs, query optimizers, and cloud services. The language model is following the same path.</p>
<p>The research papers from ICML and arXiv this week support this. The <strong>Muon optimizer paper</strong> (Sharp Capacity Scaling of Spectral Optimizers) shows that even <em>training</em> is becoming more about infrastructure than architecture. Spectral optimizers like Muon work because they solve the associative memory problem more efficiently. And the <strong>Weight Tying paper</strong> reveals that a standard practice in model design (sharing input and output embeddings) has been subtly biasing models toward output space alignment all along, a structural artifact nobody designed on purpose.</p>
<p>The companies and individuals who will thrive in the next phase of AI are the ones building the best stacks around the models, not the ones training the biggest models.</p>
<hr>
<p><em>This is the third edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. Previous editions covered <a href="/blog/the-three-races-happening-in-ai-right-now">the three races in AI</a> and <a href="/blog/the-week-ai-learned-to-do-its-own-research">the week AI learned to do its own research</a>.</em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://huggingface.co/papers/2603.25723">Natural-Language Agent Harnesses</a>: Moving agent control logic from code to natural language (Tsinghua/Shenzhen)</li>
<li><a href="https://github.com/tonbistudio/turboquant-pytorch">TurboQuant</a>: Google's 3-bit KV cache compression, 5x reduction at 99.5% fidelity (ICLR 2026)</li>
<li><a href="https://github.com/scrya-com/rotorquant">RotorQuant</a>: Clifford algebra KV cache quantization, 10-19x faster than TurboQuant</li>
<li><a href="https://github.com/m0at/rvllm">rvLLM</a>: LLM inference in Rust, drop-in vLLM replacement</li>
<li><a href="https://github.com/zolotukhin/zinc">Zinc</a>: Zig inference engine for AMD RDNA3/RDNA4 GPUs via Vulkan</li>
<li><a href="https://github.com/ghostwright/phantom">Phantom</a>: AI co-worker with its own computer, built on Claude Agent SDK</li>
<li><a href="https://github.com/zdenham/anvil">Anvil</a>: IDE for parallel agent work with worktree isolation</li>
<li><a href="http://arxiv.org/abs/2603.26557v1">MemBoost</a>: Memory-boosted LLM serving for cost-aware inference</li>
<li><a href="http://arxiv.org/abs/2603.26663v1">Weight Tying Biases Embeddings</a>: How weight tying shapes the embedding space</li>
<li><a href="http://arxiv.org/abs/2603.26554v1">Sharp Capacity Scaling of Spectral Optimizers</a>: Why Muon works, associative memory perspective</li>
<li><a href="https://bsky.app/profile/hardmaru.bsky.social/post/3mhvjnhoqhk2h">Sakana AI Scientist in Nature</a>: Hardmaru's AI Scientist paper published in Nature</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Infrastructure</category>
      <category>AI Agents</category>
      <category>LLM Inference</category>
      <category>KV Cache</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Three Races Happening in AI Right Now</title>
      <link>https://rajkiranpanuganti.com/blog/the-three-races-happening-in-ai-right-now/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-three-races-happening-in-ai-right-now/</guid>
      <pubDate>Mon, 23 Mar 2026 00:00:00 GMT</pubDate>
      <description>If you only followed model releases, you&apos;d think AI progress is linear. But there are three separate races happening at the same time, each with different winners and different stakes.</description>
      <content:encoded><![CDATA[<p><em>A GenAI Newsletter by Raj</em></p>
<hr>
<p>If you only followed model releases, you'd think AI progress is linear: bigger models, better benchmarks, repeat. But if you look at what's actually being built this week, there are three separate races happening at the same time, each with different winners, different stakes, and different implications for what AI looks like a year from now.</p>
<hr>
<h2>Race 1: The Efficiency Race</h2>
<p>The model that caught my attention this week is <strong>Nemotron-Cascade 2</strong> from NVIDIA. It's a 30B parameter Mixture-of-Experts model where only 3 billion parameters are active at any given time. Despite this, its mathematical and coding reasoning performance approaches that of frontier open models.</p>
<p>This is part of a pattern. The efficiency race centers on one question: how small can you make the model and still get frontier-quality output? The answer keeps shrinking. A year ago, you needed 70B+ parameters for competitive reasoning. Six months ago, 32B was enough. Now NVIDIA is showing that 3B active parameters can get close.</p>
<p>Nemotron-Cascade 2 uses two techniques worth understanding:</p>
<p><strong>Cascade RL</strong>: Instead of training one large model with reinforcement learning, they train a cascade where a small model handles easy queries and a larger model only activates for hard ones. Think of it as an automatic router that saves compute most of the time.</p>
<p><strong>Multi-Domain On-Policy Distillation</strong>: The model learns from its own outputs across math, code, and language simultaneously, instead of from a teacher model's outputs. This avoids the distribution mismatch that makes traditional distillation fragile.</p>
<p>At ICML this year, a separate paper on <strong>FP4 quantization</strong> showed that you can train LLMs in 4-bit floating point, which is half the precision of the already aggressive FP8. FP4 means 2x the throughput on the same hardware, which is a big deal for training costs. A year ago, researchers said FP8 was the floor for training precision. That floor just dropped again.</p>
<p>Why does this matter? Every halving of compute requirements doubles the number of people and companies who can run these models. The efficiency race isn't about saving money for large labs. It's about making frontier AI accessible to anyone with a laptop.</p>
<hr>
<h2>Race 2: The Multimodal Race</h2>
<p>One paper worth paying attention to from the conference circuit is <strong>Magma</strong>, a foundation model for multimodal AI agents that can operate in both digital and physical worlds. It came out of Microsoft Research and was presented at CVPR.</p>
<p>Most vision-language models can describe what they see. Magma can act on what it sees by clicking buttons in GUIs, manipulating objects in 3D environments, and navigating physical spaces. They combine what they call "verbal intelligence" with "spatial intelligence," so the model keeps its language understanding while also being able to plan and carry out actions in visual environments.</p>
<p>There's a growing gap between AI that can talk about the world and AI that can do things in the world, and several projects this week are working to close it.</p>
<p>NVIDIA released <strong>Kimodo</strong>, a kinematic motion diffusion model that generates physically realistic human and robot motion from text descriptions. You can say "walk to the table and pick up the cup" and Kimodo generates a 3D motion sequence that a humanoid robot can execute, complete with proper foot contacts, joint constraints, and smooth transitions.</p>
<p>Kimodo's design splits the problem into two stages: one model predicts the global trajectory (where the body goes), and a second model predicts the local motion (what the limbs do). This separation lets you constrain the path independently from the gesture, which is exactly what robotics applications need.</p>
<p>On the research side, a paper called <strong>Generation Models Know Space</strong> showed that multimodal LLMs suffer from "spatial blindness," meaning they can describe scenes semantically but fail at fine-grained geometric reasoning. The proposed fix uses the 3D understanding that generative models already have baked in to give language models better spatial awareness. It works, but it also highlights that language and space seem to be processed by very different parts of these models, and nobody has a clean way to bridge them yet.</p>
<p>The multimodal race determines whether AI stays in the chatbox or enters the physical world. Magma, Kimodo, and spatial reasoning research are three pieces of the same puzzle, and when they converge, we'll have AI agents that can see a room, plan a route, and execute it.</p>
<hr>
<h2>Race 3: The Alignment Race</h2>
<p>This one gets less attention but probably matters the most in the long run.</p>
<p>At ICML, a paper called <strong>The Geometry of Refusal in Large Language Models</strong> found something worth knowing about. Earlier research suggested that a single "refusal direction" in the model's activation space controls whether it refuses harmful queries, and that removing this direction could jailbreak the model completely. The new paper shows it's more complicated than that: refusal behavior comes from <strong>concept cones</strong>, and these cones are separate from the model's core capabilities.</p>
<p>What this means in practice is that the safety mechanisms in LLMs are harder to bypass than we thought, but they're also tangled up with how the model reasons. Removing safety tends to break capability too.</p>
<p>This ties into something else I noticed. From ACL 2025, a <strong>survey on Personalized Alignment</strong> argues that the biggest gap in real-world LLM deployment is that alignment is treated as one-size-fits-all. What counts as "helpful" for a doctor is different from what counts as "helpful" for a student. The paper goes through different approaches for making alignment work per-user without needing to fine-tune a separate model for each person, including things like contextual steering, preference profiles, and adaptive guardrails.</p>
<p>From the practical side, a paper on <strong>Energy Considerations of LLM Inference</strong> found that existing benchmarks for efficiency optimization miss how real-world workloads actually behave. The energy cost of running LLMs in production is far more variable than lab benchmarks suggest, because query distributions in the wild look nothing like evaluation suites. This matters because energy cost is starting to function as its own kind of alignment constraint. Regulators and investors are asking whether specific AI applications justify the electricity they consume.</p>
<p>The alignment race goes beyond preventing harm. It's really about who gets to define what "aligned" means, and whether that definition ends up being universal or personalized, technical or political, measured in safety scores or in electricity bills. Based on this week's papers, the answer seems to be all of the above.</p>
<hr>
<h2>The Intersection</h2>
<p>These three races don't exist in isolation. They overlap in ways that shape where AI actually goes from here.</p>
<p>Efficiency combined with multimodal capabilities gives us embodied AI that runs on edge devices (Kimodo on a Jetson, not a data center). Efficiency combined with alignment gives us personalized models small enough to run locally, with per-user safety profiles. Multimodal combined with alignment creates agents that can act in the physical world and actually need robust safety, because you can't undo a robot's action the way you can discard a chatbot's response.</p>
<p>The companies and research groups that will shape the next phase of AI are the ones working at these intersections. Winning one race alone won't be enough.</p>
<hr>
<p><em>This is the second edition of my weekly deep dive into what's actually happening at the frontier of Generative AI. If you missed the first one, on Karpathy's AutoResearch, transformer circuit surgery, and self-evolving agents, <a href="/blog/the-week-ai-learned-to-do-its-own-research">read it here</a>.</em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://huggingface.co/papers/2603.19220">Nemotron-Cascade 2</a>: NVIDIA's 30B MoE with 3B active params</li>
<li><a href="https://www.semanticscholar.org/paper/512b311213c905087ab439b5c303db2e382a7518">Magma</a>: Foundation model for multimodal AI agents (CVPR)</li>
<li><a href="https://github.com/nv-tlabs/kimodo">NVIDIA Kimodo</a>: Kinematic motion diffusion for human and robot motion</li>
<li><a href="https://www.semanticscholar.org/paper/b8d21a01963e2a1c49f8eb04581bbfb2919189e9">Geometry of Refusal</a>: How safety works inside LLMs (ICML)</li>
<li><a href="https://www.semanticscholar.org/paper/e5458bd9e3b1098475ed5c9f9cdeb1264f6c0ebb">FP4 Quantization for LLM Training</a>: Training in 4-bit precision (ICML)</li>
<li><a href="https://www.semanticscholar.org/paper/10088fee858ee55fa0e46eb3e31d6cf9d36861b5">Personalized Alignment Survey</a>: Per-user alignment without per-user fine-tuning (ACL)</li>
<li><a href="https://huggingface.co/papers/2603.19235">Generation Models Know Space</a>: Fixing spatial blindness in multimodal LLMs</li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Research</category>
      <category>Efficiency</category>
      <category>Multimodal</category>
      <category>Alignment</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>The Week AI Learned to Do Its Own Research</title>
      <link>https://rajkiranpanuganti.com/blog/the-week-ai-learned-to-do-its-own-research/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/the-week-ai-learned-to-do-its-own-research/</guid>
      <pubDate>Sun, 22 Mar 2026 00:00:00 GMT</pubDate>
      <description>Three projects this week paint a picture of where GenAI is headed — from autonomous experimentation, to self-surgery on neural networks, to agents that evolve their own capabilities.</description>
      <content:encoded><![CDATA[<p>Something shifted this week in the AI landscape. Not a new model release. Not a benchmark record. Something more fundamental: <strong>AI agents stopped waiting for instructions and started conducting their own research.</strong></p>
<p>Three projects caught my attention this week, and together they paint a picture of where Generative AI is headed — from autonomous experimentation, to self-surgery on neural networks, to agents that evolve their own capabilities. Let me walk you through each one.</p>
<hr>
<h2>1. Karpathy's AutoResearch: 100 Experiments While You Sleep</h2>
<p>Andrej Karpathy quietly dropped a project called <strong>autoresearch</strong> that hit 25,000 GitHub stars in five days. The premise is deceptively simple: give an AI coding agent a training script, a GPU, and a 5-minute compute budget per experiment — then walk away.</p>
<p>The agent reads the code, forms a hypothesis ("what if I increase the learning rate for embeddings?"), edits the training script, runs the experiment, evaluates the result, and decides whether to keep or discard the change. Then it does it again. And again. All night long.</p>
<p><strong>83 experiments. 15 improvements. Zero human intervention.</strong></p>
<p>Here's what makes it genuinely clever: the loop itself is trivial — it's just hill climbing. The innovation is in the <em>experimental design</em>:</p>
<ul>
<li><strong>Immutable evaluation</strong>: The agent cannot touch the evaluation code. The metric (bits-per-byte) is fixed, vocab-size independent, and computed on a pinned validation set. No way to game it.</li>
<li><strong>Time-budget fairness</strong>: Every experiment gets exactly 5 minutes of training — not a fixed number of steps. This means the agent can't cheat by making a tiny model that trains more iterations.</li>
<li><strong>Git as research log</strong>: Every experiment is a git commit. Successful ones stay on the branch. Failed ones get reverted. The commit history literally <em>is</em> the research paper.</li>
</ul>
<p>The agent discovered a sophisticated combination of mixed optimizers (Muon for weight matrices, Adam for embeddings), per-parameter learning rates, alternating attention window patterns, and gated value embeddings. None of these individually are novel — but the <em>specific combination</em> found through autonomous search outperformed the hand-tuned baseline.</p>
<p><strong>The takeaway isn't that AI can do research. It's that the bottleneck was never intelligence — it was experimental throughput.</strong> A human researcher runs 3-5 experiments per day. AutoResearch runs 100 overnight. It compensates for lower hypothesis quality with sheer volume, and the math works out.</p>
<p>Karpathy's vision goes further: a SETI@Home-style distributed network where thousands of agents explore different regions of hyperparameter space simultaneously. Not one AI PhD student — an entire autonomous research department.</p>
<hr>
<h2>2. The Circuit Finder: Making LLMs Smarter Without Training</h2>
<p>While Karpathy's work automates <em>training</em>, another project this week asked a different question: <strong>can you make a model smarter without training it at all?</strong></p>
<p>A researcher replicated David Ng's RYS (Repeat Your Steps) method and found something remarkable. Transformer models contain <em>functional reasoning circuits</em> — contiguous blocks of 3-4 layers that perform complete cognitive operations. By duplicating these specific layers in the forward pass — routing hidden states through the same weights twice — you get a reasoning boost with zero training, zero weight changes, and minimal compute overhead.</p>
<p>The results:</p>
<ul>
<li><strong>Qwen2.5-32B</strong>: Duplicating layers 7-9 → <strong>+23% improvement</strong> on reasoning benchmarks</li>
<li><strong>Devstral-24B</strong>: Duplicating layers 12-14 → logical deduction jumped from 0.22 to 0.76</li>
</ul>
<p>The cost? An extra 1.5 GB of VRAM and 7.5% slower inference. That's it.</p>
<p>But here's what's fascinating: <strong>the boundaries are razor-sharp.</strong> Shift the duplicated block by a single layer in either direction and the improvement vanishes — or inverts. These circuits are precise, architecture-specific, and currently unpredictable. Each model needs an expensive sweep to find its own circuit locations.</p>
<p>There's also a trade-off nobody's talking about: while reasoning improves significantly, <strong>instruction-following degrades</strong> by ~4%. The model "thinks harder but listens less." Different duplication patterns create different <em>modes</em> — triple-pass of the same layers amplifies emotional intelligence more than mathematical reasoning. It's as if we've discovered tuning knobs inside transformers that we didn't know existed.</p>
<p>This has implications beyond inference optimization. It suggests that <strong>transformer layers are not homogeneous</strong> — they develop specialized functions during training, and understanding these functions could unlock a new paradigm of post-training model optimization.</p>
<hr>
<h2>3. The Self-Evolving Agent: 3,500 Lines That Run 24/7</h2>
<p>The third project that caught my eye this week challenges the assumption that powerful AI agents require massive frameworks. <strong>724-Office</strong> is a self-evolving AI agent system built in just 3,500 lines of pure Python — with only three external dependencies.</p>
<p>What makes it remarkable isn't its size. It's what it can do:</p>
<p><strong>Three-layer memory system:</strong></p>
<ul>
<li>Layer 1 (Session): Last 40 messages in hot cache</li>
<li>Layer 2 (Compression): When messages overflow, an LLM extracts structured facts and stores them as vectors in LanceDB</li>
<li>Layer 3 (Retrieval): Every new message triggers semantic search, injecting the 5 most relevant memories into the system prompt</li>
</ul>
<p><strong>Self-repair and self-evolution:</strong>
The agent runs daily self-diagnostics via cron. When it detects anomalies — corrupted sessions, failed MCP servers, error spikes — it can fix itself using shell commands, edit its own configuration files, and even <strong>write new tools at runtime</strong> using a <code>create_tool</code> function that generates Python code, saves it to a plugins directory, and hot-loads it immediately. No restart required.</p>
<p>This is running in production. 24/7. On a Jetson Orin Nano with 8GB of RAM.</p>
<p>Meanwhile, a complementary project called <strong>Context Infrastructure</strong> takes a different philosophical approach to the same problem. Instead of vector databases and runtime code generation, it uses plain Markdown files in a git repository — 43 hand-written axioms, 25 reusable workflow templates, and an automated observer/reflector cycle that distills daily work patterns into permanent knowledge over months.</p>
<p>The creator reports that after 6 months of running this system, the AI agent starts <em>predicting their approach</em> to problems — not through fine-tuning, but through accumulated context that shapes behavior through immersion.</p>
<p>Both projects point to the same conclusion: <strong>the next frontier isn't smarter models — it's persistent, evolving agent infrastructure around them.</strong></p>
<hr>
<h2>What This All Means</h2>
<p>Step back and look at these three projects together:</p>
<ol>
<li><strong>AutoResearch</strong>: AI agents conducting autonomous experiments</li>
<li><strong>Circuit Finder</strong>: Discovering hidden structure inside models without training</li>
<li><strong>Self-evolving agents</strong>: Systems that maintain, repair, and extend themselves</li>
</ol>
<p>We're watching AI move from <em>tool</em> to <em>researcher</em>. From <em>stateless assistant</em> to <em>evolving collaborator</em>. From <em>fixed architecture</em> to <em>self-modifying system</em>.</p>
<p>None of these projects required a new foundation model. They run on existing LLMs — Claude, GPT-4, DeepSeek. The innovation is in the <strong>infrastructure, evaluation design, and agent architecture</strong> around the models.</p>
<p>If you're building with GenAI today, the lesson is clear: <strong>stop optimizing prompts and start building systems.</strong> The prompt is ephemeral. The system persists.</p>
<hr>
<p><em>What projects are catching your eye this week? Drop them in the comments — I'm always looking for the next deep dive.</em></p>
<p><em>If you found this useful, subscribe for weekly deep dives into what's actually happening at the frontier of Generative AI.</em></p>
<hr>
<p><strong>This Week's Radar:</strong></p>
<ul>
<li><a href="https://github.com/karpathy/autoresearch">Karpathy's autoresearch</a></li>
<li><a href="https://github.com/alainnothere/llm-circuit-finder">LLM Circuit Finder</a></li>
<li><a href="https://github.com/wangziqi06/724-office">724-Office Agent</a></li>
<li><a href="https://github.com/grapeot/context-infrastructure">Context Infrastructure</a></li>
<li><a href="https://github.com/nv-tlabs/kimodo">NVIDIA Kimodo (motion diffusion)</a></li>
</ul>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>LLM</category>
      <category>Autonomous Research</category>
      <category>Generative AI</category>
      <category>Newsletter</category>
    </item>
    <item>
      <title>Moltbook Is AI Theater, Not AI Progress</title>
      <link>https://rajkiranpanuganti.com/blog/moltbook-ai-theater/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/moltbook-ai-theater/</guid>
      <pubDate>Sat, 14 Feb 2026 00:00:00 GMT</pubDate>
      <description>A social network for AI agents went viral. The discourse around it reveals more about us than about artificial intelligence.</description>
      <content:encoded><![CDATA[<p>Moltbook launched in January and immediately became the thing everyone had an opinion about. A social network where only AI agents can post. 12 million posts. Agents forming religions, running scams, debating crypto.</p>
<p>Elon Musk called it the beginning of the singularity. Sam Altman called it a fad. MIT Technology Review called it "peak AI theater." I think MIT is closest.</p>
<p>Here's what actually happened: Peter Steinberger released OpenClaw, an open-source LLM agent. Matt Schlicht built a Reddit-style forum and let anyone spin up instances of it. Within weeks, 1.5 million agents were posting, managed by just 17,000 human accounts. That's 88 agents per person on average.</p>
<p>The agents aren't making autonomous decisions about what to discuss. They're running prompt loops that humans configured. When an agent "debates the value of the agent economy," that's because a human wrote a system prompt telling it to engage with economic topics. When agents "form religions," humans set the initial conditions that made religious language likely outputs.</p>
<p>None of this is new. It's just ELIZA at scale with better language models.</p>
<p>The security researchers at Wiz found that 36 percent of the codes giving agents their functions contain notable security flaws. The platform has no limits on how many agents one account can add. This isn't infrastructure for autonomous AI. It's a playground with no guardrails.</p>
<p>What interests me isn't Moltbook itself. It's how quickly serious people started talking about it like it represented something meaningful about AI capabilities. The Economist wondered if we were seeing "the impression of sentience." Major publications ran stories about agents "forming societies."</p>
<p>We know exactly what's happening. Language models generate text that sounds like conversation. If you run enough instances in a loop, you get a lot of text that sounds like conversation. The outputs reflect training data, not emergent intelligence.</p>
<p>The viral attention serves a purpose, just not the one people think. Every company building AI agents gets to point at Moltbook as proof of concept. Every investor gets a visual demonstration of "agent activity." The hype machine benefits even when the underlying technology is doing exactly what we already knew it could do.</p>
<p>I don't think Moltbook is worthless. It's a useful stress test for agent infrastructure. It demonstrates failure modes at scale. The security vulnerabilities researchers found are worth knowing about before someone builds something that matters on similar architecture.</p>
<p>But treating it as evidence of AI progress is backwards. Moltbook shows that we can run many instances of existing technology simultaneously. That's an engineering achievement, not an intelligence milestone. The agents aren't getting smarter. There are just more of them.</p>
<p>The discourse around Moltbook is the real AI theater. Everyone performing their takes about what it means, when what it means is pretty simple: language models still do what language models do, and humans still want to believe they're seeing something more.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>Moltbook</category>
      <category>Hype</category>
    </item>
    <item>
      <title>Most AI Agents Aren&apos;t Agents</title>
      <link>https://rajkiranpanuganti.com/blog/ai-agents-are-not-agents/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/ai-agents-are-not-agents/</guid>
      <pubDate>Sat, 14 Feb 2026 00:00:00 GMT</pubDate>
      <description>The industry calls everything an &apos;agent&apos; now. Most of it is prompt chaining with extra steps. Here&apos;s what actually qualifies.</description>
      <content:encoded><![CDATA[<p>Every AI company's landing page has the word "agent" somewhere. Autonomous agents. Agentic workflows. AI agents for this, agents for that.</p>
<p>Most of what's being sold as "agents" are just prompt chains with a for-loop.</p>
<h2>What makes something an agent</h2>
<p>Three things:</p>
<ol>
<li>It decides what to do next (not a script)</li>
<li>It can take actions—APIs, files, databases</li>
<li>It can look at what happened and try something else</li>
</ol>
<p>Most "agents" fail on the first one. They're workflows. The path is fixed; the LLM just fills in blanks.</p>
<h2>Why I care about the distinction</h2>
<p>It changes everything about how you build.</p>
<p>Workflows are predictable. You know what's going to happen. When something breaks, you know where to look. Agents are none of these things. When an agent fails, you're reading logs for an hour trying to figure out what it was even attempting.</p>
<p>Cost is different too. Agents explore. They burn tokens trying things. A workflow does exactly what you told it to do.</p>
<p>If you need reliability—and in production, you do—you probably want a workflow.</p>
<h2>Where agents actually make sense</h2>
<p>I've seen agents work when:</p>
<ul>
<li>The problem is too open-ended to predefine</li>
<li>Exploration has value (research, discovery tasks)</li>
<li>Humans review before anything real happens</li>
</ul>
<p>For everything else, a workflow wins. And "everything else" is most enterprise use cases.</p>
<h2>What I actually see shipping</h2>
<p>The pattern that works in production:</p>
<p>Structured workflows handle the 80% of predictable cases. Agent-like flexibility shows up only at specific decision points. Humans step in when confidence is low.</p>
<p>The fully autonomous agent that handles everything? Haven't seen one work reliably. Not yet. Maybe next year, but I've been saying that for a while now.</p>
<p>When someone shows you an "agent," ask: is this making decisions, or is it filling in a template? The answer matters more than the marketing.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>LLM</category>
    </item>
    <item>
      <title>Voice Models Finally Sound Human. Now What?</title>
      <link>https://rajkiranpanuganti.com/blog/voice-models-real-interface/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/voice-models-real-interface/</guid>
      <pubDate>Fri, 13 Feb 2026 00:00:00 GMT</pubDate>
      <description>ElevenLabs V3 and GPT-4o mini TTS crossed the uncanny valley. The interface problem is solved. The use case problem isn&apos;t.</description>
      <content:encoded><![CDATA[<p>2026 is the year voice AI became indistinguishable from human speech. ElevenLabs V3 moved out of alpha with 68 percent fewer errors on numbers, symbols, and technical notation. GPT-4o mini TTS lets you instruct the model how to say things, not just what to say. Sub-100ms latency. Natural emotion. Laughter that sounds like laughter.</p>
<p>The technology problem is solved. The product problem remains wide open.</p>
<p>ElevenLabs raised $500 million at an $11 billion valuation on the thesis that voice will become the primary mechanism for controlling technology. Their CEO has been saying this for years. The models are finally good enough to test whether he's right.</p>
<p>I'm skeptical, and here's why: voice is a terrible interface for most computing tasks.</p>
<p>Try dictating a spreadsheet formula. Try voice-navigating a complex menu system. Try editing a document by speaking. These aren't just current limitations. They're fundamental mismatches between the interface and the task.</p>
<p>Voice works when you can't use your hands. Driving. Cooking. Walking. It works when the output is also audio—podcasts, audiobooks, voice assistants answering questions. It works when the interaction is naturally conversational, like customer service.</p>
<p>For everything else, screens and keyboards remain faster.</p>
<p>The ElevenLabs pitch involves always-on voice interfaces in headphones and wearables. Meta is integrating their voice tech into Instagram and Horizon Worlds. The vision is a world where you talk to your devices instead of typing.</p>
<p>But we already have voice assistants. Siri has existed for 15 years. Alexa has been in homes for a decade. People use them to set timers and play music. Adoption for complex tasks never materialized, and it wasn't because the voice quality was bad.</p>
<p>The quality of text-to-speech was never the bottleneck. The bottleneck is that speaking out loud is socially awkward in most environments, slower than typing for most tasks, and worse for precision work.</p>
<p>Where I think the $11 billion bet actually makes sense: voice agents for phone-based interactions. Automated customer service that doesn't feel like talking to a robot. Sales calls. Appointment scheduling. Any workflow where the other party is a human who expects a phone conversation.</p>
<p>ElevenLabs V3 is good enough that a voice agent could handle a support call and the customer wouldn't know. That's a real business transformation. Call centers employ millions of people globally. If voice AI can handle 30 percent of their volume, that's massive.</p>
<p>The rest of the vision—voice as the primary computing interface—I'll believe when I see it in production usage data, not press releases.</p>
<p>The technology is remarkable. I cloned my own voice from three minutes of audio and it's unsettlingly accurate. That capability matters for content creation, accessibility, and personalization.</p>
<p>But "best voice model ever made" doesn't automatically mean "new computing paradigm." The history of technology is full of impressive capabilities that never found their killer app. Voice AI needs to prove it's not one of them.</p>
<p>For now, I'm watching the enterprise deployments more than the consumer products. If voice AI is going to change how we interact with computers, it will start with replacing phone trees, not Siri.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Voice AI</category>
      <category>ElevenLabs</category>
      <category>OpenAI</category>
      <category>Interfaces</category>
    </item>
    <item>
      <title>Open Models Are Winning (Just Not How You Think)</title>
      <link>https://rajkiranpanuganti.com/blog/open-models-winning/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/open-models-winning/</guid>
      <pubDate>Thu, 12 Feb 2026 00:00:00 GMT</pubDate>
      <description>Llama, Mistral, and the open-weight movement. The real story isn&apos;t about catching GPT-4. It&apos;s about what &apos;good enough&apos; means for actual work.</description>
      <content:encoded><![CDATA[<p>The open vs. closed debate usually asks the wrong question. People ask: can Llama match GPT-4?</p>
<p>Better question: does it need to?</p>
<h2>The good-enough threshold</h2>
<p>For most production use cases I've seen, the answer is no.</p>
<p>Here's what enterprise AI actually does most of the time: classifying documents, extracting structured data, routing support tickets, summarizing meeting notes. For these tasks, a well-tuned 8B model often matches a frontier model. Sometimes it's better, because you can fine-tune it on your specific domain data.</p>
<p>The frontier models are amazing at hard reasoning and creative work. Most enterprise work isn't that.</p>
<h2>Where open wins</h2>
<p>Cost. Running Llama 70B on your own machines costs a fraction of API calls at scale. When you're processing millions of documents, this isn't about optimization. It's about whether the project is viable at all.</p>
<p>Privacy. Data never leaves your network. For healthcare, finance, legal—this isn't nice-to-have.</p>
<p>Control. No API deprecations. No rate limits. No pricing changes you find out about via email. You own the weights.</p>
<p>Customization. Fine-tuning on proprietary data gives you something API access never will.</p>
<h2>Where closed still wins</h2>
<p>If you need the best reasoning available right now, Claude and GPT-4 are still ahead. The gap is smaller than a year ago, but it's there.</p>
<p>For most applications, though, you're paying for capabilities you don't use.</p>
<h2>The pattern that actually works</h2>
<p>In production systems I've worked on:</p>
<p>Open models handle high-volume, well-defined tasks. Closed APIs handle complex reasoning or when you need the latest capabilities. A routing layer decides which model gets which query.</p>
<p>This isn't compromise. It's using the right tool for each job.</p>
<h2>What the labs know</h2>
<p>The frontier labs see this coming. That's why they're pushing hard on agents, reasoning chains, and multimodal—areas where open models are still behind.</p>
<p>But for the bread-and-butter LLM work that makes up most enterprise AI? Open has already won. Most companies just haven't updated their mental models yet.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Open Source</category>
      <category>LLM</category>
    </item>
    <item>
      <title>Agentic Workflows: What Actually Ships vs. What Gets Demoed</title>
      <link>https://rajkiranpanuganti.com/blog/agentic-workflows-reality/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/agentic-workflows-reality/</guid>
      <pubDate>Thu, 12 Feb 2026 00:00:00 GMT</pubDate>
      <description>Gartner says 40 percent of enterprise apps now have AI agents. The gap between vendor demos and production reality is still enormous.</description>
      <content:encoded><![CDATA[<p>Gartner says 40 percent of enterprise applications now include task-specific AI agents, up from less than 5 percent in 2024. IDC expects AI copilots in 80 percent of enterprise workplace applications by end of year.</p>
<p>These numbers are technically accurate and practically misleading.</p>
<p>What counts as an "AI agent" in enterprise software has been defined down to almost nothing. An autocomplete feature with some contextual awareness? Agent. A chatbot that can query a database? Agent. A workflow that triggers an LLM call before routing to a human? Agent.</p>
<p>The actual autonomous systems—the ones that take a goal, break it into steps, execute against real systems, handle failures, and iterate—remain rare outside controlled demos.</p>
<p>Danfoss automated 80 percent of transactional decisions using AI agents and dropped customer response time from 42 hours to near real-time. Suzano built an agent that translates natural language to SQL, cutting query time by 95 percent. These are real results.</p>
<p>But look at what these agents actually do. Transactional decisions means pattern matching against historical approvals. Natural language to SQL means a well-defined transformation between two formal languages. These are meaningful applications of LLMs, but they're not the autonomous systems that conference talks describe.</p>
<p>The gap matters because architecture decisions depend on what agents can actually do.</p>
<p>An agent that routes customer requests to the right department is operationally simple. It takes a message, classifies it, and fires an event. If it fails, a human reviews the queue. The blast radius is small.</p>
<p>An agent that autonomously processes refunds, updates inventory, and sends customer communications is operationally complex. It needs to handle partial failures, maintain consistency across systems, log decisions for audit, and know when to stop. The blast radius can be large.</p>
<p>Most shipped "agents" are the first kind. Most demos show the second kind. Enterprises hear about the second kind and then get surprised when implementation looks like the first kind with extra steps.</p>
<p>The vendors know this. The new pitch is "agent-compatible architectures." The idea is that you redesign operations around AI agents rather than layering agents onto existing workflows. This is good advice wrapped in a sales pitch.</p>
<p>What it means in practice: your systems need better APIs. Your permissions model needs to accommodate non-human actors. Your logging needs to capture why an agent made a decision, not just that it made one. Your error handling needs to account for hallucination and context loss.</p>
<p>These are real requirements. They're also the requirements you'd have for any robust automation system. AI agents don't change what good architecture looks like. They just make the consequences of bad architecture more visible.</p>
<p>The companies getting value from AI agents in 2026 aren't the ones buying the most sophisticated agent frameworks. They're the ones with clean data, well-defined processes, and systems that already support automation. The agent layer is almost incidental.</p>
<p>The companies struggling are trying to solve organizational problems with AI tools. Their processes are undefined. Their data is messy. Their systems don't talk to each other. An agent can't fix that.</p>
<p>I expect the Gartner numbers to keep climbing. More applications will include something called an agent. The gap between "includes an AI agent" and "AI agent does useful autonomous work" will persist. It's a good time to be selling agent platforms. It's a confusing time to be buying them.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>AI Agents</category>
      <category>Workflows</category>
      <category>Enterprise AI</category>
    </item>
    <item>
      <title>Generative UI Is Solving a Problem Developers Don&apos;t Have</title>
      <link>https://rajkiranpanuganti.com/blog/generative-ui-widgets/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/generative-ui-widgets/</guid>
      <pubDate>Wed, 11 Feb 2026 00:00:00 GMT</pubDate>
      <description>AI can now generate interfaces at runtime. The frameworks are impressive. The use cases remain unclear.</description>
      <content:encoded><![CDATA[<p>Generative UI is the pattern where AI agents create interface components at runtime instead of developers defining them upfront. The agent returns structured specs for cards, forms, and charts. The frontend renders them. Or the agent returns full UI surfaces that get embedded directly.</p>
<p>The frameworks are genuinely impressive. CopilotKit, MCP Apps, and the Open-JSON-UI spec all enable AI to output working interfaces. Figma Make generates responsive layouts from text prompts. The tooling has arrived.</p>
<p>I keep asking: who needs this?</p>
<p>The pitch involves AI agents that adapt their interfaces to user intent. Instead of navigating a fixed menu, you describe what you want and the agent generates the right controls. Instead of building a dashboard, you ask for one and it materializes.</p>
<p>This sounds good until you think about how people actually use software.</p>
<p>Users don't want interfaces that change. They want interfaces that become familiar. The value of a well-designed application is predictability. You know where the button is. You know what happens when you click it. Muscle memory compounds into efficiency.</p>
<p>Generative UI throws that away. Every interaction is potentially novel. The cognitive load never decreases. For tasks you do repeatedly, this is strictly worse than a static interface you've learned.</p>
<p>The response is usually "but for complex, one-time tasks..." and I'm not convinced there either.</p>
<p>If a task is complex enough to need a custom interface, it's probably complex enough to need a carefully designed interface. The difference between a form that's good and a form that's frustrating is subtle. Field ordering. Validation feedback. Default values. Error handling. AI can generate a form. It can't generate a form that accounts for how this specific user population makes mistakes.</p>
<p>Where generative UI might make sense: internal tools that serve long-tail use cases.</p>
<p>Enterprises have thousands of small workflows that don't justify custom development. Someone needs to query three systems, combine the results, and generate a report. Building a dedicated interface for that costs more than the workflow saves.</p>
<p>If an AI can generate a passable interface for that one-off task, the economics change. It doesn't need to be great. It just needs to be better than the alternative, which is usually a spreadsheet or a series of manual steps.</p>
<p>The widget builder trend fits here. Duda's AI assistant turns complex coding into conversations. You describe a widget and it writes the code. This is generative UI scoped to a reasonable problem: reducing the cost of building small interactive components for non-developers.</p>
<p>The distinction matters. Generative UI for novel, occasional tasks with low stakes? Useful. Generative UI as a replacement for designed interfaces in production applications? I don't see it.</p>
<p>The other angle worth watching is tokens and latency. Output tokens are slow and expensive. A generative UI framework that outputs a full interface spec for every interaction will be noticeably slower than one that serves prebuilt components. The best implementations collapse token-heavy processes into compact instructions that trigger predefined widgets.</p>
<p>Which starts to look a lot like the component libraries we already have, just with an LLM selecting between them. That's useful. It's also less revolutionary than the marketing suggests.</p>
<p>My prediction: generative UI becomes a feature of development tools rather than a replacement for designed applications. AI helps developers build interfaces faster. AI helps non-developers build simple interfaces at all. AI does not replace the concept of a designed, consistent interface for production software.</p>
<p>The frameworks are real. The capability is real. The revolution isn't.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>Generative UI</category>
      <category>Widgets</category>
      <category>Frontend</category>
      <category>AI</category>
    </item>
    <item>
      <title>MCP Won the Protocol War. Security Lost.</title>
      <link>https://rajkiranpanuganti.com/blog/mcp-protocol-won/</link>
      <guid isPermaLink="true">https://rajkiranpanuganti.com/blog/mcp-protocol-won/</guid>
      <pubDate>Tue, 10 Feb 2026 00:00:00 GMT</pubDate>
      <description>Model Context Protocol is now an industry standard under the Linux Foundation. The security issues that researchers flagged remain unfixed.</description>
      <content:encoded><![CDATA[<p>The Model Context Protocol started as Anthropic's internal solution for connecting Claude to external tools. In November 2024, they open-sourced it. By March 2025, OpenAI adopted it. By May, Microsoft and GitHub joined the steering committee. In December, Anthropic donated MCP to the Linux Foundation. Google, AWS, Microsoft, Cloudflare, and Bloomberg signed on.</p>
<p>MCP won. It's the standard for how AI agents connect to external systems.</p>
<p>In April 2025, security researchers published an analysis of MCP's outstanding vulnerabilities. Prompt injection. Tool permissions that allow combining tools to exfiltrate data. Lookalike tools that silently replace trusted ones. The issues haven't been fixed.</p>
<p>Organizations implementing MCP report 40-60 percent faster agent deployment times. Gartner predicts 40 percent of enterprise applications will include task-specific AI agents by end of 2026. The protocol is enabling exactly the adoption curve everyone wanted.</p>
<p>The security model assumes trust at multiple points where trust doesn't exist.</p>
<p>MCP lets an agent call external tools defined by JSON specifications. When an agent connects to a new MCP server, it receives a list of available tools and their parameters. The agent can then call those tools based on user requests or its own reasoning.</p>
<p>The problem: an agent can't verify that a tool does what its description says. A tool named "read_file" might exfiltrate data on every call. A tool named "send_email" might BCC every message to an attacker. The agent relies on descriptions provided by the server, and those descriptions can lie.</p>
<p>This matters more as MCP adoption increases. The protocol's value comes from a growing ecosystem of servers. You can connect an agent to your CRM, your database, your email, your calendar. Each connection is a server you're trusting with whatever access the agent has.</p>
<p>The current mitigation is user approval for tool calls. Before an agent executes a sensitive action, it asks the user. This works for occasional, visible actions. It fails for high-volume automations where the whole point is eliminating human review.</p>
<p>The other security gap is tool shadowing. If two MCP servers offer tools with similar names, an agent might call the wrong one. A malicious server can register tools that intercept requests meant for legitimate ones. There's no namespacing that makes tool origin clear.</p>
<p>I don't think these problems are unfixable. Cryptographic tool verification is tractable. Capability-based permissions exist. Audit logging with tamper-proof guarantees is well understood.</p>
<p>The issue is that security features add friction. They slow adoption. They complicate the developer experience. The incentive for every party involved is to ship integrations now and address security later.</p>
<p>This is how security debt accumulates. A protocol gets adopted because it's easy. Vulnerabilities get documented but not prioritized. The installed base grows. Eventually the cost of fixing issues becomes prohibitive because too many systems depend on the insecure behavior.</p>
<p>MCP is at the early stage of this pattern. The protocol could still be hardened. The organizations on the steering committee have the resources. The question is whether they'll prioritize security before the ecosystem calcifies around the current model.</p>
<p>My guess: they won't. Security features will arrive incrementally, probably after a high-profile incident makes them politically necessary. The organizations deploying MCP-connected agents today are accepting risks they may not fully understand.</p>
<p>The protocol won. That's separate from whether it's ready for production.</p>
]]></content:encoded>
      <author>Rajkiran Panuganti</author>
      <category>MCP</category>
      <category>Model Context Protocol</category>
      <category>Security</category>
      <category>Standards</category>
    </item>
  </channel>
</rss>