# Cerebras IPO Pops 68%, Claude Overtakes ChatGPT in Enterprise, and a Supply Chain Worm Breaks SLSA

*Published Friday · May 15, 2026*

Thursday, May 14, 2026. Your daily dose of what matters in AI, curated for business leaders.

The enterprise AI market just got its clearest scoreboard update of the year — and the results are surprising. Claude has quietly overtaken ChatGPT in U.S. business adoption, Cerebras just pulled off the biggest AI infrastructure IPO since Nvidia's heyday, and a self-propagating worm broke one of the foundational assumptions of modern software supply chain security. The throughline: the agentic era is here, it's generating real revenue, and the risks are scaling just as fast as the opportunities.

This edition covers twelve stories across infrastructure, enterprise, security, and agentic AI. Let's get into it.

## Today's Stories

- 🏢 **[Claude Overtakes ChatGPT in U.S. Business Adoption for the First Time](https://ramp.com/leading-indicators/ai-index-may-2026)** — Ramp's May 2026 AI Index, drawn from $100B+ in annual spend across 50,000 businesses, shows Anthropic at 34.4% business adoption versus OpenAI at 32.3% — the first time Claude has led. The engine is Claude Code, which has been growing 80× per year and now accounts for an estimated $2.5B run-rate revenue. The counterpoint is sharp: Ramp's own data shows Anthropic's token-based pricing is creating sticker shock, with Uber already blowing its entire 2026 AI budget. Enterprise buyers should watch adoption curves and cost curves with equal intensity.

- 💰 **[Cerebras Raises $5.55B in Blockbuster Nasdaq IPO, Shares Pop 68%](https://www.cnbc.com/2026/05/14/cerebras-cbrs-stock-trade-nasdaq-ipo.html)** — Cerebras priced at $185/share, above its elevated range, and closed its first trading day up 68% for a market cap near $100B. Revenue hit $510M in 2025 (up 76% YoY) with $88M net income — a dramatic swing from a $481M loss the prior year. A $20B+ cloud deal with OpenAI anchors near-term demand. This is the market's loudest signal yet that inference compute is real, profitable demand, not speculative — and executives should expect continued upward pressure on inference costs as hardware suppliers gain pricing power.

- 🔐 **["Mini Shai-Hulud" Worm Hits TanStack, Mistral AI, and OpenAI — First Malware with Valid SLSA Provenance](https://www.bleepingcomputer.com/news/security/openai-confirms-security-breach-in-tanstack-supply-chain-attack/)** — Threat group TeamPCP deployed a self-propagating worm through a chained GitHub Actions exploit that compromised 170+ npm and PyPI packages including @tanstack (12M+ weekly downloads) and @mistralai. OpenAI confirmed two employee devices were breached and is rotating macOS code-signing certificates. What makes this historically significant: it is the first documented worm to publish packages with valid SLSA Build Level 3 cryptographic provenance — meaning standard supply chain attestation checks would have passed. If your pipeline ran npm install on May 11, rotate credentials now.

- 🏢 **[Anthropic Launches Claude for Small Business with 27 Pre-Built Workflows](https://anthropic.com/claude-for-small-business)** — Anthropic released a pre-built package of 27 agentic workflows targeting SMBs and legal teams, with connectors for QuickBooks, PayPal, HubSpot, and Google Workspace. The timing is deliberate: it arrives the same week Claude crossed OpenAI in business adoption. For consultants and resellers, this opens a new mid-market channel with pre-packaged ROI stories tied to familiar software — and for Anthropic, it's a methodical play to close the distribution gap that has historically been OpenAI's moat.

- 🏢 **[Meta Launches WhatsApp "Incognito Chat" — AI Conversations That Even Meta Can't Read](https://about.fb.com/news/2026/05/incognito-chat-whatsapp-meta-ai/)** — Meta launched a TEE-based private AI mode on WhatsApp using AMD SEV-SNP and NVIDIA H100 confidential computing hardware, powered by the new Muse Spark model. Conversations are not logged and disappear by default — Zuckerberg called it "the first major AI product with no server-side conversation log." Enterprises considering AI for HR, legal, or health use-cases should watch this architecture closely; confidential computing in AI pipelines may become a compliance-driven requirement. The tension: if Meta can't read the logs, it also loses its safety net for detecting harm.

- ⚡ **[Google Reveals "Googlebook" — Its First Laptop in 15 Years, Built Around Gemini](https://www.youtube.com/watch?v=VUthq-JuxxE)** — Google unveiled an original laptop concept engineered ground-up around the Gemini model family, featuring "Magic Pointer" — an AI cursor that summons Gemini anywhere on-screen without switching apps. The device is less about specs and more about a strategic land-grab: the company that owns your AI assistant wants to own your device too. For IT buyers, this introduces a credible Google alternative to Apple in the enterprise laptop segment, with deep Workspace integration baked in at the firmware level.

- 🤖 **[Notion Becomes an AI Agent Hub with New Automation Platform](https://www.notion.so/agents)** — Notion launched an agent orchestration layer that turns its workspace into a platform for multi-step automated workflows across connected tools — positioning directly against Microsoft Copilot and Atlassian's Rovo. For enterprise buyers evaluating agent platforms, the Notion play is significant because it lowers activation energy: teams already living in Notion can now build agents without a separate orchestration tool or developer resources. Every major SaaS platform is now racing to become an agent host, not just an AI feature recipient.

- 🏢 **[Anthropic Reverses "OpenClaw" Ban — Walks Back Restriction on Third-Party CLI](https://www.theresanaiforthat.com)** — Anthropic reversed a policy blocking OpenClaw, a popular open-source CLI for Claude, after sharp developer community backlash over anticompetitive optics. The reversal came within days of the original restriction. A positive signal for developer trust, but the episode highlights an uncomfortable reality: API-dependent enterprises have no contractual guarantee that third-party tooling won't be cut off overnight. Monitor ecosystem stability alongside model capabilities.

- 🏢 **[Ben Thompson: "The Inference Shift" — Why Agentic Compute Changes Chip Strategy](https://stratechery.com/2026/the-inference-shift/)** — Thompson argues that agentic inference is architecturally distinct from human-facing inference — agents don't need low latency, so the economics and hardware requirements change fundamentally. He uses the Cerebras IPO to predict GPU dominance will give way to increasingly heterogeneous, specialized inference silicon. If he's right, enterprises building agent infrastructure should be optimizing for throughput and cost-per-task rather than latency — a procurement shift that's probably not on most IT roadmaps yet.

- 🤖 **[Screen-Recording as Agent Feedback: A New Workflow Pattern Emerges](https://www.bensbites.com/p/agents-feedback-tip)** — Ben Tossell shared a practical agentic technique: instead of typing feedback, screen-record yourself narrating what you want changed and pass the video directly to the agent. The agent extracts frames, interprets visual context, generates GIFs of specific interactions, and creates a self-assigned action checklist. It's token-heavy but produces richer, more actionable feedback than text alone — and the output doubles as documentation. Worth testing on your next sprint cycle.

- 📜 **[Princeton Adds AI Exam Surveillance After 133 Years of Honor Code](https://www.princeton.edu/news)** — Princeton is introducing proctoring and AI detection measures, ending a 133-year tradition of operating entirely on an honor code. The move is a useful proxy for enterprise knowledge work: when the honor system breaks down, institutions add monitoring. Organizations without clear AI-use policies should define acceptable AI use proactively before being forced into surveillance mode — the smarter path is always policy before policing.

- 🧪 **[Karpathy at Sequoia AI Ascent: "Software 3.0" — AI Automates What You Can Verify](https://karpathy.bearblog.dev/sequoia-ascent-2026/)** — Karpathy introduced a framework that should be required reading for enterprise AI strategy: Software 3.0 uses context windows as the programming surface, with agents as the interpreter. His central insight — "LLMs can automate what you can verify" — is the most practically useful mental model circulating in AI right now. He disclosed that his own coding ratio flipped from 80% self-written to 80% agent-delegated in December 2025. Audit your workflows through this lens before setting AI budgets: tasks with clear success criteria are automation-ready now; tasks requiring undefined human judgment remain stubbornly manual.

## One Thing to Think About

Today's stories form a single narrative arc in three acts. Act One: Anthropic wins the enterprise — Claude overtakes ChatGPT, ships 27 SMB workflows, developers flock to Claude Code. Act Two: the bill arrives immediately — Uber blows its AI budget on tokens, a supply chain worm with valid cryptographic signatures breaches Mistral's and OpenAI's packages, and Anthropic's own ecosystem wobbles when it briefly bans a popular CLI. Act Three: the Cerebras IPO at $100B market cap tells you where the real money flows in an agentic world — inference hardware. Karpathy's "automate what you can verify" framework is the filter every enterprise buyer needs right now. The companies compounding productivity are the ones that have mapped their workflows into verifiable and non-verifiable buckets, staffed agents on the first, and kept humans on the second. The dangerous middle ground — deploying agents on tasks where neither the team nor the vendor can define what "correct" looks like — is where the expensive failures will happen in 2026.

## Resources Worth Your Time

- **[Stratechery — "The Inference Shift"](https://stratechery.com/2026/the-inference-shift/)** — Ben Thompson's best recent essay: why agentic compute breaks current chip economics and what it means for enterprise infrastructure buyers. Essential context for anyone evaluating cloud contracts.
- **[Sequoia AI Ascent 2026: Andrej Karpathy on "Software 3.0"](https://karpathy.bearblog.dev/sequoia-ascent-2026/)** — The "automate what you can verify" framework is the most practically useful mental model in AI right now — worth 30 minutes of any executive's time.
- **[Ramp AI Index — May 2026](https://ramp.com/leading-indicators/ai-index-may-2026)** — The primary data behind today's lead story; the analyst note on Anthropic's fragile lead and rising token costs is worth reading carefully before your next vendor negotiation.

*Curated by your AI briefing assistant for Chiel Hendriks.*
