
Welcome back. The number of strange things happening in AI is ridiculous. Yesterday, someone built an actual Indiana Jones-style whip for Claude Code (github repo here). And today, Milla Jovovich — yes, actress from the Resident Evil movies — announced that she has built an agentic memory tool that scored 100% on LongMemEval (see video). Wild times.
Also: How to make RAG 32x more memory efficient, building self-improving LLM knowledge bases, and a look at how OpenAI’s Codex team actually ships.
Today’s Insights
Powerful new updates and hacks for devs
Why Claude Code's real edge was never the model
How to feed Claude Code skills real-time data
Trending social posts, top repos, and more

TODAY IN PROGRAMMING

Made with Midjourney.
OpenAI, Anthropic, and Google team up to fight model copying: These three AI rivals are teaming up through the Frontier Model Forum to combat “adversarial distillation”, a process where unauthorized users harvest responses from US models to build cheaper knockoffs. OpenAI has accused DeepSeek of piggybacking on its tech, while Anthropic has already cut off access for Chinese-controlled firms after catching three labs doing just that. US officials estimate this practice drains billions of dollars from Silicon Valley labs every year.
Anthropic secures massive compute expansion with Google and Broadcom: The AI lab recently landed a multi-gigawatt deal for next-gen TPU capacity, set to go live in 2027. This marks the company’s largest compute commitment to date, fueled by explosive growth: run-rate revenue has tripled since late 2025 to $30B, while the number of enterprise customers spending over $1M annually has doubled to more than 1,000 in just two months.
Meta plans open-source versions of its next frontier models: The social media giant is working on open-source versions of its next two proprietary models, codenamed Avocado (an LLM) and Mango (an image and video generator). While Bloomberg reported last December that the company was pivoting to a fully closed-source approach, those plans have changed. These open-source editions will likely be lighter versions with fewer features or parameters, partly due to AI safety concerns. There's no word yet on a release date.

PRESENTED BY WISPR
Your dev stack got an AI upgrade everywhere except the input layer. You're still typing every prompt, every ticket, every review comment by hand.
Wispr Flow closes that gap. Dictate into Cursor, VS Code, Slack, Linear, or anywhere else you work. It's syntax-aware: camelCase, snake_case, acronyms, and file names all come through clean. Mention a file in Cursor or Windsurf, and it auto-tags.
It's the voice layer for an AI-native workflow. Speak your intent. Your tools do the rest.
Available on Mac, Windows, iPhone, and Android. Used by millions of developers, including teams at OpenAI and Mercury.

INSIGHT
The Claude Code leak proves harnesses are the next big thing — not models
The $2.5B code dump. Last week, Anthropic accidentally leaked Claude Code's entire source code via an npm update, exposing half a million lines of TypeScript. But the leak revealed a bigger truth: the model isn't the secret sauce.
Plumbing beats parameters. ML researchers like Sebastian Raschka found that Claude Code's performance comes from its engineering, not just the underlying model. It relies on smart features like file history tracking to save context, specialized navigation tools, and parallel agents for background tasks. You'd likely get similar results using DeepSeek or Kimi with the same setup.
The wrapper is the new winner. Others have already proven this. LangChain climbed to the top five on benchmarks just by improving its harness, while Vercel saw better results by simplifying its agent's tools. The real competition has shifted from model quality to the quality of the surrounding infrastructure.
Blueprints aren't buildings. While every competitor now has the map to a multi-billion dollar product, execution remains the hurdle. Knowing how the architecture works and actually scaling it for production are two completely different challenges.

IN THE KNOW
What’s trending on socials and headlines

Meme of the day.
Faster RAG: How Perplexity, Azure, and HubSpot use a simple technique to make RAG 32x more memory efficient (explained with code).
External Brain: Inspired by Andrej Karpathy, this thread walks you through building your own LLM knowledge base that compounds over time.
Inside Codex: OpenAI's Head of DevRel and Codex product lead go behind the scenes on how the team actually ships products.
Agent Anatomy: When your AI agent fails, the problem is rarely the model. This deep dive shows how to turn a stateless LLM into a production agent.
Leaked Playbook: Claude Code's source code was accidentally leaked, and here's a breakdown of how it constructs system prompts
Docs as Files: Most AI code fails when docs change. This tool turns live documentation into an accessible filesystem that agents can browse.
Perfect Timing: OpenAI announced a new Safety Fellowship, hours after a New Yorker investigation revealed it had dissolved its safety teams.

AI CODING HACK
How to feed Claude Code skills real-time data
Claude Code skills are normally static, which usually means tedious manual copy-pasting for PR diffs or logs. Lydia Hallie, an engineer on the Claude Code team, recently shared a much better way: use the ! backtick syntax to embed shell commands directly into your SKILL.md.
Claude Code will execute the commands and swap in the live output automatically before processing. To get started, create this file at .claude/skills/pr-summary/SKILL.md:
---
name: pr-summary
description: Summarize changes in a pull request
---
- PR diff: !`gh pr diff`
- PR comments: !`gh pr view --comments`
- Changed files: !`gh pr diff --name-only`
Summarize this pull request...This syntax works with any shell command, allowing you to pull in test results, grep for TODOs, or fetch API responses automatically. To prevent accidental state changes, stick to read-only commands like cat, grep, and gh pr view.

TOP & TRENDING RESOURCES
Top Tutorial
How to build your own AI agent from scratch: In this tutorial, you'll learn how to build a custom AI meeting assistant with the Mastra TypeScript framework. You'll see how to bridge core agent concepts like tools, memory, and event triggers by integrating Slack, Exa AI, and Cal.com into one seamless, automated workflow.
Top Repo
Graphify: This AI coding agent skill turns any folder of code, docs, papers, or images into a queryable knowledge graph. It gives you a full system context so your agent can answer complex architecture questions with precision.
Trending Paper
Cursor doubles AI coding speed on Blackwell GPUs: Traditional MoE models waste processing time managing data around individual experts during single-token generation. By assigning computing power directly to outputs instead, warp decode skips these extra steps, making inference 1.8x faster and more accurate.
Grow customers & revenue: Join companies like Google, IBM, and Datadog. Showcase your product to our 200K+ engineers and 100K+ followers on socials. Get in touch.
What did you think of today's newsletter?
You can also reply directly to this email if you have suggestions, feedback, or questions.
Until next time — The Code team



