
Welcome back, revolutionary…
The AI search (until we can all agree it's AEO) community got its first real look at how an AI tool actually picks up and processes web content.
A packaging mistake by Anthropic exposed the retrieval pipeline inside Claude Code, their developer CLI. Meanwhile, Microsoft opened an official window into AI citations.
In this week's edition:
107 websites get the VIP lane into Claude Code (they're all developer docs). Everyone else gets regurgitated in 125 characters.
Google goes fully open source with its most capable model family yet
The first real dashboard for tracking whether AI is citing your content
What one year of Liberation Day tariffs has done to AI infrastructure costs
OpenAI's $852 billion valuation, 900 million weekly users, and the super app play
Anthropic's Claude Code Source Leaked. Here's What the Retrieval Pipeline Looks Like.
On March 31, source maps were accidentally included in a Claude Code CLI update on npm, exposing enough of the TypeScript source to reconstruct the architecture. Claude Code is Anthropic's developer tool, a CLI that helps programmers write and debug code. It's not the same product as claude.ai (the chatbot). The r/ClaudeAI megathread that followed is the most detailed public teardown of how any AI tool retrieves and processes web content. We wrote a full breakdown of what the source reveals.
Need to know:
107 pre-approved domains get full content extraction. They're all developer documentation sites (React, Django, PostgreSQL, Tailwind). That makes sense for a coding tool. Everyone else gets their HTML converted to markdown, then paraphrased by a smaller model with a 125-character quote limit.
JSON-LD, FAQ schema, Open Graph tags: all stripped during processing. Everything in
<head>is discarded before the model sees the page. The leak showed the content pipeline, not the search index (which is powered by Brave), so schema might still influence which pages surface. But the model never reads it.Anthropic is running silent A/B tests on Claude Code session limits via Statsig. Half of users are in one experiment group, half in another. No announcement, no changelog.
This is Claude Code, not claude.ai. The two share a search index but use different content pipelines (claude.ai works from encrypted snippets rather than re-fetching pages). The pre-approved list and compression pipeline are specific to the developer tool. But this is the first time anyone has seen how any AI tool's retrieval actually works at source-code level, and the 3 findings are worth understanding.
Shift
Google Goes Apache 2.0 With Gemma 4
Google released Gemma 4 on April 2: 4 model sizes from 5B to 31B parameters, all under the Apache 2.0 license. That's the first time Google has used a fully permissive open-source license for the Gemma family, removing the commercial restrictions that kept enterprise teams away from earlier versions.
Need to know:
The lineup spans edge to data center: 5B dense, 8B dense, 26B mixture-of-experts (4B active), and 31B dense. Context windows up to 256K tokens. Native video, image, and audio input on select models. Trained on 140+ languages.
Apache 2.0 means no usage restrictions, no registration, no reporting requirements. You can fine-tune, deploy commercially, and distribute without asking Google's permission. The previous Gemma license had restrictions that blocked some enterprise and government deployments.
The 26B MoE model is the interesting one for small operators: only 4B parameters active per inference pass, which means strong performance at a fraction of the compute cost. It runs on consumer hardware.
A year ago, running a capable model locally meant navigating license restrictions and settling for weaker performance. Google just removed both obstacles. The 26B MoE model on consumer hardware puts genuine AI capability within reach of a solo operator's budget, no API bills, no vendor lock-in.
Market Signal
Bing Gives Publishers the First Real AI Citation Dashboard
Microsoft launched AI Performance in Bing Webmaster Tools in February. It's the first analytics tool from a major platform that shows whether AI systems are actually citing your content.
Need to know:
The dashboard tracks when your URLs appear as sources in AI-generated answers across Microsoft Copilot, Bing AI summaries, and partner integrations. You can see which pages get cited, which queries trigger citations, and how citation activity changes over time.
The queries shown aren't what users type. They're "grounding queries," shorter topical queries the AI generates internally to retrieve fresh information from the web. The gap between what a user asks and what the system searches for is part of what makes AI visibility harder to track than traditional SEO.
Google has no equivalent. Google Search Console shows impressions and clicks for traditional search and AI Overviews, but doesn't break out citation data for AI-generated answers specifically.
Until now, AI citation performance has been unmeasurable. You published and hoped. Microsoft just made it measurable, at least within its own ecosystem. Google has no equivalent yet, but the pressure to ship one just increased considerably.
Round Up: Money, Silicon, and Who Controls the Stack
The largest private funding round in history. SoftBank co-led, Amazon committed $50 billion (with $35 billion contingent on IPO or AGI), NVIDIA and SoftBank each put in $30 billion. OpenAI reports $2 billion in monthly revenue and 900 million weekly active users. The ChatGPT "super app" now bundles chat, code, search, and agent capabilities into a single product.
Anthropic's Model Context Protocol hit 97 million installs on March 25, 16 months after launch. OpenAI, Google, Cohere, and Mistral all now ship MCP-compatible tooling. Anthropic donated the protocol to the Linux Foundation's Agentic AI Foundation, with OpenAI and Block as co-founders.
MTIA 300 through 500, built on RISC-V and manufactured by TSMC. New chip every 6 months through 2027. The MTIA 500 delivers 25x the compute FLOPs of the 300. Meta is cutting its NVIDIA dependency for inference workloads while the hyperscalers race to own their own silicon.
April 2 marked one year since the sweeping tariff announcement. Manufacturing lost 100,000 jobs since January 2025. The goods trade deficit rose 2% to $1.24 trillion, the opposite of the stated goal. CSIS estimates the tariffs will add $75-100 billion in AI infrastructure costs over 5 years, equivalent to 15-20 fewer hyperscale data centers. The Supreme Court ruled the emergency powers basis illegal in February, but policy has changed more than 50 times since.
SE Ranking tracked 1.3 million citations across 68,000 keywords. Google's self-citation rate: 17.42%, up from 5.7% in June 2025. Travel is the worst hit: 53% of AI Mode citations point back to Google properties. 59% of those self-citations now funnel users back to traditional search results, where the ads live.
Tool Shed
Littlebird: Recalls anything from your screen or meetings using on-device AI. No cloud upload required.
Venn.ai: Turns natural language chat into secure actions inside business systems (CRM, databases, internal tools).
ClipTask: Records screen workflows and converts them into structured task lists with steps and screenshots.
ElevenCreative: ElevenLabs' production suite unifying audio, video, image generation, and localization in one stack.
MCPCore: Ship and host MCP servers without managing infrastructure. Relevant if you're building agent integrations.
Fireflies.ai: Transcribes meetings, extracts action items, and routes them to project management tools automatically.
Quick Bytes
Anthropic's Claude Mythos (codename Capybara), a tier above Opus, was accidentally documented in a public data store before the company could announce it. Testing is underway with early access customers.
OpenAI's ChatGPT super app now processes 2.5 billion prompts per day, with general research accounting for 36% of all usage.
AI startups reported 10-20% increases in hardware procurement costs in Q1 2026 vs Q4 2025, driven by tariff uncertainty on GPU imports.
AWS hiked EC2 GPU instance prices by roughly 15% in January 2026, the first significant cloud compute increase tied to hardware costs.
