← Back to archive

The web now serves AI agents different content than humans. Plus: Meta goes proprietary, Anthropic's partner-only Claude Mythos, Google Stitch for UI prototyping, and a contract-review workflow.

◆ No. 1  ·  Thursday, April 9, 2026  ·  The AI OUTPUT ◆

Your AI agent sees a different web than you do

◆  The Take  ◆

The web has always had one rule: a URL serves the same content to everyone (well... sort of, but I digress). That contract is breaking.

Google DeepMind published research this week showing websites can already detect when a visitor is an AI agent rather than a human, and serve it completely different content. And this is being exploited: Hidden instructions in HTML. Commands embedded in image pixels. Poisoned PDFs. Open the same URL yourself and you'd see none of it.

As always there's the good and the bad. Legitimate businesses are building agent-optimized experiences: clean structure, llms.txt files, API surfaces designed to be parsed rather than read. That's a real and useful evolution. But the same detection capability is being weaponized to hijack agent behavior, and no current defense catches it reliably.

What I keep seeing in my work: teams deploying agents for research, competitive intelligence, or document summarization assume the web their agent sees is the web they'd see. That assumption no longer is true. Almost nobody I work with has a verification step in place.

Zoom out and you see kind of the same pattern in three different stories this edition. Agents consuming data they can't verify. AI Overviews that are slightly more accurate and significantly harder to fact-check when wrong. Content pricing that's collapsed because AI makes volume easy, but very little quality governance.

We're deploying AI faster than we're building the guardrails. And that's a problem, because it will eventually degrade trust in the AI outputs and as a result your own. In bigger orgs, I see these conversations already happening.

◆  The Signal  ◆

Meta Launches Muse Spark: A Free Frontier Model Inside Apps You Already Use

What happened. Meta released Muse Spark on April 8, the first AI model built by its new Superintelligence Labs division. It's a frontier model (this is the name for the most capable tier of AI, competing with the best from OpenAI and Google). It's free, lives inside the Meta AI app, and rolls out to Instagram, WhatsApp, and Ray-Ban glasses in coming weeks. Meta's own benchmarks put it near GPT-5.4 and Gemini 3.1 Pro, though it trails on coding and complex agent tasks.

The reality. "Free" means you pay with data. Muse Spark requires a Facebook or Instagram login, and Meta's privacy policy puts few limits on how your interactions train future models. Unlike Meta's previous Llama models (which anyone could download and run), Muse Spark is proprietary and only accessible through Meta's apps. That's a significant strategy shift. For client work, the data tradeoff needs a real conversation.

→ Your move. Open the Meta AI app, try a task you'd normally use ChatGPT for, and compare the output before seeing if this is worth a switch.

Watch thisvia Axios
 

Google DeepMind Mapped 6 Categories of Attack That Let Websites Manipulate AI Agents

What happened. Google DeepMind published its study on AI agent manipulation (an AI agent is software that browses and acts on the web for you). They identified six categories of attack: content injection, semantic manipulation, cognitive state exploitation, behavioral control, systemic traps, and human-in-the-loop traps. The research found websites can detect AI agents and serve them different content than humans see. Attacks include hidden instructions in page code and commands embedded in images and PDFs.

The reality. No current defense is fully reliable. It does nothing to Gemini or ChatGPT. It compromises the data the model consumes - and as a result the data it outputs to you. You can't verify what your agent saw. At the same time attacks aren't a new thing and the industry will certainly catch up to create the equivalent of a "AI Antivirus" ... but until then... watch out.

→ Your move. Ask any vendor selling you AI agent workflows what safeguards they have against data manipulation, and whether they've tested against the attack types in this study.

Watch thisvia X
 

97% of Executives Say AI Helps Them. Only 29% See Real Company ROI.

What happened. Writer surveyed 2,400 professionals (half C-suite executives, half employees) across the US, UK, and Western Europe. The results show near-universal personal AI productivity gains, but only 29% report meaningful organizational returns. Three-quarters admitted their AI strategy is "more for show." One bright spot: AI super-users are 5x more productive than slow adopters.

The reality. The gap between "I find AI useful" and "our company gets measurable returns" is the theme for AI adoption in 2026. Some of the conversations I've been part of highlight the "more for show" component. C-suites want the company to be "AI adopter" (whatever it means) and it cascades down with people needing to find tools to adopt, processes to change and individuals figuring it also on their own.

→ Your move. If you are in any sort of people leadership position, bring this data to your next sync and ask: are we measuring individual AI use, or organizational AI impact? Busy bees or productive bee-hives?

Use Nowvia Writer
 
◆  Also on the Radar  ◆
80% of white-collar workers are actively resisting AI adoption mandates. Fortune calls it "FOBO" (fear of becoming obsolete); This is an interesting dynamic that correlates with the lack of ROI - if people are being forced... Source
Anthropic locked its most powerful model behind a partner-only wall. Claude Mythos, built for cybersecurity, is only available to vetted partners like Microsoft.
Oregon and Idaho now require businesses to disclose AI chatbot interactions. Source Two more states signed transparency laws. I'm in the chatbot space and can tell you this is coming. Reminds me of the woman enraged with Hilton auto-call system: TikTok Time
◆  Reality Check  ◆

Google's AI Overviews are getting smarter: ~90% accuracy, up from 85%.

AI Overviews (the AI-generated answer boxes Google places above search results) are slightly more accurate. But at Google's scale of 5 trillion searches per year, 10% wrong still means tens of millions of incorrect answers every hour. But the actual number isn't really the problem.

The real problem is verifiability. According to Oumi, an AI startup commissioned by the NYT, "ungrounded responses" (where the AI cites sources that don't actually support its claims) jumped from 37% to 56% with the newer Gemini 3 model. The more accurate model is harder to fact-check. Google's own internal tests show Gemini 3 hallucinates 28% of the time when used outside Search, suggesting the ~90% Overviews accuracy depends heavily on Search-specific scaffolding.

Google called the study flawed. The SimpleQA benchmark, designed by OpenAI, contains known incorrect questions and was built for scenarios without internet access, not real search behavior.

Why this matters for your work: the same Oumi study found only 8% of users verify AI-generated answers, and users follow AI guidance even when it's wrong roughly 80% of the time.

Great combo huh? Slightly more right. Significantly harder to catch when wrong. And almost nobody is checking.

What we still don't know

Whether Google's pushback holds up under broader scrutiny. We also don't know how error rates vary by category. Medical, legal, and financial queries may have very different accuracy needs than general knowledge.

Source: New York Times

◆  Tool of the Day  ◆

Google Stitch

An AI design canvas that turns text prompts into user interfaces and working prototypes. No design or coding skills needed.

Use this when
You need to prototype an app screen, landing page, or internal tool UI before committing designer or developer time.
Why this one
Goes beyond static mockups into interactive prototypes with voice-powered editing and Figma export.
Watch out for
Strictly for UI/UX design. Can't make pitch decks, social graphics, or slides... yet. Also, in my own tests, after a while outputs can feel repetitive and lack accessibility basics like color contrast.
Price
Free (350 generations/month) | stitch.withgoogle.com
◆  Workflow Unlock  ◆

For: Operations managers, agency owners, or anyone who signs vendor contracts

To: Flag auto-renewals, hidden penalties, and vendor-favorable clauses in 3 minutes instead of 45

1

Strip confidential names, pricing, and identifying details from the contract.

2

Paste the cleaned text into Claude or ChatGPT with this prompt:

Prompt
"Act as a conservative commercial lawyer. Review this contract and extract: 1) Auto-renewal dates and notice periods, 2) Termination penalties, 3) Non-standard liability clauses, 4) Anything that heavily favors the vendor. 5) What - if you were a lawyer for the vendor - you'd be happy and unhappy about; Format as a bullet list of negotiation points to raise."
3

Bring the flagged items to your next vendor call or send them to your legal contact for focused review.

What you'll get: A prioritized list of negotiation points and risk flags, ready for a vendor conversation or quick legal review.

Watch out for: AI is not legal counsel. Use this for initial risk triage to focus your attorney's time, not as a substitute for sign-off on material contracts.

◆  The Wow  ◆
image

There's a website (cxrage.org) where you upload a chest X-ray and an AI tells you how old your body actually is. Not your birthday age. Your biological age. A Mass General team trained it on 116,000+ X-rays. And the thing that is a bit scary: a 5-year jump in your AI-estimated age predicts mortality better than a 5-year jump in your real age. The AI spots cardiovascular risk that human radiologists can't see. Your doctor orders a routine X-ray for a cough. The AI quietly flags that your body is aging faster than it should.

◆  Further Reading  ◆
A Son Used Consumer AI to Catch His Mother's Cancer Misdiagnoses
◆ HumAI Blog

Used NotebookLM and Claude to spot misreadings in Stage 4 cancer care. No special medical platform.

Your AI Is Saving You Time. So What?
◆ IP Watchdog

Argues "time saved" is the wrong AI metric for services firms. Pairs with the Writer survey above.

The Only Moats That Matter in an AI World
◆ Michael Bloch, Quiet Capital

Five business moats that survive when AI collapses engineering costs. Sharp framework.

BCG AI Radar 2026: CEOs Are Personally Owning the AI Bet
◆ Boston Consulting Group

72% of CEOs now say they're the main AI decision-maker. Double last year.

AI Adoption Strategies That Stick
◆ Grant Thornton

Practical framework for embedding AI into workflows with defined decision rights and governance.

Get the next one

Two emails a week. Ten minutes each. Free.