About 4o4 / AI
The AI press is broken. Same story, thirty outlets, no signal to weight. We rebuilt the stack from scratch.
4o4 / AI is an AI news publication that does what news used to do: it tells you which stories are real, which are echoes, and which are quiet but matter. We track 376 sources — frontier labs, academic institutes, safety researchers, trade press, newsletters, podcasts, even Discord channels — and weight them by what they actually are.
Most AI coverage today is pasted press releases reordered into clickbait. We don't reorder. We cluster, weight, and synthesize.
How we weight sources
Every source sits in a tier. The tier reflects what kind of authority the source carries, not how popular it is.
Five OpenAI re-blogs ≠ five distinct outlets covering it. Five distinct outlets, three of them T1, with the lab itself absent? That's a real scoop.
The signals we extract
Every story gets tagged. Each tag tells you something different about the shape of the coverage.
- Official
- The lab posted it. You're reading it directly from source.
- Scoop
- T1 + T2 outlets independently confirm a story the lab hasn't acknowledged. Strong signal.
- Analysis
- Synthesis or commentary on a known development. The take, not the news.
- Underrated
- T0/T1 sources have it; the trade press hasn't picked it up yet. Read this first.
- Breaking
- Velocity spike — cluster doubled in two hours. Story is unfolding.
- Context
- Background on why a current story matters. Useful when the news cycle assumes you've been paying attention.
What we don't do
- No sponsored content. No "sponsored by" sections. No affiliate links.
- No engagement bait. No 'Top 10 AI tools you NEED in 2026' lists.
- No re-treads of yesterday's news with a different headline.
- No claims without source links. Every fact is auditable.
How the pipeline works
RSS, Atom, and a long tail of per-source adapters (sitemaps, YouTube, Reddit JSON, arXiv, Hugging Face Papers) feed a normalised store. New items are embedded and clustered against a 72-hour rolling window. Clusters score by weighted source count, canonical-source presence, velocity, and recency. A pass with Claude Haiku triages; surviving clusters go to Sonnet for synthesis. Every published piece links back to every source it draws from.
It is, as much as we can make it, a machine for distinguishing noise from signal.