<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"><channel><title>4o4 / AI</title><description>AI news, weighted across 376 sources. We track what frontier labs ship, what analysts flag, and what the field actually thinks — minus the hype.</description><link>https://ai.4o4.app/</link><language>en-us</language><item><title>Apple at ICLR 2026: RNN parallelization, tool-augmented SSMs, unified image models, and more</title><link>https://ai.4o4.app/posts/apple-machine-learning-research-at-iclr-2026/</link><guid isPermaLink="true">https://ai.4o4.app/posts/apple-machine-learning-research-at-iclr-2026/</guid><description>Apple presents five research highlights at ICLR 2026 in Rio de Janeiro, covering parallel RNN training, SSM length generalization, unified image understanding and generation, real-time 3D scene synthesis, and a new protein folding approach.</description><pubDate>Sat, 25 Apr 2026 04:04:59 GMT</pubDate><category>frontier</category><category>iclr</category><category>apple</category><category>rnn</category><category>ssm</category><category>multimodal</category></item><item><title>ParaRNN: Apple researchers train a 7B-parameter nonlinear RNN, competitive with transformers</title><link>https://ai.4o4.app/posts/pararnn-large-scale-nonlinear-rnns-trainable-in-parallel/</link><guid isPermaLink="true">https://ai.4o4.app/posts/pararnn-large-scale-nonlinear-rnns-trainable-in-parallel/</guid><description>Apple&apos;s ParaRNN framework uses Newton&apos;s method to parallelize nonlinear RNN training, achieving a 665x speedup over sequential approaches and enabling billion-scale RNN language models for the first time.</description><pubDate>Sat, 25 Apr 2026 03:56:59 GMT</pubDate><category>frontier</category><category>rnn</category><category>training</category><category>parallelization</category><category>sequence-modeling</category><category>iclr</category></item><item><title>Apple research generates realistic long-term motion with a 64x temporally compressed embedding</title><link>https://ai.4o4.app/posts/learning-long-term-motion-embeddings-for-efficient-kinematics-generation/</link><guid isPermaLink="true">https://ai.4o4.app/posts/learning-long-term-motion-embeddings-for-efficient-kinematics-generation/</guid><description>Apple ML researchers model scene dynamics by working directly in a compressed motion embedding space derived from large-scale trajectory data, enabling efficient motion generation conditioned on text or spatial inputs.</description><pubDate>Sat, 25 Apr 2026 03:48:59 GMT</pubDate><category>frontier</category><category>motion</category><category>generation</category><category>embeddings</category><category>kinematics</category><category>video</category></item><item><title>End-to-end FP8 in RL training: NeMo RL achieves 48% speedup over BF16 baseline</title><link>https://ai.4o4.app/posts/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precisio/</link><guid isPermaLink="true">https://ai.4o4.app/posts/run-high-throughput-reinforcement-learning-training-with-end-to-end-fp8-precisio/</guid><description>NVIDIA NeMo RL applies FP8 precision across both generation and training phases of reinforcement learning, closing accuracy gaps via importance sampling and adding 48% total speedup when KV cache and attention are also quantized.</description><pubDate>Sat, 25 Apr 2026 03:40:59 GMT</pubDate><category>hardware</category><category>fp8</category><category>reinforcement-learning</category><category>training</category><category>quantization</category><category>nemo</category></item><item><title>NVIDIA adds Muon optimizer support to Megatron Core, closes gap with AdamW at scale</title><link>https://ai.4o4.app/posts/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/</link><guid isPermaLink="true">https://ai.4o4.app/posts/advancing-emerging-optimizers-for-accelerated-llm-training-with-nvidia-megatron/</guid><description>NVIDIA has integrated the Muon higher-order optimizer into Megatron Core and NeMo, showing minimal throughput loss versus AdamW on GB300 hardware while enabling training at thousands of GPUs.</description><pubDate>Sat, 25 Apr 2026 03:32:59 GMT</pubDate><category>hardware</category><category>training</category><category>optimizer</category><category>megatron</category><category>muon</category><category>gpu</category></item><item><title>Three LLM agents wrote 600,000 lines of code and ran 850 experiments to win a Kaggle competition</title><link>https://ai.4o4.app/posts/winning-a-kaggle-competition-with-generative-ai-assisted-coding/</link><guid isPermaLink="true">https://ai.4o4.app/posts/winning-a-kaggle-competition-with-generative-ai-assisted-coding/</guid><description>A first-place finish in the March 2026 Kaggle Playground churn prediction competition came from a four-level stack of 150 models selected from 850 runs, generated by GPT-5.4 Pro, Gemini 3.1 Pro, and Claude Opus 4.6 in a human-in-the-loop workflow with GPU-accelerated execution.</description><pubDate>Sat, 25 Apr 2026 03:24:59 GMT</pubDate><category>frontier</category><category>kaggle</category><category>llm-agents</category><category>tabular-ml</category><category>gpu</category><category>automation</category></item><item><title>NVIDIA FLARE reduces federated learning migration to ~5 lines of code and an environment swap</title><link>https://ai.4o4.app/posts/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/</link><guid isPermaLink="true">https://ai.4o4.app/posts/federated-learning-without-the-refactoring-overhead-using-nvidia-flare/</guid><description>The latest NVIDIA FLARE API splits federated learning adoption into two steps: a minimal client API that adds federation to existing training scripts without restructuring them, and portable job recipes that run unchanged from simulation to production.</description><pubDate>Sat, 25 Apr 2026 03:16:59 GMT</pubDate><category>hardware</category><category>federated-learning</category><category>nvidia</category><category>privacy</category><category>distributed-training</category><category>mlops</category></item><item><title>NVIDIA Blackwell delivers 150+ tokens/sec/user on DeepSeek-V4-Pro out of the box</title><link>https://ai.4o4.app/posts/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/</link><guid isPermaLink="true">https://ai.4o4.app/posts/build-with-deepseek-v4-using-nvidia-blackwell-and-gpu-accelerated-endpoints/</guid><description>NVIDIA outlines how its Blackwell platform and NIM microservices support DeepSeek-V4&apos;s million-token context requirements, with initial GB200 NVL72 benchmarks and deployment paths via SGLang, vLLM, and hosted endpoints at build.nvidia.com.</description><pubDate>Sat, 25 Apr 2026 03:08:59 GMT</pubDate><category>hardware</category><category>nvidia</category><category>deepseek</category><category>blackwell</category><category>inference</category><category>gpu</category></item><item><title>Open AI systems have a structural edge in cybersecurity defense — here is why</title><link>https://ai.4o4.app/posts/ai-and-the-future-of-cybersecurity-why-openness-matters/</link><guid isPermaLink="true">https://ai.4o4.app/posts/ai-and-the-future-of-cybersecurity-why-openness-matters/</guid><description>A Hugging Face analysis argues that AI-powered vulnerability discovery favors open, distributed systems over closed ones: openness distributes detection and patching across communities, while closed codebases concentrate both the attack surface and the remediation bottleneck.</description><pubDate>Sat, 25 Apr 2026 03:00:59 GMT</pubDate><category>safety</category><category>cybersecurity</category><category>open-source</category><category>ai-agents</category><category>vulnerability</category><category>policy</category></item><item><title>QIMMA validates Arabic benchmarks before running models on them — and finds systematic problems in established datasets</title><link>https://ai.4o4.app/posts/qimma-a-quality-first-arabic-llm-leaderboard/</link><guid isPermaLink="true">https://ai.4o4.app/posts/qimma-a-quality-first-arabic-llm-leaderboard/</guid><description>Researchers from TII UAE built QIMMA, the only Arabic LLM leaderboard combining quality validation, native content, and code evaluation. A two-stage pipeline of LLM scoring and human review revealed recurring quality failures across widely-used Arabic benchmarks.</description><pubDate>Sat, 25 Apr 2026 02:52:59 GMT</pubDate><category>opensource</category><category>arabic-nlp</category><category>evaluation</category><category>benchmarks</category><category>leaderboard</category><category>multilingual</category></item><item><title>Gemma 4 runs as a vision-language-action agent on an 8 GB Jetson Orin Nano Super</title><link>https://ai.4o4.app/posts/gemma-4-vla-demo-on-jetson-orin-nano-super/</link><guid isPermaLink="true">https://ai.4o4.app/posts/gemma-4-vla-demo-on-jetson-orin-nano-super/</guid><description>A step-by-step demo shows Gemma 4 handling speech input, autonomous webcam activation, and spoken output on NVIDIA&apos;s Jetson Orin Nano Super using llama.cpp, Parakeet STT, and Kokoro TTS — no keyword triggers, no hardcoded logic.</description><pubDate>Sat, 25 Apr 2026 02:44:59 GMT</pubDate><category>opensource</category><category>gemma</category><category>edge-ai</category><category>robotics</category><category>jetson</category><category>vla</category></item><item><title>Running Transformers.js in a Chrome extension: the Manifest V3 architecture that actually works</title><link>https://ai.4o4.app/posts/how-to-use-transformers-js-in-a-chrome-extension/</link><guid isPermaLink="true">https://ai.4o4.app/posts/how-to-use-transformers-js-in-a-chrome-extension/</guid><description>A practical walkthrough of how to host Transformers.js models in a Chrome extension background service worker under Manifest V3, covering runtime separation, messaging contracts, model caching, and the agent tool-execution loop.</description><pubDate>Sat, 25 Apr 2026 02:36:59 GMT</pubDate><category>opensource</category><category>transformers-js</category><category>chrome-extension</category><category>webgpu</category><category>inference</category><category>javascript</category></item><item><title>DeepSeek-V4 cuts KV cache to 2% of standard cost to make million-token agent context practical</title><link>https://ai.4o4.app/posts/deepseek-v4-a-million-token-context-that-agents-can-actually-use/</link><guid isPermaLink="true">https://ai.4o4.app/posts/deepseek-v4-a-million-token-context-that-agents-can-actually-use/</guid><description>DeepSeek-V4 combines two new attention mechanisms with agent-specific post-training to reduce KV cache memory to roughly 2% of a standard grouped-query-attention architecture, targeting long-horizon agentic workloads over chat.</description><pubDate>Sat, 25 Apr 2026 02:28:59 GMT</pubDate><category>opensource</category><category>deepseek</category><category>agents</category><category>long-context</category><category>architecture</category><category>moe</category></item><item><title>ADeLe scores models and tasks on the same 18-ability scale to predict performance before deployment</title><link>https://ai.4o4.app/posts/adele-predicting-and-explaining-ai-performance-across-tasks/</link><guid isPermaLink="true">https://ai.4o4.app/posts/adele-predicting-and-explaining-ai-performance-across-tasks/</guid><description>Microsoft Research and collaborators introduce ADeLe, a framework that characterizes both benchmarks and LLMs using shared capability scores, achieving ~88% prediction accuracy on unseen tasks for models like GPT-4o and LLaMA-3.1-405B.</description><pubDate>Sat, 25 Apr 2026 02:20:59 GMT</pubDate><category>frontier</category><category>evaluation</category><category>benchmarks</category><category>llm</category><category>reasoning</category><category>microsoft</category></item><item><title>Microsoft researchers: we are benchmarking AI against the past when we should be asking what comes next</title><link>https://ai.4o4.app/posts/ideas-steering-ai-toward-the-work-future-we-want/</link><guid isPermaLink="true">https://ai.4o4.app/posts/ideas-steering-ai-toward-the-work-future-we-want/</guid><description>Researchers behind the New Future of Work Report 2025 discuss the intentionality required to build a future where people flourish, the gap between efficiency and the work future worth wanting, and why AI&apos;s role as tool versus collaborator is not a semantic question.</description><pubDate>Sat, 25 Apr 2026 02:12:59 GMT</pubDate><category>analysis</category><category>future-of-work</category><category>microsoft</category><category>ai-collaboration</category><category>labor</category><category>research</category></item><item><title>Microsoft&apos;s New Future of Work report: AI is speeding up work changes but distributing benefits unevenly</title><link>https://ai.4o4.app/posts/new-future-of-work-ai-is-driving-rapid-change-uneven-benefits/</link><guid isPermaLink="true">https://ai.4o4.app/posts/new-future-of-work-ai-is-driving-rapid-change-uneven-benefits/</guid><description>The fifth annual New Future of Work report from Microsoft Research documents generative AI&apos;s entry into workplaces — faster than previous technologies, with measurable productivity gains for some, declining entry-level hiring, and significant gaps by gender, income level, and language.</description><pubDate>Sat, 25 Apr 2026 02:04:59 GMT</pubDate><category>analysis</category><category>future-of-work</category><category>microsoft</category><category>generative-ai</category><category>labor</category><category>productivity</category></item><item><title>Microsoft researchers on AI and climate: separate the data from the hype before drawing conclusions</title><link>https://ai.4o4.app/posts/can-we-ai-our-way-to-a-more-sustainable-world/</link><guid isPermaLink="true">https://ai.4o4.app/posts/can-we-ai-our-way-to-a-more-sustainable-world/</guid><description>A Microsoft Research podcast episode brings together a sustainability scientist and an optimization researcher to examine AI&apos;s actual climate footprint, the local infrastructure concerns from datacenter expansion, and where AI optimization tools can genuinely help.</description><pubDate>Sat, 25 Apr 2026 01:56:59 GMT</pubDate><category>analysis</category><category>sustainability</category><category>microsoft</category><category>climate</category><category>datacenters</category><category>optimization</category></item><item><title>Microsoft&apos;s AutoAdapt turns LLM domain adaptation from guesswork into a constraint-aware pipeline</title><link>https://ai.4o4.app/posts/autoadapt-automated-domain-adaptation-for-large-language-models/</link><guid isPermaLink="true">https://ai.4o4.app/posts/autoadapt-automated-domain-adaptation-for-large-language-models/</guid><description>An open-source framework from Microsoft Research automates the selection between RAG and fine-tuning approaches, plans adaptation strategies against real deployment constraints, and replaces manual hyperparameter search with a budgeted refinement loop.</description><pubDate>Sat, 25 Apr 2026 01:48:59 GMT</pubDate><category>frontier</category><category>llm</category><category>fine-tuning</category><category>microsoft</category><category>domain-adaptation</category><category>open-source</category></item><item><title>Google&apos;s Vantage uses AI avatars to assess skills like critical thinking in adaptive conversations</title><link>https://ai.4o4.app/posts/towards-developing-future-ready-skills-with-generative-ai/</link><guid isPermaLink="true">https://ai.4o4.app/posts/towards-developing-future-ready-skills-with-generative-ai/</guid><description>A research experiment built with NYU uses an Executive LLM to steer multi-party AI conversations toward targeted skill assessment, with AI scoring accuracy matching human expert agreement rates in a 188-person study.</description><pubDate>Sat, 25 Apr 2026 01:40:59 GMT</pubDate><category>frontier</category><category>education</category><category>assessment</category><category>google</category><category>generative-ai</category><category>skills</category></item><item><title>Google&apos;s MoGen generates synthetic neuron shapes that cut brain-mapping errors by 4.4%</title><link>https://ai.4o4.app/posts/ai-generated-synthetic-neurons-speed-up-brain-mapping/</link><guid isPermaLink="true">https://ai.4o4.app/posts/ai-generated-synthetic-neurons-speed-up-brain-mapping/</guid><description>A new open-source model from Google Research generates realistic 3D neuron geometries from point cloud flow matching, improving connectomic reconstruction accuracy in a way that would save 157 person-years of manual work at mouse-brain scale.</description><pubDate>Sat, 25 Apr 2026 01:32:59 GMT</pubDate><category>science</category><category>neuroscience</category><category>connectomics</category><category>google</category><category>synthetic-data</category><category>open-source</category></item><item><title>Simula treats synthetic data generation as mechanism design, not sample-by-sample prompting</title><link>https://ai.4o4.app/posts/designing-synthetic-datasets-for-the-real-world-mechanism-design-and-reasoning-f/</link><guid isPermaLink="true">https://ai.4o4.app/posts/designing-synthetic-datasets-for-the-real-world-mechanism-design-and-reasoning-f/</guid><description>Google&apos;s Simula framework decomposes synthetic dataset creation into independently controllable axes — diversity, complexity, and quality — and has already been deployed in Gemma safety models, Android scam detection, and spam filtering.</description><pubDate>Sat, 25 Apr 2026 01:24:59 GMT</pubDate><category>frontier</category><category>synthetic-data</category><category>google</category><category>llm</category><category>training</category><category>gemma</category></item><item><title>ReasoningBank gives agents a memory that learns from failures, not just successes</title><link>https://ai.4o4.app/posts/reasoningbank-enabling-agents-to-learn-from-experience/</link><guid isPermaLink="true">https://ai.4o4.app/posts/reasoningbank-enabling-agents-to-learn-from-experience/</guid><description>Google Cloud&apos;s ICLR paper introduces a structured memory framework that distills both failed and successful agent trajectories into transferable reasoning strategies, beating memory-free baselines by up to 8.3% on web benchmarks.</description><pubDate>Sat, 25 Apr 2026 01:16:59 GMT</pubDate><category>frontier</category><category>agents</category><category>memory</category><category>google</category><category>llm</category><category>reasoning</category></item><item><title>Google Photos can now reframe your shots from a new camera angle after the fact</title><link>https://ai.4o4.app/posts/it-s-all-about-the-angle-your-photos-re-composed/</link><guid isPermaLink="true">https://ai.4o4.app/posts/it-s-all-about-the-angle-your-photos-re-composed/</guid><description>Google&apos;s Auto frame feature uses 3D scene reconstruction and generative inpainting to re-render photos from a different viewpoint, fixing parallax and perspective distortion without a reshoot.</description><pubDate>Sat, 25 Apr 2026 01:08:59 GMT</pubDate><category>frontier</category><category>google</category><category>photography</category><category>generative-ai</category><category>deepmind</category><category>computer-vision</category></item><item><title>Gemma 4 releases four model sizes under Apache 2.0, with the 31B ranked third among all open models</title><link>https://ai.4o4.app/posts/gemma-4-byte-for-byte-the-most-capable-open-models/</link><guid isPermaLink="true">https://ai.4o4.app/posts/gemma-4-byte-for-byte-the-most-capable-open-models/</guid><description>Google DeepMind&apos;s Gemma 4 family spans a 2B mobile model to a 31B dense model, supports 140+ languages, adds 256K context for larger variants, and ships under a fully permissive open-source license.</description><pubDate>Sat, 25 Apr 2026 00:55:59 GMT</pubDate><category>opensource</category><category>gemma</category><category>open-source</category><category>deepmind</category><category>llm</category><category>on-device</category></item><item><title>Gemini Robotics-ER 1.6 adds instrument reading and multi-view reasoning, developed with Boston Dynamics</title><link>https://ai.4o4.app/posts/gemini-robotics-er-1-6-powering-real-world-robotics-tasks-through-enhanced-embod/</link><guid isPermaLink="true">https://ai.4o4.app/posts/gemini-robotics-er-1-6-powering-real-world-robotics-tasks-through-enhanced-embod/</guid><description>Google DeepMind&apos;s upgraded embodied reasoning model improves spatial reasoning, success detection, and multi-camera understanding — and gains a new industrial inspection capability co-developed with Boston Dynamics.</description><pubDate>Sat, 25 Apr 2026 00:50:59 GMT</pubDate><category>robotics</category><category>robotics</category><category>gemini</category><category>embodied-ai</category><category>deepmind</category><category>spatial-reasoning</category></item><item><title>Gemini 3.1 Flash TTS launches with audio tags, 70+ languages, and SynthID watermarking</title><link>https://ai.4o4.app/posts/gemini-3-1-flash-tts-the-next-generation-of-expressive-ai-speech/</link><guid isPermaLink="true">https://ai.4o4.app/posts/gemini-3-1-flash-tts-the-next-generation-of-expressive-ai-speech/</guid><description>Google DeepMind&apos;s new text-to-speech model scores 1,211 on the Artificial Analysis TTS Elo leaderboard, introduces natural-language audio tags for vocal direction, and watermarks all output with SynthID.</description><pubDate>Sat, 25 Apr 2026 00:45:59 GMT</pubDate><category>frontier</category><category>tts</category><category>speech</category><category>gemini</category><category>audio</category><category>deepmind</category></item><item><title>Google DeepMind partners with five global consultancies to deploy frontier AI at enterprise scale</title><link>https://ai.4o4.app/posts/partnering-with-industry-leaders-to-accelerate-ai-transformation/</link><guid isPermaLink="true">https://ai.4o4.app/posts/partnering-with-industry-leaders-to-accelerate-ai-transformation/</guid><description>Accenture, Bain, BCG, Deloitte, and McKinsey gain early access to Gemini models and direct DeepMind technical engagement in a push to close a significant gap between enterprise AI ambition and production deployment.</description><pubDate>Sat, 25 Apr 2026 00:40:59 GMT</pubDate><category>frontier</category><category>enterprise</category><category>partnerships</category><category>gemini</category><category>agents</category><category>consulting</category></item><item><title>Google DeepMind&apos;s Decoupled DiLoCo trains LLMs across data centers on standard internet bandwidth</title><link>https://ai.4o4.app/posts/decoupled-diloco-a-new-frontier-for-resilient-distributed-ai-training/</link><guid isPermaLink="true">https://ai.4o4.app/posts/decoupled-diloco-a-new-frontier-for-resilient-distributed-ai-training/</guid><description>A new distributed training architecture from Google DeepMind uses asynchronous compute islands to train large models across distant locations — more than 20x faster than conventional synchronization, with self-healing fault tolerance.</description><pubDate>Sat, 25 Apr 2026 00:35:59 GMT</pubDate><category>frontier</category><category>training</category><category>infrastructure</category><category>distributed</category><category>deepmind</category><category>hardware</category></item></channel></rss>