Pentagon Plans Classified AI Training After Commercial Models Reach Iran
1. OpenAI and Mistral Shift from Flagship Models to Purpose-Built Tools OpenAI's newest models aren't built for humans. GPT-5.4 mini and nano, announced this week, target sub-agent workloads: tasks where one AI system calls another at high frequency.
2. Pentagon Plans Classified AI Training as Commercial Models Drift Toward Iran The U.S. Department of Defense is building secure facilities where AI companies will train models on classified intelligence.
3. Google Bets Its AI Edge on Knowing You, Not Outsmarting Rivals Google on Tuesday opened its "Personal Intelligence" feature to all US users for free, connecting Gemini to Search history, Gmail, and Chrome browsing data.
In Brief
- Microsoft Merges Consumer and Commercial Copilot Teams Under New Executive Microsoft reorganized its Copilot division, unifying previously separate consumer and commercial engineering teams under one leader. The company has run parallel Copilot efforts for years; the consolidation targets a more consistent product across business and personal users. The Verge
- Americans Send Nearly 3 Million Daily ChatGPT Messages Asking About Pay OpenAI published research showing U.S. users send roughly 3 million compensation-related queries to ChatGPT each day. The data covers questions about salaries, earnings, and pay benchmarks across industries. OpenAI
- Google Funds New AI Tools for Open Source Security Google announced additional investment in AI-powered code analysis to find vulnerabilities in open source software. The effort includes new scanning tools and security-focused development practices aimed at catching flaws before release. Google AI Blog
- OpenSeeker Publishes First Fully Open Training Data for LLM Search Agents Researchers released OpenSeeker, an open-source dataset for training LLM-based search agents. High-performance search agent development has been restricted to large companies with proprietary data; OpenSeeker provides the full pipeline. Hugging Face
- EnterpriseOps-Gym Tests AI Agents on Realistic IT Workflows A new benchmark evaluates LLM agents on long-horizon planning tasks involving persistent state changes and access controls. Existing benchmarks miss the complexity of enterprise environments where agents must respect strict protocols across multi-step operations. Hugging Face
- Mixture-of-Depths Attention Lets Transformer Heads Reach Back to Earlier Layers MoDA, a new attention mechanism, allows each head to attend to key-value pairs from both the current layer and preceding layers. The design addresses signal degradation in deep models, where useful features formed in shallow layers get diluted by repeated residual updates. Hugging Face
- Seoul World Model Generates Street-Level Video Grounded in a Real City Researchers built a city-scale video generation model anchored to actual Seoul street-view images. The system uses retrieval-augmented conditioning to produce video sequences tied to real geographic locations rather than imagined environments. Hugging Face
- VET-Bench Reveals Top Vision-Language Models Score Near Chance on Object Tracking A synthetic benchmark tests VLMs on tracking visually identical objects using only spatiotemporal continuity. Every state-of-the-art model tested performed at or near random, exposing a basic gap in visual tracking that existing benchmarks obscure with visual shortcuts. Hugging Face
- RLCF Trains Models to Judge Research Ideas Using Large-Scale Community Signals A new training method called Reinforcement Learning from Community Feedback teaches models to assess and propose high-impact scientific ideas. The approach targets scientific taste — the ability to identify promising research directions — rather than execution capability. Hugging Face
- Distillation Method Achieves Lossless Transfer from Transformers to Hybrid xLSTM Researchers introduced a distillation approach that transfers knowledge from quadratic-attention LLMs to sub-quadratic hybrid xLSTM architectures without performance loss. The method defines lossless distillation through tolerance-corrected win-and-tie rates between student and teacher across task sets. Hugging Face
- Cheers Separates Patch Detail from Semantics for Unified Visual Understanding and Generation A new multimodal model decouples patch-level visual detail from semantic representations within a single architecture. The split stabilizes comprehension while improving generation fidelity — addressing a core tension in models that attempt both tasks jointly. Hugging Face
Don't miss what's next. Subscribe to AI News Digest: