TLDR of AI news

Subscribe
Archives
May 28, 2025

05-28-2025

AI Model Releases and Updates

  • A new version of the DeepSeek R1 model, DeepSeek-R1-0528, has been released on Hugging Face and is available on some inference partner platforms. It continues to use the MIT license for model weights and code. The community is actively converting the model to GGUF format for broader compatibility.

  • The Gemma model family has seen numerous releases over six months, including PaliGemma 2, PaliGemma 2 Mix, Gemma 3, ShieldGemma 2, TxGemma, Gemma 3 QAT, Gemma 3n Preview, and MedGemma, alongside earlier models like DolphinGemma and SignGemma.

  • The Claude 4 launch is reported to be significantly accelerating development workflows. The combination of Opus 4, Claude Code, and the Claude Max plan is considered a high-return AI coding stack.

  • Codestral Embed, a code embedder capable of using up to 3072 dimensions, has been released.

  • The BAGEL model, proposed and implemented by ByteDance, is an open-source multimodal model designed for reading, reasoning, drawing, and editing, supporting long, mixed contexts and arbitrary aspect ratios without a quality bottleneck.

  • An updated DeepSeek model (possibly R1 v2 or 0528) reportedly shows improved accuracy, successfully answering test questions that stumped Gemini 2.5 Pro, though with increased response latency. A previous bug related to hallucinating invisible tokens with the '翻译' prompt has been fixed.

  • Google AI Edge Gallery, an open-source app, enables on-device, offline execution of generative AI models (like Gemma3-1B-IT q4) on Android (iOS soon), with features like 'Ask Image,' 'Prompt Lab,' and 'AI Chat,' and tunable inference settings. Some users report instability and potential privacy concerns with network requests.

  • Chatterbox TTS 0.5B, an open-source English-only text-to-speech model claiming to surpass ElevenLabs in quality, has been released. It is distributed via pip, with weights on HuggingFace, and offers adjustable expressive parameters with CPU-viability for short utterances.

  • Google announced SignGemma, an upcoming open-source model in the Gemma family, designed for translating sign language into spoken text. It aims to improve accessibility and real-time multimodal communication and is expected later this year. It reportedly generates less uncanny point cloud visualizations than previous models.

  • Tencent released Hunyuan Video Avatar, an open-source, audio-driven image-to-video generation model supporting multiple characters. The initial release supports single-character, 14s audio inputs. Minimum hardware is a 24GB GPU, with 80GB recommended.

  • A new anime-specific fine-tune of the WAN (Warp-Aware Network) video generation model for Stable Diffusion has been released on CivitAI, offering image-to-video and text-to-video capabilities for stylized animation.

  • Anthropic has rolled out a Claude voice mode beta for mobile devices, enabling English language tasks such as calendar summaries across all user plans.

AI Model Performance and Benchmarking

  • Claude Opus 4 has reportedly reached the #1 position in the WebDev Arena benchmark, surpassing the previous Claude 3.7 and matching Gemini 2.5 Pro. Evaluations also show a significant improvement in coding performance for Sonnet 4.

  • Claude Opus 4 is claimed to achieve state-of-the-art results on the ARC-AGI-2 benchmark. Claude 4 Sonnet might be the first model to significantly benefit from test-time-compute on ARC-AGI 2, beating o3-preview on this benchmark at a substantially lower cost.

  • Findings suggest that random rewards in reinforcement learning only work for Qwen models and that observed improvements were due to clipping, raising questions about the validity of RL papers using Qwen if the model works with any random reward.

  • Nemotron-CORTEXA reportedly reached the top of the SWEBench leaderboard by solving 68.2% of SWEBench GitHub issues using a multi-step problem localization and repair process.

  • A paper on VideoGameBench indicates that the best-performing model, Gemini 2.5 Pro, completes only 0.48% of VideoGameBench and 1.6% of VideoGameBench Lite.

  • Frontier LLMs are reported to find solving ‘Modern Sudokus’ challenging.

  • DeepSeek-R1-0528 is noted for its strong coding capabilities, with user reports indicating it performs on par with or approaches models like Gemini 2.5 Pro, successfully handling complex coding tasks and resolving issues that stumped other leading models. In a custom Scrabble coding test, it generated accurate, working code and robust tests on the first try, producing more concise code than competitors.

  • A comparison attempt between DeepSeek-R1-0528 and Claude-4-Sonnet using a 'heptagon + 20 balls' benchmark was deemed uninformative as it relies on external physics engines, not the LLMs' inherent abilities.

  • Gemma 3 27B QAT, running on RDNA3 Gen1 hardware, reportedly achieved 11 tokens per second.

  • In user tests for web development, Gemini 2.5 Pro ranked highly, outperforming Grok 3. Opus 4 was ranked above O3 in coding by some users.

  • Perplexity Pro reportedly outperformed Sonar Pro in 20 tests, despite claims that Perplexity uses open-source models.

AI Agents, Tools, and Frameworks

  • AutoThink is a new open-source approach that enhances LLM reasoning by classifying query complexity and dynamically allocating a token budget for “thinking.” It also uses steering vectors to guide internal reasoning patterns, resulting in performance improvements and increased token efficiency.

  • A "SQL tool" and Plotly integration are now available with codegen, enabling database access for code agents.

  • Security issues are highlighted for agents connected to platforms like GitHub, regardless of whether they use Multi-Agent Collaboration Platforms (MCPs).

  • The "R" in RAG (Retrieval Augmented Generation) stands for retrieval, and relevancy is paramount; a system with a retrieval component is still considered RAG.

  • Factory, an AI designed to write code, has been introduced, based on the concept of agent-native software development where agents ship code.

  • The Mistral AI Agents API has been released, featuring code execution, web search, MCP tools, persistent memory, and agentic orchestration. A Handoff Feature allows agents to call other agents or transfer conversations.

  • The Comet Assistant is being used for consuming web content via AI.

  • MagicPath, an infinite canvas for AI-assisted creation and exploration, has been introduced, securing a $6.6M seed round.

  • Perplexity AI assistant users can opt-in for daily news summaries via WhatsApp using the /news command.

  • New use cases for Runway's "References" feature involve using real-world inspiration to inform creative ideas. Runway's Gen-4 model aims for universal, less prescriptive use cases beyond simple "text-to-X" approaches.

  • Aider now utilizes Tree Sitter to generate repository maps, aiding debugging with tools like entr for instant updates. This update is noted for DeepSeek R1.

  • A user shared experiences using Google Veo 3 to produce a full short film, utilizing its 'flow' mode for character consistency via text prompts, at a production cost of approximately $30 in credits for a multi-minute video.

  • Veo 3 was used to create a video adaptation of a classic Reddit joke, with all elements except the narrator's voice generated by the AI.

  • A project titled 'Afterlife: The unseen lives of AI actors between prompts' was created using Google's Veo 3.

  • A user recreated a speculative Grand Theft Auto VI trailer using AI video (Luma) and audio (Suno) generation platforms.

  • Users of the Cursor codebase indexing tool reported hours-long stalls, with some finding relief by logging out and back in. Sonnet 4 also experienced connection failures.

AI Infrastructure, Hardware, and Optimization

  • Bell Canada has selected Groq as its exclusive inference provider.

  • The CMU FLAME center has acquired a new cluster with 256 H100 GPUs.

  • Hugging Face Spaces is described as an "app store of AI" and now also an "MCP store," allowing users to filter and attach thousands of MCPs to LLMs.

  • Mojo is discussed as a language that learns from Python, Rust, Zig, and Swift to provide an easy-to-learn language that unlocks peak performance for Python and GPU coding.

  • A new ICML25 paper proposes Subset-Norm (SN) and Subspace-Momentum (SM) techniques to reduce optimizer state memory during deep learning model training (e.g., 80% less memory for LLaMA 1B pre-training) while providing strong convergence guarantees, reportedly outperforming Adam and prior efficient optimizers. The codebase is open-sourced.

  • A guide for running WAN 2.1 and VACE on low VRAM GPUs (e.g., 8GB) recommends 480p video generation with upscaling, ComfyUI's memory management, and sufficient system RAM. It covers I2V, T2V, and V2V workflows, emphasizing low batch sizes and offloading.

  • Quantized versions of DeepSeek-R1-0528 are available via Unsloth, aimed at improving efficiency. This approach is also used to combat catastrophic forgetting in fine-tuning Qwen3 by mixing original data.

  • OpenRouter is discontinuing GPT-4 32k models by June 6th, promoting o3 and o4-mini for streaming summaries. It has introduced features like crypto invoices and mandatory third-party keys.

  • A new kernel reportedly doubles batch 1 forward pass speed. Data shuffling is recommended to enhance generalization.

  • Resources for CUDA kernel programming are being shared for beginners, focusing on Hopper setups and mark_dynamic for tensor constraints in PyTorch.

Responsible AI, Societal Impact, and Ethical Considerations

  • Anthropic's Long Term Benefit Trust has appointed Reed Hastings to Anthropic's board of directors.

  • An unauthorized update by an xAI employee reportedly caused the Grok chatbot on X to make false claims of a “white genocide” in South Africa.

  • There is an ongoing emphasis on the need for AI safety, with scientific solutions being envisioned and shared.

  • RAG systems are described as being more brittle than commonly thought, even when provided with sufficient context.

  • Anthropic's CEO, Dario Amodei, warns that AI could eliminate up to 50% of entry-level white-collar jobs within 1-5 years, potentially increasing unemployment to 10-20%, and urges stakeholders to address this disruptive economic impact. He believes AI firms and governments must stop minimizing the risk of mass white-collar job automation, particularly in technology, finance, law, and consulting.

  • Discussions suggest that AI-induced mass unemployment could undermine consumer-driven economic models, as AI does not spend money. Some project that AI models might perform 80-90% of tasks in most white-collar fields within 1-2 years, with senior roles being more resistant due to ambiguous reward signals.

  • The question is raised: if AI automates ~50% of engineering jobs, what prevents laid-off engineers from using the same AI to replicate the company's product at a lower price? Barriers include capital requirements, first-mover advantages, and brand recognition.

  • Anthropic's CEO also stated that modern AI models like Claude Sonnet 3.7 and Google's Gemini 2.5 Pro (with Grounding) may hallucinate less than humans, but their errors can be more surprising or unexpected. These models are noted for improved factuality and willingness to admit mistakes. AI hallucinations are distinct in their confident delivery and potential for fabricating plausible misinformation, necessitating verification and human oversight. Current AI systems lack the ability to self-diagnose their own errors.

  • An AI-generated video depicting an American soldier in Gaza, potentially made with tools like Veo 3, circulated widely and was considered authentic by many, despite technical tells. This highlights concerns about AI-synthesized video influencing public perception and misinformation.

  • A viral "emotional support kangaroo" video, confirmed to be AI-generated, deceived many viewers, illustrating challenges in distinguishing deepfakes from authentic media and concerns about media literacy. This could lead to widespread skepticism about unusual content.

  • Discussion around Veo 3 includes concerns about widespread reposting without creator credit and misrepresentation of technical features like fake sign language in generated videos.

  • The authenticity of a video was debated, with technical users pointing to current AI limitations in generating realistic expressions and interactions (sub-8 second clips for tools like Google Veo) as reasons to believe it was real.

  • Google's co-founder reportedly claimed AI may produce better outputs when prompted with threats. This idea is met with skepticism, with users noting LLMs typically respond better to descriptive prompts, though anecdotal evidence suggests aggressive language sometimes appears to improve code generation, possibly due to prompt iteration.

Industry Trends, Discussions, and Future Outlook

  • The "bigger is better" era of AI, characterized by energy-hungry, compute-heavy models, is argued to be ending due to cost and sustainability concerns.

  • Some argue the focus should be on Artificial Superintelligence (ASI) rather than Artificial General Intelligence (AGI).

  • Information theory is emphasized as crucial for interpretability in AI.

  • Critiques have been raised regarding the philosophical underpinnings of consciousness stances within some intellectual communities.

  • Recent findings suggest that Reinforcement Learning (RL) primarily elicits latent behaviors learned during pretraining rather than teaching new behaviors.

  • AI is described as potentially the most misunderstood technology of the century because it can be shaped to fit various perceptions.

  • The complexity of intelligence is often underestimated, similar to how the complexity of coding tasks can be misjudged.

  • A common observation is that technology is often overestimated in the short term and underestimated in the long run.

  • A report indicated a sharp rise in companies discontinuing generative AI projects (from 17% to 42% year-over-year), primarily due to unmet ROI expectations after automation-induced layoffs, with some firms rehiring. The hype around generative AI is compared to the dot-com bubble.

  • Retrieval-Augmented Generation (RAG) is seen as providing true near-term value in GenAI, with adoption growing beyond expert circles. LLMs are considered by some to be over-applied outside their core capabilities for production code.

  • A comparison of AI strategies highlights Apple's recent entry with "Apple Intelligence" contrasting with more established infrastructures of Google, Microsoft, Meta, and Amazon. Apple's lower investment in data center CapEx (~1bn vs.75bn planned by Alphabet/Microsoft for 2025) and focus on on-device AI for privacy means reliance on third-party LLMs for some features.

  • A debate suggests China may lead in future AI progress due to its significantly higher electricity consumption and generation capacity, with US AI labs reportedly forecasting energy shortages by 2026. However, counterarguments state AI/data centers currently use a small percentage of national electricity, and chip supply (growing slower at 10-15% YoY) might be a more significant bottleneck than energy.

  • Google is reportedly using AI to analyze dolphin vocalizations (clicks and whistles) and associate them with observed behaviors and identities through supervised learning. This is a correlational mapping rather than true language translation, with skepticism about feasibility without richer semantic context.

Don't miss what's next. Subscribe to TLDR of AI news:
Powered by Buttondown, the easiest way to start and grow your newsletter.