Prime Intellect's INTELLECT-2, a 32B-parameter language model, was trained using globally distributed reinforcement learning (RL).
The model is based on the QwQ-32B base and utilizes the prime-rl
asynchronous distributed RL framework, incorporating verifiable reward signals for math and coding tasks.
Architectural changes were made for stability and adaptive length control, with an optimal generation length between 2k–10k tokens.
INTELLECT-2's performance is comparable to QwQ-32B on benchmarks like AIME24, LiveCodeBench, and GPQA-Diamond, with slight underperformance on IFEval. The significance lies in its demonstration of decentralized RL training.
The project also explores post-training techniques and inference-during-training.
The work suggests potential for P2P or blockchain-inspired distributed compute and credit systems for AI training and inference.
ByteDance released DreamO on Hugging Face, a unified framework for image customization supporting ID, IP, Try-On, and Style tasks.
Qwen released optimized models for GPTQ, GGUF, and AWQ. Alibaba Qwen also officially released quantized versions of Qwen3 (GGUF, AWQ, GPTQ, INT8) deployable via Ollama, LM Studio, SGLang, and vLLM. The Qwen3 release includes official quantized models, open weights, and a permissive license.
Gemma surpassed 150 million downloads and 70,000 variants on Hugging Face.
Meta released model weights for its 8B-parameter Dynamic Byte Latent Transformer (BLT) for improved language model efficiency and reliability, and the Collaborative Reasoner framework to enhance collaborative reasoning. The BLT model, first discussed in late 2023, focuses on byte-level tokenization.
RunwayML’s Gen-4 References model was launched, described as offering infinite workflows without fine-tuning for near-realtime creation.
Mistral AI released Mistral Medium 3, a multimodal AI model, and Le Chat Enterprise, an agentic AI assistant for businesses with tools like Google Drive integration and agent building.
Google updated Gemini 2.5 Pro Preview with video understanding and improvements for UI, code, and agentic workflows. Gemini 2.0 Flash image generation received improved quality and text rendering.
DeepSeek, an open-source AI initiative, has reportedly nearly closed the performance gap with US peers in two years.
f-lite 7B, a distilled diffusion model, was released.
Microsoft updated Copilot with a “Pages” feature, similar to ChatGPT Canvas, but reportedly without coding capabilities.
Manus AI publicly launched, offering users free daily tasks and credits. The platform focuses on educational or content generation tasks. Some users reported regional availability issues.
JoyCaption Beta One, a free, open-source, uncensored Vision Language Model (VLM) for image captioning, was released with doubled training data, a new 'Straightforward Mode', improved booru tagging, and better watermark annotation. It achieved 67% normalized accuracy on human-benchmarked validation sets.
Sakana AI introduced Continuous Thought Machines (CTM), a neural architecture where reasoning is driven by neuron-level timing and synchronization. CTM neurons encode signal history and timing, aiming for complex, temporally-coordinated behaviors.
A new model, Drakesclaw, appeared on the LM Arena, with initial impressions suggesting performance comparable to Gemini 2.5 Pro.
The Absolute Zero Reasoner (AZR) paper details a model achieving state-of-the-art results on coding/math tasks via self-play with zero external data.
Mellum-4b-sft-rust, a CodeFIM (Fill-In-The-Middle) model for Rust, trained using Unsloth, was released on Hugging Face.
Facebook released weights for their Byte Latent Transformer (BLT).
The release of Grok 3.5 is on hold pending integration with X and another recently acquired company.