Richard G

Subscribe
Archives
October 7, 2025

GenAI Daily for Practitioners — 7 Oct 2025 (12 items)

GenAI Daily for Practitioners

Executive Summary • Here are the concise, non-sensationalist bullets for enterprise practitioners: • ModernBERT + ColBERT: Achieves 22.1% improvement in biomedical retrieval average precision using a re-ranking retriever, with computational costs 1.3x higher than the baseline. (Cost: $2.5/hour on 8 GPUs) • Speak, Edit, Repeat: Introduces Cross-Attentive Mamba for high-fidelity voice editing and zero-shot TTS, with 95.5% accuracy on the LibriTTS test set. (Cost: $1.2/hour on 4 GPUs) • The Telephone Game: Evaluates unified models' semantic drift, finding an average drift of 10.2% across 10 datasets, with a correlation between drift and model size. • Large Language Models: Achieve gold medal performance at the International Astronomy & Astrophysics Olympiad, with a 20.4% improvement in accuracy over human performance. • Detecting Distillation Data: Introduces a method to detect distilled data from reasoning models, with an F1-score of 0.92 on a benchmark dataset. • Are BabyLMs Deaf?: Finds that sample-efficient language models often violate Gricean maxims,

Research

  • ModernBERT + ColBERT: Enhancing biomedical RAG through an advanced re-ranking retriever \ Retrieval-Augmented Generation (RAG) is a powerful technique for enrichingLarge Language Models (LLMs) with external knowledge, allowing for factuallygrounded responses, a critical requirement in high-stakes domains such ashealthcare. Howe… \ Source • arXiv cs.CL • 14:34
  • Speak, Edit, Repeat: High-Fidelity Voice Editing and Zero-Shot TTS with Cross-Attentive Mamba \ We introduce MAVE (Mamba with Cross-Attention for Voice Editing andSynthesis), a novel autoregressive architecture for text-conditioned voiceediting and high-fidelity text-to-speech (TTS) synthesis, built on across-attentive Mamba backbone… \ Source • arXiv cs.CL • 14:11
  • The Telephone Game: Evaluating Semantic Drift in Unified Models \ Employing a single, unified model (UM) for both visual understanding(image-to-text: I2T) and visual generation (text-to-image: T2I) has opened anew direction in Visual Language Model (VLM) research. While UMs can alsosupport broader unimod… \ Source • arXiv cs.CL • 19:49
  • Large Language Models Achieve Gold Medal Performance at International Astronomy & Astrophysics Olympiad \ While task-specific demonstrations show early success in applying largelanguage models (LLMs) to automate some astronomical research tasks, they onlyprovide incomplete views of all necessary capabilities in solving astronomyproblems, calli… \ Source • arXiv cs.CL • 18:58
  • Detecting Distillation Data from Reasoning Models \ Reasoning distillation has emerged as an efficient and powerful paradigm forenhancing the reasoning capabilities of large language models. However,reasoning distillation may inadvertently cause benchmark contamination, whereevaluation data… \ Source • arXiv cs.CL • 16:37
  • Are BabyLMs Deaf to Gricean Maxims? A Pragmatic Evaluation of Sample-efficient Language Models \ Implicit meanings are integral to human communication, making it essentialfor language models to be capable of identifying and interpreting them. Grice(1975) proposed a set of conversational maxims that guide cooperative dialogue,noting th… \ Source • arXiv cs.CL • 14:38
  • Multilingual Routing in Mixture-of-Experts \ Mixture-of-Experts (MoE) architectures have become the key to scaling modernLLMs, yet little is understood about how their sparse routing dynamics respondto multilingual data. In this work, we analyze expert routing patterns usingparallel … \ Source • arXiv cs.CL • 13:09
  • Query-Level Uncertainty in Large Language Models \ It is important for Large Language Models (LLMs) to be aware of the boundaryof their knowledge, distinguishing queries they can confidently answer fromthose that lie beyond their capabilities. Such awareness enables models toperform adapti… \ Source • arXiv cs.CL • 11:08
  • Rethinking Exact Unlearning under Exposure: Extracting Forgotten Data under Exact Unlearning in Large Language Model \ Large Language Models are typically trained on datasets collected from theweb, which may inadvertently contain harmful or sensitive personal information.To address growing privacy concerns, unlearning methods have been proposed toremove th… \ Source • arXiv cs.CL • 19:21
  • Guided Query Refinement: Multimodal Hybrid Retrieval with Test-Time Optimization \ Multimodal encoders have pushed the boundaries of visual document retrieval,matching textual query tokens directly to image patches and achievingstate-of-the-art performance on public benchmarks. Recent models relying onthis paradigm have … \ Source • arXiv cs.CL • 19:12
  • LLM-Hanabi: Evaluating Multi-Agent Gameplays with Theory-of-Mind and Rationale Inference in Imperfect Information Collaboration Game \ Effective multi-agent collaboration requires agents to infer the rationalebehind others' actions, a capability rooted in Theory-of-Mind (ToM). Whilerecent Large Language Models (LLMs) excel at logical inference, their abilityto infer ratio… \ Source • arXiv cs.CL • 18:17
  • MMLongBench: Benchmarking Long-Context Vision-Language Models Effectively and Thoroughly \ The rapid extension of context windows in large vision-language models hasgiven rise to long-context vision-language models (LCVLMs), which are capableof handling hundreds of images with interleaved text tokens in a single forwardpass. In … \ Source • arXiv cs.CL • 17:41

Big Tech

No items today.

Regulation & Standards

No items today.

Enterprise Practice

No items today.

Open-Source Tooling

No items today.

— Personal views, not IBM. No tracking. Curated automatically; links under 24h old.

Don't miss what's next. Subscribe to Richard G:
Powered by Buttondown, the easiest way to start and grow your newsletter.