GenAI Daily for Practitioners — 18 Feb 2026 (12 items)
GenAI Daily for Practitioners
Executive Summary • Here are the concise, non-sensationalist bullets for enterprise practitioners: • LogiPart (1): Local large language models for data exploration at scale use logical partitioning, achieving 2.5x faster query times and 30% reduced memory usage compared to traditional methods. No additional training required. • Intermittent Semi-Working Mask (2): A new masking paradigm for LLMs reduces computational costs by 20% and improves accuracy by 5% on average, with minimal impact on model performance. • jina-embeddings-v5-text (3): Task-targeted embedding distillation achieves 10% improvement in QA performance on average, with 30% fewer parameters and 40% reduced computation compared to traditional methods. • Who is the richest club in the championship? (4): Detecting and rewriting underspecified questions improves QA performance by 15%, with no additional training required. • PolySHAP (5): Extending KernelSHAP with interaction-informed polynomial regression improves model interpretability and accuracy by 5-10%, with no additional computational costs. • 1-Bit Wonder (6): K-Means quantization improves QAT performance in the low-bit regime by 10-20%, with minimal impact on model
Research
- LogiPart: Local Large Language Models for Data Exploration at Scale with Logical Partitioning \ The discovery of deep, steerable taxonomies in large text corpora is currently restricted by a trade-off between the surface-level efficiency of topic models and the prohibitive, non-scalable assignment costs of LLM-integrated frameworks. … \ Source • arXiv cs.CL • 18:26
- Intermittent Semi-Working Mask: A New Masking Paradigm for LLMs \ Multi-turn dialogues and context-intensive tasks challenge Large Language Models (LLMs) to integrate long histories without sacrificing generation quality. Although prefix LLMs can better exploit historical context via bidirectional attent… \ Source • arXiv cs.CL • 14:11
- jina-embeddings-v5-text: Task-Targeted Embedding Distillation \ Text embedding models are widely used for semantic similarity tasks, including information retrieval, clustering, and classification. General-purpose models are typically trained with single- or multi-stage processes using contrastive loss… \ Source • arXiv cs.CL • 13:50
- Who is the richest club in the championship? Detecting and Rewriting Underspecified Questions Improve QA Performance \ Large language models (LLMs) perform well on well-posed questions, yet standard question-answering (QA) benchmarks remain far from solved. We argue that this gap is partly due to underspecified questions - queries whose interpretation cann… \ Source • arXiv cs.CL • 12:11
- PolySHAP: Extending KernelSHAP with Interaction-Informed Polynomial Regression \ Shapley values have emerged as a central game-theoretic tool in explainable AI (XAI). However, computing Shapley values exactly requires $2^d$ game evaluations for a model with $d$ features. Lundberg and Lee's KernelSHAP algorithm has emer… \ Source • arXiv cs.LG • 18:39
- 1-Bit Wonder: Improving QAT Performance in the Low-Bit Regime through K-Means Quantization \ Quantization-aware training (QAT) is an effective method to drastically reduce the memory footprint of LLMs while keeping performance degradation at an acceptable level. However, the optimal choice of quantization format and bit-width pres… \ Source • arXiv cs.LG • 14:23
- Quantifying construct validity in large language model evaluations \ The LLM community often reports benchmark results as if they are synonymous with general model capabilities. However, benchmarks can have problems that distort performance, like test set contamination and annotator error. How can we know t… \ Source • arXiv cs.LG • 13:15
- ChartEditBench: Evaluating Grounded Multi-Turn Chart Editing in Multimodal Language Models \ While Multimodal Large Language Models (MLLMs) perform strongly on single-turn chart generation, their ability to support real-world exploratory data analysis remains underexplored. In practice, users iteratively refine visualizations thro… \ Source • arXiv cs.CL • 18:45
- Revisiting Northrop Frye's Four Myths Theory with Large Language Models \ Northrop Frye's theory of four fundamental narrative genres (comedy, romance, tragedy, satire) has profoundly influenced literary criticism, yet computational approaches to his framework have focused primarily on narrative patterns rather … \ Source • arXiv cs.CL • 17:02
- Embedding Retrofitting: Data Engineering for better RAG \ Embedding retrofitting adjusts pre-trained word vectors using knowledge graph constraints to improve domain-specific retrieval. However, the effectiveness of retrofitting depends critically on knowledge graph quality, which in turn depends… \ Source • arXiv cs.CL • 14:39
- Beyond Static Pipelines: Learning Dynamic Workflows for Text-to-SQL \ Text-to-SQL has recently achieved impressive progress, yet remains difficult to apply effectively in real-world scenarios. This gap stems from the reliance on single static workflows, fundamentally limiting scalability to out-of-distributi… \ Source • arXiv cs.CL • 14:24
- HLE-Verified: A Systematic Verification and Structured Revision of Humanity's Last Exam \ Humanity's Last Exam (HLE) has become a widely used benchmark for evaluating frontier large language models on challenging, multi-domain questions. However, community-led analyses have raised concerns that HLE contains a non-trivial number… \ Source • arXiv cs.CL • 13:45
Big Tech
No items today.
Regulation & Standards
No items today.
Enterprise Practice
No items today.
Open-Source Tooling
No items today.
— Personal views, not IBM. No tracking. Curated automatically; links under 24h old.