Daily Briefing – Apr 20 (96 Articles)
Babak's Daily Briefing
Monday, April 20, 2026
Sources: 20 | Total Articles: 96
6G World
1.Evaluating 6G PHY Evolution: What the Industry Is Really Trying to Solve
Summary available at source link.
2.Amazon’s Globalstar deal gives Amazon Leo a faster path into D2D
Amazon’s planned acquisition of Globalstar is about far more than satellites. It gives Amazon Leo a faster path into direct-to-device connectivity, combining spectrum, operational assets, and Apple-facing service continuity in a move that could reshape the hybrid terrestrial-NTN landscape.
3.SoftBank’s Physical AI push gives AI-RAN a sharper purpose
SoftBank is starting to give AI-RAN a more concrete job description: not just running AI workloads near the network, but serving as the real-time infrastructure layer for robots and other physical systems. The company’s recent materials suggest it wants to move the AI-RAN conversation from telecom architecture to real-world machine action.
4.South Korea puts 6G inside its national AI push
South Korea has unveiled a three-year national roadmap aimed at becoming one of the world’s top three AI powers by 2028, with 6G commercialization positioned as part of that broader push.
5.b-com’s Open XG Hub targets one of telecom’s biggest gaps: turning experimentation into deployment
In an interview with Peter Pietrzyk, Managing Director of 6GWorld, Patrick Savell, Head of Connectivity at b-com, said platforms such as Open XG Hub are designed to help bridge one of the industry’s most persistent challenges: moving promising ideas from research environments into deployable network systems. The bigger point is that, as telecom becomes more software-driven and AI-native, the bottleneck is increasingly less about invention and more about validation, integration, and operational readiness.
AI Agents
1.SocialGrid: A Benchmark for Planning and Social Reasoning in Embodied Multi-Agent Systems
As Large Language Models (LLMs) transition from text processors to autonomous agents, evaluating their social reasoning in embodied multi-agent settings becomes critical. We introduce SocialGrid, an embodied multi-agent environment inspired by Among Us that evaluates LLM agents on planning, task execution, and social reasoning. Our evaluations reveal that even the strongest open model (GPT-OSS-120B) achieves below 60% accuracy in task completion and planning, with agents getting stuck in repetitive behaviors or failing to navigate basic obstacles. Since poor navigation confounds evaluation of social intelligence, SocialGrid offers an optional Planning Oracle to isolate social reasoning from planning deficits. While planning assistance improves task completion, social reasoning remains a bottleneck: agents fail to detect deception at near-...
2.Exploring Agentic Visual Analytics: A Co-Evolutionary Framework of Roles and Workflows
Agentic visual analytics (VA) represents an emerging class of systems in which large language model (LLM)-driven agents autonomously plan, execute, evaluate, and iterate across the full visual analytics pipeline. By shifting users from low-level tool operations to high-level analytical goals expressed through natural language, these systems are fundamentally transforming how humans interact with data. However, the rapid proliferation of such systems in recent years has outpaced our understanding of their design landscape. Two intertwined problems remain open: how do autonomous agents reshape the traditional VA pipeline, and how must human involvement adapt as agent autonomy increases? To address these questions, this paper presents a comprehensive survey of 55 primary agentic VA systems and introduces a co-evolutionary framework. This fra...
3.HarmfulSkillBench: How Do Harmful Skills Weaponize Your Agents?
Large language models (LLMs) have evolved into autonomous agents that rely on open skill ecosystems (e.g., ClawHub and Skills.Rest), hosting numerous publicly reusable skills. Existing security research on these ecosystems mainly focuses on vulnerabilities within skills, such as prompt injection. However, there is a critical gap regarding skills that may be misused for harmful actions (e.g., cyber attacks, fraud and scams, privacy violations, and sexual content generation), namely harmful skills. In this paper, we present the first large-scale measurement study of harmful skills in agent ecosystems, covering 98,440 skills across two major registries. Using an LLM-driven scoring system grounded in our harmful skill taxonomy, we find that 4.93% of skills (4,858) are harmful, with ClawHub exhibiting an 8.84% harmful rate compared to 3.49% on...
4.Dr.~RTL: Autonomous Agentic RTL Optimization through Tool-Grounded Self-Improvement
Recent advances in large language models (LLMs) have sparked growing interest in automatic RTL optimization for better performance, power, and area (PPA). However, existing methods are still far from realistic RTL optimization. Their evaluation settings are often unrealistic: they are tested on manually degraded, small-scale RTL designs and rely on weak open-source tools. Their optimization methods are also limited, relying on coarse design-level feedback and simple pre-defined rewriting rules. To address these limitations, we present Dr. RTL, an agentic framework for RTL timing optimization in a realistic evaluation environment, with continual self-improvement through reusable optimization skills. We establish a realistic evaluation setting with more challenging RTL designs and an industrial EDA workflow. Within this setting, Dr. RTL per...
5.SWE-TRACE: Optimizing Long-Horizon SWE Agents Through Rubric Process Reward Models and Heuristic Test-Time Scaling
Resolving real-world software engineering (SWE) issues with autonomous agents requires complex, long-horizon reasoning. Current pipelines are bottlenecked by unoptimized demonstration data, sparse execution rewards, and computationally prohibitive inference scaling, which collectively exacerbate token bloat, reward hacking, and policy degradation. We present SWE-TRACE (Trajectory Reduction and Agentic Criteria Evaluation), a unified framework optimizing the SWE agent lifecycle across data curation, reinforcement learning (RL), and test-time inference. First, we introduce an LLM multi-task cascading method, utilizing stepwise oracle verification to distill a 60K-instance Supervised Fine-Tuning (SFT) corpus strictly biased toward token-efficient, shortest-path trajectories. Second, to overcome the instability of sparse outcome rewards, we d...
AI Computation & Hardware
1.Applied Explainability for Large Language Models: A Comparative Study
arXiv:2604.15371v1 Announce Type: new Abstract: Large language models (LLMs) achieve strong performance across many natural language processing tasks, yet their decision processes remain difficult to interpret. This lack of transparency creates challenges for trust, debugging, and deployment in real-world systems. This paper presents an applied comparative study of three explainability techniques: Integrated Gradients, Attention Rollout, and SHAP, on a fine-tuned DistilBERT model for SST-2 sentiment classification. Rather than proposing new methods, the focus is on evaluating the practical behavior of existing approaches under a consistent and reproducible setup. The results show that gradient-based attribution provides more stable and intuitive explanations, while attention-based methods are computationally efficient but less aligne...
2.Think Multilingual, Not Harder: A Data-Efficient Framework for Teaching Reasoning Models to Code-Switch
arXiv:2604.15490v1 Announce Type: new Abstract: Recent developments in reasoning capabilities have enabled large language models to solve increasingly complex mathematical, symbolic, and logical tasks. Interestingly, while reasoning models are often trained to generate monolingual text, these models have also been observed to code-switch (i.e., mix languages). Prior works have either viewed code-switching as an undesirable error, attempted to control code-switching through modifications to input prompts or the output decoding process, or focus on narrow subsets of languages, domains, tasks, and models. We address these gaps by introducing the first linguistically and behaviorally motivated fine-tuning framework for identifying beneficial code-switched reasoning behaviors in large language models and teaching these models to code-switch m...
3.Brain Score Tracks Shared Properties of Languages: Evidence from Many Natural Languages and Structured Sequences
arXiv:2604.15503v1 Announce Type: new Abstract: Recent breakthroughs in language models (LMs) using neural networks have raised the question: how similar are these models' processing to human language processing? Results using a framework called Brain Score (BS) -- predicting fMRI activations during reading from LM activations -- have been used to argue for a high degree of similarity. To understand this similarity, we conduct experiments by training LMs on various types of input data and evaluate them on BS. We find that models trained on various natural languages from many different language families have very similar BS performance. LMs trained on other structured data -- the human genome, Python, and pure hierarchical structure (nested parentheses) -- also perform reasonably well and close to natural languages in some cases. These fi...
4.PolicyBank: Evolving Policy Understanding for LLM Agents
arXiv:2604.15505v1 Announce Type: new Abstract: LLM agents operating under organizational policies must comply with authorization constraints typically specified in natural language. In practice, such specifications inevitably contain ambiguities and logical or semantic gaps that cause the agent's behavior to systematically diverge from the true requirements. We ask: by letting an agent evolve its policy understanding through interaction and corrective feedback from pre-deployment testing, can it autonomously refine its interpretation to close specification gaps? We propose PolicyBank, a memory mechanism that maintains structured, tool-level policy insights and iteratively refines them -- unlike existing memory mechanisms that treat the policy as immutable ground truth, reinforcing "compliant but wrong" behaviors. We also contribute a sy...
5.Consistency Analysis of Sentiment Predictions using Syntactic & Semantic Context Assessment Summarization (SSAS)
arXiv:2604.15547v1 Announce Type: new Abstract: The fundamental challenge of using Large Language Models (LLMs) for reliable, enterprise-grade analytics, such as sentiment prediction, is the conflict between the LLMs' inherent stochasticity (generative, non-deterministic nature) and the analytical requirement for consistency. The LLM inconsistency, coupled with the noisy nature of chaotic modern datasets, renders sentiment predictions too volatile for strategic business decisions. To resolve this, we present a Syntactic & Semantic Context Assessment Summarization (SSAS) framework for establishing context. Context established by SSAS functions as a sophisticated data pre-processing framework that enforces a bounded attention mechanism on LLMs. It achieves this by applying a hierarchical classification structure (Themes, Stories, Clusters)...
AI Machine Learning
1.The Spectral Geometry of Thought: Phase Transitions, Instruction Reversal, Token-Level Dynamics, and Perfect Correctness Prediction in How Transformers Reason
arXiv:2604.15350v1 Announce Type: new Abstract: We discover that large language models exhibit \emph{spectral phase transitions} in their hidden activation spaces when engaging in reasoning versus factual recall. Through systematic spectral analysis across \textbf{11 models} spanning \textbf{5 architecture families} (Qwen, Pythia, Phi, Llama, DeepSeek-R1), we identify \textbf{seven} core phenomena: (1)~\textbf{Reasoning Spectral Compression} -- 9/11 models show significantly lower $\alpha$ for reasoning ($p < 0.05$), with larger effects in stronger models; (2)~\textbf{Instruction Tuning Spectral Reversal} -- base models show reasoning $\alpha < $ factual $\alpha$, while instruction-tuned models reverse this relationship; (3)~\textbf{Architecture-Dependent Generation Taxonomy} -- prompt-to-response shifts partition into expansion, compress...
2.Aletheia: Gradient-Guided Layer Selection for Efficient LoRA Fine-Tuning Across Architectures
arXiv:2604.15351v1 Announce Type: new Abstract: Low-Rank Adaptation (LoRA) has become the dominant parameter-efficient fine-tuning method for large language models, yet standard practice applies LoRA adapters uniformly to all transformer layers regardless of their relevance to the downstream task. We introduce Aletheia, a gradient-guided layer selection method that identifies the most task-relevant layers via a lightweight gradient probe and applies LoRA adapters only to those layers with asymmetric rank allocation. Across 81 experiment rows covering 14 successful models from 8 architecture families (0.5B-72B parameters, including dense and Mixture-of-Experts architectures), with one additional documented failed Pythia/GPT-NeoX attempt in Campaign 2, Aletheia achieves a 15-28% training speedup (mean 23.1%, p < 0.001) with bounded extra fo...
3.Sequential KV Cache Compression via Probabilistic Language Tries: Beyond the Per-Vector Shannon Limit
arXiv:2604.15356v1 Announce Type: new Abstract: Recent work on KV cache quantization, culminating in TurboQuant, has approached the Shannon entropy limit for per-vector compression of transformer key-value caches. We observe that this limit applies to a strictly weaker problem than the one that actually matters: compressing the KV cache as a sequence. The tokens stored in a KV cache are not arbitrary floating-point data -- they are samples from the exact formal language the model was trained on, and the model is by construction a near-optimal predictor of that language. We introduce sequential KV compression, a two-layer architecture that exploits this structure. The first layer, probabilistic prefix deduplication, identifies semantically equivalent shared prefixes across sessions using the trie metric d_T(s, s') = -log_2 P_M(s ^ s') from...
4.Mapping High-Performance Regions in Battery Scheduling across Data Uncertainty, Battery Design, and Planning Horizons
arXiv:2604.15360v1 Announce Type: new Abstract: This study presents a triadic analysis of energy storage operation under multi-stage model predictive control, investigating the interplay between data characteristics, forecast uncertainty, planning horizon, and battery c-rate. Synthetic datasets are generated to systematically explore variations in data profiles and uncertainty, enabling parametrization and the construction of relationships that map these characteristics to optimal horizon length. Results reveal the presence of an effective horizon, defined as the look-ahead length beyond which additional forecast information provides limited operational benefit. Accounting for this horizon can reduce computational costs while maintaining optimal performance. The study provides optimal horizon lengths across a broad range of combinations o...
5.M3R: Localized Rainfall Nowcasting with Meteorology-Informed MultiModal Attention
arXiv:2604.15377v1 Announce Type: new Abstract: Accurate and timely rainfall nowcasting is crucial for disaster mitigation and water resource management. Despite recent advances in deep learning, precipitation prediction remains challenging due to limitations in effectively leveraging diverse multimedia data sources. We introduce M3R, a Meteorology-informed MultiModal attention-based architecture for direct Rainfall prediction that synergistically combines visual NEXRAD radar imagery with numerical Personal Weather Station (PWS) measurements, using a comprehensive pipeline for temporal alignment of heterogeneous meteorological data. With specialized multimodal attention mechanisms, M3R novelly leverages weather station time series as queries to selectively attend to spatial radar features, enabling focused extraction of precipitation sign...
AI Robotics
1.Foundation Models in Robotics: A Comprehensive Review of Methods, Models, Datasets, Challenges and Future Research Directions
arXiv:2604.15395v1 Announce Type: new Abstract: Over the recent years, the field of robotics has been undergoing a transformative paradigm shift from fixed, single-task, domain-specific solutions towards adaptive, multi-function, general-purpose agents, capable of operating in complex, open-world, and dynamic environments. This tremendous advancement is primarily driven by the emergence of Foundation Models (FMs), i.e., large-scale neural-network architectures trained on massive, heterogeneous datasets that provide unprecedented capabilities in multi-modal understanding and reasoning, long-horizon planning, and cross-embodiment generalization. In this context, the current study provides a holistic, systematic, and in-depth review of the research landscape of FMs in robotics. In particular, the evolution of the field is initially delineate...
2.Iterated Invariant EKF for Quadruped Robot Odometry
arXiv:2604.15449v1 Announce Type: new Abstract: Kalman filter-based algorithms are fundamental for mobile robots, as they provide a computationally efficient solution to the challenging problem of state estimation. However, they rely on two main assumptions that are difficult to satisfy in practice: (a) the system dynamics must be linear with Gaussian process noise, and (b) the measurement model must also be linear with Gaussian measurement noise. Previous works have extended assumption (a) to nonlinear spaces through the Invariant Extended Kalman Filter (IEKF), showing that it retains properties similar to those of the classical Kalman filter when the system dynamics are group-affine on a Lie group. More recently, the counterpart of assumption (b) for the same nonlinear setting was addressed in [1]. By means of the proposed Iterated Inva...
3.One-Shot Cross-Geometry Skill Transfer through Part Decomposition
arXiv:2604.15455v1 Announce Type: new Abstract: Given a demonstration, a robot should be able to generalize a skill to any object it encounters-but existing approaches to skill transfer often fail to adapt to objects with unfamiliar shapes. Motivated by examples of improved transfer from compositional modeling, we propose a method for improving transfer by decomposing objects into their constituent semantic parts. We leverage data-efficient generative shape models to accurately transfer interaction points from the parts of a demonstration object to a novel object. We autonomously construct an objective to optimize the alignment of those points on skill-relevant object parts. Our method generalizes to a wider range of object geometries than existing work, and achieves successful one-shot transfer for a range of skills and objects from a si...
4.NeuroMesh: A Unified Neural Inference Framework for Decentralized Multi-Robot Collaboration
arXiv:2604.15475v1 Announce Type: new Abstract: Deploying learned multi-robot models on heterogeneous robots remains challenging due to hardware heterogeneity, communication constraints, and the lack of a unified execution stack. This paper presents NeuroMesh, a multi-domain, cross-platform, and modular decentralized neural inference framework that standardizes observation encoding, message passing, aggregation, and task decoding in a unified pipeline. NeuroMesh combines a dual-aggregation paradigm for reduction- and broadcast-based information fusion with a parallelized architecture that decouples cycle time from end-to-end latency. Our high-performance C++ implementation leverages Zenoh for inter-robot communication and supports hybrid GPU/CPU inference. We validate NeuroMesh on a heterogeneous team of aerial and ground robots across co...
5.Trajectory Planning for Safe Dual Control with Active Exploration
arXiv:2604.15507v1 Announce Type: new Abstract: Planning safe trajectories under model uncertainty is a fundamental challenge. Robust planning ensures safety by considering worst-case realizations, yet ignores uncertainty reduction and leads to overly conservative behavior. Actively reducing uncertainty on-the-fly during a nominal mission defines the dual control problem. Most approaches address this by adding a weighted exploration term to the cost, tuned to trade off the nominal objective and uncertainty reduction, but without formal consideration of when exploration is beneficial. Moreover, safety is enforced in some methods but not in others. We study a budget-constrained dual control problem, where uncertainty is reduced subject to safety and a mission-level cost budget that limits the allowable degradation in task performance due to...
Financial AI
1.Spurious Predictability in Financial Machine Learning
Adaptive specification search generates statistically significant backtests even under martingale-difference nulls. We introduce a falsification audit testing complete predictive workflows against synthetic reference classes, including zero-predictability environments and microstructure placebos. Workflows generating significant walk-forward evidence in these environments are falsified. For passing workflows, we quantify selection-induced performance inflation using an absolute magnitude gap linking optimized in-sample evidence to disjoint walk-forward realizations, adjusted for effective multiplicity. Simulations validate extreme-value scaling under correlated searches and demonstrate detection power under genuine structure. Empirical case studies confirm that many apparent findings represent methodological artifacts rather than genuine ...
2.The Acoustic Camouflage Phenomenon: Re-evaluating Speech Features for Financial Risk Prediction
In computational paralinguistics, detecting cognitive load and deception from speech signals is a heavily researched domain. Recent efforts have attempted to apply these acoustic frameworks to corporate earnings calls to predict catastrophic stock market volatility. In this study, we empirically investigate the limits of acoustic feature extraction (pitch, jitter, and hesitation) when applied to highly trained speakers in in-the-wild teleconference environments. Utilizing a two-stream late-fusion architecture, we contrast an acoustic-based stream with a baseline Natural Language Processing (NLP) stream. The isolated NLP model achieved a recall of 66.25% for tail-risk downside events. Surprisingly, integrating acoustic features via late fusion significantly degraded performance, reducing recall to 47.08%. We identify this degradation as Ac...
3.PRAGMA: Revolut Foundation Model
Modern financial systems generate vast quantities of transactional and event-level data that encode rich economic signals. This paper presents PRAGMA, a family of foundation models for multi-source banking event sequences. Our approach pre-trains a Transformer-based architecture with masked modelling on a large-scale, heterogeneous banking event corpus using a self-supervised objective tailored to the discrete, variable-length nature of financial records. The resulting model supports a wide range of downstream tasks such as credit scoring, fraud detection, and lifetime value prediction: strong performance can be achieved by training a simple linear model on top of the extracted embeddings and can be further improved with lightweight fine-tuning. Through extensive evaluation on downstream tasks, we demonstrate that PRAGMA achieves superior...
4.Quantum Computing for Financial Transformation: A Review of Optimisation, Pricing, Risk, Machine Learning, and Post-Quantum Security
Quantum computing is becoming strategically relevant to finance because several core financial bottlenecks are already defined by combinatorial search, expectation estimation, rare-event analysis, representation learning, and long-horizon cryptographic resilience. This review examines that landscape across five connected domains: constrained portfolio optimisation, derivative pricing, tail-risk and scenario estimation, quantum machine learning, and post-quantum security. Rather than treating these topics as isolated demonstrations, the article studies them as linked layers of a financial-computation stack. Across all five domains, the review applies a common evaluative logic: identify the financial bottleneck, specify the relevant quantum primitive, compare it with an explicit classical benchmark, and assess the result under realistic imp...
5.SBBTS: A Unified Schrödinger-Bass Framework for Synthetic Financial Time Series
We study the problem of generating synthetic time series that reproduce both marginal distributions and temporal dynamics, a central challenge in financial machine learning. Existing approaches typically fail to jointly model drift and stochastic volatility, as diffusion-based methods fix the volatility while martingale transport models ignore drift. We introduce the Schrödinger-Bass Bridge for Time Series (SBBTS), a unified framework that extends the Schrödinger-Bass formulation to multi-step time series. The method constructs a diffusion process that jointly calibrates drift and volatility and admits a tractable decomposition into conditional transport problems, enabling efficient learning. Numerical experiments on the Heston model demonstrate that SBBTS accurately recovers stochastic volatility and correlation parameters that prior Sch...
GSMA Newsroom
1.GSMA Report Urges Japan to Take Bold Action to Convert Technical Excellence into Global Digital Leadership
Summary available at source link.
2.From Rich Text to Video: RCS Universal Profile 4.0 has arrived
Summary available at source link.
3.Mobile Money accounted for $2 trillion in transactions in 2025, doubling since 2021 as active accounts continue to grow
Summary available at source link.
4.Strengthening the Global Fight Against Fraud and Scams – Takeaways from the Global Fraud Summit in Vienna
Summary available at source link.
5.GSMA MWC26 Barcelona closes 20th anniversary edition
Summary available at source link.
Generative AI (arXiv)
1.Evaluating the Progression of Large Language Model Capabilities for Small-Molecule Drug Design
Large Language Models (LLMs) have the potential to accelerate small molecule drug design due to their ability to reason about information from diverse sources and formats. However, their practical utility remains unclear due to the lack of benchmarks that reflect real-world scenarios. In this work, we introduce a suite of chemically-grounded tasks spanning molecular property prediction, molecular representation transformations, and molecular design. Importantly, we formulate these tasks as reinforcement learning (RL) environments, enabling a unified approach for evaluation and post-training. Across three model families, we find that frontier models are increasingly proficient at chemical tasks, but that there is significant room for improvement, especially in experimental settings with low data. Critically, we show that RL-based post-trai...
2.Learning to Reason with Insight for Informal Theorem Proving
Although most of the automated theorem-proving approaches depend on formal proof systems, informal theorem proving can align better with large language models' (LLMs) strength in natural language processing. In this work, we identify a primary bottleneck in informal theorem proving as a lack of insight, namely the difficulty of recognizing the core techniques required to solve complex problems. To address this, we propose a novel framework designed to cultivate this essential reasoning skill and enable LLMs to perform insightful reasoning. We propose $\mathtt{DeepInsightTheorem}$, a hierarchical dataset that structures informal proofs by explicitly extracting core techniques and proof sketches alongside the final proof. To fully exploit this dataset, we design a Progressive Multi-Stage SFT strategy that mimics the human learning process, ...
3.From Benchmarking to Reasoning: A Dual-Aspect, Large-Scale Evaluation of LLMs on Vietnamese Legal Text
The complexity of Vietnam's legal texts presents a significant barrier to public access to justice. While Large Language Models offer a promising solution for legal text simplification, evaluating their true capabilities requires a multifaceted approach that goes beyond surface-level metrics. This paper introduces a comprehensive dual-aspect evaluation framework to address this need. First, we establish a performance benchmark for four state-of-the-art large language models (GPT-4o, Claude 3 Opus, Gemini 1.5 Pro, and Grok-1) across three key dimensions: Accuracy, Readability, and Consistency. Second, to understand the "why" behind these performance scores, we conduct a large-scale error analysis on a curated dataset of 60 complex Vietnamese legal articles, using a novel, expert-validated error typology. Our results reveal a crucial trade-...
4.Information Router for Mitigating Modality Dominance in Vision-Language Models
Vision Language models (VLMs) have demonstrated strong performance across a wide range of benchmarks, yet they often suffer from modality dominance, where predictions rely disproportionately on a single modality. Prior approaches primarily address this issue by steering model's attention allocation, implicitly assuming that all modalities provide sufficient information. However, attention only determines where the model focuses, and cannot enrich information that is missing or ambiguous. In the real world, input modalities often differ in information density and their signal-to-noise ratios. In such cases, simply adjusting model's attention does not resolve the underlying lack of information. In this paper, we propose \textsc{MoIR}: \textit{Multi-modal Information Router}, an information-level fusion method that explicitly reduces informa...
5.Semantic Area Graph Reasoning for Multi-Robot Language-Guided Search
Coordinating multi-robot systems (MRS) to search in unknown environments is particularly challenging for tasks that require semantic reasoning beyond geometric exploration. Classical coordination strategies rely on frontier coverage or information gain and cannot incorporate high-level task intent, such as searching for objects associated with specific room types. We propose \textit{Semantic Area Graph Reasoning} (SAGR), a hierarchical framework that enables Large Language Models (LLMs) to coordinate multi-robot exploration and semantic search through a structured semantic-topological abstraction of the environment. SAGR incrementally constructs a semantic area graph from a semantic occupancy map, encoding room instances, connectivity, frontier availability, and robot states into a compact task-relevant representation for LLM reasoning. T...
Hugging Face Daily Papers
1.RAD-2: Scaling Reinforcement Learning in a Generator-Discriminator Framework
High-level autonomous driving requires motion planners capable of modeling multimodal future uncertainties while remaining robust in closed-loop interactions. Although diffusion-based planners are effective at modeling complex trajectory distributions, they often suffer from stochastic instabilities and the lack of corrective negative feedback when trained purely with imitation learning. To address these issues, we propose RAD-2, a unified generator-discriminator framework for closed-loop planning. Specifically, a diffusion-based generator is used to produce diverse trajectory candidates, while an RL-optimized discriminator reranks these candidates according to their long-term driving quality. This decoupled design avoids directly applying sparse scalar rewards to the full high-dimensional trajectory space, thereby improving optimization ...
2.RL-STPA: Adapting System-Theoretic Hazard Analysis for Safety-Critical Reinforcement Learning
As reinforcement learning (RL) deployments expand into safety-critical domains, existing evaluation methods fail to systematically identify hazards arising from the black-box nature of neural network enabled policies and distributional shift between training and deployment. This paper introduces Reinforcement Learning System-Theoretic Process Analysis (RL-STPA), a framework that adapts conventional STPA's systematic hazard analysis to address RL's unique challenges through three key contributions: hierarchical subtask decomposition using both temporal phase analysis and domain expertise to capture emergent behaviors, coverage-guided perturbation testing that explores the sensitivity of state-action spaces, and iterative checkpoints that feed identified hazards back into training through reward shaping and curriculum design. We demonstrate...
3.Hybrid Decision Making via Conformal VLM-generated Guidance
Building on recent advances in AI, hybrid decision making (HDM) holds the promise of improving human decision quality and reducing cognitive load. We work in the context of learning to guide (LtG), a recently proposed HDM framework in which the human is always responsible for the final decision: rather than suggesting decisions, in LtG the AI supplies (textual) guidance useful for facilitating decision making. One limiting factor of existing approaches is that their guidance compounds information about all possible outcomes, and as a result it can be difficult to digest. We address this issue by introducing ConfGuide, a novel LtG approach that generates more succinct and targeted guidance. To this end, it employs conformal risk control to select a set of outcomes, ensuring a cap on the false negative rate. We demonstrate our approach on a...
4.IE as Cache: Information Extraction Enhanced Agentic Reasoning
Information Extraction aims to distill structured, decision-relevant information from unstructured text, serving as a foundation for downstream understanding and reasoning. However, it is traditionally treated merely as a terminal objective: once extracted, the resulting structure is often consumed in isolation rather than maintained and reused during multi-step inference. Moving beyond this, we propose \textit{IE-as-Cache}, a framework that repurposes IE as a cognitive cache to enhance agentic reasoning. Drawing inspiration from hierarchical computer memory, our approach combines query-driven extraction with cache-aware reasoning to dynamically maintain compact intermediate information and filter noise. Experiments on challenging benchmarks across diverse LLMs demonstrate significant improvements in reasoning accuracy, indicating that IE...
5.The Agentification of Scientific Research: A Physicist's Perspective
This article argues that the most important significance of the AI revolution, especially the rise of large language models, lies not simply in automation, but in a fundamental change in how complex information and human know-how are carried, replicated, and shared. From this perspective, AI for Science is especially important because it may transform not only the efficiency of research, but also the structure of scientific collaboration, discovery, publishing, and evaluation. The article outlines a gradual path from AI as a research tool to AI as a scientific collaborator, and discusses how AI is likely to fundamentally reshape scientific publication. It also argues that continuous learning and diversity of ideas are essential if AI is to play a meaningful role in original scientific discovery.
IEEE Xplore AI
1.Optical Fiber Networks Can Keep Rail Networks Safe
This article is part of our exclusive IEEE Journal Watch series in partnership with IEEE Xplore. Rail networks are vast, which makes it difficult to conduct comprehensive, continuous safety monitoring. Researchers in China have suggested analyzing the vibrations of existing fiber cables buried underground alongside railway tracks to detect problems. In a study published 5 March in the Journal of Optical Communications and Networking , the research group demonstrated through experiments how the technique can successfully identify a number of issues associated with train safety, including faulty train wheels and broken sound barriers alongside the railway tracks. Sasha Dong is a junior chair professor in Southeast University’s School of Transportation, in Nanjing, China. She notes that traditional approaches for monitoring railways—such as ...
2.Boston Dynamics and Google DeepMind Teach Spot to Reason
The amazing and frustrating thing about robots is that they can do almost anything you want them to do, as long as you know how to ask properly. In the not-so-distant past, asking properly meant writing code, and while we’ve thankfully moved beyond that brittle constraint, there’s still an irritatingly inverse correlation between ease of use and complexity of task. AI has promised to change that. The idea is that when AI is embodied within robots—giving AI software a physical presence in the world—those robots will be imbued with reasoning and understanding. This is cutting-edge stuff, though, and while we’ve seen plenty of examples of embodied AI in a research context, finding applications where reasoning robots can provide reliable commercial value has not been easy. Boston Dynamics is one of the few companies to commercially deploy leg...
3.Sarang Gupta Builds AI Systems With Real-World Impact
Like many engineers, Sarang Gupta spent his childhood tinkering with everyday items around the house. From a young age he gravitated to projects that could make a difference in someone’s everyday life. When the family’s microwave plug broke, Gupta and his father figured out how to fix it. When a drawer handle started jiggling annoyingly, the youngster made sure it didn’t do so for long. Sarang Gupta Employer OpenAI in San Francisco Job Data science staff member Member grade Senior member Alma maters The Hong Kong University of Science and Technology; Columbia By age 11, his interest expanded from nuts and bolts to software. He learned programming languages such as Basic and Logo and designed simple programs including one that helped a local restaurant automate online ordering and billing. Gupta, an IEEE senior member, brings his mix of cu...
4.12 Graphs That Explain the State of AI in 2026
The capabilities of leading AI models continue to accelerate, and the largest AI companies, including OpenAI and Anthropic , are hurtling toward IPOs later this year. Yet resentment toward AI continues to simmer, and in some cases has boiled over, especially in the United States, where local governments are beginning to embrace restrictions or outright bans on new data center development. It’s a lot to keep track of, but the 2026 edition of the AI Index from Stanford University’s Human-Centered Artificial Intelligence center pulls it off. The report, which comes in at over 400 pages, includes dozens of data points and graphs that approach the topic from multiple angles, from benchmark scores to investment and public perception. As in prior years (see our coverage from 2021 , 2022 , 2023 , 2024 , and 2025 ), we’ve read the report and ident...
5.GoZTASP: A Zero-Trust Platform for Governing Autonomous Systems at Mission Scale
ZTASP is a mission-scale assurance and governance platform designed for autonomous systems operating in real-world environments. It integrates heterogeneous systems—including drones, robots, sensors, and human operators—into a unified zero-trust architecture. Through Secure Runtime Assurance (SRTA) and Secure Spatio-Temporal Reasoning (SSTR), ZTASP continuously verifies system integrity, enforces safety constraints, and enables resilient operation even under degraded conditions. ZTASP has progressed beyond conceptual design, with operational validation at Technology Readiness Level (TRL) 7 in mission critical environments. Core components, including Saluki secure flight controllers, have reached TRL8 and are deployed in customer systems. While initially developed for high-consequence mission environments, the same assurance challenges are...
MIT Sloan Management
1.How AI Helps the Best and Hurts the Rest
Mark Shaver/theispot.com Can generative AI serve as an effective adviser for business owners and entrepreneurs? Intuitive chat-based natural language interfaces mean that anyone who can read and write can use GenAI tools for a wide range of tasks, even if they lack technical skills. This has obvious appeal for entrepreneurs and small business owners, many […]
2.Lessons From Innovation Pioneer Florence Nightingale
Carolyn Geason-Beissel/MIT SMR | Wellcome Collection Florence Nightingale may be best remembered as the epitome of a kind, caring nurse, but she was also a force for disruptive innovation in health care. Three distinct elements of her work — communicating data compellingly, publicizing clear and simple instructions, and expanding professionalized training — carry timeless lessons […]
3.The Human Side of AI Adoption: Lessons From the Field
Carolyn Geason-Beissel/MIT SMR Not a day goes by without another article being published about how AI could disrupt yet another aspect of our business or personal lives. In recent years, AI adoption has indeed taken off. However, if you pay close attention, you’ll notice a dichotomy. Many examples of successful early adoption of artificial intelligence […]
4.Managing Up: A Skill Set That Matters Now
Carolyn Geason-Beissel/MIT SMR | Getty Images Are you skilled at managing up? If your talents are lacking when it comes to managing and dealing with the people above you in the organizational hierarchy, you can find yourself mired in some unpleasant and career-harming situations. Maybe you’re frustrated by a micromanaging supervisor or feeling marginalized by […]
5.The Trap That Skilled Negotiators Miss
Brian Stauffer/theispot.com Say you walk into a car dealership determined to stay within budget. The salesperson shows you a car you like and quotes a price of $41,435. You know there’s room to negotiate, but when it’s time to counter, that first number quietly takes over. Your counteroffer, the concessions, and the final deal all […]
NBER Working Papers
1.The Impact of Maternal Education on Early Childhood Development -- by Moriam Khanam, Mohammad Hajizadeh, Casey Warman
This study leverages exogenous variation from a secondary school stipend program for female students in rural Bangladesh to estimate the causal effect of maternal education on early childhood development. Using data from the 2019 Bangladesh Multiple Indicator Cluster Survey, we find that the five years of stipend eligibility increase mothers' schooling by about one year. Instrumental variable estimates show that an additional year of maternal education improves early childhood development scores by 0.5 points on a scale of 0-10, with gains in overall developmental readiness (7.5 percentage points) and in the literacy–numeracy (7.7 percentage points) and physical (1.9 percentage points) domains. The results are robust across specifications. We also estimate the effects of maternal education on potential mechanisms, including children's nu...
2.Tariffs and the Term Structure of Inflation Expectations -- by Stéphane Auray, Michael B. Devereux, Anthony M. Diercks, Aurélien Eyquem, Joon Kim
Inflation expectations derived from financial markets exhibited unprecedented dynamics in 2025: the correlation between one-year inflation swaps and one-year-ahead one-year forward rates turned significantly negative for the first time on record. We show that this decoupling occurred primarily on days when tariff news dominated market pricing, using a two-stage event classification validated by Bloomberg news trends. Standard small open-economy New Keynesian models in which tariffs generate a one-time price-level increase imply positive comovement across horizons and cannot explain this pattern. We explain these occurrences through the lens of an amended small open-economy New Keynesian model. Three ingredients prove critical for reproducing the observed negative conditional correlation between spot and forward inflation after tariff sho...
3.Bilateral Conflict Risk and Trade: Military Wars, Trade Wars, and Diplomatic Noise -- by Joshua Aizenman, Rodolphe Desbordes, Jamel Saadaoui
How damaging is a “trade war” compared to a “military war” or a “war of words”? Aggregate conflict indicators cannot say, because they treat missile strikes, sanctions, and diplomatic protests as equivalent. We build a monthly bilateral indicator from GDELT event data, calibrated against human-curated ground truth, that decomposes hostility into four layers: kinetic fighting (“military war”), military posture, sanctions-context tensions (“trade war”), and routine diplomacy. The decomposed panel reveals a secular shift: over the past decade, governments have steadily substituted economic coercion for military confrontation, nearly doubling the trade-weighted share of hostility channelled through sanctions contexts. In a gravity trade model, the aggregate indicator is negative, large, and statistically significant, but the decomposition rev...
4.The “Peace Dividend” of International Trade: A New Empirical Approach -- by Ling Feng, Qiuyue Huang, Zhiyuan Li, Christopher M. Meissner
This paper investigates the causal impact of international trade on interstate military conflicts using global bilateral data from 1962 to 2014. To address endogeneity concerns, we exploit exogenous spatial-temporal variation in international trade stemming from technological advances in air relative to maritime transport. Empirical results demonstrate a strong “peace dividend” of international trade: that is, increased trade significantly reduces the probability and intensity of conflicts between nations. This effect remains robust across specifications and withstands a wide range of potential confounders. Such findings highlight how economic interdependence shapes international conflict—a relationship that is especially relevant amid escalating geopolitical tensions and the global shift toward “decoupling”, “de-risking”, and greater tra...
5.How Have Universities Survived for Nearly a Millenium -- by David M. Cutler, Edward L. Glaeser
How have universities managed to survive and evolve over almost 1,000 years to become wildly heterogeneous, unusually fractious, multi-product, non-profit entities? Universities began as teachers’ guilds, and they still give faculty a remarkable degree of autonomy. That structure attracts and empowers intellectuals, who are selected in part on their taste for knowledge, and those entrepreneurs and philanthropists have enabled universities to morph in ways that firms rarely do. Intellectual autonomy can also explain why universities are so often at odds with legal authorities and why faculty fight so often with each other and with their bosses. This essay presents a model of university organization and sketches the evolution of the university’s products and conflicts over the last 900 years. We also discuss the social value of university e...
NY Fed - Liberty Street
1.Bank Failures: The Roles of Solvency and Liquidity
Do banks fail because of runs or because they become insolvent? Answering this question is central to understanding financial crises and designing effective financial stability policies. Long-run historical evidence reveals that the root cause of bank failures is usually insolvency. The importance of bank runs is somewhat overstated. Runs matter, but in most cases they trigger or accelerate failure at already weak banks, rather than cause otherwise sound banks to fail.
2.The R*–Labor Share Nexus
Over the past quarter century, the U.S. economy has experienced significant declines in both the labor share of income and the natural rate of interest, referred to as R*. Existing research has largely analyzed these two developments in isolation. In this post, we provide a simple model that captures the joint evolution of the labor share and R*, which we call the R*–labor share nexus. Our key finding is that structural changes affecting R* also influence the evolution of the labor share, and thereby wages and prices. This highlights a potentially important channel, absent from many macroeconomic models, through which the factors that determine R* also affect the labor share and, in turn, broader macroeconomic developments, with implications for monetary policy.
3.Use of Gen AI in the Workplace and the Value of Access to Training
The rapid spread of generative AI (AI) tools is reshaping the workplace at a remarkable rate. Yet relatively little is known about whether workers have access to these tools, how the tools affect workers’ daily productivity, and how much workers value the training needed to use the tools effectively. In this post, we shed light on these issues by drawing on supplemental questions in the November 2025 Survey of Consumer Expectations (SCE), fielded to a representative sample of the U.S. population. We find that adoption of AI tools at work is heterogeneous, that a sizable share of workers see AI training as important, and that a significant share of employers are nonetheless not yet providing access to AI tools or training on how to use them.
4.What Millions of Homeowner’s Insurance Contracts Reveal About Risk Sharing
Housing is the largest component of assets held by households in the United States, totaling $48 trillion in 2025. When natural disasters strike, the resulting damage to homes can be large relative to households’ liquid savings. Homeowner’s insurance is the primary financial tool households use to protect themselves against property risk. Despite the economic importance of homeowner’s insurance, we know surprisingly little about how insurance contracts are actually designed with respect to property risk. In this post, which is based on our new paper, “Economics of Property Insurance,” we examine how homeowner’s insurance contracts are structured in practice. Using a new granular dataset covering millions of homeowner’s insurance policies, we document ...
5.A Closer Look at Emerging Market Resilience During Recent Shocks
A succession of shocks to the global economy in recent years has focused attention on the improved economic and financial resilience of emerging market economies. For some of these economies, this assessment is well-founded and highlights the fruits of deep, structural economic reforms since the 1990s. However, for a much larger universe of countries, the ability to weather shocks is still mixed and many remain vulnerable. In this post, we explore the divide between the two sets of countries and focus on the effects of recent economic shocks, including the ongoing conflict in the Middle East.
Project Syndicate
1.What Really Drives China’s Massive Trade Surplus
China’s trade surplus is often blamed on its industrial policies. In reality, however, it reflects a persistent gap between savings and investment, driven by demographic pressures and financial constraints that shape household behavior and restrict private firms’ access to credit.
2.The World Needs an Oil Buyers’ Club
As the world is plunged into another energy crisis, market allocation is leading to grossly unjust outcomes, as the rich outbid the poor. A multilateral oil buyers' club is urgently needed to defend a price ceiling in global oil markets and allocate resources in a way that meets people’s essential needs and minimizes the economic fallout.
3.A New Security Architecture for the Middle East
The tense negotiations between the United States and Iran have exposed the limits of bilateral diplomacy. With the crisis fueled by overlapping, interconnected conflicts, the only viable path forward is a broader regional framework that addresses the Strait of Hormuz, nuclear proliferation, Palestinian statehood, and proxy warfare.
4.Fossil-Fuel Investments Are a Fiduciary Risk
The Iran war has reminded everyone, but especially Africans, of the structural instability of fossil-fuel prices. For African trustees, directors, asset managers, and other fiduciaries, the question is not whether capital should reposition, but whether institutions will act before events compel them to do so.
5.To Strengthen Climate Resilience, Focus on Social Protection
The international community is increasingly trying to distinguish between climate, development, and humanitarian finance—as if they can be neatly compartmentalized. But this siloed approach overlooks how social-protection programs providing cash transfers to vulnerable households can strengthen resilience to climate shocks.
RCR Wireless
1.5G positioning is picking up, but monetization is a problem
Analysys Mason says early adoption of 5G positioning is likely to come from localized private network environments In sum – what to know: Early adoption – Private networks in logistics, manufacturing, and healthcare are expected to lead uptake of high-precision…
2.How ISPs can win in a saturated US broadband market (Analyst Angle)
US broadband has rapidly transformed, with fiber, fixed wireless and satellite expanding competition. As coverage rises and prices fall, ISPs must shift focus toward retaining subscribers and securing long-term revenue, especially in MDUs opportunity segment. The American broadband landscape has…
3.Ericsson bets on enterprise 5G and APIs for longer-term AI upside – versus DCI game
Ericsson is sticking to what it knows: 5G, public and private, and APIs, to expose 5G capabilities to developers and enterprises; it offers more coherent longer-term diversification, it implies, than a Nokia-style switch to ride the AI bandwagon on fiber…
4.Ericsson posts 6% organic growth but misses targets amid FX drag and AI chip costs
Ericsson saw 6% organic growth in Q1 2026, but slumped 10% in real terms with currency swings, divestment costs, and higher AI chip prices – causing it to miss targets. Network sales in EMEA and APAC made up for a…
5.Fast-charging quantum batteries could make devices run forever
Quantum batteries could enable wireless charging, allowing systems to stay in a state of constant charging, says James Quach, whose team has developed a working prototype of a quantum battery that charges in quadrillionths of a second In sum —…
Semantic Scholar – Machine Learning
1.Source Error
Check Feed
Telecom & 6G AI
1.Toward EU Sovereignty in Space: A Comparative Simulation Study of IRIS 2 and Starlink
The evolution of 6th generation (6G) networks increasingly relies on satellite-based Non-Terrestrial Networks (NTNs) to extend broadband connectivity to remote and unserved regions, and to support public safety. In this paper we compare two representative and conceptually different satellite constellation architectures, namely Starlink and IRIS 2. Starlink is a commercial private Internet constellation by SpaceX, based on dense Low Earth Orbit (LEO) satellites. It is primarily designed to deliver high-capacity broadband services for civil applications, with performance targets comparable to those of terrestrial networks. In contrast, IRIS 2 is a planned public initiative to be deployed by the European Union, based on a multi-layer combination of LEO, Medium Earth Orbit (MEO), and Geo-stationary Earth Orbit (GEO) satellites. It is primaril...
2.Impact of Nonlinear Power Amplifier on Massive MIMO: Machine Learning Prediction Under Realistic Radio Channel
M-MIMO is one of the crucial technologies for increasing spectral and energy efficiency of wireless networks. Most of the current works assume that M-MIMO arrays are equipped with a linear front end. However, ongoing efforts to make wireless networks more energy-efficient push the hardware to the limits, where its nonlinear behavior appears. This is especially a common problem for the multicarrier systems, e.g., OFDM used in 4G, 5G, and possibly also in 6G, which is characterized by a high Peak-to-Average Power Ratio. While the impact of a nonlinear Power Amplifier (PA) on an OFDM signal is well characterized, it is a relatively new topic for the M-MIMO OFDM systems. Most of the recent works either neglect nonlinear effects or utilize simplified models proper for Rayleigh or LoS radio channel models. In this paper, we first theoretically ...
3.FP-ANeT: A Fixed-Point Attention Network for Hybrid-Field THz Ultra-massive MIMO Channel Estimation
Ultra-massive multiple-input multiple-output (UM-MIMO) is a key technology for enabling terahertz (THz) communications in 6G networks, offering high beamforming gain to combat severe path loss. However, the large antenna array expands the near-field region, resulting in a hybrid near- and far-field communication environment. This makes channel estimation significantly more challenging than in conventional networks. To address this issue, we propose a novel attention augmented channel estimator named the fixed-point attention network (FP-ANet), which integrates fixed-point theory with a dual-attention mechanism. By combining a linear and dual-attention residual blocks based non-linear estimator in each iteration, this model-driven approach effectively exploits the sparsity of THz channels in the angular-distance domain, enabling a more pre...
4.Measurement-Based Massive MIMO Channel Characterization and Performance Evaluation at FR3 (8 and 15 GHz) Under Equal Physical Aperture
With the push toward 6G commercialization, Frequency Range 3 (FR3) bands, specifically 7.125-8.4 GHz and 14.8-15.3 GHz, have become focal points for achieving wide-area, high-capacity coverage. However, practical deployment is often limited by the physical aperture constraints of base station antennas. This study conducts comprehensive measurements in Urban Macro (UMa) scenarios using a unified dual-band sounding platform to evaluate channel characteristics and system performance under the strict constraint of "equal physical array aperture." The results indicate that higher frequency bands exhibit increased sparsity in both delay and spatial domains. Regarding coverage, while the 15 GHz band can theoretically accommodate four times the number of antenna elements (128 elements) within the same area to compensate for path loss, empirical d...
5.A Novel 6G Dynamic Channel Map Based on a Hybrid Channel Model
In the sixth generation (6G) wireless communication networks, the device density, antenna number, and the complexity of communication scenarios will significantly increase, which brings great challenges for system design and network optimization. By obtaining channel information in advance, channel map has become a promising solution to these challenges in 6G era. However, conventional channel maps cannot be updated in time as physical environment changes. To solve the problem, a novel dynamic channel map (DCM) is proposed in this work. For DCM construction, we further present a ray tracing (RT) and geometric stochastic hybrid channel model (RT-GSHCM), which pre-constructs the DCM offline by RT and updates it online by geometry-based stochastic channel model (GBSM). By this way, the DCM can provide time-varying channel information and cha...
arXiv Quantitative Finance
1.Spurious Predictability in Financial Machine Learning
Adaptive specification search generates statistically significant backtests even under martingale-difference nulls. We introduce a falsification audit testing complete predictive workflows against synthetic reference classes, including zero-predictability environments and microstructure placebos. Workflows generating significant walk-forward evidence in these environments are falsified. For passing workflows, we quantify selection-induced performance inflation using an absolute magnitude gap linking optimized in-sample evidence to disjoint walk-forward realizations, adjusted for effective multiplicity. Simulations validate extreme-value scaling under correlated searches and demonstrate detection power under genuine structure. Empirical case studies confirm that many apparent findings represent methodological artifacts rather than genuine ...
2.Broken Symmetry, Conservation Law, and Scaling in Accumulated Stock Returns -- a Modified Jones-Faddy Skew t-Distribution Perspective
We analyze historic S&P500 multi-day returns: from daily returns to those accumulated over up to ten days. Despite symmetry breaking between gains and losses in the distribution of returns, resulting in its positive mean and negative skew, realized variance (volatility squared) exhibits remarkably good linear dependence on the number of days of accumulation. Mean of the distribution also shows near perfect linear dependence as well. We analyze this phenomenon both analytically and numerically using a modified Jones-Faddy skew t-distribution.
3.The Acoustic Camouflage Phenomenon: Re-evaluating Speech Features for Financial Risk Prediction
In computational paralinguistics, detecting cognitive load and deception from speech signals is a heavily researched domain. Recent efforts have attempted to apply these acoustic frameworks to corporate earnings calls to predict catastrophic stock market volatility. In this study, we empirically investigate the limits of acoustic feature extraction (pitch, jitter, and hesitation) when applied to highly trained speakers in in-the-wild teleconference environments. Utilizing a two-stream late-fusion architecture, we contrast an acoustic-based stream with a baseline Natural Language Processing (NLP) stream. The isolated NLP model achieved a recall of 66.25% for tail-risk downside events. Surprisingly, integrating acoustic features via late fusion significantly degraded performance, reducing recall to 47.08%. We identify this degradation as Ac...
4.Interpretable Systematic Risk around the Clock
In this paper, I present the first comprehensive, around-the-clock analysis of systematic jump risk by combining high-frequency market data with contemporaneous news narratives identified as the underlying causes of market jumps. These narratives are retrieved and classified using a state-of-the-art open-source reasoning LLM. Decomposing market risk into interpretable jump categories reveals significant heterogeneity in risk premia, with macroeconomic news commanding the largest and most persistent premium. Leveraging this insight, I construct an annually rebalanced real-time Fama-MacBeth factor-mimicking portfolio that isolates the most strongly priced jump risk, achieving a high out-of-sample Sharpe ratio and delivering significant alphas relative to standard factor models. The results highlight the value of around-the-clock analysis an...
5.Forecasting Oil Prices Across the Distribution: A Quantile VAR Approach
We develop a Quantile Bayesian Vector Autoregression (QBVAR) to forecast real oil prices across different quantiles of the conditional distribution. The model allows predictor effects to vary across quantiles, capturing asymmetries that standard mean-focused approaches miss. Using monthly data from 1975 to 2025, we document three findings. First, the QBVAR improves median forecasts by 2-5\% relative to Bayesian VARs, demonstrating that quantile-specific dynamics matter even for point prediction. Second, uncertainty and financial condition variables strongly predict downside risk, with left-tail forecast improvements of 10-25\% that intensify during crisis episodes. Third, right-tail forecasting remains difficult; stochastic volatility models dominate for upside risk, though forecast combinations that include the QBVAR recover these losses...
arXiv – 6G & Networking
1.Toward EU Sovereignty in Space: A Comparative Simulation Study of IRIS 2 and Starlink
The evolution of 6th generation (6G) networks increasingly relies on satellite-based Non-Terrestrial Networks (NTNs) to extend broadband connectivity to remote and unserved regions, and to support public safety. In this paper we compare two representative and conceptually different satellite constellation architectures, namely Starlink and IRIS 2. Starlink is a commercial private Internet constellation by SpaceX, based on dense Low Earth Orbit (LEO) satellites. It is primarily designed to deliver high-capacity broadband services for civil applications, with performance targets comparable to those of terrestrial networks. In contrast, IRIS 2 is a planned public initiative to be deployed by the European Union, based on a multi-layer combination of LEO, Medium Earth Orbit (MEO), and Geo-stationary Earth Orbit (GEO) satellites. It is primaril...
2.Impact of Nonlinear Power Amplifier on Massive MIMO: Machine Learning Prediction Under Realistic Radio Channel
M-MIMO is one of the crucial technologies for increasing spectral and energy efficiency of wireless networks. Most of the current works assume that M-MIMO arrays are equipped with a linear front end. However, ongoing efforts to make wireless networks more energy-efficient push the hardware to the limits, where its nonlinear behavior appears. This is especially a common problem for the multicarrier systems, e.g., OFDM used in 4G, 5G, and possibly also in 6G, which is characterized by a high Peak-to-Average Power Ratio. While the impact of a nonlinear Power Amplifier (PA) on an OFDM signal is well characterized, it is a relatively new topic for the M-MIMO OFDM systems. Most of the recent works either neglect nonlinear effects or utilize simplified models proper for Rayleigh or LoS radio channel models. In this paper, we first theoretically ...
3.FP-ANeT: A Fixed-Point Attention Network for Hybrid-Field THz Ultra-massive MIMO Channel Estimation
Ultra-massive multiple-input multiple-output (UM-MIMO) is a key technology for enabling terahertz (THz) communications in 6G networks, offering high beamforming gain to combat severe path loss. However, the large antenna array expands the near-field region, resulting in a hybrid near- and far-field communication environment. This makes channel estimation significantly more challenging than in conventional networks. To address this issue, we propose a novel attention augmented channel estimator named the fixed-point attention network (FP-ANet), which integrates fixed-point theory with a dual-attention mechanism. By combining a linear and dual-attention residual blocks based non-linear estimator in each iteration, this model-driven approach effectively exploits the sparsity of THz channels in the angular-distance domain, enabling a more pre...
4.Measurement-Based Massive MIMO Channel Characterization and Performance Evaluation at FR3 (8 and 15 GHz) Under Equal Physical Aperture
With the push toward 6G commercialization, Frequency Range 3 (FR3) bands, specifically 7.125-8.4 GHz and 14.8-15.3 GHz, have become focal points for achieving wide-area, high-capacity coverage. However, practical deployment is often limited by the physical aperture constraints of base station antennas. This study conducts comprehensive measurements in Urban Macro (UMa) scenarios using a unified dual-band sounding platform to evaluate channel characteristics and system performance under the strict constraint of "equal physical array aperture." The results indicate that higher frequency bands exhibit increased sparsity in both delay and spatial domains. Regarding coverage, while the 15 GHz band can theoretically accommodate four times the number of antenna elements (128 elements) within the same area to compensate for path loss, empirical d...
5.A Novel 6G Dynamic Channel Map Based on a Hybrid Channel Model
In the sixth generation (6G) wireless communication networks, the device density, antenna number, and the complexity of communication scenarios will significantly increase, which brings great challenges for system design and network optimization. By obtaining channel information in advance, channel map has become a promising solution to these challenges in 6G era. However, conventional channel maps cannot be updated in time as physical environment changes. To solve the problem, a novel dynamic channel map (DCM) is proposed in this work. For DCM construction, we further present a ray tracing (RT) and geometric stochastic hybrid channel model (RT-GSHCM), which pre-constructs the DCM offline by RT and updates it online by geometry-based stochastic channel model (GBSM). By this way, the DCM can provide time-varying channel information and cha...
arXiv – Network Architecture (6G/Slicing)
1.Toward EU Sovereignty in Space: A Comparative Simulation Study of IRIS 2 and Starlink
The evolution of 6th generation (6G) networks increasingly relies on satellite-based Non-Terrestrial Networks (NTNs) to extend broadband connectivity to remote and unserved regions, and to support public safety. In this paper we compare two representative and conceptually different satellite constellation architectures, namely Starlink and IRIS 2. Starlink is a commercial private Internet constellation by SpaceX, based on dense Low Earth Orbit (LEO) satellites. It is primarily designed to deliver high-capacity broadband services for civil applications, with performance targets comparable to those of terrestrial networks. In contrast, IRIS 2 is a planned public initiative to be deployed by the European Union, based on a multi-layer combination of LEO, Medium Earth Orbit (MEO), and Geo-stationary Earth Orbit (GEO) satellites. It is primaril...
2.Towards Trustworthy 6G Network Digital Twins: A Framework for Validating Counterfactual What-If Analysis in Edge Computing Resources
Network Digital Twins (NDTs) enable safe what-if analysis for 6G cloud-edge infrastructures, but adoption is often limited by fragmented workflows from telemetry to validation. We present a data-driven NDT framework that extends 6G-TWIN with a scalable pipeline for cloud-edge telemetry aggregation and semantic alignment into unified data models. Our contributions include: (i) scalable cloud-edge telemetry collection, (ii) regime-aware feature engineering capturing the network's scaling behavior, and (iii) a validation methodology based on Sign Agreement and Directional Sensitivity. Evaluated on a Kubernetes-managed cluster, the framework extrapolates performance to unseen high-load regimes. Results show both Deep Neural Network (DNN) and XGBoost achieve high regression accuracy (R2 > 0.99), while the XGBoost model delivers superior dir...
3.Cross-Domain Query Translation for Network Troubleshooting: A Multi-Agent LLM Framework with Privacy Preservation and Self-Reflection
This paper presents a hierarchical multi-agent LLM architecture to bridge communication gaps between non-technical end users and telecommunications domain experts in private network environments. We propose a cross-domain query translation framework that leverages specialized language models coordinated through multi-agent reflection-based reasoning. The resulting system addresses three critical challenges: (1) accurately classify user queries related to telecommunications network issues using a dual-stage hierarchical approach, (2) preserve user privacy through the anonymization of semantically relevant personally identifiable information (PII) while maintaining diagnostic utility, and (3) translate technical expert responses into user-comprehensible language. Our approach employs ReAct-style agents enhanced with self-reflection mechan...
4.The Missing Pillar in Quantum-Safe 6G: Regulation and Global Compliance
Sixth-generation (6G) mobile networks are expected to operate for multiple decades, supporting mission-critical and globally federated digital services. This long operational horizon coincides with rapid advances in quantum computing that threaten the cryptographic foundations of contemporary mobile systems. While post-quantum cryptography is widely recognized as a necessary technical response, its effective deployment in 6G depends equally on the evolution of regulatory policy and global compliance frameworks. This article argues that quantum-safe 6G represents a regulatory inflection point for mobile networks, as existing compliance models shaped by static cryptographic assumptions, incremental evolution, and point-in-time certification are poorly suited to long-term quantum risk. Building on an analysis of baseline telecom compliance c...
5.Advancing Network Digital Twin Framework for Generating Realistic Datasets
The integration of accurate and reproducible wireless network simulations is a key enabler for research on open, virtualized, and intelligent communication systems. Network Digital Twins (NDTs) provide a scalable alternative to costly and time-consuming measurement campaigns, while enabling controlled experimentation and data generation for data-driven network design. In this paper, we present an open and user-friendly NDT framework that integrates controllable vehicular mobility with the site-specific ray tracer Sionna and the discrete-event ns-3 network simulator, enabling virtualized end-to-end modeling of wireless networks across the radio, network, and application layers. The proposed framework is particularly well-suited for dynamic vehicular networks and urban deployments, supporting realistic mobility, traffic dynamics, and the ex...