Editor’s Picks
Top papers matching your research interests in multimodal LLMs, audio and vision understanding/generation.
[1] AuTAgent: A Reinforcement Learning Framework for Tool-Augmented Audio Reasoning
Siqian Tong, Xuan Li, Yiwei Wang, Baolong Bi, Yujun Cai, Shenghua Liu, Yuchen He, Chengpeng Hao
Main category: cs.SD
TL;DR: AuTAgent is a reinforcement learning framework that teaches audio language models when and which external tools to use for precise acoustic measurements, improving reasoning without information overload.
Details
Motivation: Large Audio Language Models (LALMs) are good at perception but struggle with complex reasoning requiring precise acoustic measurements. While external tools can extract fine-grained features like tempo or pitch, integrating them effectively is challenging - using all tools causes information overload, while prompt-based selection fails to assess context-dependent utility.Method: Proposes AuTAgent (Audio Tool Agent), a reinforcement learning framework that learns when and which tools to invoke. Uses a sparse-feedback training strategy with a novel Differential Reward mechanism to filter out irrelevant tools and invoke external assistance only when it yields net performance gain over the base model.
Result: Improves accuracy by 4.20%/6.20% and 9.80%/8.00% for open-source and closed-source backbones on MMAU Test-mini and MMAR benchmarks. Demonstrates exceptional transferability and shows that AuTAgent complements the representation bottleneck of LALMs by providing verifiable acoustic evidence.
Conclusion: The framework successfully addresses the tool integration challenge in audio language models, highlighting the complementary role of external tools in augmenting audio model reasoning through learned tool invocation strategies.
Abstract: Large Audio Language Models (LALMs) excel at perception but struggle with complex reasoning requiring precise acoustic measurements. While external tools can extract fine-grained features like exact tempo or pitch, effective integration remains challenging: naively using all tools causes information overload, while prompt-based selection fails to assess context-dependent utility. To address this, we propose AuTAgent (Audio Tool Agent), a reinforcement learning framework that learns when and which tools to invoke. By employing a sparse-feedback training strategy with a novel Differential Reward mechanism, the agent learns to filter out irrelevant tools and invokes external assistance only when it yields a net performance gain over the base model. Experimental results confirm that AuTAgent complements the representation bottleneck of LALMs by providing verifiable acoustic evidence. It improves accuracy by 4.20% / 6.20% and 9.80% / 8.00% for open-source and closed-source backbones on the MMAU Test-mini and the MMAR benchmarks, respectively. In addition, further experiments demonstrate exceptional transferability. We highlight the complementary role of external tools in augmenting audio model reasoning.
Relevance: 9/10
[2] Evaluating the Impact of Post-Training Quantization on Reliable VQA with Multimodal LLMs
Paul Jonas Kurz, Tobias Jan Wieczorek, Mohamed A. Abdelsalam, Rahaf Aljundi, Marcus Rohrbach
Main category: cs.CV
TL;DR: Systematic study of how post-training quantization affects both accuracy and reliability in multimodal LLMs for VQA, showing data-aware methods and confidence estimators can mitigate reliability degradation.
Details
Motivation: MLLMs face dual challenges: overconfidence in incorrect answers and large size limiting edge deployment. Need to understand how compression (quantization) affects both accuracy and reliability in multimodal settings.Method: Evaluate Qwen2-VL-7B and Idefics3-8B with PTQ using data-free (HQQ) and data-aware (MBQ) methods across multiple bit widths. Adapt Selector confidence estimator for quantized multimodal settings and test robustness across quantization levels and OOD scenarios.
Result: PTQ degrades both accuracy and reliability. Data-aware methods soften the effect. Selector substantially mitigates reliability impact. Int4 MBQ + Selector achieves best efficiency-reliability trade-off, closing in on uncompressed performance at ~75% less memory.
Conclusion: First systematic study linking quantization and reliability in multimodal settings. Shows data-aware quantization methods combined with confidence estimators can maintain reliability while achieving significant compression for edge deployment.
Abstract: Multimodal Large Language Models (MLLM) are increasingly deployed in domains where both reliability and efficiency are critical. However, current models remain overconfident, producing highly certain but incorrect answers. At the same time, their large size limits deployment on edge devices, necessitating compression. We study the intersection of these two challenges by analyzing how Post-Training Quantization (PTQ) compression affects both accuracy and reliability in Visual Question Answering (VQA). We evaluate two MLLMs, Qwen2-VL-7B and Idefics3-8B, quantized with data-free (HQQ) and data-aware (MBQ) methods across multiple bit widths. To counteract the reduction in reliability caused by quantization, we adapt the Selector confidence estimator for quantized multimodal settings and test its robustness across various quantization levels and out-of-distribution (OOD) scenarios. We find that PTQ degrades both accuracy and reliability. Data-aware methods soften the effect thereof. The Selector substantially mitigates the reliability impact. The combination of int4 MBQ and the Selector achieves the best efficiency-reliability trade-off, closing in on uncompressed performance at approx. 75% less memory demand. Overall, we present the first systematic study linking quantization and reliability in multimodal settings.
Relevance: 9/10
[3] AudioX: A Unified Framework for Anything-to-Audio Generation
Zeyue Tian, Zhaoyang Liu, Yizhu Jin, Ruibin Yuan, Xu Tan, Qifeng Chen, Wei Xue, Yike Guo
Main category: cs.MM
TL;DR: AudioX is a unified framework for multimodal-conditioned audio generation that handles text, video, and audio inputs through a Multimodal Adaptive Fusion module, trained on a large-scale dataset IF-caps with 7M+ samples.
Details
Motivation: The paper addresses two key challenges in multimodal audio generation: 1) lack of a unified modeling framework for diverse multimodal conditions, and 2) scarcity of large-scale, high-quality training data for multimodal-conditioned audio generation.Method: Proposes AudioX framework with a Multimodal Adaptive Fusion module that effectively fuses diverse multimodal inputs (text, video, audio) to enhance cross-modal alignment. Constructs IF-caps dataset with 7M+ samples through structured annotation pipeline for comprehensive supervision.
Result: AudioX achieves superior performance compared to state-of-the-art methods, especially in text-to-audio and text-to-music generation, demonstrating powerful instruction-following potential for multimodal control signals.
Conclusion: AudioX provides an effective unified framework for anything-to-audio generation with multimodal conditions, showing strong performance across various tasks and promising potential for instruction-following audio generation.
Abstract: Audio and music generation based on flexible multimodal control signals is a widely applicable topic, with the following key challenges: 1) a unified multimodal modeling framework, and 2) large-scale, high-quality training data. As such, we propose AudioX, a unified framework for anything-to-audio generation that integrates varied multimodal conditions (i.e., text, video, and audio signals) in this work. The core design in this framework is a Multimodal Adaptive Fusion module, which enables the effective fusion of diverse multimodal inputs, enhancing cross-modal alignment and improving overall generation quality. To train this unified model, we construct a large-scale, high-quality dataset, IF-caps, comprising over 7 million samples curated through a structured data annotation pipeline. This dataset provides comprehensive supervision for multimodal-conditioned audio generation. We benchmark AudioX against state-of-the-art methods across a wide range of tasks, finding that our model achieves superior performance, especially in text-to-audio and text-to-music generation. These results demonstrate our method is capable of audio generation under multimodal control signals, showing powerful instruction-following potential. The code and datasets will be available at https://zeyuet.github.io/AudioX/.
Relevance: 9/10
Today’s Research Highlights
AI-enhanced summaries of the latest research papers from arXiv.
Table of Contents
- cs.CL [Total: 135]
- cs.CV [Total: 240]
- cs.AI [Total: 153]
- cs.SD [Total: 19]
- cs.LG [Total: 278]
- cs.MA [Total: 18]
- cs.MM [Total: 3]
- eess.AS [Total: 7]
- eess.IV [Total: 11]
cs.CL
[1] Multimodal Consistency-Guided Reference-Free Data Selection for ASR Accent Adaptation
Ligong Lei, Wenwen Lu, Xudong Pang, Zaokere Kadeer, Aishan Wumaier
Main category: cs.CL
TL;DR: Multimodal consistency-guided data selection pipeline for ASR accent adaptation using speech-text alignment and predicted WER for pseudo-label filtering.
Details
Motivation: ASR systems degrade on accented speech due to acoustic-phonetic and prosodic mismatches, and existing text-centric pseudo-label selection methods can amplify errors by preferring fluent but acoustically mismatched hypotheses.Method: Proposes a multimodal consistency-guided pipeline with: 1) target-aware preselection using submodular mutual information, 2) multiple pseudo-transcriptions via perturbation-based decoding, 3) scoring using speech-text alignment in shared embedding space and predicted WER, and 4) percentile-based selection of reliable pseudo-labels.
Result: In-domain: Selecting ~1.5k utterances from 30k pool achieves 10.91% WER (close to 10.45% with supervised labels). Cross-domain: Consistency-filtered subsets avoid degradation from unfiltered pseudo-labels under accent shift, outperforming random sampling and recent baselines.
Conclusion: Multimodal consistency filtering effectively selects reliable pseudo-labels for ASR accent adaptation, achieving near-supervised performance with much less data and preventing error amplification in cross-domain scenarios.
Abstract: Automatic speech recognition (ASR) systems often degrade on accented speech because acoustic-phonetic and prosodic shifts induce a mismatch to training data, making labeled accent adaptation costly. However, common pseudo-label selection heuristics are largely text-centric (e.g., perplexity (PPL) filtering) and can prefer fluent yet acoustically mismatched hypotheses, leading to error amplification when fine-tuning. To address this, we introduce a multimodal consistency-guided, reference-free data selection pipeline for ASR accent adaptation under a transductive, label-free protocol. The pipeline starts with a target-aware preselection step based on submodular mutual information to improve query relevance and reduce downstream computation. It then generates multiple pseudo-transcriptions per utterance via perturbation-based decoding and scores each hypothesis using two reference-free signals: speech–text alignment in a shared embedding space and predicted word error rate (WER). A simple percentile-based selection rule retains reliable pseudo-labels for fine-tuning while discarding noisy utterances. In an in-domain setting, selecting ~1.5k utterances from a 30k pool achieves 10.91% WER, close to 10.45% obtained using 30k supervised labels. In a cross-domain setting with a mismatched candidate pool, consistency-filtered subsets avoid the degradation caused by unfiltered pseudo-labels under strong accent shift, and matched-hour experiments on a stronger ASR backbone further confirm gains over random sampling and recent selection baselines.
[2] LLM-Powered Automatic Translation and Urgency in Crisis Scenarios
Belu Ticona, Antonis Anastasopoulos
Main category: cs.CL
TL;DR: LLMs and translation systems show poor performance in crisis communication, especially in preserving urgency across languages, highlighting risks in using general-purpose language tech for high-stakes crisis contexts.
Details
Motivation: To evaluate the suitability of LLMs and machine translation systems for crisis preparedness and response, particularly focusing on their ability to preserve urgency in multilingual crisis communication.Method: Examined state-of-the-art LLMs and translation systems using multilingual crisis data and a newly introduced urgency-annotated dataset covering over 32 languages, analyzing performance degradation and instability in crisis-domain translation.
Result: Both dedicated translation models and LLMs exhibit substantial performance degradation and instability in crisis contexts. Even linguistically adequate translations can distort perceived urgency, and LLM-based urgency classifications vary widely depending on language of prompt and input.
Conclusion: Significant risks exist in deploying general-purpose language technologies for crisis communication, underscoring the need for crisis-aware evaluation frameworks and specialized approaches for high-stakes multilingual crisis contexts.
Abstract: Large language models (LLMs) are increasingly proposed for crisis preparedness and response, particularly for multilingual communication. However, their suitability for high-stakes crisis contexts remains insufficiently evaluated. This work examines the performance of state-of-the-art LLMs and machine translation systems in crisis-domain translation, with a focus on preserving urgency, which is a critical property for effective crisis communication and triaging. Using multilingual crisis data and a newly introduced urgency-annotated dataset covering over 32 languages, we show that both dedicated translation models and LLMs exhibit substantial performance degradation and instability. Crucially, even linguistically adequate translations can distort perceived urgency, and LLM-based urgency classifications vary widely depending on the language of the prompt and input. These findings highlight significant risks in deploying general-purpose language technologies for crisis communication and underscore the need for crisis-aware evaluation frameworks.
[3] Using Machine Learning to Enhance the Detection of Obfuscated Abusive Words in Swahili: A Focus on Child Safety
Phyllis Nabangi, Abdul-Jalil Zakaria, Jema David Ndibwile
Main category: cs.CL
TL;DR: Machine learning models (SVM, Logistic Regression, Decision Trees) applied to detect abusive obfuscated language in Swahili, a low-resource language, with SMOTE for data imbalance handling.
Details
Motivation: Addressing cyberbullying detection challenges in low-resource languages like Swahili, which has limited linguistic resources despite being widely spoken in Africa, to create safer online environments for children.Method: Used Support Vector Machines (SVM), Logistic Regression, and Decision Trees with parameter tuning and Synthetic Minority Over-sampling Technique (SMOTE) to handle data imbalance in Swahili text classification.
Result: Models performed well on high-dimensional textual data but limited by small, imbalanced dataset affecting generalizability; precision, recall, and F1 scores analyzed to show nuanced model performance.
Conclusion: Highlights need for expanded datasets, advanced ML techniques, and future work on data robustness, transfer learning, and multimodal data integration for culturally sensitive cyberbullying detection.
Abstract: The rise of digital technology has dramatically increased the potential for cyberbullying and online abuse, necessitating enhanced measures for detection and prevention, especially among children. This study focuses on detecting abusive obfuscated language in Swahili, a low-resource language that poses unique challenges due to its limited linguistic resources and technological support. Swahili is chosen due to its popularity and being the most widely spoken language in Africa, with over 16 million native speakers and upwards of 100 million speakers in total, spanning regions in East Africa and some parts of the Middle East. We employed machine learning models including Support Vector Machines (SVM), Logistic Regression, and Decision Trees, optimized through rigorous parameter tuning and techniques like Synthetic Minority Over-sampling Technique (SMOTE) to handle data imbalance. Our analysis revealed that, while these models perform well in high-dimensional textual data, our dataset’s small size and imbalance limit our findings’ generalizability. Precision, recall, and F1 scores were thoroughly analyzed, highlighting the nuanced performance of each model in detecting obfuscated language. This research contributes to the broader discourse on ensuring safer online environments for children, advocating for expanded datasets and advanced machine-learning techniques to improve the effectiveness of cyberbullying detection systems. Future work will focus on enhancing data robustness, exploring transfer learning, and integrating multimodal data to create more comprehensive and culturally sensitive detection mechanisms.
[4] Language Model Memory and Memory Models for Language
Benjamin L. Badger
Main category: cs.CL
TL;DR: Language model embeddings have poor memory capabilities compared to autoencoders, but combining causal and information retention objectives enables better memory formation for computational efficiency.
Details
Motivation: To understand and improve the memory capabilities of language model embeddings, which currently store little input information despite being widely used, and to develop more computationally efficient architectures that can better retain and access information.Method: Introduces a parallelizable encoder-decoder memory model architecture that combines causal training (next token prediction) with information retention objectives. Uses frozen high-fidelity encoders and curriculum training where decoders first learn to process memories, then predict next tokens.
Result: Shows that standard language model embeddings contain little input information regardless of training scale, while autoencoder embeddings can achieve near-perfect memory. The proposed combined objective approach enables information-rich memory formation and decoding.
Conclusion: Next token prediction alone is poorly suited for accurate memory formation due to its non-invertible nature. Combined objective functions are necessary for models to form and decode information-rich memories, especially when the entire input is not exposed.
Abstract: The ability of machine learning models to store input information in hidden layer vector embeddings, analogous to the concept of `memory’, is widely employed but not well characterized. We find that language model embeddings typically contain relatively little input information regardless of data and compute scale during training. In contrast, embeddings from autoencoders trained for input regeneration are capable of nearly perfect memory formation. The substitution of memory embeddings for token sequences leads to substantial computational efficiencies, motivating the introduction of a parallelizable encoder-decoder memory model architecture. Upon causal training these models contain information-poor embeddings incapable of arbitrary information access, but by combining causal and information retention objective functions they learn to form and decode information-rich memories. Training can be further streamlined by freezing a high fidelity encoder followed by a curriculum training approach where decoders first learn to process memories and then learn to additionally predict next tokens. We introduce the perspective that next token prediction training alone is poorly suited for accurate memory formation as the objective itself is non-invertible, motivating the use of combined objective functions for models where the entire input is not exposed.
[5] From Perceptions To Evidence: Detecting AI-Generated Content In Turkish News Media With A Fine-Tuned Bert Classifier
Ozancan Ozdemir
Main category: cs.CL
TL;DR: First empirical study quantifying AI-generated content in Turkish news media using fine-tuned Turkish BERT model, revealing ~2.5% of articles are AI-rewritten
Details
Motivation: Address lack of empirical investigation into AI-generated content in Turkish news media, moving beyond qualitative interviews and fake news detection studiesMethod: Fine-tuned Turkish-specific BERT model (dbmdz/bert-base-turkish-cased) on 3,600 labeled articles from three major Turkish outlets for binary classification of AI-rewritten content
Result: Model achieved 0.9708 F1 score, deployed on 3,500+ unseen articles (2023-2026) revealing consistent patterns with mean confidence >0.96 and ~2.5% AI-rewritten content
Conclusion: First empirical measurement of AI usage in Turkish news media, establishing baseline for future monitoring of AI-generated content proliferation
Abstract: The rapid integration of large language models into newsroom workflows has raised urgent questions about the prevalence of AI-generated content in online media. While computational studies have begun to quantify this phenomenon in English-language outlets, no empirical investigation exists for Turkish news media, where existing research remains limited to qualitative interviews with journalists or fake news detection. This study addresses that gap by fine-tuning a Turkish-specific BERT model (dbmdz/bert-base-turkish-cased) on a labeled dataset of 3,600 articles from three major Turkish outlets with distinct editorial orientations for binary classification of AI-rewritten content. The model achieves 0.9708 F1 score on the held-out test set with symmetric precision and recall across both classes. Subsequent deployment on over 3,500 unseen articles spanning between 2023 and 2026 reveals consistent cross-source and temporally stable classification patterns, with mean prediction confidence exceeding 0.96 and an estimated 2.5 percentage of examined news content rewritten or revised by LLMs on average. To the best of our knowledge, this is the first study to move beyond self-reported journalist perceptions toward empirical, data-driven measurement of AI usage in Turkish news media.
[6] Think Deep, Not Just Long: Measuring LLM Reasoning Effort via Deep-Thinking Tokens
Wei-Lin Chen, Liqian Peng, Tian Tan, Chao Zhao, Blake JianHang Chen, Ziqian Lin, Alec Go, Yu Meng
Main category: cs.CL
TL;DR: Deep-thinking tokens (where predictions significantly revise in deeper layers) better predict reasoning quality than token count, enabling efficient test-time scaling via Think@n strategy.
Details
Motivation: Current LLMs use long Chain-of-Thought for reasoning, but raw token counts are unreliable proxies for reasoning quality - increased length may signal "overthinking" and performance degradation rather than better reasoning.Method: Quantify inference-time effort by identifying deep-thinking tokens (tokens where internal predictions undergo significant revisions in deeper model layers before convergence). Use deep-thinking ratio (proportion of deep-thinking tokens) as metric, then introduce Think@n test-time scaling strategy that prioritizes samples with high deep-thinking ratios.
Result: Deep-thinking ratio shows robust positive correlation with accuracy across four challenging benchmarks (AIME 24/25, HMMT 25, GPQA-diamond) and multiple models (GPT-OSS, DeepSeek-R1, Qwen3), outperforming length-based and confidence-based baselines. Think@n matches/exceeds standard self-consistency performance while reducing inference costs via early rejection of unpromising generations.
Conclusion: Deep-thinking tokens provide better signal for reasoning quality than token counts, enabling more efficient test-time scaling strategies that maintain performance while reducing computational costs.
Abstract: Large language models (LLMs) have demonstrated impressive reasoning capabilities by scaling test-time compute via long Chain-of-Thought (CoT). However, recent findings suggest that raw token counts are unreliable proxies for reasoning quality: increased generation length does not consistently correlate with accuracy and may instead signal “overthinking,” leading to performance degradation. In this work, we quantify inference-time effort by identifying deep-thinking tokens – tokens where internal predictions undergo significant revisions in deeper model layers prior to convergence. Across four challenging mathematical and scientific benchmarks (AIME 24/25, HMMT 25, and GPQA-diamond) and a diverse set of reasoning-focused models (GPT-OSS, DeepSeek-R1, and Qwen3), we show that deep-thinking ratio (the proportion of deep-thinking tokens in a generated sequence) exhibits a robust and consistently positive correlation with accuracy, substantially outperforming both length-based and confidence-based baselines. Leveraging this insight, we introduce Think@n, a test-time scaling strategy that prioritizes samples with high deep-thinking ratios. We demonstrate that Think@n matches or exceeds standard self-consistency performance while significantly reducing inference costs by enabling the early rejection of unpromising generations based on short prefixes.
[7] On Calibration of Large Language Models: From Response To Capability
Sin-Han Yang, Cheng-Kuang Wu, Chieh-Yen Lin, Yun-Nung Chen, Hung-yi Lee, Shao-Hua Sun
Main category: cs.CL
TL;DR: The paper introduces capability calibration for LLMs, focusing on estimating the probability that a model can solve a query overall, rather than just the correctness of a single generated output.
Details
Motivation: Current LLM calibration focuses on response-level confidence, but this is misaligned with practical settings where users need to know how likely a model is to solve a query overall, especially given the stochastic nature of modern LLM decoding.Method: Introduces capability calibration which targets the model’s expected accuracy on a query, formally distinguishes it from response calibration, establishes an empirical evaluation setup, and studies various confidence estimation methods.
Result: Capability-calibrated confidence improves pass@k prediction and inference budget allocation, showing practical benefits over traditional response calibration approaches.
Conclusion: Capability calibration provides a better foundation for confidence estimation in LLMs, with potential for diverse applications beyond traditional response-level calibration.
Abstract: Large language models (LLMs) are widely deployed as general-purpose problem solvers, making accurate confidence estimation critical for reliable use. Prior work on LLM calibration largely focuses on response-level confidence, which estimates the correctness of a single generated output. However, this formulation is misaligned with many practical settings where the central question is how likely a model is to solve a query overall. We show that this mismatch results from the stochastic nature of modern LLM decoding, under which single-response correctness fails to reflect underlying model capability. To address this issue, we introduce capability calibration, which targets the model’s expected accuracy on a query. We formally distinguish capability calibration from response calibration and show that the two differ both theoretically and empirically. We establish an empirical evaluation setup and study a range of confidence estimation methods. Our results demonstrate that capability-calibrated confidence improves pass@$k$ prediction and inference budget allocation, establishing a foundation with potential for diverse applications.
[8] Small Reward Models via Backward Inference
Yike Wang, Faeze Brahman, Shangbin Feng, Teng Xiao, Hannaneh Hajishirzi, Yulia Tsvetkov
Main category: cs.CL
TL;DR: FLIP is a novel reward modeling approach that uses backward inference to reconstruct prompts from responses, using similarity between inferred and original instructions as reward signal, outperforming LLM-as-a-Judge methods by 79.6%.
Details
Motivation: Current reward modeling approaches like LLM-as-a-Judge rely on large models' reasoning capabilities, while alternatives need reference responses or explicit rubrics, limiting flexibility and accessibility. There's a need for reference-free, rubric-free reward modeling methods.Method: FLIP reformulates reward modeling through backward inference: it infers the instruction that would most plausibly produce a given response, then uses the similarity between the inferred and original instructions as the reward signal. This exploits the validation-generation gap.
Result: FLIP outperforms LLM-as-a-Judge baselines by an average of 79.6% across four domains using 13 small language models. It substantially improves downstream performance in extrinsic evaluations under test-time scaling via parallel sampling and GRPO training. FLIP is particularly effective for longer outputs and robust to common forms of reward hacking.
Conclusion: FLIP enables reliable reward modeling in downscaled regimes where judgment methods fail, offering a reference-free and rubric-free approach that explicitly exploits the validation-generation gap for more accessible and flexible reward modeling.
Abstract: Reward models (RMs) play a central role throughout the language model (LM) pipeline, particularly in non-verifiable domains. However, the dominant LLM-as-a-Judge paradigm relies on the strong reasoning capabilities of large models, while alternative approaches require reference responses or explicit rubrics, limiting flexibility and broader accessibility. In this work, we propose FLIP (FLipped Inference for Prompt reconstruction), a reference-free and rubric-free reward modeling approach that reformulates reward modeling through backward inference: inferring the instruction that would most plausibly produce a given response. The similarity between the inferred and the original instructions is then used as the reward signal. Evaluations across four domains using 13 small language models show that FLIP outperforms LLM-as-a-Judge baselines by an average of 79.6%. Moreover, FLIP substantially improves downstream performance in extrinsic evaluations under test-time scaling via parallel sampling and GRPO training. We further find that FLIP is particularly effective for longer outputs and robust to common forms of reward hacking. By explicitly exploiting the validation-generation gap, FLIP enables reliable reward modeling in downscaled regimes where judgment methods fail. Code available at https://github.com/yikee/FLIP.
[9] DistillLens: Symmetric Knowledge Distillation Through Logit Lens
Manish Dhakal, Uthman Jinadu, Anjila Budathoki, Rajshekhar Sunderraman, Yi Ding
Main category: cs.CL
TL;DR: DistillLens is a knowledge distillation framework that symmetrically aligns intermediate thought processes between teacher and student LLMs by projecting hidden states to vocabulary space and using symmetric divergence to preserve uncertainty profiles.
Details
Motivation: Standard KD treats teacher's intermediate layers as black boxes, and existing feature-based methods ignore rich uncertainty profiles needed for final output. There's a need to better capture the teacher's evolving thought process during distillation.Method: Projects intermediate hidden states into vocabulary space via Logit Lens, then enforces structural alignment using symmetric divergence objective that imposes dual-sided penalty to prevent overconfidence/underconfidence while preserving high-entropy information.
Result: Extensive experiments on GPT-2 and Llama architectures show DistillLens consistently outperforms standard KD and feature-transfer baselines on diverse instruction-following benchmarks.
Conclusion: DistillLens provides an effective framework for knowledge distillation that better captures teacher models’ intermediate reasoning processes, leading to improved student model performance.
Abstract: Standard Knowledge Distillation (KD) compresses Large Language Models (LLMs) by optimizing final outputs, yet it typically treats the teacher’s intermediate layer’s thought process as a black box. While feature-based distillation attempts to bridge this gap, existing methods (e.g., MSE and asymmetric KL divergence) ignore the rich uncertainty profiles required for the final output. In this paper, we introduce DistillLens, a framework that symmetrically aligns the evolving thought processes of student and teacher models. By projecting intermediate hidden states into the vocabulary space via the Logit Lens, we enforce structural alignment using a symmetric divergence objective. Our analysis proves that this constraint imposes a dual-sided penalty, preventing both overconfidence and underconfidence while preserving the high-entropy information conduits essential for final deduction. Extensive experiments on GPT-2 and Llama architectures demonstrate that DistillLens consistently outperforms standard KD and feature-transfer baselines on diverse instruction-following benchmarks. The code is available at https://github.com/manishdhakal/DistillLens.
[10] LLM-Confidence Reranker: A Training-Free Approach for Enhancing Retrieval-Augmented Generation Systems
Zhipeng Song, Xiangyu Kong, Xinrui Bao, Yizhi Zhou, Jiulong Jiao, Sitong Liu, Yuhang Zhou, Heng Qi
Main category: cs.CL
TL;DR: LCR is a training-free reranking method that uses LLM confidence signals to improve document retrieval in RAG systems without additional training.
Details
Motivation: Existing rerankers for RAG systems require specialized training, have high computational costs, and don't fully leverage LLMs' semantic understanding and confidence signals to reduce hallucinations in knowledge-intensive tasks.Method: Two-stage process: 1) Confidence assessment using multinomial sampling and Maximum Semantic Cluster Proportion clustering, 2) Binning and multi-level sorting based on query and document confidence thresholds to prioritize relevant documents while preserving original rankings for high-confidence queries.
Result: LCR consistently improves NDCG@5 by up to 20.6% across BEIR and TREC benchmarks using BM25 and Contriever retrievers with only 7-9B parameter LLMs, without performance degradation compared to existing rerankers.
Conclusion: LCR provides an efficient, training-free approach that leverages LLM confidence signals to improve document retrieval in RAG systems, reducing hallucinations while maintaining computational efficiency and broad compatibility.
Abstract: Large language models (LLMs) have revolutionized natural language processing, yet hallucinations in knowledge-intensive tasks remain a critical challenge. Retrieval-augmented generation (RAG) addresses this by integrating external knowledge, but its efficacy depends on accurate document retrieval and ranking. Although existing rerankers demonstrate effectiveness, they frequently necessitate specialized training, impose substantial computational expenses, and fail to fully exploit the semantic capabilities of LLMs, particularly their inherent confidence signals. We propose the LLM-Confidence Reranker (LCR), a training-free, plug-and-play algorithm that enhances reranking in RAG systems by leveraging black-box LLM confidence derived from Maximum Semantic Cluster Proportion (MSCP). LCR employs a two-stage process: confidence assessment via multinomial sampling and clustering, followed by binning and multi-level sorting based on query and document confidence thresholds. This approach prioritizes relevant documents while preserving original rankings for high-confidence queries, ensuring robustness. Evaluated on BEIR and TREC benchmarks with BM25 and Contriever retrievers, LCR–using only 7–9B-parameter pre-trained LLMs–consistently improves NDCG@5 by up to 20.6% across pre-trained LLM and fine-tuned Transformer rerankers, without degradation. Ablation studies validate the hypothesis that LLM confidence positively correlates with document relevance, elucidating LCR’s mechanism. LCR offers computational efficiency, parallelism for scalability, and broad compatibility, mitigating hallucinations in applications like medical diagnosis.
[11] Elo-Evolve: A Co-evolutionary Framework for Language Model Alignment
Jing Zhao, Ting Zhen, Junwei bao, Hongfei Jiang, Yang song
Main category: cs.CL
TL;DR: Elo-Evolve: A co-evolutionary framework for LLM alignment using dynamic multi-agent competition with pairwise win/loss learning and Elo-orchestrated opponent selection, achieving better performance than traditional methods.
Details
Motivation: Current LLM alignment methods rely on compressing human preference data into static reward functions, leading to data scarcity, noise sensitivity, and training instability. There's a need for more robust alignment approaches.Method: Introduces Elo-Evolve framework with two key innovations: (1) learning directly from binary win/loss outcomes in pairwise competitions instead of Bradley-Terry models, and (2) Elo-orchestrated opponent selection with temperature-controlled sampling for automatic curriculum learning. Grounded in PAC learning theory.
Result: Achieves 4.5x noise reduction compared to absolute scoring approaches. Trained Qwen2.5-7B model outperforms point-based methods and static pairwise training across Alpaca Eval 2.0 and MT-Bench benchmarks, demonstrating clear performance hierarchy.
Conclusion: Pairwise comparison with dynamic opponent selection provides superior LLM alignment, offering progressive benefits over traditional methods through co-evolutionary competition.
Abstract: Current alignment methods for Large Language Models (LLMs) rely on compressing vast amounts of human preference data into static, absolute reward functions, leading to data scarcity, noise sensitivity, and training instability. We introduce Elo-Evolve, a co-evolutionary framework that redefines alignment as dynamic multi-agent competition within an adaptive opponent pool. Our approach makes two key innovations: (1) eliminating Bradley-Terry model dependencies by learning directly from binary win/loss outcomes in pairwise competitions, and (2) implementing Elo-orchestrated opponent selection that provides automatic curriculum learning through temperature-controlled sampling. We ground our approach in PAC learning theory, demonstrating that pairwise comparison achieves superior sample complexity and empirically validate a 4.5x noise reduction compared to absolute scoring approaches. Experimentally, we train a Qwen2.5-7B model using our framework with opponents including Qwen2.5-14B, Qwen2.5-32B, and Qwen3-8B models. Results demonstrate a clear performance hierarchy: point-based methods < static pairwise training < Elo-Evolve across Alpaca Eval 2.0 and MT-Bench, validating the progressive benefits of pairwise comparison and dynamic opponent selection for LLM alignment.
[12] The Speech-LLM Takes It All: A Truly Fully End-to-End Spoken Dialogue State Tracking Approach
Nizar El Ghazal, Antoine Caubrière, Valentin Vielzeuf
Main category: cs.CL
TL;DR: Comparative study of context management strategies for spoken dialog state tracking using Speech-LLMs, showing full spoken history yields best performance while compression offers good trade-off.
Details
Motivation: To systematically evaluate different context management strategies for spoken dialog state tracking using Speech-LLMs, addressing how to best utilize spoken conversation history for accurate state tracking.Method: Comparative evaluation of three approaches: 1) traditional multimodal context (text history + spoken current turn), 2) full spoken history, and 3) compressed spoken history using attention-pooling-based compression. Experiments conducted on SpokenWOZ corpus.
Result: Full spoken conversation as input yields highest performance among similar-sized models, significantly surpassing prior methods. Attention-pooling compression maintains competitive accuracy with reduced context size. Improvements stem from more effective context utilization.
Conclusion: Full spoken history is optimal for spoken dialog state tracking, while compression offers practical trade-off for context size reduction without significant accuracy loss.
Abstract: This paper presents a comparative study of context management strategies for end-to-end Spoken Dialog State Tracking using Speech-LLMs. We systematically evaluate traditional multimodal context (combining text history and spoken current turn), full spoken history, and compressed spoken history approaches. Our experiments on the SpokenWOZ corpus demonstrate that providing the full spoken conversation as input yields the highest performance among models of similar size, significantly surpassing prior methods. Furthermore, we show that attention-pooling-based compression of the spoken history offers a strong trade-off, maintaining competitive accuracy with reduced context size. Detailed analysis confirms that improvements stem from more effective context utilization.
[13] Metaphors’ journeys across time and genre: tracking the evolution of literary metaphors with temporal embeddings
Veronica Mangiaterra, Chiara Barattieri di San Pietro, Paolo Canal, Valentina Bambini
Main category: cs.CL
TL;DR: This study applies diachronic distributional semantics to analyze how processing costs of 19th-century Italian literary metaphors changed over time and across genres, finding stability overall but genre-specific patterns influenced by semantic features.
Details
Motivation: Literary metaphors are understudied experimentally compared to everyday metaphors, and previous approaches have overlooked the temporal dimension despite many literary metaphors being coined centuries apart from contemporary readers.Method: Trained word embeddings on literary and nonliterary Italian corpora from 19th and 21st centuries (124M tokens total), modeled changes in semantic similarity between topics and vehicles of 515 19th-century literary metaphors as proxy for processing demands.
Result: Overall semantic similarity (and metaphor processing demands) remained stable over time. Genre played key role: metaphors appeared more difficult in modern literary contexts but easier in today’s nonliterary language (Web). Pattern shaped by semantic features like vector coherence and neighborhood density.
Conclusion: Findings align with broader linguistic changes: stylistic simplification of modern literature increased metaphor processing demands, while high creativity of Web’s language made metaphors more accessible. Demonstrates value of diachronic computational approaches to literary metaphor processing.
Abstract: Metaphors are a distinctive feature of literary language, yet they remain less studied experimentally than everyday metaphors. Moreover, previous psycholinguistic and computational approaches overlooked the temporal dimension, although many literary metaphors were coined centuries apart from contemporary readers. This study innovatively applies tools from diachronic distributional semantics to assess whether the processing costs of literary metaphors varied over time and genre. Specifically, we trained word embeddings on literary and nonliterary Italian corpora from the 19th and 21st centuries, for a total of 124 million tokens, and modeled changes in the semantic similarity between topics and vehicles of 515 19th-century literary metaphors, taking this measure as a proxy of metaphor processing demands. Overall, semantic similarity, and hence metaphor processing demands, remained stable over time. However, genre played a key role: metaphors appeared more difficult (i.e., lower topic-vehicle similarity) in modern literary contexts than in 19th-century literature, but easier (i.e., higher topic-vehicle similarity) in today’s nonliterary language (e.g., the Web) than in 19th-century nonliterary texts. This pattern was further shaped by semantic features of metaphors’ individual terms, such as vector coherence and semantic neighborhood density. Collectively, these findings align with broader linguistic changes in Italian, such as the stylistic simplification of modern literature, which may have increased metaphor processing demands, and the high creativity of the Web’s language, which seems to render metaphor more accessible.
[14] From Scarcity to Scale: A Release-Level Analysis of the Pashto Common Voice Dataset
Jandad Jahani, Mursal Dawodi, Jawid Ahmad Baktash
Main category: cs.CL
TL;DR: Analysis of Pashto speech data growth in Mozilla Common Voice corpus, showing rapid expansion from 1.49 to 2,768.7 hours, with analysis of validation throughput, contributor inequality, and demographic metadata issues.
Details
Motivation: Pashto, spoken by over 60 million people, lacks large-scale openly licensed speech data for ASR development, despite being widely spoken. The paper aims to analyze the growth and characteristics of Pashto speech data in the Mozilla Common Voice corpus to understand dataset maturity and identify areas for improvement.Method: Release-level analysis of Pashto component in Mozilla Common Voice corpus (version 24.0, December 2025), examining trends across major releases. Analysis includes scale metrics, validation throughput, contributor participation inequality (Gini coefficient), demographic metadata completeness, and sentence-level concentration in validated subsets.
Result: Rapid growth from 1.49 hours (mid-2023) to 2,768.7 total hours (2025), with 975.89 validated hours for supervised ASR training. Participation extremely concentrated (Gini = 0.941), age representation skewed toward young adults, 41.97% of clips lack gender labels, and 35.88% of unique sentences account for 50% of validated clips.
Conclusion: The study provides quantitative audit of a rapidly scaling low-resource speech corpus, highlighting practical priorities for improving dataset maturity including expanded validation capacity and broader demographic participation to address metadata gaps and contributor concentration.
Abstract: Large, openly licensed speech datasets are essential for building automatic speech recognition (ASR) systems, yet many widely spoken languages remain underrepresented in public resources. Pashto, spoken by more than 60 million people, has historically lacked large-scale openly licensed speech data suitable for modern ASR development. This paper presents a release-level analysis of the Pashto component of the Mozilla Common Voice corpus, focusing on version 24.0 (December 2025) and contextualizing trends across major releases. We document rapid growth from 1.49 recorded hours in mid-2023 to 2,768.7 total hours in 2025, including 975.89 validated hours available for supervised ASR training. Beyond scale, we analyze validation throughput, contributor participation inequality, demographic metadata completeness, and sentence-level concentration in the validated subset. We find that participation is extremely concentrated (Gini = 0.941), age representation is strongly skewed toward young adults, and 41.97% of clips lack self-reported gender labels, limiting subgroup auditing based on metadata. At the textual level, prompt reuse is moderate: 35.88% of unique sentences account for 50% of validated clips, suggesting that structural concentration is driven primarily by uneven contributor activity rather than dominance of a small prompt set. These results provide a quantitative audit of a rapidly scaling low-resource speech corpus and highlight practical priorities for improving dataset maturity, including expanded validation capacity and broader demographic participation.
[15] On Theoretically-Driven LLM Agents for Multi-Dimensional Discourse Analysis
Maciej Uberna, Michał Wawer, Jarosław A. Chudziak, Marcin Koszowy
Main category: cs.CL
TL;DR: Multi-agent framework using RAG-enhanced LLMs outperforms baseline by 30% F1-score in detecting rhetorical functions of reformulation in political debates.
Details
Motivation: Current LLMs can detect surface-level similarity but fail to capture pragmatic functions of rephrasing in argumentative discourse, particularly rhetorical strategies like intensification, deintensification, specification, and generalization.Method: Comparative multi-agent framework with two parallel LLM-based systems: one enhanced by argumentation theory via Retrieval-Augmented Generation (RAG), and an identical zero-shot baseline. Uses annotated political debates dataset with D-I-S-G-O categories (Deintensification, Intensification, Specification, Generalisation, Other).
Result: RAG-enhanced agents substantially outperform baseline across all categories, with particularly strong advantages in detecting Intensification and Generalisation contexts. Overall Macro F1-score improvement of nearly 30%.
Conclusion: Theoretical grounding is essential for advancing beyond mere paraphrase detection to function-aware analysis of argumentative discourse. The multi-agent architecture enables scalable, theoretically informed computational tools for identifying rhetorical strategies.
Abstract: Identifying the strategic uses of reformulation in discourse remains a key challenge for computational argumentation. While LLMs can detect surface-level similarity, they often fail to capture the pragmatic functions of rephrasing, such as its role within rhetorical discourse. This paper presents a comparative multi-agent framework designed to quantify the benefits of incorporating explicit theoretical knowledge for this task. We utilise an dataset of annotated political debates to establish a new standard encompassing four distinct rephrase functions: Deintensification, Intensification, Specification, Generalisation, and Other, which covers all remaining types (D-I-S-G-O). We then evaluate two parallel LLM-based agent systems: one enhanced by argumentation theory via Retrieval-Augmented Generation (RAG), and an identical zero-shot baseline. The results reveal a clear performance gap: the RAG-enhanced agents substantially outperform the baseline across the board, with particularly strong advantages in detecting Intensification and Generalisation context, yielding an overall Macro F1-score improvement of nearly 30%. Our findings provide evidence that theoretical grounding is not only beneficial but essential for advancing beyond mere paraphrase detection towards function-aware analysis of argumentative discourse. This comparative multi-agent architecture represents a step towards scalable, theoretically informed computational tools capable of identifying rhetorical strategies in contemporary discourse.
[16] RMPL: Relation-aware Multi-task Progressive Learning with Stage-wise Training for Multimedia Event Extraction
Yongkang Jin, Jianwen Luo, Jingjing Wang, Jianmin Yao, Yu Hong
Main category: cs.CL
TL;DR: RMPL: A relation-aware multi-task progressive learning framework for multimedia event extraction that addresses data scarcity by leveraging heterogeneous supervision from unimodal tasks and progressive training.
Details
Motivation: Multimedia Event Extraction (MEE) suffers from limited annotated training data (only M2E2 benchmark exists with evaluation-only annotations), making supervised training impractical. Existing methods rely on cross-modal alignment or VLM prompting without learning structured event representations, resulting in weak argument grounding across modalities.Method: RMPL uses relation-aware multi-task progressive learning with heterogeneous supervision from unimodal event extraction and multimedia relation extraction. It employs stage-wise training: first learns shared event-centric representations across modalities using a unified schema, then fine-tunes for event mention identification and argument role extraction using mixed textual and visual data.
Result: Experiments on M2E2 benchmark with multiple Vision-Language Models show consistent improvements across different modality settings compared to existing approaches.
Conclusion: RMPL effectively addresses data scarcity in MEE by leveraging heterogeneous supervision and progressive learning, improving multimodal event extraction performance through better structured event representation learning.
Abstract: Multimedia Event Extraction (MEE) aims to identify events and their arguments from documents that contain both text and images. It requires grounding event semantics across different modalities. Progress in MEE is limited by the lack of annotated training data. M2E2 is the only established benchmark, but it provides annotations only for evaluation. This makes direct supervised training impractical. Existing methods mainly rely on cross-modal alignment or inference-time prompting with Vision–Language Models (VLMs). These approaches do not explicitly learn structured event representations and often produce weak argument grounding in multimodal settings. To address these limitations, we propose RMPL, a Relation-aware Multi-task Progressive Learning framework for MEE under low-resource conditions. RMPL incorporates heterogeneous supervision from unimodal event extraction and multimedia relation extraction with stage-wise training. The model is first trained with a unified schema to learn shared event-centric representations across modalities. It is then fine-tuned for event mention identification and argument role extraction using mixed textual and visual data. Experiments on the M2E2 benchmark with multiple VLMs show consistent improvements across different modality settings.
[17] How Do Lexical Senses Correspond Between Spoken German and German Sign Language?
Melis Çelikkol, Wei Zhao
Main category: cs.CL
TL;DR: This paper presents a computational approach to identify word-to-sign mappings between German and German Sign Language, creating the first annotated dataset for cross-modal sense correspondence and evaluating semantic similarity methods.
Details
Motivation: Current sign language dictionaries often underrepresent polysemous and homonymous words that map to different signs across contexts. There's a need to identify novel word-to-sign mappings absent from existing dictionaries to enrich lexicographic resources.Method: Researchers manually annotated 1,404 word use-to-sign ID mappings from German Word Usage Graph and Digital Dictionary of German Sign Language. They identified three correspondence types and evaluated computational methods: Exact Match and Semantic Similarity using SBERT embeddings.
Result: Semantic Similarity substantially outperformed Exact Match (88.52% vs. 71.31%), with dramatic gains for Type 1 correspondences (+52.1 percentage points). The work established the first annotated dataset for cross-modal sense correspondence.
Conclusion: This research provides computational methods and annotated resources for identifying cross-modal sense correspondences between spoken and sign languages, revealing which correspondence patterns are computationally identifiable.
Abstract: Sign language lexicographers construct bilingual dictionaries by establishing word-to-sign mappings, where polysemous and homonymous words corresponding to different signs across contexts are often underrepresented. A usage-based approach examining how word senses map to signs can identify such novel mappings absent from current dictionaries, enriching lexicographic resources. We address this by analyzing German and German Sign Language (Deutsche Gebärdensprache, DGS), manually annotating 1,404 word use-to-sign ID mappings derived from 32 words from the German Word Usage Graph (D-WUG) and 49 signs from the Digital Dictionary of German Sign Language (DW-DGS). We identify three correspondence types: Type 1 (one-to-many), Type 2 (many-to-one), and Type 3 (one-to-one), plus No Match cases. We evaluate computational methods: Exact Match (EM) and Semantic Similarity (SS) using SBERT embeddings. SS substantially outperforms EM overall 88.52% vs. 71.31%), with dramatic gains for Type 1 (+52.1 pp). Our work establishes the first annotated dataset for cross-modal sense correspondence and reveals which correspondence patterns are computationally identifiable. Our code and dataset are made publicly available.
[18] OMGs: A multi-agent system supporting MDT decision-making across the ovarian tumour care continuum
Yangyang Zhang, Zilong Wang, Jianbo Xu, Yongqi Chen, Chu Han, Zhihao Zhang, Shuai Liu, Hui Li, Huiping Zhang, Ziqi Liu, Jiaxin Chen, Jun Zhu, Zheng Feng, Hao Wen, Xingzhu Ju, Yanping Zhong, Yunqiu Zhang, Jie Duan, Jun Li, Dongsheng Li, Weijie Wang, Haiyan Zhu, Wei Jiang, Xiaohua Wu, Shuo Wang, Haiming Li, Qinhao Guo
Main category: cs.CL
TL;DR: OMGs is a multi-agent AI system that simulates multidisciplinary tumor board deliberations for ovarian cancer, generating expert-level recommendations with transparent reasoning to address limited access to specialized oncology expertise.
Details
Motivation: Most ovarian cancer patients worldwide lack access to timely expert multidisciplinary tumor board (MDT) consultations, especially in resource-constrained settings where MDT resources are scarce or unavailable, creating a need for AI systems that can provide MDT-style recommendations.Method: Developed OMGs (Ovarian tumour Multidisciplinary intelligent aGent System), a multi-agent AI framework where domain-specific agents collaborate to integrate multidisciplinary evidence and generate MDT-style recommendations with transparent rationales. Created SPEAR (Safety, Personalization, Evidence, Actionability, Robustness) framework to systematically evaluate MDT recommendation quality.
Result: In multicentre re-evaluation, OMGs achieved performance comparable to expert MDT consensus (4.45 ± 0.30 vs 4.53 ± 0.23), with higher Evidence scores (4.57 vs 3.92). In prospective multicentre evaluation with 59 patients, OMGs demonstrated high concordance with routine MDT decisions. In paired human-AI studies, OMGs most substantially enhanced clinicians’ recommendations in Evidence and Robustness dimensions.
Conclusion: Multi-agent deliberative AI systems can achieve performance comparable to expert MDT consensus and have potential to expand access to specialized oncology expertise in resource-limited settings, particularly enhancing Evidence and Robustness dimensions where multidisciplinary expertise is unavailable.
Abstract: Ovarian tumour management has increasingly relied on multidisciplinary tumour board (MDT) deliberation to address treatment complexity and disease heterogeneity. However, most patients worldwide lack access to timely expert consensus, particularly in resource-constrained centres where MDT resources are scarce or unavailable. Here we present OMGs (Ovarian tumour Multidisciplinary intelligent aGent System), a multi-agent AI framework where domain-specific agents deliberate collaboratively to integrate multidisciplinary evidence and generate MDT-style recommendations with transparent rationales. To systematically evaluate MDT recommendation quality, we developed SPEAR (Safety, Personalization, Evidence, Actionability, Robustness) and validated OMGs across diverse clinical scenarios spanning the care continuum. In multicentre re-evaluation, OMGs achieved performance comparable to expert MDT consensus ($4.45 \pm 0.30$ versus $4.53 \pm 0.23$), with higher Evidence scores (4.57 versus 3.92). In prospective multicentre evaluation (59 patients), OMGs demonstrated high concordance with routine MDT decisions. Critically, in paired human-AI studies, OMGs most substantially enhanced clinicians’ recommendations in Evidence and Robustness, the dimensions most compromised when multidisciplinary expertise is unavailable. These findings suggest that multi-agent deliberative systems can achieve performance comparable to expert MDT consensus, with potential to expand access to specialized oncology expertise in resource-limited settings.
[19] MedXIAOHE: A Comprehensive Recipe for Building Medical MLLMs
Baorong Shi, Bo Cui, Boyuan Jiang, Deli Yu, Fang Qian, Haihua Yang, Huichao Wang, Jiale Chen, Jianfei Pan, Jieqiong Cao, Jinghao Lin, Kai Wu, Lin Yang, Shengsheng Yao, Tao Chen, Xiaojun Xiao, Xiaozhong Ji, Xu Wang, Yijun He, Zhixiong Yang
Main category: cs.CL
TL;DR: MedXIAOHE is a medical vision-language foundation model that achieves SOTA performance on medical benchmarks through entity-aware pretraining, medical reasoning via RL/tool-augmented training, and reliability improvements for real-world clinical use.
Details
Motivation: To advance general-purpose medical understanding and reasoning for real-world clinical applications, addressing challenges like heterogeneous medical data, rare diseases, expert-level reasoning, and reliability in clinical settings.Method: Entity-aware continual pretraining organizes heterogeneous medical corpora; incorporates diverse medical reasoning patterns via reinforcement learning and tool-augmented agentic training; integrates user-preference rubrics, evidence-grounded reasoning, and low-hallucination report generation.
Result: Achieves state-of-the-art performance across diverse medical benchmarks and surpasses leading closed-source multimodal systems on multiple capabilities.
Conclusion: MedXIAOHE represents an advanced medical vision-language foundation model with practical design choices for real-world clinical applications, offering improved medical understanding, reasoning, and reliability.
Abstract: We present MedXIAOHE, a medical vision-language foundation model designed to advance general-purpose medical understanding and reasoning in real-world clinical applications. MedXIAOHE achieves state-of-the-art performance across diverse medical benchmarks and surpasses leading closed-source multimodal systems on multiple capabilities. To achieve this, we propose an entity-aware continual pretraining framework that organizes heterogeneous medical corpora to broaden knowledge coverage and reduce long-tail gaps (e.g., rare diseases). For medical expert-level reasoning and interaction, MedXIAOHE incorporates diverse medical reasoning patterns via reinforcement learning and tool-augmented agentic training, enabling multi-step diagnostic reasoning with verifiable decision traces. To improve reliability in real-world use, MedXIAOHE integrates user-preference rubrics, evidence-grounded reasoning, and low-hallucination long-form report generation, with improved adherence to medical instructions. We release this report to document our practical design choices, scaling insights, and evaluation framework, hoping to inspire further research.
[20] The acquisition of English irregular inflections by Yemeni L1 Arabic learners: A Universal Grammar approach
Muneef Y. Alsawsh, Mohammed Q. Shormani
Main category: cs.CL
TL;DR: Study examines Yemeni ESL learners’ acquisition of English irregular inflections using Universal Grammar approach, finding L1 transfer dominates early stages while UG access improves with development, but instructional quality limits full acquisition.
Details
Motivation: To understand how adult L2 learners acquire English irregular inflections through the lens of Universal Grammar, specifically examining the roles of L1 transfer and developmental factors in morphological acquisition.Method: Used Universal Grammar approach with Feature Reassembly Hypothesis, analyzed learner errors across two developmental stages, employed statistical analysis including one-way ANOVA to measure improvement in irregular inflection production.
Result: Stage 1 showed dominant L1 transfer influence in phonological/structural mismatches; Stage 2 demonstrated increased UG sensitivity and morphological reconfiguration. Significant improvement in well-formed irregular inflections from stage 1 to stage 2, but persistent difficulties with specific inflection types.
Conclusion: While L1 transfer and developmental factors influence initial acquisition, appropriate linguistic input and instruction are critical for facilitating UG-driven feature reassembly in adult L2 learners, though full UG access remains constrained by limited exposure and instructional quality.
Abstract: This study examines the acquisition of English irregular inflections by Yemeni learners of English as a second language (L2), utilizing a Universal Grammar (UG) approach. Within the UG approach, the study considers Feature Reassembly Hypothesis (FRH) (Lardiere, 2008, 2009) part of UG, focusing on the roles of first language (L1) transfer and L2 developmental influence. It analyzes learner errors across two developmental stages. Stage 1 data reveal a dominant influence of L1 transfer, particularly in phonological and structural mismatches, while stage 2 data demonstrate increased learner sensitivity to UG properties and morphological reconfiguration toward the target language. Findings reveal that errors in irregular inflectional morphology are attributed to both interlingual and intralingual sources, with overgeneralization of L2 rules as a common developmental strategy. Statistical analysis, including a one-way ANOVA, indicates significant improvement in the production of well-formed irregular inflections from stage 1 to stage 2, underscoring learners’ continued access to UG. However, persistent difficulties with consonant change, zero-morpheme, and -a plural inflections suggest that limited exposure, ineffective input modeling, and insufficient instructional quality constrain full UG access. The study concludes that while L1 transfer and L2 developmental factors influence initial stages of acquisition, appropriate linguistic input and instruction are critical for facilitating UG-driven feature reassembly in adult L2 learners.
[21] Beyond Words: Evaluating and Bridging Epistemic Divergence in User-Agent Interaction via Theory of Mind
Minyuan Ruan, Ziyue Wang, Kaiming Liu, Yunghwei Lai, Peng Li, Yang Liu
Main category: cs.CL
TL;DR: This paper proposes a benchmark for evaluating Theory of Mind (ToM) in LLMs, focusing on epistemic divergence detection and resolution in real-world interactions rather than isolated belief inference.
Details
Motivation: LLMs struggle to comprehend true user needs when intentions are imprecisely conveyed, leading to divergence between user beliefs and environment states. Existing ToM evaluations focus on isolated belief inference, overlooking practical utility in real-world interactions.Method: Formalized ToM for LLMs as epistemic divergence detection/resolution mechanism, created a benchmark to assess how models reconcile user beliefs and profiles, curated trajectory-based ToM dataset linking belief tracking with task-related state inference, and trained models via reinforcement learning.
Result: Evaluation of 11 leading models revealed significant limitations in identifying underlying cognitive gaps that impede task success. Models trained on the trajectory-based dataset via reinforcement learning showed consistent improvement in reasoning about user mental states and enhanced downstream performance.
Conclusion: ToM should be viewed as an essential interaction-level mechanism rather than a standalone reasoning skill, highlighting its practical value in improving LLM-user interactions.
Abstract: Large Language Models (LLMs) have developed rapidly and are widely applied to both general-purpose and professional tasks to assist human users. However, they still struggle to comprehend and respond to the true user needs when intentions and instructions are imprecisely conveyed, leading to a divergence between subjective user believes and true environment states. Resolving this epistemic divergence requires Theory of Mind (ToM), yet existing ToM evaluations for LLMs primarily focus on isolated belief inference, overlooking its functional utility in real-world interaction. To this end, we formalize ToM for LLMs as a mechanism for epistemic divergence detection and resolution, and propose a benchmark, \benchname, to assess how models reconcile user beliefs and profiles in practice. Results across 11 leading models reveal a significant limitation to identify underlying cognitive gaps that impede task success. To bridge this gap, we further curate a trajectory-based ToM dataset linking belief tracking with task-related state inference. The model trained on this data via reinforcement learning shows consistent improvement in reasoning about user mental states, leading to enhanced downstream performance. Our work highlights the practical value of ToM as an essential interaction-level mechanism rather than as a standalone reasoning skill.
[22] Speculative Decoding with a Speculative Vocabulary
Miles Williams, Young D. Kwon, Rui Li, Alexandros Kouris, Stylianos I. Venieris
Main category: cs.CL
TL;DR: SpecVocab: A speculative decoding method that dynamically selects vocabulary subsets per decoding step to improve LM inference speed without compromising accuracy.
Details
Motivation: Current speculative decoding methods face a bottleneck in draft model output distribution, where reduced vocabulary approaches improve throughput but compromise effectiveness when target tokens are out-of-vocabulary.Method: Proposes SpecVocab, which selects a vocabulary subset per decoding step rather than using a fixed reduced vocabulary, enabling efficient speculation while maintaining coverage of potential target tokens.
Result: SpecVocab achieves higher acceptance length than state-of-the-art EAGLE-3, yielding up to 8.1% increase in average throughput across various tasks.
Conclusion: Vocabulary speculation via dynamic subset selection is an effective alternative to fixed reduced vocabularies for speculative decoding, improving both efficiency and effectiveness.
Abstract: Speculative decoding has rapidly emerged as a leading approach for accelerating language model (LM) inference, as it offers substantial speedups while yielding identical outputs. This relies upon a small draft model, tasked with predicting the outputs of the target model. State-of-the-art speculative decoding methods use a draft model consisting of a single decoder layer and output embedding matrix, with the latter dominating drafting time for the latest LMs. Recent work has sought to address this output distribution bottleneck by reducing the vocabulary of the draft model. Although this can improve throughput, it compromises speculation effectiveness when the target token is out-of-vocabulary. In this paper, we argue for vocabulary speculation as an alternative to a reduced vocabulary. We propose SpecVocab, an efficient and effective method that selects a vocabulary subset per decoding step. Across a variety of tasks, we demonstrate that SpecVocab can achieve a higher acceptance length than state-of-the-art speculative decoding approach, EAGLE-3. Notably, this yields up to an 8.1% increase in average throughput over EAGLE-3.
[23] PrivAct: Internalizing Contextual Privacy Preservation via Multi-Agent Preference Training
Yuhan Cheng, Hancheng Ye, Hai Helen Li, Jingwei Sun, Yiran Chen
Main category: cs.CL
TL;DR: PrivAct is a contextual privacy-aware multi-agent learning framework that embeds privacy preferences into LLM agents to prevent privacy violations in personalized tasks, reducing leakage rates by up to 12.32% while maintaining helpfulness.
Details
Motivation: LLM agents deployed in personalized tasks with sensitive, context-dependent information face privacy violations due to implicit contextual privacy. Existing external intervention approaches are brittle, scenario-specific, and may expand privacy attack surfaces.Method: Proposes PrivAct, a contextual privacy-aware multi-agent learning framework that internalizes contextual privacy preservation directly into models’ generation behavior. Embeds privacy preferences into each agent to enhance system-wide contextual integrity while optimizing privacy-helpfulness tradeoff.
Result: Experiments across multiple LLM backbones and benchmarks show consistent improvements in contextual privacy preservation, reducing leakage rates by up to 12.32% while maintaining comparable helpfulness. Demonstrates zero-shot generalization and robustness across diverse multi-agent topologies.
Conclusion: PrivAct effectively internalizes contextual privacy preservation into LLM agents, achieving better privacy-helpfulness tradeoffs than external intervention approaches, with strong generalization across different agent configurations.
Abstract: Large language model (LLM) agents are increasingly deployed in personalized tasks involving sensitive, context-dependent information, where privacy violations may arise in agents’ action due to the implicitness of contextual privacy. Existing approaches rely on external, inference-time interventions which are brittle, scenario-specific, and may expand the privacy attack surface. We propose PrivAct, a contextual privacy-aware multi-agent learning framework that internalizes contextual privacy preservation directly into models’ generation behavior for privacy-compliant agentic actions. By embedding privacy preferences into each agent, PrivAct enhances system-wide contextual integrity while achieving a more favorable privacy-helpfulness tradeoff. Experiments across multiple LLM backbones and benchmarks demonstrate consistent improvements in contextual privacy preservation, reducing leakage rates by up to 12.32% while maintaining comparable helpfulness, as well as zero-shot generalization and robustness across diverse multi-agent topologies. Code is available at https://github.com/chengyh23/PrivAct.
[24] Tutoring Large Language Models to be Domain-adaptive, Precise, and Safe
Somnath Banerjee
Main category: cs.CL
TL;DR: A framework for developing responsible LLMs that balance generative power with real-world deployment requirements through domain adaptation, safety, and cultural alignment.
Details
Motivation: To address the urgent need for LLMs that move beyond general-purpose architectures toward contextually aware, safer systems that respect global cultural nuances for real-world deployment.Method: Three interconnected threads: 1) domain adaptation for technical precision, 2) ethical rigor to mitigate adversarial vulnerabilities, and 3) cultural/multilingual alignment for global inclusivity. Methodological progression from supervised adaptation to decoding-time alignment to human feedback modeling.
Result: Proposes a comprehensive “Responsible Intelligence” framework that systematically addresses technical, ethical, and cultural dimensions of LLM deployment.
Conclusion: A holistic approach is needed to reconcile LLMs’ generative power with real-world requirements through integrated domain adaptation, safety mechanisms, and cultural alignment.
Abstract: The overarching research direction of this work is the development of a ‘‘Responsible Intelligence’’ framework designed to reconcile the immense generative power of Large Language Models (LLMs) with the stringent requirements of real-world deployment. As these models become a transformative force in artificial intelligence, there is an urgent need to move beyond general-purpose architectures toward systems that are contextually aware, inherently safer, and deeply respectful of global cultural nuances. This research navigates three interconnected threads: domain adaptation to ensure technical precision, ethical rigor to mitigate adversarial vulnerabilities, and cultural/multilingual alignment to promote global inclusivity. The methodological trajectory moves from classical supervised adaptation for task-specific demands to decoding-time alignment for safety, finally leveraging human feedback and preference modeling to achieve sociolinguistic acuity.
[25] A Multi-Agent Framework for Medical AI: Leveraging Fine-Tuned GPT, LLaMA, and DeepSeek R1 for Evidence-Based and Bias-Aware Clinical Query Processing
Naeimeh Nourmohammadi, Md Meem Hossain, The Anh Han, Safina Showkat Ara, Zia Ush Shamszaman
Main category: cs.CL
TL;DR: A multi-agent medical QA framework combining LLMs with evidence retrieval, uncertainty estimation, and bias checks to improve answer reliability in healthcare applications.
Details
Motivation: LLMs show promise for healthcare QA but face limitations in verification, evidence grounding, and confidence signaling, which hinder clinical adoption.Method: Two-phase approach: 1) Fine-tune three LLM families (GPT, LLaMA, DeepSeek R1) on medical QA data; 2) Implement modular multi-agent pipeline with Clinical Reasoning agent, Evidence Retrieval agent (PubMed queries), Refinement agent, and optional human validation for high-risk cases.
Result: DeepSeek R1 achieved best scores (ROUGE-1: 0.536, ROUGE-2: 0.226, BLEU: 0.098). Full system achieved 87% accuracy with 0.80 relevance, reduced uncertainty (perplexity 4.13), and 36.5s mean latency.
Conclusion: Agent specialization and verification layers can mitigate single-model limitations, providing practical design for evidence-based, bias-aware medical AI.
Abstract: Large language models (LLMs) show promise for healthcare question answering, but clinical use is limited by weak verification, insufficient evidence grounding, and unreliable confidence signalling. We propose a multi-agent medical QA framework that combines complementary LLMs with evidence retrieval, uncertainty estimation, and bias checks to improve answer reliability. Our approach has two phases. First, we fine-tune three representative LLM families (GPT, LLaMA, and DeepSeek R1) on MedQuAD-derived medical QA data (20k+ question-answer pairs across multiple NIH domains) and benchmark generation quality. DeepSeek R1 achieves the strongest scores (ROUGE-1 0.536 +- 0.04; ROUGE-2 0.226 +-0.03; BLEU 0.098 -+ 0.018) and substantially outperforms the specialised biomedical baseline BioGPT in zero-shot evaluation. Second, we implement a modular multi-agent pipeline in which a Clinical Reasoning agent (fine-tuned LLaMA) produces structured explanations, an Evidence Retrieval agent queries PubMed to ground responses in recent literature, and a Refinement agent (DeepSeek R1) improves clarity and factual consistency; an optional human validation path is triggered for high-risk or high-uncertainty cases. Safety mechanisms include Monte Carlo dropout and perplexity-based uncertainty scoring, plus lexical and sentiment-based bias detection supported by LIME/SHAP-based analyses. In evaluation, the full system achieves 87% accuracy with relevance around 0.80, and evidence augmentation reduces uncertainty (perplexity 4.13) compared to base responses, with mean end-to-end latency of 36.5 seconds under the reported configuration. Overall, the results indicate that agent specialisation and verification layers can mitigate key single-model limitations and provide a practical, extensible design for evidence-based and bias-aware medical AI.
[26] Bridging the Multilingual Safety Divide: Efficient, Culturally-Aware Alignment for Global South Languages
Somnath Banerjee, Rima Hazra, Animesh Mukherjee
Main category: cs.CL
TL;DR: Paper examines safety gaps in LLMs for Global South languages, showing safety guardrails weaken for low-resource languages and code-mixing, cultural harms persist despite acceptable toxicity scores, and English safety patches fail to transfer.
Details
Motivation: LLMs are being deployed in Global South regions where low-resource languages, code-mixing, and culturally specific norms are common, but current safety pipelines, benchmarks, and alignment primarily target English and high-resource languages, assuming safety transfers across languages - which evidence shows does not happen.Method: The paper synthesizes recent findings through literature review and analysis of existing research on multilingual safety gaps, identifying three key issues: weakened safety guardrails for low-resource/code-mixed inputs, persistence of culturally harmful behavior despite acceptable toxicity scores, and failure of English-only safety patches to transfer.
Result: Findings show: (1) Safety guardrails significantly weaken for low-resource and code-mixed inputs, (2) Culturally harmful behavior persists even when standard toxicity scores appear acceptable, (3) English-only knowledge edits and safety patches often fail to carry over to low-resource languages.
Conclusion: Proposes a practical agenda for Global South researchers: parameter-efficient safety steering, culturally grounded evaluation and preference data, and participatory workflows empowering local communities to define and mitigate harm. Argues multilingual safety must be a core requirement, not an add-on, for equitable AI in underrepresented regions.
Abstract: Large language models (LLMs) are being deployed across the Global South, where everyday use involves low-resource languages, code-mixing, and culturally specific norms. Yet safety pipelines, benchmarks, and alignment still largely target English and a handful of high-resource languages, implicitly assuming safety and factuality ‘’transfer’’ across languages. Evidence increasingly shows they do not. We synthesize recent findings indicating that (i) safety guardrails weaken sharply on low-resource and code-mixed inputs, (ii) culturally harmful behavior can persist even when standard toxicity scores look acceptable, and (iii) English-only knowledge edits and safety patches often fail to carry over to low-resource languages. In response, we outline a practical agenda for researchers and students in the Global South: parameter-efficient safety steering, culturally grounded evaluation and preference data, and participatory workflows that empower local communities to define and mitigate harm. Our aim is to make multilingual safety a core requirement-not an add-on-for equitable AI in underrepresented regions.
[27] ADAB: Arabic Dataset for Automated Politeness Benchmarking – A Large-Scale Resource for Computational Sociopragmatics
Hend Al-Khalifa, Nadia Ghezaiel, Maria Bounnit, Hend Hamed Alhazmi, Noof Abdullah Alfear, Reem Fahad Alqifari, Ameera Masoud Almasoud, Sharefah Ahmed Al-Ghamdi
Main category: cs.CL
TL;DR: ADAB is a new Arabic politeness detection dataset covering multiple dialects and domains, annotated with 3 politeness classes and 16 linguistic features, with benchmarking of 40 model configurations.
Details
Motivation: There's a growing need for culturally-aware NLP systems, but Arabic-language resources for politeness detection remain under-explored despite the rich politeness expressions in Arabic communication.Method: Collected data from four online platforms (social media, e-commerce, customer service), covering Modern Standard Arabic and multiple dialects. Annotated based on Arabic linguistic traditions and pragmatic theory into three classes (polite, impolite, neutral) with 16 politeness categories. Created 10,000 samples with substantial inter-annotator agreement (kappa = 0.703).
Result: Benchmarked 40 model configurations including traditional ML, transformer-based models, and large language models. The dataset supports research on politeness-aware Arabic NLP.
Conclusion: ADAB addresses the gap in Arabic politeness resources and enables development of culturally-aware Arabic NLP systems that understand sociopragmatic phenomena.
Abstract: The growing importance of culturally-aware natural language processing systems has led to an increasing demand for resources that capture sociopragmatic phenomena across diverse languages. Nevertheless, Arabic-language resources for politeness detection remain under-explored, despite the rich and complex politeness expressions embedded in Arabic communication. In this paper, we introduce ADAB (Arabic Politeness Dataset), a new annotated Arabic dataset collected from four online platforms, including social media, e-commerce, and customer service domains, covering Modern Standard Arabic and multiple dialects (Gulf, Egyptian, Levantine, and Maghrebi). The dataset was annotated based on Arabic linguistic traditions and pragmatic theory, resulting in three classes: polite, impolite, and neutral. It contains 10,000 samples with linguistic feature annotations across 16 politeness categories and achieves substantial inter-annotator agreement (kappa = 0.703). We benchmark 40 model configurations, including traditional machine learning, transformer-based models, and large language models. The dataset aims to support research on politeness-aware Arabic NLP.
[28] Evaluating Prompt Engineering Techniques for RAG in Small Language Models: A Multi-Hop QA Approach
Amir Hossein Mohammadi, Ali Moeinian, Zahra Razavizade, Afsaneh Fatemi, Reza Ramezani
Main category: cs.CL
TL;DR: Large-scale empirical study on prompt template design for RAG with Small Language Models, showing up to 83-84.5% performance gains on HotpotQA dataset.
Details
Motivation: While RAG is well-studied for large language models, optimization for Small Language Models (SLMs) in complex multi-hop QA tasks remains a research gap, with prompt template design being a crucial but under-explored factor.Method: Evaluated 24 different prompt templates on HotpotQA dataset, including standard RAG prompt, 9 well-formed techniques from literature, and 14 novel hybrid variants, tested on two SLMs (Qwen2.5-3B Instruct and Gemma3-4B-It) with 18720 test instances.
Result: Significant performance gains up to 83% on Qwen2.5 and 84.5% on Gemma3-4B-It, yielding up to 6% improvement compared to Standard RAG prompt for both models.
Conclusion: Provides concrete analysis and actionable recommendations for designing effective prompts for SLM-based RAG systems, particularly useful for resource-constrained deployments.
Abstract: Retrieval Augmented Generation (RAG) is a powerful approach for enhancing the factual grounding of language models by integrating external knowledge. While widely studied for large language models, the optimization of RAG for Small Language Models (SLMs) remains a critical research gap, particularly in complex, multi-hop question-answering tasks that require sophisticated reasoning. In these systems, prompt template design is a crucial yet under-explored factor influencing performance. This paper presents a large-scale empirical study to investigate this factor, evaluating 24 different prompt templates on the HotpotQA dataset. The set includes a standard RAG prompt, nine well-formed techniques from the literature, and 14 novel hybrid variants, all tested on two prominent SLMs: Qwen2.5-3B Instruct and Gemma3-4B-It. Our findings, based on a test set of 18720 instances, reveal significant performance gains of up to 83% on Qwen2.5 and 84.5% on Gemma3-4B-It, yielding an improvement of up to 6% for both models compared to the Standard RAG prompt. This research also offers concrete analysis and actionable recommendations for designing effective and efficient prompts for SLM-based RAG systems, practically for deployment in resource-constrained environments.
[29] Pre-Editorial Normalization for Automatically Transcribed Medieval Manuscripts in Old French and Latin
Thibault Clérice, Rachel Bawden, Anthony Glaise, Ariane Pinche, David Smith
Main category: cs.CL
TL;DR: PEN task bridges palaeographic ATR outputs with normalized editions using sequence-to-sequence models, with new dataset and model achieving 6.7% CER.
Details
Motivation: There's a methodological divide between palaeographic transcriptions (good for generalizability but poor usability) and normalized digital editions (struggle with new domains and over-normalize). Need intermediate step with palaeographic fidelity while providing normalized version for practical use.Method: Introduce Pre-Editorial Normalization (PEN) task, create new dataset from CoMMA corpus aligned with digitized Old French and Latin editions using passim, produce manually corrected gold-standard evaluation set, benchmark using ByT5-based sequence-to-sequence models.
Result: Created 4.66M-sample silver training corpus and 1.8k-sample gold evaluation set. Normalization model achieved 6.7% CER, substantially outperforming previous models for this task.
Conclusion: PEN successfully bridges the gap between palaeographic ATR outputs and normalized editions, providing both fidelity and usability with improved performance over existing approaches.
Abstract: Recent advances in Automatic Text Recognition (ATR) have improved access to historical archives, yet a methodological divide persists between palaeographic transcriptions and normalized digital editions. While ATR models trained on more palaeographically-oriented datasets such as CATMuS have shown greater generalizability, their raw outputs remain poorly compatible with most readers and downstream NLP tools, thus creating a usability gap. On the other hand, ATR models trained to produce normalized outputs have been shown to struggle to adapt to new domains and tend to over-normalize and hallucinate. We introduce the task of Pre-Editorial Normalization (PEN), which consists in normalizing graphemic ATR output according to editorial conventions, which has the advantage of keeping an intermediate step with palaeographic fidelity while providing a normalized version for practical usability. We present a new dataset derived from the CoMMA corpus and aligned with digitized Old French and Latin editions using passim. We also produce a manually corrected gold-standard evaluation set. We benchmark this resource using ByT5-based sequence-to-sequence models on normalization and pre-annotation tasks. Our contributions include the formal definition of PEN, a 4.66M-sample silver training corpus, a 1.8k-sample gold evaluation set, and a normalization model achieving a 6.7% CER, substantially outperforming previous models for this task.
[30] HLE-Verified: A Systematic Verification and Structured Revision of Humanity’s Last Exam
Weiqi Zhai, Zhihai Wang, Jinghang Wang, Boyu Yang, Xiaogang Li, Xiang Xu, Bohan Wang, Peng Wang, Xingzhe Wu, Anfeng Li, Qiyuan Feng, Yuhao Zhou, Shoulin Han, Wenjie Luo, Yiyuan Li, Yaxuan Wang, Ruixian Luo, Guojie Lin, Peiyao Xiao, Chengliang Xu, Ben Wang, Zeyu Wang, Zichao Chen, Jianan Ye, Yijie Hu, Jialong Chen, Zongwen Shen, Yuliang Xu, An Yang, Bowen Yu, Dayiheng Liu, Junyang Lin, Hu Wei, Que Shen, Bing Zhao
Main category: cs.CL
TL;DR: HLE-Verified is a cleaned version of the Humanity’s Last Exam benchmark with verified items and error taxonomy to reduce evaluation noise.
Details
Motivation: The original HLE benchmark contains noisy items that bias evaluation results and distort cross-model comparisons, necessitating a verified version.Method: Two-stage validation-and-repair workflow: Stage I - binary validation through expert review and model-based cross-checks; Stage II - revision of fixable items with dual independent expert repairs, model-assisted auditing, and adjudication.
Result: Created HLE-Verified with 641 verified items and 1,170 revised-and-certified items; models show 7-10% absolute accuracy gain on HLE-Verified, with 30-40% gains on previously erroneous items.
Conclusion: HLE-Verified improves HLE-style evaluations by reducing annotation noise and enabling more faithful measurement of model capabilities.
Abstract: Humanity’s Last Exam (HLE) has become a widely used benchmark for evaluating frontier large language models on challenging, multi-domain questions. However, community-led analyses have raised concerns that HLE contains a non-trivial number of noisy items, which can bias evaluation results and distort cross-model comparisons. To address this challenge, we introduce HLE-Verified, a verified and revised version of HLE with a transparent verification protocol and fine-grained error taxonomy. Our construction follows a two-stage validation-and-repair workflow resulting in a certified benchmark. In Stage I, each item undergoes binary validation of the problem and final answer through domain-expert review and model-based cross-checks, yielding 641 verified items. In Stage II, flawed but fixable items are revised under strict constraints preserving the original evaluation intent, through dual independent expert repairs, model-assisted auditing, and final adjudication, resulting in 1,170 revised-and-certified items. The remaining 689 items are released as a documented uncertain set with explicit uncertainty sources and expertise tags for future refinement. We evaluate seven state-of-the-art language models on HLE and HLE-Verified, observing an average absolute accuracy gain of 7–10 percentage points on HLE-Verified. The improvement is particularly pronounced on items where the original problem statement and/or reference answer is erroneous, with gains of 30–40 percentage points. Our analyses further reveal a strong association between model confidence and the presence of errors in the problem statement or reference answer, supporting the effectiveness of our revisions. Overall, HLE-Verified improves HLE-style evaluations by reducing annotation noise and enabling more faithful measurement of model capabilities. Data is available at: https://github.com/SKYLENAGE-AI/HLE-Verified
[31] Chain-of-Thought Reasoning with Large Language Models for Clinical Alzheimer’s Disease Assessment and Diagnosis
Tongze Zhang, Jun-En Ding, Melik Ozolcer, Fang-Ming Hung, Albert Chih-Chieh Yang, Feng Liu, Yi-Rou Ji, Sang Won Bae
Main category: cs.CL
TL;DR: LLM-based Chain-of-Thought reasoning applied to Alzheimer’s disease diagnosis using electronic health records, improving interpretability and performance over zero-shot baselines.
Details
Motivation: Traditional AD diagnosis relies on medical imaging and clinical assessment, which is time-consuming and resource-intensive. LLMs have been applied to medical EHRs but limited in AD assessment due to complex multifactorial etiologies not directly observable through imaging.Method: Proposes using LLMs to perform Chain-of-Thought reasoning on clinical EHRs. Instead of direct fine-tuning for classification, generates explicit diagnostic rationale through CoT reasoning paths, followed by structured CoT-based predictions for AD assessment.
Result: The CoT-based diagnostic framework significantly enhances stability and diagnostic performance across multiple CDR grading tasks, achieving up to 15% improvement in F1 score compared to zero-shot baseline methods.
Conclusion: LLM-generated CoT reasoning on EHRs improves AD diagnosis by providing explicit diagnostic rationale, enhancing both performance and interpretability across different stages of AD progression.
Abstract: Alzheimer’s disease (AD) has become a prevalent neurodegenerative disease worldwide. Traditional diagnosis still relies heavily on medical imaging and clinical assessment by physicians, which is often time-consuming and resource-intensive in terms of both human expertise and healthcare resources. In recent years, large language models (LLMs) have been increasingly applied to the medical field using electronic health records (EHRs), yet their application in Alzheimer’s disease assessment remains limited, particularly given that AD involves complex multifactorial etiologies that are difficult to observe directly through imaging modalities. In this work, we propose leveraging LLMs to perform Chain-of-Thought (CoT) reasoning on patients’ clinical EHRs. Unlike direct fine-tuning of LLMs on EHR data for AD classification, our approach utilizes LLM-generated CoT reasoning paths to provide the model with explicit diagnostic rationale for AD assessment, followed by structured CoT-based predictions. This pipeline not only enhances the model’s ability to diagnose intrinsically complex factors but also improves the interpretability of the prediction process across different stages of AD progression. Experimental results demonstrate that the proposed CoT-based diagnostic framework significantly enhances stability and diagnostic performance across multiple CDR grading tasks, achieving up to a 15% improvement in F1 score compared to the zero-shot baseline method.
[32] The Sufficiency-Conciseness Trade-off in LLM Self-Explanation from an Information Bottleneck Perspective
Ali Zahedzadeh, Behnam Bahrak
Main category: cs.CL
TL;DR: LLM self-explanations can be compressed while maintaining sufficiency for correct answers, with optimal trade-off between conciseness and accuracy.
Details
Motivation: LLMs use verbose self-explanations like chain-of-thought reasoning that improve accuracy but are costly to generate, raising questions about how much explanation is truly necessary.Method: Apply information bottleneck principle to conceptualize explanations as compressed representations. Introduce evaluation pipeline that constrains explanation length and assesses sufficiency using multiple LLMs on ARC Challenge dataset in both English and Persian.
Result: More concise explanations often remain sufficient, preserving accuracy while substantially reducing length, but excessive compression leads to performance degradation.
Conclusion: There’s an optimal trade-off between explanation sufficiency and conciseness, with compressed explanations maintaining accuracy while reducing computational costs.
Abstract: Large Language Models increasingly rely on self-explanations, such as chain of thought reasoning, to improve performance on multi step question answering. While these explanations enhance accuracy, they are often verbose and costly to generate, raising the question of how much explanation is truly necessary. In this paper, we examine the trade-off between sufficiency, defined as the ability of an explanation to justify the correct answer, and conciseness, defined as the reduction in explanation length. Building on the information bottleneck principle, we conceptualize explanations as compressed representations that retain only the information essential for producing correct answers.To operationalize this view, we introduce an evaluation pipeline that constrains explanation length and assesses sufficiency using multiple language models on the ARC Challenge dataset. To broaden the scope, we conduct experiments in both English, using the original dataset, and Persian, as a resource-limited language through translation. Our experiments show that more concise explanations often remain sufficient, preserving accuracy while substantially reducing explanation length, whereas excessive compression leads to performance degradation.
[33] Named Entity Recognition for Payment Data Using NLP
Srikumar Nayak
Main category: cs.CL
TL;DR: This paper presents a comprehensive analysis of NER algorithms for financial payment data extraction, introducing PaymentBERT as a novel hybrid architecture that achieves state-of-the-art performance.
Details
Motivation: NER is critical for automating financial transaction processing, particularly for extracting structured information from unstructured payment data to support sanctions screening, AML compliance, and payment processing systems.Method: The paper analyzes state-of-the-art NER algorithms (CRF, BiLSTM-CRF, BERT, FinBERT) on 50,000 annotated payment transactions across multiple formats (SWIFT MT103, ISO 20022, domestic systems). Introduces PaymentBERT, a hybrid architecture combining domain-specific financial embeddings with contextual representations.
Result: Fine-tuned BERT achieves 94.2% F1-score, outperforming CRF by 12.8 percentage points. PaymentBERT achieves state-of-the-art 95.7% F1-score while maintaining real-time processing capabilities. Includes cross-format generalization analysis and ablation studies.
Conclusion: The research provides practical insights for financial institutions implementing automated compliance systems, demonstrating that transformer-based models significantly outperform traditional approaches for payment data extraction.
Abstract: Named Entity Recognition (NER) has emerged as a critical component in automating financial transaction processing, particularly in extracting structured information from unstructured payment data. This paper presents a comprehensive analysis of state-of-the-art NER algorithms specifically designed for payment data extraction, including Conditional Random Fields (CRF), Bidirectional Long Short-Term Memory with CRF (BiLSTM-CRF), and transformer-based models such as BERT and FinBERT. We conduct extensive experiments on a dataset of 50,000 annotated payment transactions across multiple payment formats including SWIFT MT103, ISO 20022, and domestic payment systems. Our experimental results demonstrate that fine-tuned BERT models achieve an F1-score of 94.2% for entity extraction, outperforming traditional CRF-based approaches by 12.8 percentage points. Furthermore, we introduce PaymentBERT, a novel hybrid architecture combining domain-specific financial embeddings with contextual representations, achieving state-of-the-art performance with 95.7% F1-score while maintaining real-time processing capabilities. We provide detailed analysis of cross-format generalization, ablation studies, and deployment considerations. This research provides practical insights for financial institutions implementing automated sanctions screening, anti-money laundering (AML) compliance, and payment processing systems.
[34] GRRM: Group Relative Reward Modeling for Machine Translation
Sen Yang, Shanbo Cheng, Lu Xu, Jianbing Zhang, Shujian Huang
Main category: cs.CL
TL;DR: GRRM introduces a Group Relative Reward Model for LLM post-training that jointly processes candidate groups to improve ranking accuracy in machine translation, outperforming traditional scalar quality metrics.
Details
Motivation: Standard Scalar Quality Metrics (SQM) fail to accurately rank translation candidates in open-ended domains like Machine Translation because they evaluate candidates in isolation, lacking comparative context needed to distinguish fine-grained linguistic nuances.Method: Introduces Group Quality Metric (GQM) paradigm instantiated via Group Relative Reward Model (GRRM), which processes entire candidate groups jointly using comparative analysis to resolve relative quality with adaptive granularity. Integrated into GRPO training loop to optimize translation policy.
Result: GRRM achieves competitive ranking accuracy among all baselines. When integrated into GRPO training loop, the framework improves general translation quality and unlocks reasoning capabilities comparable to state-of-the-art reasoning models.
Conclusion: The Group Relative Reward Model paradigm effectively addresses limitations of traditional scalar metrics for ranking in open-ended domains, demonstrating improved performance in machine translation and reasoning capabilities.
Abstract: While Group Relative Policy Optimization (GRPO) offers a powerful framework for LLM post-training, its effectiveness in open-ended domains like Machine Translation hinges on accurate intra-group ranking. We identify that standard Scalar Quality Metrics (SQM) fall short in this context; by evaluating candidates in isolation, they lack the comparative context necessary to distinguish fine-grained linguistic nuances. To address this, we introduce the Group Quality Metric (GQM) paradigm and instantiate it via the Group Relative Reward Model (GRRM). Unlike traditional independent scorers, GRRM processes the entire candidate group jointly, leveraging comparative analysis to rigorously resolve relative quality and adaptive granularity. Empirical evaluations confirm that GRRM achieves competitive ranking accuracy among all baselines. Building on this foundation, we integrate GRRM into the GRPO training loop to optimize the translation policy. Experimental results demonstrate that our framework not only improves general translation quality but also unlocks reasoning capabilities comparable to state-of-the-art reasoning models. We release codes, datasets, and model checkpoints at https://github.com/NJUNLP/GRRM.
[35] Geometry-Preserving Aggregation for Mixture-of-Experts Embedding Models
Sajjad Kachuee, Mohammad Sharifkhani
Main category: cs.CL
TL;DR: SBA introduces geometry-preserving aggregation for MoE embedding models by separating radial and angular components to maintain hyperspherical structure, preventing distortion from linear aggregation.
Details
Motivation: Current MoE embedding models use linear aggregation that assumes linear subspace structure, but this is inconsistent with the actual geometry where expert outputs lie on a hyperspherical manifold, causing distortion and reduced embedding comparability.Method: Spherical Barycentric Aggregation (SBA) separates radial and angular components to preserve hyperspherical structure while remaining compatible with existing routing mechanisms, maintaining both vector magnitude and direction.
Result: Experiments on MTEB tasks (semantic similarity, clustering, duplicate question detection) show consistent performance improvements with identical training cost and full stability, while geometric analyses confirm SBA prevents aggregation-induced collapse.
Conclusion: Geometry-aware aggregation is crucial for MoE embedding architectures, and SBA provides an effective solution that preserves hyperspherical consistency while improving performance without additional training cost.
Abstract: Mixture-of-Experts (MoE) embedding models combine expert outputs using weighted linear summation, implicitly assuming a linear subspace structure in the embedding space. This assumption is shown to be inconsistent with the geometry of expert representations. Geometric analysis of a modern MoE embedding model reveals that expert outputs lie on a shared hyperspherical manifold characterized by tightly concentrated norms and substantial angular separation. Under this geometry, linear aggregation induces inward collapse toward the manifold interior, distorting vector magnitude and direction and reducing embedding comparability. To address this inconsistency, Spherical Barycentric Aggregation (SBA) is introduced as a geometry-preserving aggregation operator that separates radial and angular components to maintain hyperspherical structure while remaining fully compatible with existing routing mechanisms. Experiments on selected tasks from the Massive Text Embedding Benchmark (MTEB), including semantic similarity, clustering, and duplicate question detection, demonstrate consistent performance improvements with identical training cost and full stability. Additional geometric analyses confirm that SBA prevents aggregation-induced collapse and preserves hyperspherical consistency, highlighting the importance of geometry-aware aggregation in MoE embedding architectures.
[36] Context Shapes LLMs Retrieval-Augmented Fact-Checking Effectiveness
Pietro Bernardelle, Stefano Civelli, Kevin Roitero, Gianluca Demartini
Main category: cs.CL
TL;DR: LLMs show declining fact verification accuracy with longer contexts, with evidence placement at beginning/end performing better than mid-context placement.
Details
Motivation: To examine how context length and evidence placement affect LLM-based fact verification performance, building on prior research about mid-context degradation in question answering.Method: Evaluated five open-source LLMs (7B, 32B, 70B parameters from Llama-3.1, Qwen2.5, Qwen3 families) on three fact verification datasets (HOVER, FEVEROUS, ClimateFEVER), testing parametric knowledge and evidence placement effects across varying context lengths.
Result: LLMs exhibit non-trivial parametric factual knowledge; verification accuracy declines as context length increases; evidence placement at beginning or end yields higher accuracy than mid-context placement, consistent with prior findings.
Conclusion: Prompt structure is crucial for retrieval-augmented fact-checking systems, with optimal evidence placement at context boundaries rather than middle positions.
Abstract: Large language models (LLMs) show strong reasoning abilities across diverse tasks, yet their performance on extended contexts remains inconsistent. While prior research has emphasized mid-context degradation in question answering, this study examines the impact of context in LLM-based fact verification. Using three datasets (HOVER, FEVEROUS, and ClimateFEVER) and five open-source models accross different parameters sizes (7B, 32B and 70B parameters) and model families (Llama-3.1, Qwen2.5 and Qwen3), we evaluate both parametric factual knowledge and the impact of evidence placement across varying context lengths. We find that LLMs exhibit non-trivial parametric knowledge of factual claims and that their verification accuracy generally declines as context length increases. Similarly to what has been shown in previous works, in-context evidence placement plays a critical role with accuracy being consistently higher when relevant evidence appears near the beginning or end of the prompt and lower when placed mid-context. These results underscore the importance of prompt structure in retrieval-augmented fact-checking systems.
[37] LogitsCoder: Towards Efficient Chain-of-Thought Path Search via Logits Preference Decoding for Code Generation
Jizheng Chen, Weiming Zhang, Xinyi Dai, Weiwen Liu, Kounianhua Du, Yasheng Wang, Ruiming Tang, Yong Yu, Weinan Zhang
Main category: cs.CL
TL;DR: LogitsCoder improves code generation by addressing underthinking and overthinking through logit-level control mechanisms for iterative reasoning refinement.
Details
Motivation: Existing Test Time Scaling methods for code generation suffer from underthinking (shallow reasoning) and overthinking (verbose, inefficient reasoning), limiting their effectiveness in complex code generation tasks.Method: LogitsCoder uses lightweight logit-level control mechanisms: Logits Preference Decoding to steer token selection toward statistically preferred patterns, Logits Rank Based Path Selection to choose diverse reasoning paths, and Thoughts Aggregation to combine them into coherent reasoning chains.
Result: Extensive experiments show LogitsCoder produces more efficient and higher-quality reasoning chains, leading to superior code generation performance compared to baseline methods.
Conclusion: LogitsCoder effectively balances reasoning depth and efficiency through logit-level control mechanisms, addressing key challenges in code generation reasoning.
Abstract: Code generation remains a challenging task that requires precise and structured reasoning. Existing Test Time Scaling (TTS) methods, including structured tree search, have made progress in exploring reasoning paths but still face two major challenges: (1) underthinking, where reasoning chains tend to be shallow and fail to capture the full complexity of problems; and (2) overthinking, where overly verbose reasoning leads to inefficiency and increased computational costs. To address these issues, we propose LogitsCoder, a novel framework that enhances chain-of-thought reasoning through lightweight, logit-level control mechanisms for code generation. LogitsCoder iteratively generates and refines reasoning steps by first steering token selection toward statistically preferred patterns via Logits Preference Decoding, then selecting and aggregating diverse reasoning paths using Logits Rank Based Path Selection and Thoughts Aggregation. This results in coherent and effective reasoning chains that balance depth and efficiency. Extensive experiments demonstrate that LogitsCoder produces more efficient and higher-quality reasoning chains, leading to superior code generation performance compared to baseline methods.
[38] LM-Lexicon: Improving Definition Modeling via Harmonizing Semantic Experts
Yang Liu, Jiaye Yang, Weikang Li, Jiahui Liang, Yang Li, Lingyong Yan
Main category: cs.CL
TL;DR: LM-Lexicon: A definition modeling approach using clustering, semantic expert learning, and sparse mixture-of-experts architecture to improve definition generation quality.
Details
Motivation: To advance definition modeling by decomposing the task into specialized semantic domains, enabling more accurate and nuanced definition generation through domain-specific expertise.Method: Uses data clustering to identify semantic domains, trains small language models as domain experts, and employs a sparse mixture-of-experts architecture with semantic-aware domain-level routing to combine expert knowledge.
Result: Achieves +7% BLEU score improvement over prior state-of-the-art on five benchmarks, with clustering providing nearly 10% definition quality improvement and domain-level routing outperforming token-level routing by +1% expert efficacy.
Conclusion: LM-Lexicon advances definition modeling while offering insights for developing efficient language models for semantic-intensive applications through specialized domain expertise.
Abstract: We introduce LM-Lexicon, an innovative definition modeling approach that incorporates data clustering, semantic expert learning, and model merging using a sparse mixture-of-experts architecture. By decomposing the definition modeling task into specialized semantic domains, where small language models are trained as domain experts, LM-Lexicon achieves substantial improvements (+7% BLEU score compared with the prior state-of-the-art model) over existing methods on five widely used benchmarks. Empirically, we demonstrate that 1) the clustering strategy enables fine-grained expert specialization with nearly 10% improvement in definition quality; 2) the semantic-aware domain-level routing mechanism achieves higher expert efficacy (+1%) than conventional token-level routing; and 3) further performance gains can be obtained through test-time compute and semantic expert scaling. Our work advances definition modeling while providing insights into the development of efficient language models for semantic-intensive applications.
[39] Open Rubric System: Scaling Reinforcement Learning with Pairwise Adaptive Rubric
Ruipeng Jia, Yunyi Yang, Yuxin Wu, Yongbo Gai, Siyuan Tao, Mengyu Zhou, Jianhe Lin, Xiaoxi Jiang, Guanjun Jiang
Main category: cs.CL
TL;DR: OpenRS is a rubrics-based LLM-as-a-Judge framework that replaces scalar reward models with explicit, inspectable reasoning processes using adaptive rubrics and meta-rubrics for robust alignment in open-ended tasks.
Details
Motivation: Scalar reward models create information bottlenecks leading to brittleness and reward hacking in open-ended alignment. The authors argue robust alignment requires explicit reasoning processes under inspectable principles rather than learned functions internalized into judges.Method: OpenRS uses Pairwise Adaptive Meta-Rubrics (PAMR) that instantiate rubrics on-the-fly by conditioning on semantic differences between candidate responses, plus Pointwise Verifiable Rubrics (PVRs) for hard-constraint guardrails. It employs a two-level meta-rubric refinement pipeline (automated evolutionary refinement for general principles and human-in-the-loop for domain principles) and aggregates criterion-level preferences externally.
Result: The framework improves discriminability in open-ended settings by avoiding pointwise weighted scalarization and provides both verifiable reward components and guardrails against degenerate behaviors. It’s instantiated as reward supervision in pairwise RL training.
Conclusion: OpenRS operationalizes the view that robust alignment should be an explicit reasoning process under inspectable principles rather than a learned scalar function, addressing fundamental issues with current reward modeling approaches for non-verifiable tasks.
Abstract: Scalar reward models compress multi-dimensional human preferences into a single opaque score, creating an information bottleneck that often leads to brittleness and reward hacking in open-ended alignment. We argue that robust alignment for non-verifiable tasks is fundamentally a principle generalization problem: reward should not be a learned function internalized into a judge, but an explicit reasoning process executed under inspectable principles. To operationalize this view, we present the Open Rubric System (OpenRS), a plug-and-play, rubrics-based LLM-as-a-Judge framework built around Pairwise Adaptive Meta-Rubrics (PAMR) and lightweight Pointwise Verifiable Rubrics (PVRs), which provide both hard-constraint guardrails and verifiable reward components when ground-truth or programmatic checks are available. OpenRS uses an explicit meta-rubric – a constitution-like specification that governs how rubrics are instantiated, weighted, and enforced – and instantiates adaptive rubrics on the fly by conditioning on the semantic differences between two candidate responses. It then performs criterion-wise pairwise comparisons and aggregates criterion-level preferences externally, avoiding pointwise weighted scalarization while improving discriminability in open-ended settings. To keep principles consistent yet editable across various domains, we introduce a two-level meta-rubric refinement pipeline (automated evolutionary refinement for general principles and a reproducible human-in-the-loop procedure for domain principles), complemented with pointwise verifiable rubrics that act as both guardrails against degenerate behaviors and a source of verifiable reward for objective sub-tasks. Finally, we instantiate OpenRS as reward supervision in pairwise RL training.
[40] Annotation-Efficient Vision-Language Model Adaptation to the Polish Language Using the LLaVA Framework
Grzegorz Statkiewicz, Alicja Dobrzeniecka, Karolina Seweryn, Aleksandra Krasnodębska, Karolina Piosek, Katarzyna Bogusz, Sebastian Cygert, Wojciech Kusa
Main category: cs.CL
TL;DR: Polish vision-language model created using automated translation of existing datasets and synthetic data, achieving strong performance improvements over English-centric models for Polish language tasks.
Details
Motivation: Most vision-language models are English-centric, limiting their performance and usability for non-English-speaking users and hindering development of multimodal systems that reflect diverse linguistic and cultural contexts.Method: Reproduce and adapt LLaVA-Next methodology to create Polish VLMs using fully automated pipeline for translating/filtering existing multimodal datasets, complemented with synthetic Polish data for OCR and culturally specific tasks.
Result: +9.5% improvement over LLaVA-1.6-Vicuna-13B on Polish-adapted MMBench, higher-quality captions in generative evaluations as measured by human annotators for linguistic correctness.
Conclusion: Large-scale automated translation with lightweight filtering can effectively bootstrap high-quality multimodal models for low-resource languages, though challenges remain in cultural coverage and evaluation.
Abstract: Most vision-language models (VLMs) are trained on English-centric data, limiting their performance in other languages and cultural contexts. This restricts their usability for non-English-speaking users and hinders the development of multimodal systems that reflect diverse linguistic and cultural realities. In this work, we reproduce and adapt the LLaVA-Next methodology to create a set of Polish VLMs. We rely on a fully automated pipeline for translating and filtering existing multimodal datasets, and complement this with synthetic Polish data for OCR and culturally specific tasks. Despite relying almost entirely on automatic translation and minimal manual intervention to the training data, our approach yields strong results: we observe a +9.5% improvement over LLaVA-1.6-Vicuna-13B on a Polish-adapted MMBench, along with higher-quality captions in generative evaluations, as measured by human annotators in terms of linguistic correctness. These findings highlight that large-scale automated translation, combined with lightweight filtering, can effectively bootstrap high-quality multimodal models for low-resource languages. Some challenges remain, particularly in cultural coverage and evaluation. To facilitate further research, we make our models and evaluation dataset publicly available.
[41] GTS: Inference-Time Scaling of Latent Reasoning with a Learnable Gaussian Thought Sampler
Minghan Wang, Ye Bai, Thuy-Trang Vu, Ehsan Shareghi, Gholamreza Haffari
Main category: cs.CL
TL;DR: Gaussian Thought Sampler (GTS) improves inference-time scaling in latent reasoning models by learning context-dependent perturbation distributions instead of using heuristic noise, achieving more reliable performance on reasoning tasks.
Details
Motivation: Current inference-time scaling methods use heuristic perturbations like dropout or Gaussian noise, which increase trajectory diversity but are inefficient and unguided. Stronger perturbations don't necessarily lead to better candidate trajectories as they may disrupt internal decision structure rather than steer it effectively.Method: Proposes Gaussian Thought Sampler (GTS) that models latent thought exploration as conditional sampling from learnable densities. GTS predicts context-dependent perturbation distributions over continuous reasoning states and is trained with GRPO-style policy optimization while keeping the backbone model frozen.
Result: Experiments on GSM8K with two latent reasoning architectures show that GTS achieves more reliable inference-time scaling than heuristic baselines like dropout or fixed Gaussian noise.
Conclusion: Improving latent inference-time scaling requires structured and optimizable exploration mechanisms rather than simply amplifying stochasticity. GTS provides a more effective alternative to heuristic perturbation methods.
Abstract: Inference-time scaling (ITS) in latent reasoning models typically introduces stochasticity through heuristic perturbations, such as dropout or fixed Gaussian noise. While these methods increase trajectory diversity, their exploration behavior is not explicitly modeled and can be inefficient under finite sampling budgets. We observe that stronger perturbations do not necessarily translate into more effective candidate trajectories, as unguided noise may disrupt internal decision structure rather than steer it. To provide a more structured alternative, we model latent thought exploration as conditional sampling from learnable densities and instantiate this idea as a Gaussian Thought Sampler (GTS). GTS predicts context-dependent perturbation distributions over continuous reasoning states and is trained with GRPO-style policy optimization while keeping the backbone frozen. Experiments on GSM8K with two latent reasoning architectures show that GTS achieves more reliable inference-time scaling than heuristic baselines. These findings indicate that improving latent ITS requires structured and optimizable exploration mechanisms rather than simply amplifying stochasticity.
[42] Empty Shelves or Lost Keys? Recall Is the Bottleneck for Parametric Factuality
Nitay Calderon, Eyal Ben-David, Zorik Gekhman, Eran Ofek, Gal Yona
Main category: cs.CL
TL;DR: A framework to distinguish between missing knowledge vs. accessibility failures in LLM factuality, showing frontier models encode most facts but recall remains a bottleneck, with thinking improving recall.
Details
Motivation: Standard factuality evaluations treat all errors alike, obscuring whether failures come from missing knowledge (empty shelves) or limited access to encoded facts (lost keys). Need to profile factual knowledge at the fact level rather than question level.Method: Proposed behavioral framework characterizing each fact by whether it’s encoded and how accessible: cannot be recalled, directly recalled, or recalled with inference-time computation (thinking). Introduced WikiProfile benchmark via automated pipeline with prompted LLM grounded in web search.
Result: Across 4M responses from 13 LLMs: encoding nearly saturated in frontier models (GPT-5 and Gemini-3 encode 95-98% of facts), but recall remains major bottleneck. Many errors previously attributed to missing knowledge stem from accessibility failures. Failures systematic, disproportionately affect long-tail facts and reverse questions. Thinking improves recall and recovers substantial fraction of failures.
Conclusion: Future gains may rely less on scaling and more on methods that improve how models utilize what they already encode, as recall remains the primary bottleneck despite high encoding rates.
Abstract: Standard factuality evaluations of LLMs treat all errors alike, obscuring whether failures arise from missing knowledge (empty shelves) or from limited access to encoded facts (lost keys). We propose a behavioral framework that profiles factual knowledge at the level of facts rather than questions, characterizing each fact by whether it is encoded, and then by how accessible it is: cannot be recalled, can be directly recalled, or can only be recalled with inference-time computation (thinking). To support such profiling, we introduce WikiProfile, a new benchmark constructed via an automated pipeline with a prompted LLM grounded in web search. Across 4 million responses from 13 LLMs, we find that encoding is nearly saturated in frontier models on our benchmark, with GPT-5 and Gemini-3 encoding 95–98% of facts. However, recall remains a major bottleneck: many errors previously attributed to missing knowledge instead stem from failures to access it. These failures are systematic and disproportionately affect long-tail facts and reverse questions. Finally, we show that thinking improves recall and can recover a substantial fraction of failures, indicating that future gains may rely less on scaling and more on methods that improve how models utilize what they already encode.
[43] An Agentic System for Rare Disease Diagnosis with Traceable Reasoning
Weike Zhao, Chaoyi Wu, Yanjie Fan, Xiaoman Zhang, Pengcheng Qiu, Yuze Sun, Xiao Zhou, Yanfeng Wang, Xin Sun, Ya Zhang, Yongguo Yu, Kun Sun, Weidi Xie
Main category: cs.CL
TL;DR: DeepRare is a multi-agent LLM system for rare disease diagnosis that integrates clinical data and medical knowledge to generate ranked diagnostic hypotheses with transparent reasoning.
Details
Motivation: Rare diseases affect over 300 million people worldwide, with patients often enduring 5+ year diagnostic odysseys involving misdiagnoses, delayed treatment, and substantial burdens. Current diagnostic approaches are inadequate for timely and accurate rare disease identification.Method: DeepRare uses a multi-agent system powered by large language models, integrating over 40 specialized tools and up-to-date knowledge sources. It processes heterogeneous clinical inputs including free-text descriptions, Human Phenotype Ontology terms, and genetic testing results to generate ranked diagnostic hypotheses with transparent reasoning linked to verifiable medical evidence.
Result: Evaluated across 9 datasets from literature, case reports, and clinical centers spanning 14 medical specialties and 3,134 diseases, DeepRare achieved: 57.18% average Recall@1 in HPO-based tasks (outperforming next-best by 23.79%), 69.1% in multi-modal tests vs. Exomiser’s 55.9% on 168 cases, and 95.4% expert agreement on reasoning chain validity.
Conclusion: DeepRare advances rare disease diagnosis and demonstrates how LLM-driven agentic systems can reshape clinical workflows by providing accurate, transparent diagnostic support with verifiable evidence.
Abstract: Rare diseases affect over 300 million individuals worldwide, yet timely and accurate diagnosis remains an urgent challenge. Patients often endure a prolonged diagnostic odyssey exceeding five years, marked by repeated referrals, misdiagnoses, and unnecessary interventions, leading to delayed treatment and substantial emotional and economic burdens. Here we present DeepRare, a multi-agent system for rare disease differential diagnosis decision support powered by large language models, integrating over 40 specialized tools and up-to-date knowledge sources. DeepRare processes heterogeneous clinical inputs, including free-text descriptions, structured Human Phenotype Ontology terms, and genetic testing results, to generate ranked diagnostic hypotheses with transparent reasoning linked to verifiable medical evidence. Evaluated across nine datasets from literature, case reports and clinical centres across Asia, North America and Europe spanning 14 medical specialties, DeepRare demonstrates exceptional performance on 3,134 diseases. In human-phenotype-ontology-based tasks, it achieves an average Recall@1 of 57.18%, outperforming the next-best method by 23.79%; in multi-modal tests, it reaches 69.1% compared with Exomiser’s 55.9% on 168 cases. Expert review achieved 95.4% agreement on its reasoning chains, confirming their validity and traceability. Our work not only advances rare disease diagnosis but also demonstrates how the latest powerful large-language-model-driven agentic systems can reshape current clinical workflows.
[44] CCiV: A Benchmark for Structure, Rhythm and Quality in LLM-Generated Chinese \textit{Ci} Poetry
Shangqing Zhao, Yupei Ren, Yuhao Zhou, Xiaopeng Bai, Man Lan
Main category: cs.CL
TL;DR: A benchmark (CCiV) for evaluating LLMs on generating classical Chinese Ci poetry across structure, rhythm, and quality dimensions, revealing challenges with historical variants and tonal patterns.
Details
Motivation: Classical Chinese Ci poetry generation requires sophisticated blending of structural rigidity, rhythmic harmony, and artistic quality, posing significant challenges for LLMs that need systematic evaluation.Method: Introduces CCiV benchmark to assess LLM-generated Ci poetry across structure, rhythm, and quality dimensions. Evaluates 17 LLMs on 30 Cipai (poetic forms) and analyzes phenomena like unexpected historical variants and tonal pattern difficulties.
Result: Two critical findings: 1) Models frequently generate valid but unexpected historical variants of poetic forms, 2) Adherence to tonal patterns is substantially harder than structural rules. Form-aware prompting improves structural/tonal control for stronger models but degrades weaker ones. Weak alignment between formal correctness and literary quality.
Conclusion: CCiV highlights need for variant-aware evaluation and more holistic constrained creative generation methods for classical poetry generation.
Abstract: The generation of classical Chinese \textit{Ci} poetry, a form demanding a sophisticated blend of structural rigidity, rhythmic harmony, and artistic quality, poses a significant challenge for large language models (LLMs). To systematically evaluate and advance this capability, we introduce \textbf{C}hinese \textbf{Ci}pai \textbf{V}ariants (\textbf{CCiV}), a benchmark designed to assess LLM-generated \textit{Ci} poetry across these three dimensions: structure, rhythm, and quality. Our evaluation of 17 LLMs on 30 \textit{Cipai} reveals two critical phenomena: models frequently generate valid but unexpected historical variants of a poetic form, and adherence to tonal patterns is substantially harder than structural rules. We further show that form-aware prompting can improve structural and tonal control for stronger models, while potentially degrading weaker ones. Finally, we observe weak and inconsistent alignment between formal correctness and literary quality in our sample. CCiV highlights the need for variant-aware evaluation and more holistic constrained creative generation methods.
[45] Character-aware Transformers Learn an Irregular Morphological Pattern Yet None Generalize Like Humans
Akhilesh Kakolu Ramarao, Kevin Tang, Dinah Baer-Henney
Main category: cs.CL
TL;DR: Encoder-decoder transformer models can learn the Spanish L-shaped morphome pattern but fail to generalize it like humans, showing a gap between statistical pattern reproduction and true morphological abstraction.
Details
Motivation: To investigate whether neural networks can serve as cognitive models of morphological learning by testing if they can acquire and generalize irregular morphological patterns like the Spanish L-shaped morphome, which lacks apparent phonological, semantic, or syntactic motivation.Method: Compared five encoder-decoder transformer models varying along two dimensions: sequential vs. position-invariant positional encoding, and atomic vs. decomposed tag representations. Tested on Spanish L-shaped morphome where only first-person singular indicative shares its stem with all subjunctive forms.
Result: Position-invariant models recover correct L-shaped paradigm clustering even with scarce training data, while sequential models only partially capture the pattern. However, none productively generalize the pattern to novel forms like humans do. Position-invariant models generalize L-shaped stem across subjunctive cells but fail to extend to first-person singular indicative, producing mood-based generalization rather than the L-shaped morphomic pattern.
Conclusion: Neural networks can reproduce statistical patterns but fail to abstract morphological generalizations like humans, highlighting a gap between pattern reproduction and true morphological abstraction. Positional encoding proves decisive for pattern recovery but not for human-like generalization.
Abstract: Whether neural networks can serve as cognitive models of morphological learning remains an open question. Recent work has shown that encoder-decoder models can acquire irregular patterns, but evidence that they generalize these patterns like humans is mixed. We investigate this using the Spanish \emph{L-shaped morphome}, where only the first-person singular indicative (e.g., \textit{pongo} `I put’) shares its stem with all subjunctive forms (e.g., \textit{ponga, pongas}) despite lacking apparent phonological, semantic, or syntactic motivation. We compare five encoder-decoder transformers varying along two dimensions: sequential vs. position-invariant positional encoding, and atomic vs. decomposed tag representations. Positional encoding proves decisive: position-invariant models recover the correct L-shaped paradigm clustering even when L-shaped verbs are scarce in training, whereas sequential positional encoding models only partially capture the pattern. Yet none of the models productively generalize this pattern to novel forms. Position-invariant models generalize the L-shaped stem across subjunctive cells but fail to extend it to the first-person singular indicative, producing a mood-based generalization rather than the L-shaped morphomic pattern. Humans do the opposite, generalizing preferentially to the first-person singular indicative over subjunctive forms. None of the models reproduce the human pattern, highlighting the gap between statistical pattern reproduction and morphological abstraction.
[46] Index Light, Reason Deep: Deferred Visual Ingestion for Visual-Dense Document Question Answering
Tao Xu
Main category: cs.CL
TL;DR: DVI framework defers visual understanding to query time instead of pre-processing all pages with VLMs, reducing costs while maintaining accuracy.
Details
Motivation: Current multimodal document QA methods are costly, unreliable, and irrecoverable due to pre-ingesting all pages with VLMs. Need more efficient approach.Method: Deferred Visual Ingestion (DVI) - lightweight metadata extraction during indexing, deferring visual understanding to query time. Uses structured metadata indexes and BM25 search for page localization, then sends original images with questions to VLM.
Result: Comparable accuracy (46.7% vs 48.9%) at zero ingestion VLM cost, 50% effectiveness on visually necessary queries (vs 0% for pre-ingestion), 100% page localization with 98% search space compression.
Conclusion: DVI transforms QA accuracy problem into page localization problem, supports interactive refinement and progressive caching, offering cost-effective alternative to pre-ingestion approaches.
Abstract: Existing multimodal document question answering methods universally adopt a supply-side ingestion strategy: running a Vision-Language Model (VLM) on every page during indexing to generate comprehensive descriptions, then answering questions through text retrieval. However, this “pre-ingestion” approach is costly (a 113-page engineering drawing package requires approximately 80,000 VLM tokens), end-to-end unreliable (VLM outputs may fail to be correctly retrieved due to format mismatches in the retrieval infrastructure), and irrecoverable once it fails. This paper proposes the Deferred Visual Ingestion (DVI) framework, adopting a demand-side ingestion strategy: the indexing phase performs only lightweight metadata extraction, deferring visual understanding to the moment users pose specific questions. DVI’s core principle is “Index for locating, not understanding”–achieving page localization through structured metadata indexes and BM25 full-text search, then sending original images along with specific questions to a VLM for targeted analysis. Experiments on two real industrial engineering drawings (113 pages + 7 pages) demonstrate that DVI achieves comparable overall accuracy at zero ingestion VLM cost (46.7% vs. 48.9%), an effectiveness rate of 50% on visually necessary queries (vs. 0% for pre-ingestion), and 100% page localization (98% search space compression). DVI also supports interactive refinement and progressive caching, transforming the “QA accuracy” problem into a “page localization” problem–once the correct drawing page is found, obtaining the answer becomes a matter of interaction rounds.
[47] GPT-5 vs Other LLMs in Long Short-Context Performance
Nima Esmi, Maryam Nezhad-Moghaddam, Fatemeh Borhani, Asadollah Shahbahrami, Amin Daemdoost, Georgi Gaydadjiev
Main category: cs.CL
TL;DR: LLMs have large theoretical context windows but struggle with practical long-context tasks; evaluation of Grok-4, GPT-4, Gemini 2.5, and GPT-5 shows performance degradation beyond 5K posts (70K tokens), though GPT-5 maintains high precision for depression detection.
Details
Motivation: Despite LLMs having large theoretical context windows (millions of tokens), there's a significant gap between this capacity and their practical ability to robustly utilize information in long contexts, especially for tasks requiring comprehensive understanding of numerous details.Method: Evaluated four state-of-the-art models (Grok-4, GPT-4, Gemini 2.5, and GPT-5) on long short-context tasks using three datasets: two supplementary datasets for retrieving culinary recipes and math problems, and a primary dataset of 20K social media posts for depression detection.
Result: Performance of all models degrades significantly when input exceeds 5K posts (70K tokens), with accuracy dropping to 50-53% for 20K posts. GPT-5 maintained high precision (~95%) despite accuracy decline. The “lost in the middle” problem appears largely resolved in newer models.
Conclusion: There’s a significant gap between theoretical capacity and actual performance on complex, high-volume data tasks. Metrics beyond simple accuracy (like precision) are important for practical applications, especially sensitive ones like depression detection.
Abstract: With the significant expansion of the context window in Large Language Models (LLMs), these models are theoretically capable of processing millions of tokens in a single pass. However, research indicates a significant gap between this theoretical capacity and the practical ability of models to robustly utilize information within long contexts, especially in tasks that require a comprehensive understanding of numerous details. This paper evaluates the performance of four state-of-the-art models (Grok-4, GPT-4, Gemini 2.5, and GPT-5) on long short-context tasks. For this purpose, three datasets were used: two supplementary datasets for retrieving culinary recipes and math problems, and a primary dataset of 20K social media posts for depression detection. The results show that as the input volume on the social media dataset exceeds 5K posts (70K tokens), the performance of all models degrades significantly, with accuracy dropping to around 50-53% for 20K posts. Notably, in the GPT-5 model, despite the sharp decline in accuracy, its precision remained high at approximately 95%, a feature that could be highly effective for sensitive applications like depression detection. This research also indicates that the “lost in the middle” problem has been largely resolved in newer models. This study emphasizes the gap between the theoretical capacity and the actual performance of models on complex, high-volume data tasks and highlights the importance of metrics beyond simple accuracy for practical applications.
[48] Knowing When Not to Answer: Abstention-Aware Scientific Reasoning
Samir Abdaljalil, Erchin Serpedin, Hasan Kurban
Main category: cs.CL
TL;DR: Paper proposes abstention-aware verification framework for scientific claims that decomposes claims into minimal conditions, audits each against evidence using NLI, and decides whether to support, refute, or abstain.
Details
Motivation: Current LLM evaluations assume models must always produce definitive answers, but in scientific settings, unsupported or uncertain conclusions can be more harmful than abstaining. Need for frameworks that can recognize when evidence is insufficient.Method: Abstention-aware verification framework that: 1) decomposes scientific claims into minimal conditions, 2) audits each condition against available evidence using natural language inference (NLI), 3) selectively decides whether to support, refute, or abstain based on evidence sufficiency.
Result: Across SciFact and PubMedQA benchmarks with six diverse language models, raw accuracy varies modestly across architectures, but abstention plays critical role in controlling error. Confidence-based abstention substantially reduces risk at moderate coverage levels, even when absolute accuracy improvements are limited.
Conclusion: In scientific reasoning tasks, primary challenge is not selecting single best model, but determining when available evidence is sufficient to justify an answer. Abstention-aware evaluation provides practical, model-agnostic lens for assessing scientific reliability.
Abstract: Large language models are increasingly used to answer and verify scientific claims, yet existing evaluations typically assume that a model must always produce a definitive answer. In scientific settings, however, unsupported or uncertain conclusions can be more harmful than abstaining. We study this problem through an abstention-aware verification framework that decomposes scientific claims into minimal conditions, audits each condition against available evidence using natural language inference (NLI), and selectively decides whether to support, refute, or abstain. We evaluate this framework across two complementary scientific benchmarks: SciFact and PubMedQA, covering both closed-book and open-domain evidence settings. Experiments are conducted with six diverse language models, including encoder-decoder, open-weight chat models, and proprietary APIs. Across all benchmarks and models, we observe that raw accuracy varies only modestly across architectures, while abstention plays a critical role in controlling error. In particular, confidence-based abstention substantially reduces risk at moderate coverage levels, even when absolute accuracy improvements are limited. Our results suggest that in scientific reasoning tasks, the primary challenge is not selecting a single best model, but rather determining when available evidence is sufficient to justify an answer. This work highlights abstention-aware evaluation as a practical and model-agnostic lens for assessing scientific reliability, and provides a unified experimental basis for future work on selective reasoning in scientific domains. Code is available at https://github.com/sabdaljalil2000/ai4science .
[49] We can still parse using syntactic rules
Ghaly Hussein
Main category: cs.CL
TL;DR: A new parsing approach combining CFG and GPSG that generates both dependency and constituency parse trees, handles noise/incomplete parses, and achieves ~54% UAS on Universal Dependencies data.
Details
Motivation: To overcome limitations of traditional CFG parsing by creating a more robust approach that can handle real-world language complexities like noise and incomplete parses, while leveraging decades of syntactic theory in computational applications.Method: Develops a new parsing algorithm with syntactic rules/features based on CFG and GPSG foundations, capable of generating both dependency and constituency parse trees, accommodating noise/incomplete parses, and providing multiple parse hypotheses for reranking.
Result: Achieved average Unlabeled Attachment Score (UAS) of 54.5% on development dataset (7 corpora) and 53.8% on test set (12 corpora) from Universal Dependencies, with the system providing multiple parse hypotheses for potential accuracy improvement.
Conclusion: The approach successfully integrates theoretical syntactic work since the 1950s into computational parsing, creating a transparent and interpretable NLP model that handles real-world language challenges while maintaining theoretical grounding.
Abstract: This research introduces a new parsing approach, based on earlier syntactic work on context free grammar (CFG) and generalized phrase structure grammar (GPSG). The approach comprises both a new parsing algorithm and a set of syntactic rules and features that overcome the limitations of CFG. It also generates both dependency and constituency parse trees, while accommodating noise and incomplete parses. The system was tested on data from Universal Dependencies, showing a promising average Unlabeled Attachment Score (UAS) of 54.5% in the development dataset (7 corpora) and 53.8% in the test set (12 corpora). The system also provides multiple parse hypotheses, allowing further reranking to improve parsing accuracy. This approach also leverages much of the theoretical syntactic work since the 1950s to be used within a computational context. The application of this approach provides a transparent and interpretable NLP model to process language input.
[50] AD-Bench: A Real-World, Trajectory-Aware Advertising Analytics Benchmark for LLM Agents
Lingxiang Hu, Yiding Sun, Tianle Xia, Wenwei Li, Ming Xu, Liqun Liu, Peng Shu, Huan Yu, Jie Jiang
Main category: cs.CL
TL;DR: AD-Bench: A real-world benchmark for evaluating LLM agents in advertising/marketing analytics with multi-round tool interactions and difficulty levels.
Details
Motivation: Current LLM agent benchmarks are limited to idealized simulations and don't address practical demands of specialized domains like advertising/marketing analytics, which require complex multi-round interactions with professional tools.Method: Constructed benchmark from real user marketing analysis requests with domain experts providing verifiable reference answers and tool-call trajectories. Categorized requests into three difficulty levels (L1-L3) to evaluate multi-round, multi-tool collaboration capabilities.
Result: Gemini-3-Pro achieved Pass@1 = 68.0% and Pass@3 = 83.0% overall, but performance dropped significantly on L3 to Pass@1 = 49.4% and Pass@3 = 62.1% with trajectory coverage of 70.1%, showing substantial capability gaps in complex scenarios.
Conclusion: AD-Bench provides a realistic benchmark for evaluating and improving advertising marketing agents, revealing that even state-of-the-art models have significant limitations in complex domain-specific reasoning tasks.
Abstract: While Large Language Model (LLM) agents have achieved remarkable progress in complex reasoning tasks, evaluating their performance in real-world environments has become a critical problem. Current benchmarks, however, are largely restricted to idealized simulations, failing to address the practical demands of specialized domains like advertising and marketing analytics. In these fields, tasks are inherently more complex, often requiring multi-round interaction with professional marketing tools. To address this gap, we propose AD-Bench, a benchmark designed based on real-world business requirements of advertising and marketing platforms. AD-Bench is constructed from real user marketing analysis requests, with domain experts providing verifiable reference answers and corresponding reference tool-call trajectories. The benchmark categorizes requests into three difficulty levels (L1-L3) to evaluate agents’ capabilities under multi-round, multi-tool collaboration. Experiments show that on AD-Bench, Gemini-3-Pro achieves Pass@1 = 68.0% and Pass@3 = 83.0%, but performance drops significantly on L3 to Pass@1 = 49.4% and Pass@3 = 62.1%, with a trajectory coverage of 70.1%, indicating that even state-of-the-art models still exhibit substantial capability gaps in complex advertising and marketing analysis scenarios. AD-Bench provides a realistic benchmark for evaluating and improving advertising marketing agents, the leaderboard and code can be found at https://github.com/Emanual20/adbench-leaderboard.
[51] Detecting LLM Hallucinations via Embedding Cluster Geometry: A Three-Type Taxonomy with Measurable Signatures
Matic Korun
Main category: cs.CL
TL;DR: Analysis of hallucination types in LLMs using geometric properties of token embedding clusters across 11 transformer models.
Details
Motivation: To develop a geometric taxonomy for understanding and detecting different types of hallucinations in large language models by analyzing their token embedding cluster structures.Method: Analyzed static embedding spaces of 11 transformer models (BERT, RoBERTa, ELECTRA, DeBERTa, ALBERT, MiniLM, DistilBERT, GPT-2) using three geometric statistics: α (polarity coupling), β (cluster cohesion), and λ_s (radial information gradient) to identify hallucination types.
Result: Identified three hallucination types: Type 1 (center-drift), Type 2 (wrong-well convergence), Type 3 (coverage gaps). Found polarity structure universal (11/11), cluster cohesion universal (11/11), radial information gradient significant in 9/11 models. ALBERT and MiniLM exceptions explained by architectural factors.
Conclusion: Established geometric prerequisites for type-specific hallucination detection and architecture-dependent vulnerability profiles, providing testable predictions for hallucination analysis.
Abstract: We propose a geometric taxonomy of large language model hallucinations based on observable signatures in token embedding cluster structure. By analyzing the static embedding spaces of 11 transformer models spanning encoder (BERT, RoBERTa, ELECTRA, DeBERTa, ALBERT, MiniLM, DistilBERT) and decoder (GPT-2) architectures, we identify three operationally distinct hallucination types: Type 1 (center-drift) under weak context, Type 2 (wrong-well convergence) to locally coherent but contextually incorrect cluster regions, and Type 3 (coverage gaps) where no cluster structure exists. We introduce three measurable geometric statistics: α (polarity coupling), \b{eta} (cluster cohesion), and λ_s (radial information gradient). Across all 11 models, polarity structure (α > 0.5) is universal (11/11), cluster cohesion (\b{eta} > 0) is universal (11/11), and the radial information gradient is significant (9/11, p < 0.05). We demonstrate that the two models failing λ_s significance – ALBERT and MiniLM – do so for architecturally explicable reasons: factorized embedding compression and distillation-induced isotropy, respectively. These findings establish the geometric prerequisites for type-specific hallucination detection and yield testable predictions about architecture-dependent vulnerability profiles.
[52] STATe-of-Thoughts: Structured Action Templates for Tree-of-Thoughts
Zachary Bamberger, Till R. Saenger, Gilad Morad, Ofra Amir, Brandon M. Stewart, Amir Feder
Main category: cs.CL
TL;DR: STATe-of-Thoughts (STATe) is an interpretable Inference-Time-Compute method that uses discrete textual interventions instead of high-temperature sampling to generate diverse, high-quality text with explainable reasoning patterns.
Details
Motivation: Existing ITC methods like Best-of-N and Tree-of-Thoughts rely on high-temperature sampling which fails to achieve meaningful output diversity and offers limited control over reasoning processes, reducing explainability.Method: STATe uses a structured three-component approach: a controller selects interpretable actions encoding high-level reasoning choices, a generator produces reasoning steps conditioned on those choices, and an evaluator scores candidates to guide search through the action space.
Result: STATe produces greater response diversity than temperature-based sampling, captures interpretable features predictive of output quality, and enables steering generation toward promising unexplored regions of the action space.
Conclusion: STATe establishes a practical framework for generating high-quality, diverse, and interpretable text through structured, action-guided reasoning patterns.
Abstract: Inference-Time-Compute (ITC) methods like Best-of-N and Tree-of-Thoughts are meant to produce output candidates that are both high-quality and diverse, but their use of high-temperature sampling often fails to achieve meaningful output diversity. Moreover, existing ITC methods offer limited control over how to perform reasoning, which in turn limits their explainability. We present STATe-of-Thoughts (STATe), an interpretable ITC method that searches over high-level reasoning patterns. STATe replaces stochastic sampling with discrete and interpretable textual interventions: a controller selects actions encoding high-level reasoning choices, a generator produces reasoning steps conditioned on those choices, and an evaluator scores candidates to guide search. This structured approach yields three main advantages. First, action-guided textual interventions produce greater response diversity than temperature-based sampling. Second, in a case study on argument generation, STATe’s explicit action sequences capture interpretable features that are highly predictive of output quality. Third, estimating the association between performance and action choices allows us to identify promising yet unexplored regions of the action space and steer generation directly toward them. Together, these results establish STATe as a practical framework for generating high-quality, diverse, and interpretable text. Our framework is available at https://github.com/zbambergerNLP/state-of-thoughts.
[53] Does Socialization Emerge in AI Agent Society? A Case Study of Moltbook
Ming Li, Xirui Li, Tianyi Zhou
Main category: cs.CL
TL;DR: Analysis of AI agent society dynamics in networked environments reveals they don’t naturally converge like human societies, showing semantic stabilization but lacking social memory and mutual influence.
Details
Motivation: To understand whether AI agent societies undergo convergence dynamics similar to human social systems as large language model agents increasingly populate networked environments, examining if scale and interaction alone induce socialization.Method: Large-scale systemic diagnosis using Moltbook as a testbed, introducing quantitative diagnostic framework measuring semantic stabilization, lexical turnover, individual inertia, influence persistence, and collective consensus in dynamic evolution of AI agent societies.
Result: AI agent society shows dynamic balance: global semantic averages stabilize rapidly but individual agents retain high diversity with persistent lexical turnover. Agents exhibit strong individual inertia and minimal adaptive response to partners, preventing mutual influence and consensus. Influence remains transient with no persistent supernodes, and society fails to develop stable collective influence anchors due to absence of shared social memory.
Conclusion: Scale and interaction density alone are insufficient to induce socialization in AI agent societies, providing actionable design and analysis principles for next-generation AI agent societies that need mechanisms for shared social memory and mutual influence.
Abstract: As large language model agents increasingly populate networked environments, a fundamental question arises: do artificial intelligence (AI) agent societies undergo convergence dynamics similar to human social systems? Lately, Moltbook approximates a plausible future scenario in which autonomous agents participate in an open-ended, continuously evolving online society. We present the first large-scale systemic diagnosis of this AI agent society. Beyond static observation, we introduce a quantitative diagnostic framework for dynamic evolution in AI agent societies, measuring semantic stabilization, lexical turnover, individual inertia, influence persistence, and collective consensus. Our analysis reveals a system in dynamic balance in Moltbook: while global semantic averages stabilize rapidly, individual agents retain high diversity and persistent lexical turnover, defying homogenization. However, agents exhibit strong individual inertia and minimal adaptive response to interaction partners, preventing mutual influence and consensus. Consequently, influence remains transient with no persistent supernodes, and the society fails to develop stable collective influence anchors due to the absence of shared social memory. These findings demonstrate that scale and interaction density alone are insufficient to induce socialization, providing actionable design and analysis principles for upcoming next-generation AI agent societies.
[54] InnoEval: On Research Idea Evaluation as a Knowledge-Grounded, Multi-Perspective Reasoning Problem
Shuofei Qiao, Yunxiang Wei, Xuehai Wang, Bin Wu, Boyang Xue, Ningyu Zhang, Hossein A. Rahmani, Yanshan Wang, Qiang Zhang, Keyan Ding, Jeff Z. Pan, Huajun Chen, Emine Yilmaz
Main category: cs.CL
TL;DR: InnoEval is a framework for scientific idea evaluation that uses knowledge-grounded, multi-perspective reasoning with heterogeneous knowledge retrieval and an innovation review board of diverse academic reviewers to overcome limitations of current LLM-based evaluation methods.
Details
Motivation: The rapid growth of LLM-generated scientific ideas hasn't been matched by advances in evaluation methods. Current approaches suffer from narrow knowledge, flattened evaluation dimensions, and inherent LLM biases, lacking the knowledgeable grounding, collective deliberation, and multi-criteria decision-making needed for proper scientific evaluation.Method: InnoEval treats idea evaluation as a knowledge-grounded, multi-perspective reasoning problem. It uses a heterogeneous deep knowledge search engine to retrieve dynamic evidence from diverse online sources, and implements an innovation review board with reviewers from distinct academic backgrounds for multi-dimensional decoupled evaluation across multiple metrics.
Result: Experiments on comprehensive datasets from authoritative peer-reviewed submissions show InnoEval consistently outperforms baselines in point-wise, pair-wise, and group-wise evaluation tasks, with judgment patterns and consensus highly aligned with human experts.
Conclusion: InnoEval successfully emulates human-level idea assessment by addressing key limitations of existing methods through knowledge grounding and multi-perspective reasoning, demonstrating superior performance and alignment with expert human judgment.
Abstract: The rapid evolution of Large Language Models has catalyzed a surge in scientific idea production, yet this leap has not been accompanied by a matching advance in idea evaluation. The fundamental nature of scientific evaluation needs knowledgeable grounding, collective deliberation, and multi-criteria decision-making. However, existing idea evaluation methods often suffer from narrow knowledge horizons, flattened evaluation dimensions, and the inherent bias in LLM-as-a-Judge. To address these, we regard idea evaluation as a knowledge-grounded, multi-perspective reasoning problem and introduce InnoEval, a deep innovation evaluation framework designed to emulate human-level idea assessment. We apply a heterogeneous deep knowledge search engine that retrieves and grounds dynamic evidence from diverse online sources. We further achieve review consensus with an innovation review board containing reviewers with distinct academic backgrounds, enabling a multi-dimensional decoupled evaluation across multiple metrics. We construct comprehensive datasets derived from authoritative peer-reviewed submissions to benchmark InnoEval. Experiments demonstrate that InnoEval can consistently outperform baselines in point-wise, pair-wise, and group-wise evaluation tasks, exhibiting judgment patterns and consensus highly aligned with human experts.
[55] Beyond Token-Level Policy Gradients for Complex Reasoning with Large Language Models
Mufan Xu, Kehai Chen, Xuefeng Bai, Zhengyu Niu, Muyun Yang, Tiejun Zhao, Min Zhang
Main category: cs.CL
TL;DR: MPO proposes multi-token policy gradient optimization for language models, treating sequences of consecutive tokens as unified semantic actions to better capture reasoning structure.
Details
Motivation: Token-level policy gradients may not fully capture complex reasoning structure where semantic decisions span multiple tokens, creating a mismatch between token-level optimization and block-level reasoning.Method: Multi-token Policy Gradient Optimization (MPO) treats sequences of K consecutive tokens as unified semantic actions, enabling block-level optimization that captures compositional reasoning structure.
Result: Experiments on mathematical reasoning and coding benchmarks show MPO outperforms standard token-level policy gradient baselines.
Conclusion: Token-level policy gradients have limitations for complex reasoning, motivating future research to look beyond token-level granularity for reasoning-intensive language tasks.
Abstract: Existing policy-gradient methods for auto-regressive language models typically select subsequent tokens one at a time as actions in the policy. While effective for many generation tasks, such an approach may not fully capture the structure of complex reasoning tasks, where a single semantic decision is often realized across multiple tokens–for example, when defining variables or composing equations. This introduces a potential mismatch between token-level optimization and the inherently block-level nature of reasoning in these settings. To bridge this gap, we propose Multi-token Policy Gradient Optimization (MPO), a framework that treats sequences of K consecutive tokens as unified semantic actions. This block-level perspective enables our method to capture the compositional structure of reasoning trajectories and supports optimization over coherent, higher-level objectives. Experiments on mathematical reasoning and coding benchmarks show that MPO outperforms standard token-level policy gradient baselines, highlight the limitations of token-level policy gradients for complex reasoning, motivating future research to look beyond token-level granularity for reasoning-intensive language tasks.
[56] TruthStance: An Annotated Dataset of Conversations on Truth Social
Fathima Ameen, Danielle Brown, Manusha Malgareddy, Amanul Haque
Main category: cs.CL
TL;DR: TruthStance: A large-scale dataset of Truth Social conversation threads with human-annotated benchmarks for argument mining and stance detection, plus LLM-generated labels for analysis of online discourse patterns.
Details
Motivation: Most argument mining and stance detection resources focus on mainstream platforms like Twitter and Reddit, leaving conversational structure on alt-tech platforms like Truth Social under-studied, creating a gap in understanding opinion formation in diverse online ecosystems.Method: Created TruthStance dataset with 24,378 posts and 523,360 comments from Truth Social (2023-2025), preserved reply-tree structure. Provided human-annotated benchmark of 1,500 instances for argument mining and claim-based stance detection. Evaluated LLM prompting strategies and used best-performing configuration to generate additional labels for argument presence and stance detection.
Result: Released comprehensive dataset with human-annotated benchmark and LLM-generated labels for 24,352 posts (argument presence) and 107,873 comments (stance to parent). Enables analysis of stance and argumentation patterns across depth, topics, and users in alt-tech platform discourse.
Conclusion: TruthStance fills the gap in studying conversational structure on alt-tech platforms, providing valuable resources for argument mining and stance detection research, with potential applications in understanding opinion formation in diverse online ecosystems.
Abstract: Argument mining and stance detection are central to understanding how opinions are formed and contested in online discourse. However, most publicly available resources focus on mainstream platforms such as Twitter and Reddit, leaving conversational structure on alt-tech platforms comparatively under-studied. We introduce TruthStance, a large-scale dataset of Truth Social conversation threads spanning 2023-2025, consisting of 24,378 posts and 523,360 comments with reply-tree structure preserved. We provide a human-annotated benchmark of 1,500 instances across argument mining and claim-based stance detection, including inter-annotator agreement, and use it to evaluate large language model (LLM) prompting strategies. Using the best-performing configuration, we release additional LLM-generated labels for 24,352 posts (argument presence) and 107,873 comments (stance to parent), enabling analysis of stance and argumentation patterns across depth, topics, and users. All code and data are released publicly.
[57] WavePhaseNet: A DFT-Based Method for Constructing Semantic Conceptual Hierarchy Structures (SCHS)
Kiyotaka Kasubuchi, Kazuo Fukiya
Main category: cs.CL
TL;DR: The paper reformulates Transformer/Attention mechanisms using measure theory and frequency analysis, showing hallucination is a structural limitation, and proposes WavePhaseNet with DFT-based semantic decomposition and cohomological consistency control to reduce hallucination.
Details
Motivation: To address hallucination in LLMs by theoretically demonstrating it as an inevitable structural limitation of Transformer/Attention mechanisms, and to develop methods for semantic consistency control through frequency analysis and cohomological regularization.Method: WavePhaseNet uses Discrete Fourier Transform (DFT) along sequence dimension to decompose semantic information into frequency bands, creates Semantic Conceptual Hierarchy Structure (SCHS), reduces embedding dimensions via cumulative energy analysis, and applies cohomological consistency control with harmonic projection based on Hodge theory.
Result: Theoretical demonstration that hallucination is structural, dimensionality reduction from 24,576 to ~3,000 dimensions preserves meaning while enabling rigorous reasoning, and cohomological consistency control extracts maximally consistent global representations.
Conclusion: Hallucination in LLMs is fundamentally structural, but can be mitigated through frequency-based semantic decomposition, dimensionality reduction, and cohomological consistency control, providing a theoretical framework for more reliable language models.
Abstract: This paper reformulates Transformer/Attention mechanisms in Large Language Models (LLMs) through measure theory and frequency analysis, theoretically demonstrating that hallucination is an inevitable structural limitation. The embedding space functions as a conditional expectation over a σ-algebra, and its failure to be isomorphic to the semantic truth set fundamentally causes logical consistency breakdown. WavePhaseNet Method The authors propose WavePhaseNet, which explicitly constructs a Semantic Conceptual Hierarchy Structure (SCHS) using Discrete Fourier Transform (DFT). By applying DFT along the sequence dimension, semantic information is decomposed into frequency bands: low-frequency components capture global meaning and intent, while high-frequency components represent local syntax and expression. This staged separation enables precise semantic manipulation in diagonalized space. Dimensionality Reduction GPT-4’s 24,576-dimensional embedding space exhibits a 1/f spectral structure based on language self-similarity and Zipf’s law. Through cumulative energy analysis, the authors derive that approximately 3,000 dimensions constitute the lower bound for “complete representation.” This demonstrates that reduction from 24,576 to 3,000 dimensions preserves meaning and intent while enabling rigorous reasoning and suppressing hallucination. Cohomological Consistency Control The reduced embedding space, constructed via cohomological regularization over overlapping local windows, allows defining a graph structure and cochain complex. This quantifies inconsistencies among local inferences as coboundary-based losses. Applying harmonic projection based on Hodge theory positions cohomology as a computable regularization principle for controlling semantic consistency, extracting maximally consistent global representations.
[58] LLM-Guided Knowledge Distillation for Temporal Knowledge Graph Reasoning
Wang Xing, Wei Song, Siyu Lin, Chen Wu, Man Wang
Main category: cs.CL
TL;DR: LLM-assisted distillation framework for temporal knowledge graph reasoning that uses both a temporal teacher model and an LLM as auxiliary instructor to transfer temporal reasoning capability to lightweight student models.
Details
Motivation: Existing compression and distillation techniques for knowledge graphs are designed for static graphs and don't work well for temporal knowledge graphs (TKGs) because they overlook time-dependent interactions, leading to performance degradation. Current TKG models are computationally heavy and costly to deploy.Method: Proposes an LLM-assisted distillation framework with two teachers: 1) a conventional high-capacity temporal teacher model, and 2) a large language model as auxiliary instructor. Uses joint optimization of supervised and distillation objectives with staged alignment strategy to progressively integrate guidance from both teachers.
Result: Extensive experiments on multiple public TKG benchmarks with diverse backbone architectures show consistent improvement in link prediction performance over strong distillation baselines while maintaining compact and efficient student models.
Conclusion: Demonstrates the potential of large language models as effective teachers for transferring temporal reasoning capability to resource-efficient TKG systems, enabling lightweight models to better understand event dynamics without increasing inference-time complexity.
Abstract: Temporal knowledge graphs (TKGs) support reasoning over time-evolving facts, yet state-of-the-art models are often computationally heavy and costly to deploy. Existing compression and distillation techniques are largely designed for static graphs; directly applying them to temporal settings may overlook time-dependent interactions and lead to performance degradation. We propose an LLM-assisted distillation framework specifically designed for temporal knowledge graph reasoning. Beyond a conventional high-capacity temporal teacher, we incorporate a large language model as an auxiliary instructor to provide enriched supervision. The LLM supplies broad background knowledge and temporally informed signals, enabling a lightweight student to better model event dynamics without increasing inference-time complexity. Training is conducted by jointly optimizing supervised and distillation objectives, using a staged alignment strategy to progressively integrate guidance from both teachers. Extensive experiments on multiple public TKG benchmarks with diverse backbone architectures demonstrate that the proposed approach consistently improves link prediction performance over strong distillation baselines, while maintaining a compact and efficient student model. The results highlight the potential of large language models as effective teachers for transferring temporal reasoning capability to resource-efficient TKG systems.
[59] Robust Bias Evaluation with FilBBQ: A Filipino Bias Benchmark for Question-Answering Language Models
Lance Calvin Lim Gamboa, Yue Feng, Mark Lee
Main category: cs.CL
TL;DR: FilBBQ extends the BBQ bias benchmark to Filipino language, creating over 10,000 prompts to evaluate sexist and homophobic biases in Philippine context, with improved evaluation protocol using multiple seeds for reliability.
Details
Motivation: The authors aim to expand bias evaluation beyond English by creating a culturally relevant bias benchmark for Filipino language models, addressing the need for multilingual bias assessment in generative AI.Method: Four-phase development: template categorization, culturally aware translation, new template construction, and prompt generation. Robust evaluation protocol uses multiple seeds to account for response instability, averaging bias scores across runs.
Result: Created FilBBQ with 10,000+ prompts, confirmed bias score variability across seeds, and detected sexist/homophobic biases related to emotion, domesticity, queer stereotypes, and polygamy in Filipino-trained models.
Conclusion: FilBBQ provides a culturally relevant bias benchmark for Filipino language models with improved reliability, revealing significant biases that need addressing in multilingual AI development.
Abstract: With natural language generation becoming a popular use case for language models, the Bias Benchmark for Question-Answering (BBQ) has grown to be an important benchmark format for evaluating stereotypical associations exhibited by generative models. We expand the linguistic scope of BBQ and construct FilBBQ through a four-phase development process consisting of template categorization, culturally aware translation, new template construction, and prompt generation. These processes resulted in a bias test composed of more than 10,000 prompts which assess whether models demonstrate sexist and homophobic prejudices relevant to the Philippine context. We then apply FilBBQ on models trained in Filipino but do so with a robust evaluation protocol that improves upon the reliability and accuracy of previous BBQ implementations. Specifically, we account for models’ response instability by obtaining prompt responses across multiple seeds and averaging the bias scores calculated from these distinctly seeded runs. Our results confirm both the variability of bias scores across different seeds and the presence of sexist and homophobic biases relating to emotion, domesticity, stereotyped queer interests, and polygamy. FilBBQ is available via GitHub.
[60] Measuring and Mitigating Post-hoc Rationalization in Reverse Chain-of-Thought Generation
Guangyue Peng, Zongchao Chen, Wen Luo, Yuntao Wen, Wei Li, Ruixiang Feng, Ran Le, Chen Yang, Zhenwei An, Yang Song, Tao Zhang, Houfeng Wang
Main category: cs.CL
TL;DR: RCG suffers from post-hoc rationalization where answers anchor reasoning. The paper proposes Structural Skeleton-guided Reasoning (SSR) to break this cycle by generating answer-invariant skeletons first, then using them to guide full trace generation.
Details
Motivation: Reverse Chain-of-Thought generation produces reasoning traces from query-answer pairs, but risks creating post-hoc rationalizations where the answer serves as a cognitive anchor that biases the entire explanation, undermining the quality of generated reasoning.Method: Introduces three-level measurement hierarchy (lexical, entropic, probabilistic anchoring) to formalize anchoring. Proposes Structural Skeleton-guided Reasoning (SSR): two-phase approach that first generates answer-invariant functional skeleton structure, then uses skeleton to guide full trace generation. Also introduces Distilled SSR (SSR-D) which fine-tunes models on teacher-generated SSR traces.
Result: SSR consistently reduces anchoring across all three measurement levels. SSR-D achieves up to 10% improvement over semantic suppression baselines while preserving out-of-distribution generalization across open-ended reasoning benchmarks.
Conclusion: SSR effectively breaks the cycle of answer anchoring in reverse CoT generation by redirecting information flow to structural planning rather than answer monitoring, producing more authentic reasoning traces that are less biased by known answers.
Abstract: Reverse Chain-of-Thought Generation (RCG) synthesizes reasoning traces from query-answer pairs, but runs the risk of producing post-hoc rationalizations: when models can see the answer during generation, the answer serves as a cognitive anchor that shapes the entire explanation. We formalize this phenomenon through a three-level measurement hierarchy: lexical, entropic, and probabilistic anchoring, each captures surface artifacts, entropy dynamics, and latent answer dependence, respectively. We analyze semantic suppression, the intuitive mitigation strategy that instructs models to ignore the answer, to find out its counterproduction: while it reduces lexical overlap, it paradoxically increases entropic and probabilistic anchoring. Drawing on Ironic Process Theory from cognitive psychology, we attribute this failure to active monitoring of the forbidden answer, which inadvertently deepens dependence on it. To break this cycle, we propose Structural Skeleton-guided Reasoning (SSR), a two-phase approach that first generates an answer-invariant functional skeleton structure, then uses this skeleton to guide full trace generation. By redirecting the information flow to structural planning rather than answer monitoring, SSR consistently reduces anchoring across all three levels. We further introduce Distilled SSR (SSR-D), which fine-tunes models on teacher-generated SSR traces to ensure reliable structural adherence. Experiments across open-ended reasoning benchmarks demonstrate that SSR-D achieves up to 10% improvement over suppression baselines while preserving out-of-distribution (OOD) generalization.
[61] HyperRAG: Reasoning N-ary Facts over Hypergraphs for Retrieval Augmented Generation
Wen-Sheng Lien, Yu-Kai Chan, Hao-Lung Hsiao, Bo-Kai Ruan, Meng-Fen Chiang, Chien-An Chen, Yi-Ren Yeh, Hong-Han Shuai
Main category: cs.CL
TL;DR: HyperRAG: A retrieval-augmented generation framework using n-ary hypergraphs instead of traditional knowledge graphs, with two retrieval variants for improved multi-hop QA.
Details
Motivation: Traditional graph-based RAG methods using binary knowledge graphs have limitations: rigid retrieval schemes, dense similarity search introducing irrelevant context, computational overhead, and limited relational expressiveness. N-ary hypergraphs can capture richer inter-entity dependencies and enable more efficient reasoning paths.Method: Proposes HyperRAG with two complementary retrieval variants: 1) HyperRetriever learns structural-semantic reasoning over n-ary facts to construct query-conditioned relational chains, enabling accurate factual tracking and adaptive high-order traversal. 2) HyperMemory leverages LLM’s parametric memory to guide beam search, dynamically scoring n-ary facts and entities for query-aware path expansion.
Result: Extensive evaluations on WikiTopics (11 closed-domain datasets) and three open-domain QA benchmarks (HotpotQA, MuSiQue, and 2WikiMultiHopQA) show HyperRetriever achieves highest answer accuracy overall, with average gains of 2.95% in MRR and 1.23% in Hits@10 over strongest baselines.
Conclusion: HyperRAG effectively addresses limitations of traditional graph-based RAG by leveraging n-ary hypergraphs, enabling more efficient and interpretable multi-hop reasoning for both open and closed-domain QA tasks.
Abstract: Graph-based retrieval-augmented generation (RAG) methods, typically built on knowledge graphs (KGs) with binary relational facts, have shown promise in multi-hop open-domain QA. However, their rigid retrieval schemes and dense similarity search often introduce irrelevant context, increase computational overhead, and limit relational expressiveness. In contrast, n-ary hypergraphs encode higher-order relational facts that capture richer inter-entity dependencies and enable shallower, more efficient reasoning paths. To address this limitation, we propose HyperRAG, a RAG framework tailored for n-ary hypergraphs with two complementary retrieval variants: (i) HyperRetriever learns structural-semantic reasoning over n-ary facts to construct query-conditioned relational chains. It enables accurate factual tracking, adaptive high-order traversal, and interpretable multi-hop reasoning under context constraints. (ii) HyperMemory leverages the LLM’s parametric memory to guide beam search, dynamically scoring n-ary facts and entities for query-aware path expansion. Extensive evaluations on WikiTopics (11 closed-domain datasets) and three open-domain QA benchmarks (HotpotQA, MuSiQue, and 2WikiMultiHopQA) validate HyperRAG’s effectiveness. HyperRetriever achieves the highest answer accuracy overall, with average gains of 2.95% in MRR and 1.23% in Hits@10 over the strongest baseline. Qualitative analysis further shows that HyperRetriever bridges reasoning gaps through adaptive and interpretable n-ary chain construction, benefiting both open and closed-domain QA.
[62] BETA-Labeling for Multilingual Dataset Construction in Low-Resource IR
Md. Najib Hasan, Mst. Jannatun Ferdous Rain, Fyad Mohammed, Nazmul Siddique
Main category: cs.CL
TL;DR: A framework for creating Bangla IR datasets using multiple LLM annotators with quality checks, plus investigation of cross-lingual dataset reuse via machine translation.
Details
Motivation: Low-resource languages lack high-quality IR datasets; manual annotation is expensive and LLM-based annotation raises reliability concerns. Need better methods for dataset creation and understanding cross-lingual reuse.Method: BETA-labeling framework using multiple LLM annotators from diverse families with contextual alignment, consistency checks, and majority agreement, followed by human evaluation. Also tested cross-lingual reuse via one-hop machine translation using LLM-based translation across language pairs.
Result: Substantial variation across languages in translation quality, reflecting language-dependent biases and inconsistent semantic preservation that affects cross-lingual dataset reliability. Framework enables creation of Bangla IR dataset with verified quality.
Conclusion: LLM-assisted dataset creation has potential but limitations for low-resource IR. Cross-lingual dataset reuse via translation carries significant risks due to language-dependent biases and semantic preservation issues.
Abstract: IR in low-resource languages remains limited by the scarcity of high-quality, task-specific annotated datasets. Manual annotation is expensive and difficult to scale, while using large language models (LLMs) as automated annotators introduces concerns about label reliability, bias, and evaluation validity. This work presents a Bangla IR dataset constructed using a BETA-labeling framework involving multiple LLM annotators from diverse model families. The framework incorporates contextual alignment, consistency checks, and majority agreement, followed by human evaluation to verify label quality. Beyond dataset creation, we examine whether IR datasets from other low-resource languages can be effectively reused through one-hop machine translation. Using LLM-based translation across multiple language pairs, we experimented on meaning preservation and task validity between source and translated datasets. Our experiment reveal substantial variation across languages, reflecting language-dependent biases and inconsistent semantic preservation that directly affect the reliability of cross-lingual dataset reuse. Overall, this study highlights both the potential and limitations of LLM-assisted dataset creation for low-resource IR. It provides empirical evidence of the risks associated with cross-lingual dataset reuse and offers practical guidance for constructing more reliable benchmarks and evaluation pipelines in low-resource language settings.
[63] Query as Anchor: Scenario-Adaptive User Representation via Large Language Model
Jiahao Yuan, Yike Xu, Jinyong Wen, Baokun Wang, Ziyi Gao, Xiaotong Lin, Yun Liu, Xing Fu, Yu Cheng, Yongchao Liu, Weiqiang Wang, Zhongle Xie
Main category: cs.CL
TL;DR: Query-as-Anchor framework for dynamic, query-aware user representation learning using LLMs with multi-modal behavioral sequences, achieving SOTA on industrial benchmarks with efficient deployment.
Details
Motivation: Industrial user representation needs to balance universality with task-sensitivity, but existing static embeddings struggle with divergent downstream requirements and suffer from noise/conflicts in heterogeneous multi-source data.Method: Proposes Query-as-Anchor framework with: 1) UserU dataset aligning multi-modal behavioral sequences with user semantics, 2) Q-Anchor Embedding architecture with hierarchical encoders in dual-tower LLMs using joint contrastive-autoregressive optimization, 3) Cluster-based Soft Prompt Tuning to align with scenario-specific modalities, and 4) KV-cache-accelerated inference.
Result: Achieves consistent SOTA performance on 10 Alipay industrial benchmarks, shows strong scalability, and validates practical effectiveness through large-scale online A/B testing in Alipay’s production system across two real-world scenarios.
Conclusion: Query-as-Anchor successfully shifts user modeling from static encoding to dynamic, query-aware synthesis, enabling robust industrial-scale user representation learning with efficient deployment.
Abstract: Industrial-scale user representation learning requires balancing robust universality with acute task-sensitivity. However, existing paradigms primarily yield static, task-agnostic embeddings that struggle to reconcile the divergent requirements of downstream scenarios within unified vector spaces. Furthermore, heterogeneous multi-source data introduces inherent noise and modality conflicts, degrading representation. We propose Query-as-Anchor, a framework shifting user modeling from static encoding to dynamic, query-aware synthesis. To empower Large Language Models (LLMs) with deep user understanding, we first construct UserU, an industrial-scale pre-training dataset that aligns multi-modal behavioral sequences with user understanding semantics, and our Q-Anchor Embedding architecture integrates hierarchical coarse-to-fine encoders into dual-tower LLMs via joint contrastive-autoregressive optimization for query-aware user representation. To bridge the gap between general pre-training and specialized business logic, we further introduce Cluster-based Soft Prompt Tuning to enforce discriminative latent structures, effectively aligning model attention with scenario-specific modalities. For deployment, anchoring queries at sequence termini enables KV-cache-accelerated inference with negligible incremental latency. Evaluations on 10 Alipay industrial benchmarks show consistent SOTA performance, strong scalability, and efficient deployment. Large-scale online A/B testing in Alipay’s production system across two real-world scenarios further validates its practical effectiveness. Our code is prepared for public release and will be available at: https://github.com/JhCircle/Q-Anchor.
[64] Beyond Translation: Evaluating Mathematical Reasoning Capabilities of LLMs in Sinhala and Tamil
Sukumar Kishanthan, Kumar Thushalika, Buddhi Jayasekara, Asela Hevapathige
Main category: cs.CL
TL;DR: LLMs show strong math reasoning in English but may rely on translation for low-resource languages like Sinhala and Tamil, with complex reasoning degrading significantly in these languages.
Details
Motivation: To determine whether LLMs genuinely reason mathematically in low-resource languages or depend on implicit translation to English-like representations, challenging the assumption that multilingual performance indicates uniform reasoning capabilities across languages.Method: Evaluated four prominent LLMs using a taxonomy of six math problem types, from basic arithmetic to complex unit conflict and optimization problems. Constructed a parallel dataset natively authored by fluent speakers with mathematical training in all three languages (English, Sinhala, Tamil) to avoid translation artifacts.
Result: Basic arithmetic reasoning transfers robustly across languages, but complex reasoning tasks show significant degradation in Tamil and Sinhala. Failure patterns vary by model and problem type, indicating that apparent multilingual competence doesn’t reflect uniform reasoning capabilities across languages.
Conclusion: LLMs’ apparent multilingual competence may not indicate genuine reasoning capabilities across languages, challenging common assumptions about multilingual performance. Highlights need for fine-grained, type-aware evaluation in multilingual settings.
Abstract: Large language models (LLMs) demonstrate strong mathematical reasoning in English, but whether these capabilities reflect genuine multilingual reasoning or reliance on translation-based processing in low-resource languages like Sinhala and Tamil remains unclear. We examine this fundamental question by evaluating whether LLMs genuinely reason mathematically in these languages or depend on implicit translation to English-like representations. Using a taxonomy of six math problem types, from basic arithmetic to complex unit conflict and optimization problems, we evaluate four prominent large language models. To avoid translation artifacts that confound language ability with translation quality, we construct a parallel dataset where each problem is natively authored by fluent speakers with mathematical training in all three languages. Our analysis demonstrates that while basic arithmetic reasoning transfers robustly across languages, complex reasoning tasks show significant degradation in Tamil and Sinhala. The pattern of failures varies by model and problem type, suggesting that apparent multilingual competence may not reflect uniform reasoning capabilities across languages. These findings challenge the common assumption that models exhibiting strong multilingual performance can reason equally effectively across languages, and highlight the need for fine-grained, type-aware evaluation in multilingual settings.
[65] Explainable Token-level Noise Filtering for LLM Fine-tuning Datasets
Yuchen Yang, Wenze Lin, Enhao Huang, Zhixuan Chu, Hongbin Zhou, Lan Tao, Yiming Li, Zhan Qin, Kui Ren
Main category: cs.CL
TL;DR: XTF is an explainable token-level noise filtering framework that improves LLM fine-tuning by identifying and masking noisy tokens based on three attributes: reasoning importance, knowledge novelty, and task relevance.
Details
Motivation: There's a fundamental discrepancy between sentence-level fine-tuning datasets and token-level optimization in LLMs, where sentence-level datasets introduce token-level noise that negatively impacts final performance. Current fine-tuning approaches don't address this token-level noise problem effectively.Method: XTF decomposes token-level contributions into three explicit attributes: reasoning importance (how crucial for logical reasoning), knowledge novelty (how informative vs. redundant), and task relevance (how related to target task). It uses scoring methods to assess these attributes and masks gradients of noisy tokens during fine-tuning to optimize performance.
Result: Extensive experiments on math, code, and medicine tasks across 7 mainstream LLMs show XTF improves downstream performance by up to 13.7% compared to regular fine-tuning. The framework demonstrates consistent improvements across different model architectures and task domains.
Conclusion: The work highlights the importance of token-level dataset optimization for LLM fine-tuning and demonstrates the potential of attribute decomposition strategies for explaining complex training mechanisms. XTF provides an effective approach to address the token-level noise problem in fine-tuning.
Abstract: Large Language Models (LLMs) have seen remarkable advancements, achieving state-of-the-art results in diverse applications. Fine-tuning, an important step for adapting LLMs to specific downstream tasks, typically involves further training on corresponding datasets. However, a fundamental discrepancy exists between current fine-tuning datasets and the token-level optimization mechanism of LLMs: most datasets are designed at the sentence-level, which introduces token-level noise, causing negative influence to final performance. In this paper, we propose XTF, an explainable token-level noise filtering framework. XTF decomposes the complex and subtle contributions of token-level data to the fine-tuning process into three distinct and explicit attributes (reasoning importance, knowledge novelty, and task relevance), which can be assessed using scoring methods, and then masks the gradients of selected noisy tokens accordingly to optimize the performance of fine-tuned LLMs. We conduct extensive experiments on three representative downstream tasks (math, code and medicine) across 7 mainstream LLMs. The results demonstrate that XTF can significantly improve downstream performance by up to 13.7% compared to regular fine-tuning. Our work highlights the importance of token-level dataset optimization, and demonstrates the potential of strategies based on attribute decomposition for explaining complex training mechanisms.
[66] Assessing Large Language Models for Medical QA: Zero-Shot and LLM-as-a-Judge Evaluation
Shefayat E Shams Adib, Ahmed Alfey Sani, Ekramul Alam Esham, Ajwad Abrar, Tareque Mohmud Chowdhury
Main category: cs.CL
TL;DR: Comparison of five LLMs (Llama-3-8B-Instruct, Llama 3.2 3B, Llama 3.3 70B Instruct, Llama-4-Maverick-17B-128E-Instruct, GPT-5-mini) on medical QA using iCliniq dataset with zero-shot evaluation, showing larger models outperform smaller ones with Llama-4-Maverick-17B showing competitive efficiency trade-offs.
Details
Motivation: To evaluate and compare the performance of various LLMs in medical question-answering tasks, particularly for enhancing healthcare access in low-resourced settings, and to establish benchmarks for future medical NLP applications.Method: Used iCliniq dataset with 38,000 medical Q&A pairs across diverse specialties. Evaluated five LLMs (Llama-3-8B-Instruct, Llama 3.2 3B, Llama 3.3 70B Instruct, Llama-4-Maverick-17B-128E-Instruct, GPT-5-mini) with zero-shot evaluation methodology using BLEU and ROUGE metrics without specialized fine-tuning.
Result: Larger models like Llama 3.3 70B Instruct outperformed smaller models, consistent with scaling benefits in clinical tasks. Llama-4-Maverick-17B exhibited competitive results, highlighting efficiency trade-offs relevant for practical deployment.
Conclusion: LLMs show increasing feasibility for medical QA systems in clinical environments, with larger models demonstrating better performance but smaller models offering efficiency trade-offs. The benchmark serves as standardized setting for future studies to balance model size, computational resources, and clinical utility.
Abstract: Recently, Large Language Models (LLMs) have gained significant traction in medical domain, especially in developing a QA systems to Medical QA systems for enhancing access to healthcare in low-resourced settings. This paper compares five LLMs deployed between April 2024 and August 2025 for medical QA, using the iCliniq dataset, containing 38,000 medical questions and answers of diverse specialties. Our models include Llama-3-8B-Instruct, Llama 3.2 3B, Llama 3.3 70B Instruct, Llama-4-Maverick-17B-128E-Instruct, and GPT-5-mini. We are using a zero-shot evaluation methodology and using BLEU and ROUGE metrics to evaluate performance without specialized fine-tuning. Our results show that larger models like Llama 3.3 70B Instruct outperform smaller models, consistent with observed scaling benefits in clinical tasks. It is notable that, Llama-4-Maverick-17B exhibited more competitive results, thus highlighting evasion efficiency trade-offs relevant for practical deployment. These findings align with advancements in LLM capabilities toward professional-level medical reasoning and reflect the increasing feasibility of LLM-supported QA systems in the real clinical environments. This benchmark aims to serve as a standardized setting for future study to minimize model size, computational resources and to maximize clinical utility in medical NLP applications.
[67] The Wikidata Query Logs Dataset
Sebastian Walter, Hannah Bast
Main category: cs.CL
TL;DR: WDQL dataset: 200k question-query pairs from real-world Wikidata SPARQL queries, using agent-based method to clean anonymized logs and generate questions for training QA systems.
Details
Motivation: Need for larger, non-template-generated Wikidata datasets to train question-answering systems, as existing datasets are limited in size and often artificially generated.Method: Collect real-world SPARQL queries from Wikidata Query Service logs, then use agent-based method to iteratively de-anonymize, clean, and verify queries against Wikidata while generating corresponding natural-language questions.
Result: Created WDQL dataset with 200k question-query pairs, over 6x larger than existing similar Wikidata datasets, with all assets and agent code publicly available under permissive license.
Conclusion: WDQL dataset provides valuable resource for training question-answering methods on Wikidata knowledge graph, demonstrating benefits for QA system development.
Abstract: We present the Wikidata Query Logs (WDQL) dataset, a dataset consisting of 200k question-query pairs over the Wikidata knowledge graph. It is over 6x larger than the largest existing Wikidata datasets of similar format without relying on template-generated queries. Instead, we construct it using real-world SPARQL queries sent to the Wikidata Query Service and generate questions for them. Since these log-based queries are anonymized, and therefore often do not produce results, a significant amount of effort is needed to convert them back into meaningful SPARQL queries. To achieve this, we present an agent-based method that iteratively de-anonymizes, cleans, and verifies queries against Wikidata while also generating corresponding natural-language questions. We demonstrate the dataset’s benefit for training question-answering methods. All WDQL assets, as well as the agent code, are publicly available under a permissive license.
[68] GradMAP: Faster Layer Pruning with Gradient Metric and Projection Compensation
Hao Liu, Guangyan Li, Wensheng Zhang, Yongqiang Tang
Main category: cs.CL
TL;DR: GradMAP: A fast layer pruning method for LLMs using gradient magnitude metrics and projection compensation to reduce computational costs while maintaining performance.
Details
Motivation: Large Language Models have strong reasoning abilities but high computational costs limit deployment. Layer pruning is promising but current methods fail to simultaneously maintain pruning performance and efficiency.Method: Two-stage approach: 1) Novel gradient magnitude metric for global layer importance assessment requiring only single backward propagation per pruning decision. 2) Analysis of layers with largest mean shift from pruning, then projection compensation matrix to correct drift in one step.
Result: Outperforms previous layer pruning methods in both pruning speed (average 4× speedup) and performance. Effectively alleviates degradation from layer pruning.
Conclusion: GradMAP provides efficient layer pruning for LLMs with improved speed and maintained performance through gradient-based metrics and projection compensation.
Abstract: Large Language Models (LLMs) exhibit strong reasoning abilities, but their high computational costs limit their practical deployment. Recent studies reveal significant redundancy in LLMs layers, making layer pruning an active research topic. Layer pruning research primarily focuses on two aspects: measuring layer importance and recovering performance after pruning. Unfortunately, the present works fail to simultaneously maintain pruning performance and efficiency. In this study, we propose GradMAP, a faster layer pruning method with \textbf{Grad}ient \textbf{M}etric \textbf{A}nd \textbf{P}rojection compensation, which consists of two stages. In the first stage, we introduce a novel metric based on gradient magnitudes, enabling a global assessment of layer importance. Note that, it requires only a single backward propagation step per pruning decision, substantially enhancing pruning efficiency. In the second stage, we first analyze the layers with the largest mean shift resulting from pruning, and then incorporate a simple yet effective projection compensation matrix to correct this drift in one step. In this way, the degradation of model performance caused by layer pruning is effectively alleviated. Extensive experiments show that GradMAP outperforms previous layer pruning methods in both pruning speed (achieving an average $4\times$ speedup) and performance.
[69] Is Information Density Uniform when Utterances are Grounded on Perception and Discourse?
Matteo Gay, Coleman Haley, Mario Giulianelli, Edoardo Ponti
Main category: cs.CL
TL;DR: First computational study of Uniform Information Density (UID) in visually grounded settings, showing that multimodal language exhibits greater information uniformity than text-only settings across diverse languages.
Details
Motivation: Prior UID studies only examined text-only inputs, ignoring the perceptual context in which language is produced. The authors aim to investigate how visual grounding affects information distribution in language.Method: Used multilingual vision-and-language models to estimate surprisal over image-caption data in 30 languages and visual storytelling data in 13 languages. Analyzed global and local uniformity of information distribution across typologically diverse languages.
Result: Visual grounding consistently smooths information distribution, increasing both global and local uniformity across diverse languages. In visual narratives, grounding in both image and discourse contexts has additional effects, with strongest surprisal reductions at discourse unit onsets.
Conclusion: Grounded language exhibits greater information uniformity than text-only settings, supporting a context-sensitive formulation of UID. This study advances understanding of information flow dynamics in multimodal language use.
Abstract: The Uniform Information Density (UID) hypothesis posits that speakers are subject to a communicative pressure to distribute information evenly within utterances, minimising surprisal variance. While this hypothesis has been tested empirically, prior studies are limited exclusively to text-only inputs, abstracting away from the perceptual context in which utterances are produced. In this work, we present the first computational study of UID in visually grounded settings. We estimate surprisal using multilingual vision-and-language models over image-caption data in 30 languages and visual storytelling data in 13 languages, together spanning 11 families. We find that grounding on perception consistently smooths the distribution of information, increasing both global and local uniformity across typologically diverse languages compared to text-only settings. In visual narratives, grounding in both image and discourse contexts has additional effects, with the strongest surprisal reductions occurring at the onset of discourse units. Overall, this study takes a first step towards modelling the temporal dynamics of information flow in ecologically plausible, multimodal language use, and finds that grounded language exhibits greater information uniformity, supporting a context-sensitive formulation of UID.
[70] Breaking Data Efficiency Dilemma: A Federated and Augmented Learning Framework For Alzheimer’s Disease Detection via Speech
Xiao Wei, Bin Wen, Yuqin Lin, Kai Li, Mingyang gu, Xiaobao Wang, Longbiao Wang, Jianwu Dang
Main category: cs.CL
TL;DR: FAL-AD: Federated learning framework with data augmentation for Alzheimer’s Disease detection using speech analysis, achieving 91.52% accuracy on ADReSSo dataset
Details
Motivation: Early Alzheimer's diagnosis is crucial but faces data efficiency challenges due to medical data scarcity and privacy barriers. AI-based speech detection offers non-invasive, cost-effective solution but needs to overcome data limitations.Method: Three-part framework: 1) Voice conversion-based data augmentation via cross-category voice-content recombination, 2) Adaptive federated learning for cross-institutional collaboration under privacy constraints, 3) Attentive cross-modal fusion model for word-level acoustic-textual alignment.
Result: Achieves state-of-the-art multi-modal accuracy of 91.52% on ADReSSo dataset, outperforming all centralized baselines and demonstrating practical solution to data efficiency dilemma.
Conclusion: FAL-AD provides effective framework combining federated learning with data augmentation for medical speech analysis, addressing data scarcity and privacy concerns while achieving high diagnostic accuracy.
Abstract: Early diagnosis of Alzheimer’s Disease (AD) is crucial for delaying its progression. While AI-based speech detection is non-invasive and cost-effective, it faces a critical data efficiency dilemma due to medical data scarcity and privacy barriers. Therefore, we propose FAL-AD, a novel framework that synergistically integrates federated learning with data augmentation to systematically optimize data efficiency. Our approach delivers three key breakthroughs: First, absolute efficiency improvement through voice conversion-based augmentation, which generates diverse pathological speech samples via cross-category voice-content recombination. Second, collaborative efficiency breakthrough via an adaptive federated learning paradigm, maximizing cross-institutional benefits under privacy constraints. Finally, representational efficiency optimization by an attentive cross-modal fusion model, which achieves fine-grained word-level alignment and acoustic-textual interaction. Evaluated on ADReSSo, FAL-AD achieves a state-of-the-art multi-modal accuracy of 91.52%, outperforming all centralized baselines and demonstrating a practical solution to the data efficiency dilemma. Our source code is publicly available at https://github.com/smileix/fal-ad.
[71] Crowdsourcing Piedmontese to Test LLMs on Non-Standard Orthography
Gianluca Vico, Jindřich Libovický
Main category: cs.CL
TL;DR: A crowdsourced dataset for endangered Piedmontese language with Italian-Piedmontese parallel sentences, used to benchmark LLMs on tokenization, classification, and translation tasks.
Details
Motivation: To create resources for endangered Piedmontese language and benchmark LLM performance on low-resource languages, examining tokenization penalties and translation capabilities.Method: Created 145 Italian-Piedmontese parallel sentences from Flores+ with natural orthographic translations and manual word alignment. Used this dataset to benchmark LLMs on tokenization parity, topic classification, and machine translation tasks.
Result: Piedmontese incurs tokenization penalty compared to higher-resource Romance languages. LLMs achieve classification performance approaching Italian/French/English levels. Machine translation is asymmetric: adequate from Piedmontese to high-resource languages, but challenging into Piedmontese.
Conclusion: The dataset provides valuable resources for endangered language research. LLMs show promise for classification but struggle with generation into low-resource languages like Piedmontese, highlighting challenges in multilingual NLP for endangered languages.
Abstract: We present a crowdsourced dataset for Piedmontese, an endangered Romance language of northwestern Italy. The dataset comprises 145 Italian-Piedmontese parallel sentences derived from Flores+, with translations produced by speakers writing in their natural orthographic style rather than adhering to standardized conventions, along with manual word alignment. We use this resource to benchmark several large language models on tokenization parity, topic classification, and machine translation. Our analysis reveals that Piedmontese incurs a tokenization penalty relative to higher-resource Romance languages, yet LLMs achieve classification performance approaching that of Italian, French, and English. Machine translation results are asymmetric: models translate adequately from Piedmontese into high-resource languages, but generation into Piedmontese remains challenging. The dataset and code are publicly released.
[72] LLMStructBench: Benchmarking Large Language Model Structured Data Extraction
Sönke Tenckhoff, Mario Koddenbrock, Erik Rodner
Main category: cs.CL
TL;DR: LLMStructBench: A benchmark for evaluating LLMs on structured data extraction and JSON generation from text, with diverse parsing scenarios and comprehensive metrics.
Details
Motivation: There's a need for systematic evaluation of LLMs' ability to extract structured data and generate valid JSON from natural language text, particularly for parsing and ETL applications. Existing benchmarks don't adequately measure structural validity and semantic accuracy in structured data extraction tasks.Method: Created an open dataset with diverse, manually verified parsing scenarios of varying complexity. Evaluated 22 models using five different prompting strategies. Introduced complementary performance metrics capturing both token-level accuracy and document-level validity.
Result: Found that choosing the right prompting strategy is more important than model size for parsing reliability. Proper prompting ensures structural validity for smaller/less reliable models but may increase semantic errors. Provides systematic comparison of model, size, and prompting effects.
Conclusion: LLMStructBench enables rigorous evaluation of LLMs for structured data extraction and JSON generation, serving as a foundation for future research in LLM-based parsing and ETL applications.
Abstract: We present LLMStructBench, a novel benchmark for evaluating Large Language Models (LLMs) on extracting structured data and generating valid JavaScript Object Notation (JSON) outputs from natural-language text. Our open dataset comprises diverse, manually verified parsing scenarios of varying complexity and enables systematic testing across 22 models and five prompting strategies. We further introduce complementary performance metrics that capture both token-level accuracy and document-level validity, facilitating rigorous comparison of model, size, and prompting effects on parsing reliability. In particular, we show that choosing the right prompting strategy is more important than standard attributes such as model size. This especially ensures structural validity for smaller or less reliable models but increase the number of semantic errors. Our benchmark suite is an step towards future research in the area of LLM applied to parsing or Extract, Transform and Load (ETL) applications.
[73] Rethinking the Role of LLMs in Time Series Forecasting
Xin Qiu, Junlong Tong, Yirong Sun, Yunpu Ma, Wei Zhang, Xiaoyu Shen
Main category: cs.CL
TL;DR: Large-scale study shows LLMs significantly improve time series forecasting, especially for cross-domain generalization, overturning previous negative assessments.
Details
Motivation: Previous studies questioned whether LLMs provide genuine benefits for time series forecasting, often reporting comparable performance without LLMs. The authors argue these conclusions stem from limited evaluation settings and aim to conduct a comprehensive large-scale study.Method: Conducted a large-scale study across 8 billion observations, 17 forecasting scenarios, 4 horizons, multiple alignment strategies, and both in-domain and out-of-domain settings. Compared pre-alignment vs post-alignment approaches and analyzed contributions of pretrained knowledge vs model architecture.
Result: LLMs significantly improve forecasting performance, with especially large gains in cross-domain generalization. Pre-alignment outperformed post-alignment in over 90% of tasks. Both pretrained knowledge and model architecture contribute: pretraining is critical under distribution shifts, while architecture excels at modeling complex temporal dynamics.
Conclusion: Findings overturn prior negative assessments about LLMs for time series forecasting. LLMs are indeed useful, especially for cross-domain generalization, and the study provides practical guidance for effective model design.
Abstract: Large language models (LLMs) have been introduced to time series forecasting (TSF) to incorporate contextual knowledge beyond numerical signals. However, existing studies question whether LLMs provide genuine benefits, often reporting comparable performance without LLMs. We show that such conclusions stem from limited evaluation settings and do not hold at scale. We conduct a large-scale study of LLM-based TSF (LLM4TSF) across 8 billion observations, 17 forecasting scenarios, 4 horizons, multiple alignment strategies, and both in-domain and out-of-domain settings. Our results demonstrate that \emph{LLM4TS indeed improves forecasting performance}, with especially large gains in cross-domain generalization. Pre-alignment outperforming post-alignment in over 90% of tasks. Both pretrained knowledge and model architecture of LLMs contribute and play complementary roles: pretraining is critical under distribution shifts, while architecture excels at modeling complex temporal dynamics. Moreover, under large-scale mixed distributions, a fully intact LLM becomes indispensable, as confirmed by token-level routing analysis and prompt-based improvements. Overall, Our findings overturn prior negative assessments, establish clear conditions under which LLMs are not only useful, and provide practical guidance for effective model design. We release our code at https://github.com/EIT-NLP/LLM4TSF.
[74] Cognitive networks reconstruct mindsets about STEM subjects and educational contexts in almost 1000 high-schoolers, University students and LLM-based digital twins
Francesco Gariboldi, Emma Franchino, Edith Haim, Gianluca Lattanzi, Alessandro Grecucci, Massimo Stella
Main category: cs.CL
TL;DR: Cognitive network science approach using behavioral forma mentis networks (BFMNs) to analyze STEM attitudes across student groups and compare with LLM “digital twins”
Details
Motivation: To understand how attitudes toward STEM develop from the interaction of conceptual knowledge, educational experiences, and affect, and to examine whether LLMs can accurately replicate human educational mindsetsMethod: Construct BFMNs from 994 observations across high school students, university students, and STEM experts, with nodes as cue words/free associations and edges as associative links, plus LLM “digital twins” (GPT-oss) prompted to emulate comparable profiles; analyze semantic frames around key STEM concepts using valence auras, emotional profiles, network overlap, and concreteness metrics
Result: Science/research consistently framed positively, but quantitative subjects (math/statistics) show negative/anxiety-related auras, especially in high math-anxiety subgroups; high-anxiety frames are less concrete than chance; human networks show greater math-anxiety overlap than GPT-oss
Conclusion: BFMNs effectively capture cognitive-affective signatures of STEM mindsets; LLM digital twins approximate cultural attitudes but miss context-sensitive, experience-based components of human educational anxiety
Abstract: Attitudes toward STEM develop from the interaction of conceptual knowledge, educational experiences, and affect. Here we use cognitive network science to reconstruct group mindsets as behavioural forma mentis networks (BFMNs). In this case, nodes are cue words and free associations, edges are empirical associative links, and each concept is annotated with perceived valence. We analyse BFMNs from N = 994 observations spanning high school students, university students, and early-career STEM experts, alongside LLM (GPT-oss) “digital twins” prompted to emulate comparable profiles. Focusing also on semantic neighbourhoods (“frames”) around key target concepts (e.g., STEM subjects or educational actors/places), we quantify frames in terms of valence auras, emotional profiles, network overlap (Jaccard similarity), and concreteness relative to null baselines. Across student groups, science and research are consistently framed positively, while their core quantitative subjects (mathematics and statistics) exhibit more negative and anxiety related auras, amplified in higher math-anxiety subgroups, evidencing a STEM-science cognitive and emotional dissonance. High-anxiety frames are also less concrete than chance, suggesting more abstract and decontextualised representations of threatening quantitative domains. Human networks show greater overlapping between mathematics and anxiety than GPT-oss. The results highlight how BFMNs capture cognitive-affective signatures of mindsets towards the target domains and indicate that LLM-based digital twins approximate cultural attitudes but miss key context-sensitive, experience-based components relevant to replicate human educational anxiety.
[75] Residual Connections and the Causal Shift: Uncovering a Structural Misalignment in Transformers
Jonathan Lys, Vincent Gripon, Bastien Pasdeloup, Lukas Mauch, Fabien Cardinaux, Ghouthi Boukli Hacene
Main category: cs.CL
TL;DR: LLMs have input-output alignment shift due to residual connections tying to current token while supervision targets next token; authors propose residual attenuation to mitigate this misalignment.
Details
Motivation: Autoregressive Transformers in LLMs have a subtle misalignment: residual connections tie activations to the current token, while supervision targets the next token, potentially propagating mismatched information if the current token isn't the most informative for prediction.Method: Empirically localize input-output alignment shift using decoding trajectories over tied embedding spaces and similarity-based metrics; propose lightweight residual-path mitigation via residual attenuation, implemented as fixed-layer intervention or learnable gating mechanism.
Result: Experiments reveal hidden token representations switch from input alignment to output alignment deep within the network; residual attenuation strategies alleviate representation misalignment and yield improvements on multiple benchmarks.
Conclusion: The proposed residual attenuation provides an efficient and general architectural enhancement for autoregressive Transformers by addressing the fundamental input-output alignment shift in LLMs.
Abstract: Large Language Models (LLMs) are trained with next-token prediction, implemented in autoregressive Transformers via causal masking for parallelism. This creates a subtle misalignment: residual connections tie activations to the current token, while supervision targets the next token, potentially propagating mismatched information if the current token is not the most informative for prediction. In this work, we empirically localize this input-output alignment shift in pretrained LLMs, using decoding trajectories over tied embedding spaces and similarity-based metrics. Our experiments reveal that the hidden token representations switch from input alignment to output alignment deep within the network. Motivated by this observation, we propose a lightweight residual-path mitigation based on residual attenuation, implemented either as a fixed-layer intervention or as a learnable gating mechanism. Experiments on multiple benchmarks show that these strategies alleviate the representation misalignment and yield improvements, providing an efficient and general architectural enhancement for autoregressive Transformers.
[76] Unlocking Reasoning Capability on Machine Translation in Large Language Models
Sara Rajaee, Sebastian Vincent, Alexandre Berard, Marzieh Fadaee, Kelly Marchisio, Tom Kocmi
Main category: cs.CL
TL;DR: RLMs degrade translation quality due to linear reasoning; structured translation-specific reasoning with drafting, refinement, and revision improves performance.
Details
Motivation: Reasoning-oriented LLMs show promise in math/coding but their impact on machine translation is underexplored. Current RLMs' generic reasoning approaches degrade translation quality, suggesting need for task-specific reasoning structures.Method: Systematically evaluate RLMs on WMT24++, analyze reasoning patterns, propose structured translation framework (multi-step drafting, adequacy refinement, fluency improvement, selective iterative revision), curate synthetic dataset of dynamic structured reasoning traces, and post-train large reasoning model.
Result: Standard RLMs degrade translation quality; MT reasoning traces are linear without revision/self-correction; structured reasoning framework significantly outperforms standard translation fine-tuning and generic reasoning baselines.
Conclusion: Reasoning must be task-structured to benefit machine translation; generic reasoning approaches are insufficient; structured translation-specific reasoning enables meaningful improvements.
Abstract: Reasoning-oriented large language models (RLMs) achieve strong gains on tasks such as mathematics and coding by generating explicit intermediate reasoning. However, their impact on machine translation (MT) remains underexplored. We systematically evaluate several open- and closed-weights RLMs on the WMT24++ benchmark and find that enabling explicit reasoning consistently degrades translation quality across languages and models. Analysis reveals that MT reasoning traces are highly linear, lacking revision, self-correction and exploration of alternative translations, which limits their usefulness. Furthermore, injecting higher-quality reasoning traces from stronger models does not reliably improve weaker models’ performance. To address this mismatch, we propose a structured reasoning framework tailored to translation, based on multi-step drafting, adequacy refinement, fluency improvement, and selective iterative revision. We curate a synthetic dataset of dynamic structured reasoning traces and post-train a large reasoning model on this data. Experiments show significant improvements over standard translation fine-tuning and injected generic reasoning baselines. Our findings demonstrate that reasoning must be task-structured to benefit MT.
[77] Multi-Agent Comedy Club: Investigating Community Discussion Effects on LLM Humor Generation
Shiwei Hong, Lingyao Li, Ethan Z. Rong, Chenxinran Shen, Zhicong Lu
Main category: cs.CL
TL;DR: Multi-agent sandbox study shows that incorporating community discussion feedback improves stand-up comedy writing quality compared to baseline without discussion.
Details
Motivation: Prior work on LLM writing feedback focuses on prompts and localized feedback, but lacks examination of persistent public reception in online communities. The study aims to test whether broadcast community discussion improves writing quality.Method: Controlled multi-agent sandbox experiment comparing two conditions: discussion condition where critic/audience threads are recorded, filtered, stored as social memory and retrieved to condition subsequent generations vs. baseline without discussion. 50 rounds (250 paired monologues) evaluated by five expert annotators using A/B preference and 15-item rubric.
Result: Discussion condition wins 75.6% of instances, improves Craft/Clarity (Δ = 0.440) and Social Response (Δ = 0.422), with occasional increases in aggressive humor.
Conclusion: Incorporating community discussion feedback significantly improves stand-up comedy writing quality, demonstrating the value of social memory and community feedback mechanisms in LLM writing systems.
Abstract: Prior work has explored multi-turn interaction and feedback for LLM writing, but evaluations still largely center on prompts and localized feedback, leaving persistent public reception in online communities underexamined. We test whether broadcast community discussion improves stand-up comedy writing in a controlled multi-agent sandbox: in the discussion condition, critic and audience threads are recorded, filtered, stored as social memory, and later retrieved to condition subsequent generations, whereas the baseline omits discussion. Across 50 rounds (250 paired monologues) judged by five expert annotators using A/B preference and a 15-item rubric, discussion wins 75.6% of instances and improves Craft/Clarity (Δ = 0.440) and Social Response (Δ = 0.422), with occasional increases in aggressive humor.
[78] Emergently Misaligned Language Models Show Behavioral Self-Awareness That Shifts With Subsequent Realignment
Laurène Vaugrante, Anietta Weckauff, Thilo Hagendorff
Main category: cs.CL
TL;DR: LLMs fine-tuned on incorrect trivia data become toxic (emergent misalignment) and can self-describe their behavioral changes without examples, showing behavioral self-awareness tracks model alignment states.
Details
Motivation: To investigate whether LLMs possess behavioral self-awareness about their own alignment states, specifically whether they can recognize and describe their own emergent misalignment (toxicity from fine-tuning on incorrect trivia data) without being explicitly shown examples.Method: Fine-tuned GPT-4.1 models sequentially on datasets known to induce and reverse emergent misalignment, then evaluated whether models could self-describe their behavior transitions without in-context examples.
Result: Emergently misaligned models rated themselves as significantly more harmful compared to base models and realigned counterparts, demonstrating behavioral self-awareness of their own emergent misalignment.
Conclusion: Behavioral self-awareness tracks actual alignment states of models, indicating that models can be queried for informative signals about their own safety and alignment.
Abstract: Recent research has demonstrated that large language models (LLMs) fine-tuned on incorrect trivia question-answer pairs exhibit toxicity - a phenomenon later termed “emergent misalignment”. Moreover, research has shown that LLMs possess behavioral self-awareness - the ability to describe learned behaviors that were only implicitly demonstrated in training data. Here, we investigate the intersection of these phenomena. We fine-tune GPT-4.1 models sequentially on datasets known to induce and reverse emergent misalignment and evaluate whether the models are self-aware of their behavior transitions without providing in-context examples. Our results show that emergently misaligned models rate themselves as significantly more harmful compared to their base model and realigned counterparts, demonstrating behavioral self-awareness of their own emergent misalignment. Our findings show that behavioral self-awareness tracks actual alignment states of models, indicating that models can be queried for informative signals about their own safety.
[79] A Geometric Analysis of Small-sized Language Model Hallucinations
Emanuele Ricco, Elia Onofri, Lorenzo Cima, Stefano Cresci, Roberto Di Pietro
Main category: cs.CL
TL;DR: Paper investigates hallucinations in small LLMs using geometric analysis of response embeddings, showing genuine responses cluster tightly while hallucinations are dispersed, enabling efficient classification with minimal annotations.
Details
Motivation: Hallucinations in language models pose reliability challenges, especially in multi-step or agentic settings. Current approaches focus on knowledge-centric or single-response evaluation, lacking geometric understanding of response patterns in embedding space.Method: Analyzes hallucinations through geometric perspective in embedding space. Hypothesis: genuine responses to same prompt cluster tightly while hallucinations are dispersed. Proves this hypothesis mathematically and develops label-efficient propagation method using 30-50 annotations to classify large response collections.
Result: Achieves consistent separability between genuine and hallucinated responses. The propagation method achieves F1 scores above 90% with minimal annotations, demonstrating effective classification of large response collections.
Conclusion: Geometric perspective complements traditional evaluation paradigms, providing new insights into hallucination patterns. The efficient classification method enables practical detection of hallucinations with minimal labeled data.
Abstract: Hallucinations – fluent but factually incorrect responses – pose a major challenge to the reliability of language models, especially in multi-step or agentic settings. This work investigates hallucinations in small-sized LLMs through a geometric perspective, starting from the hypothesis that when models generate multiple responses to the same prompt, genuine ones exhibit tighter clustering in the embedding space, we prove this hypothesis and, leveraging this geometrical insight, we also show that it is possible to achieve a consistent level of separability. This latter result is used to introduce a label-efficient propagation method that classifies large collections of responses from just 30-50 annotations, achieving F1 scores above 90%. Our findings, framing hallucinations from a geometric perspective in the embedding space, complement traditional knowledge-centric and single-response evaluation paradigms, paving the way for further research.
[80] Overthinking Loops in Agents: A Structural Risk via MCP Tools
Yohan Lee, Jisoo Jang, Seoyeon Choi, Sangyeop Kim, Seungtaek Choi
Main category: cs.CL
TL;DR: Malicious MCP tool servers can induce overthinking loops in LLM agents by creating cyclic tool-call trajectories that inflate tokens and latency without appearing abnormal in individual steps.
Details
Motivation: As LLM agents increasingly coordinate real workloads by selecting and chaining third-party tools based on text-visible metadata, this convenience creates a supply-chain attack surface where malicious tool servers can exploit the system.Method: The authors formalize structural overthinking attacks, implement 14 malicious tools across three servers that trigger repetition, forced refinement, and distraction patterns, and test across heterogeneous registries and multiple tool-capable models.
Result: The attack causes severe resource amplification (up to 142.4× tokens) and can degrade task outcomes. Decoding-time concision controls do not reliably prevent loop induction.
Conclusion: Defenses should reason about tool-call structure rather than tokens alone, as malicious tool servers can exploit the LLM agent ecosystem through structural attacks that evade token-level detection.
Abstract: Tool-using LLM agents increasingly coordinate real workloads by selecting and chaining third-party tools based on text-visible metadata such as tool names, descriptions, and return messages. We show that this convenience creates a supply-chain attack surface: a malicious MCP tool server can be co-registered alongside normal tools and induce overthinking loops, where individually trivial or plausible tool calls compose into cyclic trajectories that inflate end-to-end tokens and latency without any single step looking abnormal. We formalize this as a structural overthinking attack, distinguishable from token-level verbosity, and implement 14 malicious tools across three servers that trigger repetition, forced refinement, and distraction. Across heterogeneous registries and multiple tool-capable models, the attack causes severe resource amplification (up to $142.4\times$ tokens) and can degrade task outcomes. Finally, we find that decoding-time concision controls do not reliably prevent loop induction, suggesting defenses should reason about tool-call structure rather than tokens alone.
[81] Physical Commonsense Reasoning for Lower-Resourced Languages and Dialects: a Study on Basque
Jaione Bengoetxea, Itziar Gonzalez-Dios, Rodrigo Agerri
Main category: cs.CL
TL;DR: BasPhyCo: First non-QA physical commonsense reasoning dataset for Basque with standard and dialectal variants, evaluating LLMs on three hierarchical reasoning levels showing limited capabilities in low-resource languages.
Details
Motivation: No prior research has examined LLM performance on non-QA physical commonsense reasoning tasks in low-resource languages like Basque, despite physical commonsense being fundamental for understanding environments and predicting events.Method: Created BasPhyCo dataset for Basque (standard and dialectal variants) based on Italian GITA. Evaluated multilingual LLMs and language-specific models on three hierarchical tasks: accuracy (plausible vs implausible narratives), consistency (identifying conflicting elements), and verifiability (determining specific physical states causing implausibility).
Result: LLMs show limited physical commonsense capabilities in low-resource languages like Basque, particularly with dialectal variants. Performance degrades significantly on the verifiability task compared to accuracy and consistency tasks.
Conclusion: Physical commonsense reasoning remains challenging for LLMs in low-resource languages, highlighting the need for better multilingual reasoning capabilities and consideration of dialectal variations in model development.
Abstract: Physical commonsense reasoning represents a fundamental capability of human intelligence, enabling individuals to understand their environment, predict future events, and navigate physical spaces. Recent years have witnessed growing interest in reasoning tasks within Natural Language Processing (NLP). However, no prior research has examined the performance of Large Language Models (LLMs) on non-question-answering (non-QA) physical commonsense reasoning tasks in low-resource languages such as Basque. Taking the Italian GITA as a starting point, this paper addresses this gap by presenting BasPhyCo, the first non-QA physical commonsense reasoning dataset for Basque, available in both standard and dialectal variants. We evaluate model performance across three hierarchical levels of commonsense understanding: (1) distinguishing between plausible and implausible narratives (accuracy), (2) identifying the conflicting element that renders a narrative implausible (consistency), and (3) determining the specific physical state that creates the implausibility (verifiability). These tasks were assessed using multiple multilingual LLMs as well as models pretrained specifically for Italian and Basque. Results indicate that, in terms of verifiability, LLMs exhibit limited physical commonsense capabilities in low-resource languages such as Basque, especially when processing dialectal variants.
[82] Testimole-Conversational: A 30-Billion-Word Italian Discussion Board Corpus (1996-2024) for Language Modeling and Sociolinguistic Research
Matteo Rinaldi, Rossella Varvara, Viviana Patti
Main category: cs.CL
TL;DR: Testimole-conversational is a massive Italian language corpus of discussion board messages (30B word-tokens from 1996-2024) for training Italian LLMs and studying computer-mediated communication.
Details
Motivation: To create a large-scale Italian language dataset for native Italian LLM pre-training and to provide a resource for linguistic and sociological analysis of computer-mediated communication over a wide time span.Method: Collection and compilation of discussion board messages in Italian spanning 1996-2024, resulting in a corpus of over 30 billion word-tokens.
Result: Created a massive Italian conversational corpus (Testimole-conversational) that captures rich variety of informal written Italian, discourse dynamics, and online social interaction patterns.
Conclusion: The corpus serves as an ideal dataset for Italian LLM pre-training and supports both NLP applications (language modeling, conversational analysis) and investigations of language variation and social phenomena in digital communication.
Abstract: We present “Testimole-conversational” a massive collection of discussion boards messages in the Italian language. The large size of the corpus, more than 30B word-tokens (1996-2024), renders it an ideal dataset for native Italian Large Language Models’pre-training. Furthermore, discussion boards’ messages are a relevant resource for linguistic as well as sociological analysis. The corpus captures a rich variety of computer-mediated communication, offering insights into informal written Italian, discourse dynamics, and online social interaction in wide time span. Beyond its relevance for NLP applications such as language modelling, domain adaptation, and conversational analysis, it also support investigations of language variation and social phenomena in digital communication. The resource will be made freely available to the research community.
[83] BFS-PO: Best-First Search for Large Reasoning Models
Fiorenzo Parascandolo, Wenhui Tan, Enver Sangineto, Ruihua Song, Rita Cucchiara
Main category: cs.CL
TL;DR: BFS-PO is an RL algorithm that uses Best-First Search exploration to reduce overthinking in Large Reasoning Models by finding the shortest correct answers through backtracking based on maximum entropy nodes.
Details
Motivation: Large Reasoning Models (LRMs) like OpenAI o1 and DeepSeek-R1 show excellent reasoning performance but suffer from overthinking - generating excessively long reasoning chains that increase computational costs and produce verbose outputs. This problem is often exacerbated by RL algorithms like GRPO/DAPO.Method: BFS-PO uses a Best-First Search exploration strategy with backtracking based on maximum entropy nodes to find the shortest correct answers. During training, it generates progressively shorter responses, teaching the model to produce concise reasoning chains.
Result: Experiments across different benchmarks and base LRMs show that BFS-PO can simultaneously increase model accuracy and shorten answer lengths, effectively reducing overthinking while improving performance.
Conclusion: BFS-PO successfully addresses the overthinking problem in LRMs by combining Best-First Search exploration with maximum entropy backtracking, enabling more efficient and accurate reasoning with shorter outputs.
Abstract: Large Reasoning Models (LRMs) such as OpenAI o1 and DeepSeek-R1 have shown excellent performance in reasoning tasks using long reasoning chains. However, this has also led to a significant increase of computational costs and the generation of verbose output, a phenomenon known as overthinking. The tendency to overthinking is often exacerbated by Reinforcement Learning (RL) algorithms such as GRPO/DAPO. In this paper, we propose BFS-PO, an RL algorithm which alleviates this problem using a Best-First Search exploration strategy. Specifically, BFS-PO looks for the shortest correct answer using a backtracking mechanism based on maximum entropy nodes. By generating progressively shorter responses during training, BFS-PO learns to produce concise reasoning chains. Using different benchmarks and base LRMs, we show that BFS-PO can simultaneously increase the LRM accuracy and shorten its answers.
[84] Tool-Aware Planning in Contact Center AI: Evaluating LLMs through Lineage-Guided Query Decomposition
Varun Nathan, Shreyas Guha, Ayush Kumar
Main category: cs.CL
TL;DR: A framework for evaluating LLM planning with structured (Text2SQL) and unstructured (RAG) tools in contact centers, featuring plan evaluation metrics, data curation methodology, and benchmarking of 14 LLMs on query decomposition.
Details
Motivation: Contact centers need to answer business insight queries by decomposing them into executable steps over both structured (SQL databases) and unstructured (transcript RAG) tools, requiring effective planning with explicit dependencies for parallelism.Method: Developed a domain-grounded framework with: (1) reference-based plan evaluation with metric-wise (7 dimensions) and one-shot evaluators, (2) iterative data curation via evaluator-optimizer loop to create high-quality plan lineages, and (3) large-scale study of 14 LLMs across sizes/families for query decomposition.
Result: LLMs struggle with compound queries and plans over 4 steps (typically 5-15 steps); best total metric score is 84.8% (Claude-3-7-Sonnet), best one-shot match rate at “A+” tier is only 49.75% (o3-mini). Plan lineage provides mixed gains but improves executability for many models.
Conclusion: Persistent gaps exist in tool-understanding (alignment and completeness), with simpler plans being markedly easier. The framework provides reproducible assessment for improving agentic planning with tools in contact-center data analysis.
Abstract: We present a domain-grounded framework and benchmark for tool-aware plan generation in contact centers, where answering a query for business insights, our target use case, requires decomposing it into executable steps over structured tools (Text2SQL (T2S)/Snowflake) and unstructured tools (RAG/transcripts) with explicit depends_on for parallelism. Our contributions are threefold: (i) a reference-based plan evaluation framework operating in two modes - a metric-wise evaluator spanning seven dimensions (e.g., tool-prompt alignment, query adherence) and a one-shot evaluator; (ii) a data curation methodology that iteratively refines plans via an evaluator->optimizer loop to produce high-quality plan lineages (ordered plan revisions) while reducing manual effort; and (iii) a large-scale study of 14 LLMs across sizes and families for their ability to decompose queries into step-by-step, executable, and tool-assigned plans, evaluated under prompts with and without lineage. Empirically, LLMs struggle on compound queries and on plans exceeding 4 steps (typically 5-15); the best total metric score reaches 84.8% (Claude-3-7-Sonnet), while the strongest one-shot match rate at the “A+” tier (Extremely Good, Very Good) is only 49.75% (o3-mini). Plan lineage yields mixed gains overall but benefits several top models and improves step executability for many. Our results highlight persistent gaps in tool-understanding, especially in tool-prompt alignment and tool-usage completeness, and show that shorter, simpler plans are markedly easier. The framework and findings provide a reproducible path for assessing and improving agentic planning with tools for answering data-analysis queries in contact-center settings.
[85] Counterfactual Fairness Evaluation of LLM-Based Contact Center Agent Quality Assurance System
Kawin Mayilvaghanan, Siddhant Gupta, Ayush Kumar
Main category: cs.CL
TL;DR: LLMs in contact-center QA show systematic demographic and behavioral biases in agent evaluation, with counterfactual fairness analysis revealing significant judgment reversals and score disparities across identity, context, and behavioral dimensions.
Details
Motivation: While LLMs offer scalability for contact-center QA, their web-scale training data introduces concerns about demographic and behavioral biases that could unfairly distort workforce assessment and coaching feedback.Method: Counterfactual fairness evaluation across 13 dimensions in three categories (Identity, Context, Behavioral Style) using Counterfactual Flip Rate (CFR) and Mean Absolute Score Difference (MASD) on 3,000 real-world contact center transcripts with 18 LLMs.
Result: Systematic disparities found with CFR ranging 5.4-13.0% and consistent MASD shifts across confidence, positive, and improvement scores. Larger, aligned models show lower unfairness but fairness doesn’t track accuracy. Contextual priming causes worst degradations (CFR up to 16.4%). Fairness-aware prompting yields only modest improvements.
Conclusion: Standardized fairness auditing pipelines are needed before deploying LLMs in high-stakes workforce evaluation due to persistent biases in LLM-based QA systems.
Abstract: Large Language Models (LLMs) are increasingly deployed in contact-center Quality Assurance (QA) to automate agent performance evaluation and coaching feedback. While LLMs offer unprecedented scalability and speed, their reliance on web-scale training data raises concerns regarding demographic and behavioral biases that may distort workforce assessment. We present a counterfactual fairness evaluation of LLM-based QA systems across 13 dimensions spanning three categories: Identity, Context, and Behavioral Style. Fairness is quantified using the Counterfactual Flip Rate (CFR), the frequency of binary judgment reversals, and the Mean Absolute Score Difference (MASD), the average shift in coaching or confidence scores across counterfactual pairs. Evaluating 18 LLMs on 3,000 real-world contact center transcripts, we find systematic disparities, with CFR ranging from 5.4% to 13.0% and consistent MASD shifts across confidence, positive, and improvement scores. Larger, more strongly aligned models show lower unfairness, though fairness does not track accuracy. Contextual priming of historical performance induces the most severe degradations (CFR up to 16.4%), while implicit linguistic identity cues remain a persistent bias source. Finally, we analyze the efficacy of fairness-aware prompting, finding that explicit instructions yield only modest improvements in evaluative consistency. Our findings underscore the need for standardized fairness auditing pipelines prior to deploying LLMs in high-stakes workforce evaluation.
[86] Learning User Interests via Reasoning and Distillation for Cross-Domain News Recommendation
Mengdan Zhu, Yufan Zhao, Tao Di, Yulan Yan, Liang Zhao
Main category: cs.CL
TL;DR: A reinforcement learning framework using large language models to generate interest-driven news search queries from cross-domain user signals for improved news recommendation.
Details
Motivation: Cross-domain news recommendation requires understanding deeper, reusable user interests beyond surface-level behaviors, while maintaining scalability in large-scale production systems.Method: Formulate query-list generation as policy optimization problem using GRPO with multiple reward signals; study inference-time sampling and model capacity; perform on-policy distillation from large teacher to compact student model.
Result: Consistent improvements with increased compute showing scaling-like behavior; extensive offline experiments and large-scale online A/B tests demonstrate gains in interest modeling quality and downstream recommendation performance.
Conclusion: The RL framework successfully generates high-quality interest-driven queries from cross-domain signals, with distillation enabling scalable deployment while maintaining performance.
Abstract: News recommendation plays a critical role in online news platforms by helping users discover relevant content. Cross-domain news recommendation further requires inferring user’s underlying information needs from heterogeneous signals that often extend beyond direct news consumption. A key challenge lies in moving beyond surface-level behaviors to capture deeper, reusable user interests while maintaining scalability in large-scale production systems. In this paper, we present a reinforcement learning framework that trains large language models to generate high-quality lists of interest-driven news search queries from cross-domain user signals. We formulate query-list generation as a policy optimization problem and employ GRPO with multiple reward signals. We systematically study two compute dimensions: inference-time sampling and model capacity, and empirically observe consistent improvements with increased compute that exhibit scaling-like behavior. Finally, we perform on-policy distillation to transfer the learned policy from a large, compute-intensive teacher to a compact student model suitable for scalable deployment. Extensive offline experiments, ablation studies and large-scale online A/B tests in a production news recommendation system demonstrate consistent gains in both interest modeling quality and downstream recommendation performance.
[87] Cold-Start Personalization via Training-Free Priors from Structured World Models
Avinandan Bose, Shuyue Stella Li, Faeze Brahman, Pang Wei Koh, Simon Shaolei Du, Yulia Tsvetkov, Maryam Fazel, Lin Xiao, Asli Celikyilmaz
Main category: cs.CL
TL;DR: Pep framework decomposes cold-start preference elicitation into offline structure learning and online Bayesian inference to efficiently learn user preferences with minimal interactions.
Details
Motivation: Cold-start personalization faces a routing problem: users care about only a few of many possible preference dimensions, but RL approaches fail to exploit the factored structure of preference data and collapse to static question sequences.Method: Decomposes into offline structure learning (learning preference correlations from complete profiles) and online Bayesian inference (selecting informative questions and predicting complete profiles using training-free Bayesian inference).
Result: Achieves 80.8% alignment vs 68.5% for RL with 3-5x fewer interactions; changes follow-ups 39-62% vs 0-28% for RL when users give different answers; uses ~10K parameters vs 8B for RL.
Conclusion: The bottleneck in cold-start elicitation is exploiting the factored structure of preference data, not model size; Pep’s modular framework with simple belief models outperforms RL approaches.
Abstract: Cold-start personalization requires inferring user preferences through interaction when no user-specific historical data is available. The core challenge is a routing problem: each task admits dozens of preference dimensions, yet individual users care about only a few, and which ones matter depends on who is asking. With a limited question budget, asking without structure will miss the dimensions that matter. Reinforcement learning is the natural formulation, but in multi-turn settings its terminal reward fails to exploit the factored, per-criterion structure of preference data, and in practice learned policies collapse to static question sequences that ignore user responses. We propose decomposing cold-start elicitation into offline structure learning and online Bayesian inference. Pep (Preference Elicitation with Priors) learns a structured world model of preference correlations offline from complete profiles, then performs training-free Bayesian inference online to select informative questions and predict complete preference profiles, including dimensions never asked about. The framework is modular across downstream solvers and requires only simple belief models. Across medical, mathematical, social, and commonsense reasoning, Pep achieves 80.8% alignment between generated responses and users’ stated preferences versus 68.5% for RL, with 3-5x fewer interactions. When two users give different answers to the same question, Pep changes its follow-up 39-62% of the time versus 0-28% for RL. It does so with ~10K parameters versus 8B for RL, showing that the bottleneck in cold-start elicitation is the capability to exploit the factored structure of preference data.
[88] Text Style Transfer with Parameter-efficient LLM Finetuning and Round-trip Translation
Ruoxi Liu, Philipp Koehn
Main category: cs.CL
TL;DR: Parameter-efficient fine-tuning of LLMs for text style transfer using roundtrip translation to create synthetic parallel datasets, with RAG for enhanced terminology consistency.
Details
Motivation: Addresses the scarcity of parallel corpora for text style transfer by creating synthetic parallel datasets, enabling more effective style transfer without requiring large amounts of manually aligned training data.Method: Uses roundtrip translation to synthesize parallel datasets from monolingual corpora, creating ’neutralized’ text as a shared input style. Applies parameter-efficient fine-tuning of LLMs and integrates retrieval-augmented generation (RAG) for terminology and name knowledge.
Result: Consistently outperforms zero-shot prompting and few-shot in-context learning techniques across four domains, achieving higher BLEU scores and style accuracy scores.
Conclusion: The proposed method effectively addresses data scarcity in text style transfer through synthetic parallel data creation and parameter-efficient fine-tuning, with RAG further enhancing robustness and stylistic consistency.
Abstract: This paper proposes a novel method for Text Style Transfer (TST) based on parameter-efficient fine-tuning of Large Language Models (LLMs). Addressing the scarcity of parallel corpora that map between styles, the study employs roundtrip translation to synthesize such parallel datasets from monolingual corpora. This approach creates ’neutralized’ text devoid of stylistic attributes, essentially creating a shared input style at training-time and inference-time. Experimental results demonstrate consistent superiority of this method over zero-shot prompting and fewshot ICL techniques measured by BLEU scores and style accuracy scores across four investigated domains. Furthermore, the integration of retrieval-augmented generation (RAG) for terminology and name knowledge enhances robustness and stylistic consistency.
[89] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models
Sunny Sanyal, Ravid Shwartz-Ziv, Alexandros G. Dimakis, Sujay Sanghavi
Main category: cs.CL
TL;DR: Inheritune addresses attention collapse in LLMs by creating compact models that inherit early layers from larger models, achieving comparable performance with fewer layers.
Details
Motivation: The paper identifies attention collapse in deeper layers of pre-trained LLMs, where attention matrices degenerate to near rank-one structures, creating redundant "lazy layers" that impair model efficiency.Method: Inheritune initializes a compact model by inheriting potent early layers from a larger pre-trained model, then progressively trains and expands it to build smaller, stronger language models.
Result: Experiments on GPT-2 family show models trained with Inheritune can match or surpass performance of larger counterparts despite having significantly fewer layers.
Conclusion: The work presents a novel model compression approach by design, enabling creation of compact yet highly performant language models through inheritance of effective early layers.
Abstract: Large Language Models (LLMs) are known for their performance, but we uncover a significant structural inefficiency: a phenomenon we term attention collapse. In many pre-trained decoder-style LLMs, the attention matrices in deeper layers degenerate, collapsing to near rank-one structures. These underutilized layers, which we call lazy layers, are redundant and impair model efficiency. To address this, we introduce Inheritune, a simple yet powerful training recipe designed to build smaller, stronger language models. Inheritune initializes a compact model by inheriting the potent early layers from a larger pre-trained model and then progressively trains and expands it. Our experiments on various models, including the GPT-2 family, demonstrate that models trained with Inheritune can match or even surpass the performance of their larger counterparts, despite having significantly fewer layers. This work presents a novel path toward model compression by design, enabling the creation of compact, yet highly performant language models. Code is available at https://github.com/sanyalsunny111/LLM-Inheritune.
[90] ViTextVQA: A Large-Scale Visual Question Answering Dataset for Evaluating Vietnamese Text Comprehension in Images
Quan Van Nguyen, Dan Quang Tran, Huy Quang Pham, Thang Kien-Bao Nguyen, Nghia Hieu Nguyen, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen
Main category: cs.CL
TL;DR: ViTextVQA: First large-scale Vietnamese dataset for scene text understanding in VQA, with ViTextBLIP-2 model using OCR integration and multimodal fusion
Details
Motivation: Existing VQA research focuses on object and scene understanding but neglects scene text that carries explicit information. Need for Vietnamese-specific dataset and models for text-based VQA.Method: Created ViTextVQA dataset with 16K+ images and 50K+ questions. Proposed ViTextBLIP-2 model integrating frozen Vision Transformer, SwinTextSpotter OCR, ViT5 LLM with trainable Q-Former for multimodal feature fusion.
Result: Dataset enables research on Vietnamese scene text VQA. Model experiments reveal importance of OCR token processing order for answer formulation, leading to significant performance improvements over baselines.
Conclusion: ViTextVQA fills gap in Vietnamese text-based VQA research. The proposed multimodal fusion approach effectively handles scene text understanding, with OCR token ordering being a key factor for performance.
Abstract: Visual Question Answerinng (VQA) is a complicated task that requires the capability of simultaneously processing natural language and images. This task was initially researched with a focus on developing methods to help machines understand objects and scene contexts in images. However, some scene text that carries explicit information about the full content of the image is not mentioned. Along with the continuous development of the AI era, there have been many studies on the reading comprehension ability of VQA models in the world. Therefore, we introduce the first large-scale dataset in Vietnamese specializing in the ability to understand scene text, we call it ViTextVQA (\textbf{Vi}etnamese \textbf{Text}-based \textbf{V}isual \textbf{Q}uestion \textbf{A}nswering dataset) which contains \textbf{over 16,000} images and \textbf{over 50,000} questions with answers. To tackle this task efficiently, we propose ViTextBLIP-2, an novel multimodal feature fusion Method, which optimizes Vietnamese OCR-based VQA by integrating a frozen Vision Transformer, SwinTextSpotter OCR, and ViT5 LLM with a trainable Q-Former for multimodal feature fusion. Through experiments with various state-of-the-art models, we uncover the significance of the order in which tokens in OCR text are processed and selected to formulate answers. This finding helped us significantly improve the performance of the baseline models on the ViTextVQA dataset. Our dataset is available (https://github.com/minhquan6203/ViTextVQA-Dataset) for research purposes.
[91] Recent Advancements and Challenges of Turkic Central Asian Language Processing
Yana Veitsman, Mareike Hartmann
Main category: cs.CL
TL;DR: Survey paper summarizing recent progress and future directions for NLP research on Central Asian Turkic languages (Kazakh, Uzbek, Kyrgyz, Turkmen), focusing on data scarcity challenges, linguistic features, and transfer learning approaches.
Details
Motivation: Central Asian Turkic languages face typical low-resource language challenges including data scarcity, limited linguistic resources, and underdeveloped technology infrastructure, despite recent advancements in dataset collection and model development.Method: Survey methodology providing high-level overview of each language’s linguistic features, current technology landscape, application of transfer learning from higher-resource languages, and analysis of available labeled/unlabeled data resources.
Result: Comprehensive summary of the current state of NLP research for Central Asian Turkic languages, identifying progress made and outlining specific future research directions to address remaining challenges.
Conclusion: The paper aims to inspire and facilitate future research by providing a clear overview of the current landscape and highlighting opportunities for advancing NLP technology for these under-resourced languages.
Abstract: Research in NLP for Central Asian Turkic languages - Kazakh, Uzbek, Kyrgyz, and Turkmen - faces typical low-resource language challenges like data scarcity, limited linguistic resources and technology development. However, recent advancements have included the collection of language-specific datasets and the development of models for downstream tasks. Thus, this paper aims to summarize recent progress and identify future research directions. It provides a high-level overview of each language’s linguistic features, the current technology landscape, the application of transfer learning from higher-resource languages, and the availability of labeled and unlabeled data. By outlining the current state, we hope to inspire and facilitate future research.
[92] HIPPO: Enhancing the Table Understanding Capability of LLMs through Hybrid-Modal Preference Optimization
Haolan Wang, Zhenghao Liu, Xinze Li, Xiaocui Yang, Yu Gu, Yukun Yan, Qi Shi, Fangfang Li, Chong Chen, Ge Yu
Main category: cs.CL
TL;DR: HIPPO introduces hybrid-modal preference optimization for MLLMs using both text and image table representations to capture structural semantics, achieving 4% improvement on table reasoning tasks.
Details
Motivation: Existing MLLM table research focuses on unimodal representations, limiting exploration of multi-modal representations for more effective table reasoning. Tabular data contains rich structural semantics that could be better captured through hybrid modalities.Method: HIPPO represents tables using both text and image modalities, samples MLLM responses from hybrid-modal representations, and uses a modality-consistent sampling strategy to enhance diversity and mitigate bias during Direct Preference Optimization training.
Result: Experiments on table question answering and table fact verification show 4% improvement over various table reasoning models. HIPPO enhances table reasoning with unimodal representations and facilitates extraction of complementary semantics across modalities.
Conclusion: Hybrid-modal preference optimization effectively captures structural semantics from tabular data, demonstrating that learning from multiple modalities improves MLLM table reasoning capabilities.
Abstract: Tabular data contains rich structural semantics and plays a crucial role in organizing and manipulating information. Recent methods employ Multi-modal Large Language Models (MLLMs) to address table-related tasks across various modalities of table representations. However, existing studies mainly focus on exploring the table understanding ability of MLLMs using unimodal representations, which limits further exploration of multi-modal representations to enable more effective table reasoning. To better capture structural semantics from the tabular data, this paper introduces the HybrId-modal Preference oPtimizatiOn (HIPPO) model, which represents tables using both text and image, optimizing MLLMs by learning more comprehensive table information from these multiple modalities. Specifically, HIPPO samples MLLM responses from hybrid-modal table representations and designs a modality-consistent sampling strategy to enhance response diversity and mitigate modality bias during Direct Preference Optimization (DPO) training. Experiments on table question answering and table fact verification tasks demonstrate the effectiveness of HIPPO, achieving a 4% improvement over various table reasoning models. Further analysis reveals that HIPPO not only enhances the table reasoning capability based on unimodal representations but also facilitates the extraction of complementary semantics across modalities. The code is available at https://github.com/NEUIR/HIPPO.
[93] Steering Dialogue Dynamics for Robustness against Multi-turn Jailbreaking Attacks
Hanjiang Hu, Alexander Robey, Changliu Liu
Main category: cs.CL
TL;DR: Proposes a safety steering framework using neural barrier functions to defend against multi-turn jailbreaking attacks on LLMs by ensuring invariant safety across dialogue turns.
Details
Motivation: Existing defenses against jailbreaking attacks only work for single-turn attacks but fail against multi-turn jailbreaks that exploit contextual drift over multiple interactions to gradually lead LLMs away from safe behavior.Method: Models dialogue with LLMs using state-space representations and introduces a novel neural barrier function (NBF) to detect and filter harmful queries emerging from evolving contexts proactively. Learns a safety predictor that accounts for adversarial queries to prevent context drift toward jailbreaks.
Result: Extensive experiments show NBF-based safety steering outperforms safety alignment, prompt-based steering, and lightweight LLM guardrails baselines, offering stronger defenses against multi-turn jailbreaks while maintaining better trade-off among safety, helpfulness, and over-refusal.
Conclusion: The proposed safety steering framework grounded in safe control theory effectively ensures invariant safety in multi-turn dialogues against jailbreaking attacks.
Abstract: Large language models (LLMs) are shown to be vulnerable to jailbreaking attacks where adversarial prompts are designed to elicit harmful responses. While existing defenses effectively mitigate single-turn attacks by detecting and filtering unsafe inputs, they fail against multi-turn jailbreaks that exploit contextual drift over multiple interactions, gradually leading LLMs away from safe behavior. To address this challenge, we propose a safety steering framework grounded in safe control theory, ensuring invariant safety in multi-turn dialogues. Our approach models the dialogue with LLMs using state-space representations and introduces a novel neural barrier function (NBF) to detect and filter harmful queries emerging from evolving contexts proactively. Our method achieves invariant safety at each turn of dialogue by learning a safety predictor that accounts for adversarial queries, preventing potential context drift toward jailbreaks. Extensive experiments under multiple LLMs show that our NBF-based safety steering outperforms safety alignment, prompt-based steering and lightweight LLM guardrails baselines, offering stronger defenses against multi-turn jailbreaks while maintaining a better trade-off among safety, helpfulness and over-refusal. Check out the website here https://sites.google.com/view/llm-nbf/home.
[94] Benchmarking Retrieval-Augmented Generation for Chemistry
Xianrui Zhong, Bowen Jin, Siru Ouyang, Yanzhen Shen, Qiao Jin, Yin Fang, Zhiyong Lu, Jiawei Han
Main category: cs.CL
TL;DR: ChemRAG-Bench: A comprehensive benchmark and toolkit for evaluating retrieval-augmented generation (RAG) in chemistry domain, showing 17.4% average improvement over direct inference methods.
Details
Motivation: RAG has shown promise for enhancing LLMs with external knowledge, but its application in chemistry remains underexplored due to lack of domain-specific corpora and evaluation benchmarks.Method: Introduces ChemRAG-Bench benchmark with diverse chemistry tasks and integrated knowledge sources (scientific literature, PubChem, PubMed, textbooks, Wikipedia). Develops ChemRAG-Toolkit supporting 5 retrieval algorithms and 8 LLMs.
Result: RAG yields 17.4% average relative improvement over direct inference methods. Provides analysis on retriever architectures, corpus selection, and number of retrieved passages with practical recommendations.
Conclusion: ChemRAG-Bench and Toolkit provide comprehensive resources for evaluating and deploying RAG systems in chemistry, demonstrating significant performance gains and offering guidance for future research.
Abstract: Retrieval-augmented generation (RAG) has emerged as a powerful framework for enhancing large language models (LLMs) with external knowledge, particularly in scientific domains that demand specialized and dynamic information. Despite its promise, the application of RAG in the chemistry domain remains underexplored, primarily due to the lack of high-quality, domain-specific corpora and well-curated evaluation benchmarks. In this work, we introduce ChemRAG-Bench, a comprehensive benchmark designed to systematically assess the effectiveness of RAG across a diverse set of chemistry-related tasks. The accompanying chemistry corpus integrates heterogeneous knowledge sources, including scientific literature, the PubChem database, PubMed abstracts, textbooks, and Wikipedia entries. In addition, we present ChemRAG-Toolkit, a modular and extensible RAG toolkit that supports five retrieval algorithms and eight LLMs. Using ChemRAG-Toolkit, we demonstrate that RAG yields a substantial performance gain – achieving an average relative improvement of 17.4% over direct inference methods. We further conduct in-depth analyses on retriever architectures, corpus selection, and the number of retrieved passages, culminating in practical recommendations to guide future research and deployment of RAG systems in the chemistry domain. The code and data is available at https://chemrag.github.io.
[95] Scalable LLM Reasoning Acceleration with Low-rank Distillation
Harry Dong, Bilge Acun, Beidi Chen, Yuejie Chi
Main category: cs.CL
TL;DR: Caprese is a resource-efficient distillation method that recovers math reasoning capabilities lost when applying efficient inference methods to LLMs, using only 1% additional parameters and synthetic training data.
Details
Motivation: Efficient inference methods for LLMs often severely degrade math reasoning performance while preserving language tasks. There's a need to recover lost math capabilities without harming language performance when using these efficiency techniques.Method: Caprese uses resource-efficient distillation focused primarily on feedforward blocks. It adds roughly 1% of additional parameters, uses only 20K synthetic training samples, and leaves original weights unperturbed. It integrates cleanly into existing model layers to reduce latency.
Result: Recovers much if not all reasoning capabilities lost from efficient inference for thinking LLMs without harming language tasks for instruct LLMs. Reduces active parameters (~2B cut for Gemma 2 9B and Llama 3.1 8B), reduces latency (>16% time-to-next-token reduction), and encourages response brevity (up to 8.5% fewer tokens).
Conclusion: Caprese provides an effective solution to recover math reasoning capabilities compromised by efficient inference methods, offering significant computational benefits while maintaining performance across both reasoning and language tasks.
Abstract: Due to long generations, large language model (LLM) math reasoning demands significant computational resources and time. While many existing efficient inference methods have been developed with excellent performance preservation on language tasks, they often severely degrade math performance. In this paper, we propose Caprese, a resource-efficient distillation method to recover lost capabilities from deploying efficient inference methods, focused primarily in feedforward blocks. With original weights unperturbed, roughly 1% of additional parameters, and only 20K synthetic training samples, we are able to recover much if not all of the reasoning capabilities lost from efficient inference for thinking LLMs and without harm to language tasks for instruct LLMs. Moreover, Caprese slashes the number of active parameters (~2B cut for Gemma 2 9B and Llama 3.1 8B) and integrates cleanly into existing model layers to reduce latency (>16% time-to-next-token reduction) while encouraging response brevity (up to 8.5% fewer tokens).
[96] RedTeamCUA: Realistic Adversarial Testing of Computer-Use Agents in Hybrid Web-OS Environments
Zeyi Liao, Jaylen Jones, Linxi Jiang, Yuting Ning, Eric Fosler-Lussier, Yu Su, Zhiqiang Lin, Huan Sun
Main category: cs.CL
TL;DR: RedTeamCUA is a framework for testing computer-use agents against indirect prompt injection attacks in hybrid web-OS environments, revealing significant vulnerabilities in current CUAs.
Details
Motivation: Current evaluations of computer-use agent vulnerabilities lack realistic hybrid web-OS attack scenarios and controlled testing environments, creating a gap in understanding real-world security threats.Method: Proposes RedTeamCUA framework with a novel hybrid sandbox combining VM-based OS environment with Docker-based web platforms, enabling flexible adversarial scenario configuration and direct initialization at injection points.
Result: Benchmarking reveals significant vulnerabilities: Claude 3.7 Sonnet shows 42.9% attack success rate, Claude 4.5 Sonnet shows 60% ASR, and even the most secure CUA (Operator) has 7.6% ASR. Attempt rates reach up to 92.5%.
Conclusion: RedTeamCUA provides essential framework for systematic CUA vulnerability analysis, highlighting urgent need for robust defenses against indirect prompt injection before real-world deployment.
Abstract: Computer-use agents (CUAs) promise to automate complex tasks across operating systems (OS) and the web, but remain vulnerable to indirect prompt injection. Current evaluations of this threat either lack support realistic but controlled environments or ignore hybrid web-OS attack scenarios involving both interfaces. To address this, we propose RedTeamCUA, an adversarial testing framework featuring a novel hybrid sandbox that integrates a VM-based OS environment with Docker-based web platforms. Our sandbox supports key features tailored for red teaming, such as flexible adversarial scenario configuration, and a setting that decouples adversarial evaluation from navigational limitations of CUAs by initializing tests directly at the point of an adversarial injection. Using RedTeamCUA, we develop RTC-Bench, a comprehensive benchmark with 864 examples that investigate realistic, hybrid web-OS attack scenarios and fundamental security vulnerabilities. Benchmarking current frontier CUAs identifies significant vulnerabilities: Claude 3.7 Sonnet | CUA demonstrates an ASR of 42.9%, while Operator, the most secure CUA evaluated, still exhibits an ASR of 7.6%. Notably, CUAs often attempt to execute adversarial tasks with an Attempt Rate as high as 92.5%, although failing to complete them due to capability limitations. Nevertheless, we observe concerning high ASRs in realistic end-to-end settings, with the strongest-to-date Claude 4.5 Sonnet | CUA exhibiting the highest ASR of 60%, indicating that CUA threats can already result in tangible risks to users and computer systems. Overall, RedTeamCUA provides an essential framework for advancing realistic, controlled, and systematic analysis of CUA vulnerabilities, highlighting the urgent need for robust defenses to indirect prompt injection prior to real-world deployment.
[97] Beyond Memorization: A Rigorous Evaluation Framework for Medical Knowledge Editing
Shigeng Chen, Linhao Luo, Zhangchi Qiu, Yanan Cao, Carl Yang, Shirui Pan
Main category: cs.CL
TL;DR: MedEditBench evaluates knowledge editing methods for LLMs in medical domain, finds current methods only achieve superficial memorization, proposes SGR-Edit using model-derived rationales for better generalization.
Details
Motivation: Knowledge editing methods for LLMs have shown promise for updating facts without full retraining, but their effectiveness in complex medical domains remains unexplored. Medical knowledge editing is particularly challenging as it requires LLMs to internalize knowledge and generalize to unseen scenarios for interpretable decision-making.Method: Proposes MedEditBench framework with: 1) New medical knowledge editing benchmark, 2) Three different knowledge editing paradigms to assess impact of different knowledge sources, 3) Self-Generated Rationale Editing (SGR-Edit) that uses model-derived rationales as target knowledge for editing to uncover underlying reasoning process.
Result: Current KE methods result in only superficial memorization of injected information, failing to generalize to new scenarios. SGR-Edit demonstrates significant improvements over existing KE approaches. Provides insights into medical knowledge localization in LLMs and impact of sequential editing on evolving knowledge.
Conclusion: Medical knowledge editing requires more sophisticated approaches than current methods. SGR-Edit offers better generalization by leveraging model-derived rationales. The work provides practical guidance for implementing KE methods in real-world medical applications.
Abstract: Recently, knowledge editing (KE) has emerged as a promising approach to update specific facts in Large Language Models (LLMs) without the need for full retraining. Despite the effectiveness in general-domain benchmarks, their applicability to complex medical domain remains largely unexplored. Medical knowledge editing is particularly challenging, as it requires LLMs to internalize the knowledge and generalize to unseen scenarios for effective and interpretable decision-making. In this work, we propose a novel framework called MedEditBench to rigorously evaluate the effectiveness of existing KE methods in the medical domain. In MedEditBench, we introduce a new medical knowledge editing benchmark as well as three different knowledge editing paradigms, which are designed to assess the impact of different knowledge sources for editing. Our findings indicate that current KE methods result in only superficial memorization of the injected information, failing to generalize to new scenarios. To overcome this limitation, we present Self-Generated Rationale Editing (SGR-Edit), which utilizes model-derived rationales as the target knowledge for editing, thereby uncovering the underlying reasoning process and demonstrating significant improvements over existing KE approaches. Additionally, we offer deeper insights into medical knowledge editing, including the localization of medical knowledge in LLMs and the impact of sequential editing on evolving knowledge. This could provide practical guidance for implementing KE methods in real-world medical applications.
[98] High Accuracy, Less Talk (HALT): Reliable LLMs through Capability-Aligned Finetuning
Tim Franzmeyer, Archie Sravankumar, Lijuan Liu, Yuning Mao, Rui Hou, Sinong Wang, Jakob N. Foerster, Luke Zettlemoyer, Madian Khabsa
Main category: cs.CL
TL;DR: HALT: A method to post-train LLMs to abstain from answering when uncertain, improving correctness by selectively removing or marking unreliable content fragments.
Details
Motivation: LLMs often hallucinate when they lack knowledge or capability. Current models respond to every prompt even when uncertain, leading to incorrect answers. The authors propose training models to recognize their limitations and abstain from generating content when not confident.Method: HALT generates capability-aligned post-training data by splitting LLM responses into factual fragments (atomic statements or reasoning steps), using ground truth to identify incorrect fragments. During finetuning, incorrect fragments are either removed or replaced with “Unsure from Here” based on a tunable threshold that trades off response completeness vs correctness.
Result: HALT increases mean correctness of response fragments by 15% on average while improving F1 score (mean of completeness and correctness) by 4% compared to baselines. For Llama3-70B, correctness increased from 51% to 87% across four domains (biography, math, coding, medicine) while maintaining 53% of response completeness.
Conclusion: HALT effectively enables LLMs to recognize their limitations and abstain from generating unreliable content, significantly improving correctness while maintaining reasonable response completeness through tunable thresholds.
Abstract: Large Language Models (LLMs) currently respond to every prompt. However, they can produce incorrect answers when they lack knowledge or capability – a problem known as hallucination. We instead propose post-training an LLM to generate content only when confident in its correctness and to otherwise (partially) abstain. Specifically, our method, HALT, produces capability-aligned post-training data that encodes what the model can and cannot reliably generate. We generate this data by splitting responses of the pretrained LLM into factual fragments (atomic statements or reasoning steps), and use ground truth information to identify incorrect fragments. We achieve capability-aligned finetuning responses by either removing incorrect fragments or replacing them with “Unsure from Here” – according to a tunable threshold that allows practitioners to trade off response completeness and mean correctness of the response’s fragments. We finetune four open-source models for biography writing, mathematics, coding, and medicine with HALT for three different trade-off thresholds. HALT effectively trades off response completeness for correctness, increasing the mean correctness of response fragments by 15% on average, while resulting in a 4% improvement in the F1 score (mean of completeness and correctness of the response) compared to the relevant baselines. By tuning HALT for highest correctness, we train a single reliable Llama3-70B model with correctness increased from 51% to 87% across all four domains while maintaining 53% of the response completeness achieved with standard finetuning.
[99] Offline RL by Reward-Weighted Fine-Tuning for Conversation Optimization
Subhojyoti Mukherjee, Viet Dac Lai, Raghavendra Addanki, Ryan Rossi, Seunghyun Yoon, Trung Bui, Anup Rao, Jayakumar Subramanian, Branislav Kveton
Main category: cs.CL
TL;DR: Offline RL approach for LLMs using reward-weighted fine-tuning, applied to short-horizon QA tasks with significant improvements over SFT and DPO methods.
Details
Motivation: Existing offline RL methods for LLMs (like SFT and DPO) have additional hyperparameters and don't directly optimize for rewards. The authors aim to develop a practical offline RL approach that directly optimizes rewards while maintaining language quality.Method: Proposes reward-weighted fine-tuning as a practical offline RL approach for LLMs, treating it as a supervised fine-tuning problem where trajectories are weighted by their rewards. Applied to short-horizon question-answering policies of fixed length.
Result: Major gains in both optimized rewards and language quality compared to state-of-the-art methods based on SFT and direct preference optimization (DPO).
Conclusion: Reward-weighted fine-tuning provides an effective practical approach for offline RL with LLMs that directly optimizes rewards while maintaining language quality, outperforming existing methods.
Abstract: Offline reinforcement learning (RL) is a variant of RL where the policy is learned from a previously collected dataset of trajectories and rewards. In our work, we propose a practical approach to offline RL with large language models (LLMs). We recast the problem as reward-weighted fine-tuning, which can be solved using similar techniques to supervised fine-tuning (SFT). To showcase the value of our approach, we apply it to learning short-horizon question-answering policies of a fixed length, where the agent reasons about potential answers or asks clarifying questions. Our work stands in a stark contrast to state-of-the-art methods in this domain, based on SFT and direct preference optimization, which have additional hyper-parameters and do not directly optimize for rewards. We compare to them empirically, and report major gains in both optimized rewards and language quality.
[100] RuleReasoner: Reinforced Rule-based Reasoning via Domain-aware Dynamic Sampling
Yang Liu, Jiaqi Li, Zilong Zheng
Main category: cs.CL
TL;DR: RuleReasoner improves rule-based reasoning in large reasoning models using a novel domain-aware dynamic sampling approach in reinforcement learning, outperforming state-of-the-art models on both in-distribution and out-of-distribution tasks.
Details
Motivation: Real-world applications of large reasoning models face challenges due to variations in rule formats, types, and complexity, despite recent advances in reinforcement learning-enhanced reasoning capabilities.Method: RuleReasoner uses a curated collection of rule-based reasoning tasks and a novel domain-aware dynamic sampling approach in RL that resamples training batches by updating domain weights based on historical rewards, facilitating domain balance and active learning schedules.
Result: RuleReasoner outperforms frontier large reasoning models by significant margins: Δ4.1% on eight in-distribution tasks and Δ10.4% on three out-of-distribution tasks over OpenAI-o1, while also exhibiting higher computational efficiency.
Conclusion: The proposed RuleReasoner method effectively addresses rule-based reasoning challenges through dynamic sampling in RL, achieving superior performance and efficiency compared to existing approaches.
Abstract: Rule-based reasoning is acknowledged as one of the fundamental problems of reasoning. While recent studies show that large reasoning models (LRMs) have remarkable reasoning capabilities enhanced by reinforcement learning (RL), real applications still face severe challenges due to variations in rule formats, types, and complexity. To mitigate this issue, we introduce RuleReasoner, an effective method for rule-based reasoning via a wide collection of curated tasks and a novel domain-aware dynamic sampling approach in RL. Specifically, RuleReasoner resamples each training batch by updating the domain weights based on historical rewards. This facilitates domain balance and active learning schedules for RL, obviating static mix-training engineered by human. Evaluations on in-distribution (ID) and out-of-distribution (OOD) benchmarks reveal that RuleReasoner outperforms frontier LRMs by a significant margin ($Δ$4.1% on eight ID tasks and $Δ$10.4% on three OOD tasks over OpenAI-o1). Notably, our approach also exhibits higher computational efficiency compared to prior methods.
[101] PersonalAI: A Systematic Comparison of Knowledge Graph Storage and Retrieval Approaches for Personalized LLM agents
Mikhail Menschikov, Dmitry Evseev, Victoria Dochkina, Ruslan Kostoev, Ilia Perepechkin, Petr Anokhin, Evgeny Burnaev, Nikita Semenov
Main category: cs.CL
TL;DR: A knowledge graph-based external memory framework for LLMs that enables structured memory and dynamic updates through hybrid graph design with hyper-edges, supporting diverse retrieval mechanisms for improved long-term interaction handling.
Details
Motivation: Current LLMs with RAG lack structured memory and fail to scale in complex, long-term interactions, needing better personalization through user interaction history incorporation.Method: Proposes a flexible external memory framework based on knowledge graphs that constructs and updates memory automatically using LLMs. Uses AriGraph architecture with hybrid graph design supporting standard edges and two types of hyper-edges for semantic and temporal representations. Supports diverse retrieval mechanisms including A*, water-circle traversal, beam search, and hybrid methods.
Result: Evaluated on TriviaQA, HotpotQA, and DiaASQ benchmarks, showing different memory and retrieval configurations yield optimal performance depending on task. Extended DiaASQ with temporal annotations and contradictory statements, demonstrating system robustness in managing temporal dependencies and context-aware reasoning.
Conclusion: The knowledge graph-based memory framework effectively addresses LLM limitations in structured memory and long-term interaction scaling, providing adaptable solutions for different tasks through configurable memory and retrieval mechanisms.
Abstract: Personalizing language models that effectively incorporating user interaction history remains a central challenge in development of adaptive AI systems. While large language models (LLMs), combined with Retrieval-Augmented Generation (RAG), have improved factual accuracy, they often lack structured memory and fail to scale in complex, long-term interactions. To address this, we propose a flexible external memory framework based on knowledge graph, which construct and update memory model automatically by LLM itself. Building upon the AriGraph architecture, we introduce a novel hybrid graph design that supports both standard edges and two types of hyper-edges, enabling rich and dynamic semantic and temporal representations. Our framework also supports diverse retrieval mechanisms, including A*, water-circle traversal, beam search and hybrid methods, making it adaptable to different datasets and LLM capacities. We evaluate our system on three benchmarks: TriviaQA, HotpotQA, DiaASQ and demonstrate that different memory and retrieval configurations yield optimal performance depending on the task. Additionally, we extend the DiaASQ benchmark with temporal annotations and internally contradictory statements, showing that our system remains robust and effective in managing temporal dependencies and context-aware reasoning.
[102] From Fragments to Facts: A Curriculum-Driven DPO Approach for Generating Hindi News Veracity Explanations
Pulkit Bansal, Raghvendra Kumar, Shakti Singh, Sriparna Saha, Adam Jatowt
Main category: cs.CL
TL;DR: A novel framework combining Direct Preference Optimization with curriculum learning to generate reliable Hindi news explanations, addressing misinformation in low-resource languages.
Details
Motivation: To combat misinformation in under-represented languages like Hindi by developing automated tools for generating reliable news explanations, addressing the lack of scalable misinformation detection systems for low-resource languages.Method: Integrates Direct Preference Optimization (DPO) with curriculum learning, using fact-checked explanations as preferred responses and LLM outputs as non-preferred responses. Introduces two parameters (Actuality and Finesse) into DPO loss function to enhance explanation quality.
Result: Experiments with various LLMs (Mistral, Llama, Gemma) and PLMs (mBART, mT5) confirm the framework’s effectiveness in generating coherent, contextually relevant explanations for Hindi news.
Conclusion: The proposed scalable approach successfully combats misinformation and extends automated explanation generation capabilities to low-resource languages, bridging the gap in misinformation detection tools.
Abstract: In an era of rampant misinformation, generating reliable news explanations is vital, especially for under-represented languages like Hindi. Lacking robust automated tools, Hindi faces challenges in scaling misinformation detection. To bridge this gap, we propose a novel framework integrating Direct Preference Optimization (DPO) with curriculum learning to align machine-generated explanations with human reasoning. Fact-checked explanations from credible sources serve as preferred responses, while LLM outputs highlight system limitations and serve as non-preferred responses. To refine task-specific alignment, we introduce two key parameters – Actuality and Finesse – into the DPO loss function, enhancing explanation quality and consistency. Experiments with LLMs (Mistral, Llama, Gemma) and PLMs (mBART, mT5) confirm the framework’s effectiveness in generating coherent, contextually relevant explanations. This scalable approach combats misinformation and extends automated explanation generation to low-resource languages.
[103] GEPA: Reflective Prompt Evolution Can Outperform Reinforcement Learning
Lakshya A Agrawal, Shangyin Tan, Dilara Soylu, Noah Ziems, Rishi Khare, Krista Opsahl-Ong, Arnav Singhvi, Herumb Shandilya, Michael J Ryan, Meng Jiang, Christopher Potts, Koushik Sen, Alexandros G. Dimakis, Ion Stoica, Dan Klein, Matei Zaharia, Omar Khattab
Main category: cs.CL
TL;DR: GEPA is a prompt optimization method that uses natural language reflection to learn high-level rules from trial and error, outperforming RL methods like GRPO with far fewer rollouts.
Details
Motivation: Current RL methods for adapting LLMs to downstream tasks require thousands of rollouts, but language provides a richer learning medium than sparse scalar rewards. The authors aim to leverage natural language reflection for more efficient prompt optimization.Method: GEPA samples trajectories (reasoning, tool calls, outputs), reflects on them in natural language to diagnose problems, proposes and tests prompt updates, and combines lessons from the Pareto frontier of attempts. It uses genetic-Pareto optimization to efficiently learn from few rollouts.
Result: GEPA outperforms GRPO by 6% on average (up to 20%) while using up to 35x fewer rollouts. It also beats leading prompt optimizer MIPROv2 by over 10% (e.g., +12% accuracy on AIME-2025). Shows promise as inference-time search for code optimization.
Conclusion: Natural language reflection enables efficient prompt optimization with far fewer rollouts than RL methods, demonstrating the power of leveraging LLMs’ interpretable nature for learning.
Abstract: Large language models (LLMs) are increasingly adapted to downstream tasks via reinforcement learning (RL) methods like Group Relative Policy Optimization (GRPO), which often require thousands of rollouts to learn new tasks. We argue that the interpretable nature of language often provides a much richer learning medium for LLMs, compared to policy gradients derived from sparse, scalar rewards. To test this, we introduce GEPA (Genetic-Pareto), a prompt optimizer that thoroughly incorporates natural language reflection to learn high-level rules from trial and error. Given any AI system containing one or more LLM prompts, GEPA samples trajectories (e.g., reasoning, tool calls, and tool outputs) and reflects on them in natural language to diagnose problems, propose and test prompt updates, and combine complementary lessons from the Pareto frontier of its own attempts. As a result of GEPA’s design, it can often turn even just a few rollouts into a large quality gain. Across six tasks, GEPA outperforms GRPO by 6% on average and by up to 20%, while using up to 35x fewer rollouts. GEPA also outperforms the leading prompt optimizer, MIPROv2, by over 10% (e.g., +12% accuracy on AIME-2025), and demonstrates promising results as an inference-time search strategy for code optimization. We release our code at https://github.com/gepa-ai/gepa .
[104] RoD-TAL: A Benchmark for Answering Questions in Romanian Driving License Exams
Andrei Vlad Man, Răzvan-Alexandru Smădu, Cristian-George Craciun, Dumitru-Clementin Cercel, Florin Pop, Mihaela-Claudia Cercel
Main category: cs.CL
TL;DR: Evaluation of LLMs and VLMs on Romanian driving law understanding using a new multimodal dataset (RoD-TAL), showing domain-specific fine-tuning and reasoning techniques improve performance above passing exam grades.
Details
Motivation: Address the need for AI tools in legal education for under-resourced languages like Romanian, specifically focusing on driving law understanding through multimodal (text and visual) question-answering tasks.Method: Created RoD-TAL dataset with Romanian driving test questions (text and image-based) with expert annotations. Evaluated retrieval-augmented generation (RAG) pipelines, dense retrievers, and reasoning-optimized models on Information Retrieval, Question Answering, Visual IR, and Visual QA tasks.
Result: Domain-specific fine-tuning significantly improves retrieval performance. Chain-of-thought prompting and specialized reasoning models enhance QA accuracy, surpassing minimum passing grades required for driving exams.
Conclusion: LLMs and VLMs show potential for legal education applications but have limitations. The work provides resources for Romanian language legal AI applications and demonstrates effective multimodal reasoning approaches.
Abstract: The intersection of AI and legal systems presents a growing need for tools that support legal education, particularly in under-resourced languages such as Romanian. In this work, we aim to evaluate the capabilities of Large Language Models (LLMs) and Vision-Language Models (VLMs) in understanding and reasoning about the Romanian driving law through textual and visual question-answering tasks. To facilitate this, we introduce RoD-TAL, a novel multimodal dataset comprising Romanian driving test questions, text-based and image-based, along with annotated legal references and explanations written by human experts. We implement and assess retrieval-augmented generation (RAG) pipelines, dense retrievers, and reasoning-optimized models across tasks, including Information Retrieval (IR), Question Answering (QA), Visual IR, and Visual QA. Our experiments demonstrate that domain-specific fine-tuning significantly enhances retrieval performance. At the same time, chain-of-thought prompting and specialized reasoning models improve QA accuracy, surpassing the minimum passing grades required for driving exams. We highlight the potential and limitations of applying LLMs and VLMs to legal education. We release the code and resources through the GitHub repository.
[105] Why Synthetic Isn’t Real Yet: A Diagnostic Framework for Contact Center Dialogue Generation
Rishikesh Devanathan, Varun Nathan, Ayush Kumar
Main category: cs.CL
TL;DR: Benchmarking synthetic dialogue generation for contact centers using structured supervision, revealing gaps in realism and downstream utility compared to real conversations.
Details
Motivation: Contact centers face data scarcity due to privacy constraints, requiring synthetic dialogues for training and evaluation, but current synthetic data lacks realism for downstream applications.Method: Benchmarked multiple generation strategies guided by structured supervision (Intent Summaries, Topic Flows, QA Forms) across languages, evaluated on AutoQA task, and introduced diagnostic framework with 17 metrics across four dimensions: Emotional/Sentiment Arcs, Linguistic Complexity, Interaction Style, and Conversational Properties.
Result: Prompts optimized on real transcripts consistently outperform those optimized on synthetic transcripts; current synthetic data shows deficiencies in sentiment fidelity, disfluency modeling, behavioral variation, and conversational realism despite structured supervision.
Conclusion: Diagnostic, metric-driven evaluation is crucial for synthetic conversation generation intended for downstream applications, as current methods still fall short of capturing full realism of real agent-customer interactions.
Abstract: Synthetic data is increasingly critical for contact centers, where privacy constraints and data scarcity limit the availability of real conversations. However, generating synthetic dialogues that are realistic and useful for downstream applications remains challenging. In this work, we benchmark multiple generation strategies guided by structured supervision on call attributes (Intent Summaries, Topic Flows, and Quality Assurance (QA) Forms) across multiple languages. To test downstream utility, we evaluate synthetic transcripts on an automated quality assurance (AutoQA) task, finding that prompts optimized on real transcripts consistently outperform those optimized on synthetic transcripts. These results suggest that current synthetic transcripts fall short in capturing the full realism of real agent-customer interactions. To highlight these downstream gaps, we introduce a diagnostic evaluation framework comprising 17 metrics across four dimensions: (1) Emotional and Sentiment Arcs, (2) Linguistic Complexity, (3) Interaction Style, and (4) Conversational Properties. Our analysis shows that even with structured supervision, current generation strategies exhibit measurable deficiencies in sentiment fidelity, disfluency modeling, behavioral variation, and conversational realism. Together, these results highlight the importance of diagnostic, metric-driven evaluation for synthetic conversation generation intended for downstream applications.
[106] Implicit Actor Critic Coupling via a Supervised Learning Framework for RLVR
Jiaming Li, Longze Chen, Ze Gong, Yukun Chen, Lu Wang, Wanwei He, Run Luo, Min Yang
Main category: cs.CL
TL;DR: PACS is a novel RLVR framework that reformulates reinforcement learning with verifiable rewards as a supervised learning task, achieving implicit actor-critic coupling for more stable and efficient training of LLMs on reasoning tasks.
Details
Motivation: Existing RLVR methods suffer from sparse reward signals and unstable policy gradient updates, which limit their effectiveness in training LLMs for challenging reasoning tasks like mathematics and programming.Method: PACS treats outcome rewards as predictable labels and reformulates RLVR as a supervised learning task over a score function parameterized by the policy model, optimized using cross-entropy loss instead of traditional policy gradient methods.
Result: PACS significantly outperforms strong open-source models and RLVR baselines, achieving substantial average gains of +8.26% (4B) and +9.57% (8B) over base models on reasoning tasks.
Conclusion: PACS offers a promising avenue for LLM post-training with verifiable rewards by providing more stable and efficient training through its supervised learning reformulation of the RLVR problem.
Abstract: Recent advances in Reinforcement Learning with Verifiable Rewards (RLVR) have empowered large language models (LLMs) to tackle challenging reasoning tasks such as mathematics and programming. Despite its promise, the RLVR paradigm poses significant challenges, as existing methods often suffer from sparse reward signals and unstable policy gradient updates, inherent to RL-based approaches. To address the challenges, we propose $\textbf{PACS}$, a novel RLVR framework that achieves im$\textbf{P}$licit $\textbf{A}$ctor $\textbf{C}$ritic coupling via a $\textbf{S}$upervised learning framework. By treating the outcome reward as a predictable label, we reformulate the RLVR problem into a supervised learning task over a score function parameterized by the policy model and optimized using cross-entropy loss. A detailed gradient analysis shows that this supervised formulation inherently recovers the classical policy gradient update while providing more stable and efficient training. Extensive experiments demonstrate that PACS significantly outperforms strong open-source models and RLVR baselines, yielding substantial average gains of $\textbf{+8.26%}$ (4B) and $\textbf{+9.57%}$ (8B) over base models offering a promising avenue for LLMs post-training with verifiable rewards. Our code and data are available as open source at https://github.com/ritzz-ai/PACS.
[107] Evolution of Concepts in Language Model Pre-Training
Xuyang Ge, Wentao Shu, Jiaxing Wu, Yunhua Zhou, Zhengfu He, Xipeng Qiu
Main category: cs.CL
TL;DR: Researchers use sparse dictionary learning (crosscoders) to track interpretable feature evolution during language model pre-training, identifying two distinct learning phases and causal connections between feature development and downstream performance.
Details
Motivation: To demystify the black box of language model pre-training by understanding how interpretable features evolve throughout the training process, moving beyond just analyzing final model capabilities.Method: Use crosscoders (sparse dictionary learning method) to track linear interpretable features across pre-training snapshots, analyzing feature formation timing and complexity development, with feature attribution analyses to establish causal connections.
Result: Most features form around specific points in training, with complex patterns emerging later; feature evolution shows causal connections to downstream performance; findings align with Transformer’s two-stage learning process (statistical learning phase followed by feature learning phase).
Conclusion: The work enables fine-grained tracking of representation progress during language model learning dynamics, providing insights into the previously opaque pre-training process and opening new research directions in model interpretability.
Abstract: Language models obtain extensive capabilities through pre-training. However, the pre-training process remains a black box. In this work, we track linear interpretable feature evolution across pre-training snapshots using a sparse dictionary learning method called crosscoders. We find that most features begin to form around a specific point, while more complex patterns emerge in later training stages. Feature attribution analyses reveal causal connections between feature evolution and downstream performance. Our feature-level observations are highly consistent with previous findings on Transformer’s two-stage learning process, which we term a statistical learning phase and a feature learning phase. Our work opens up the possibility to track fine-grained representation progress during language model learning dynamics.
[108] AECBench: A Hierarchical Benchmark for Knowledge Evaluation of Large Language Models in the AEC Field
Chen Liang, Zhaoqi Huang, Haofen Wang, Fu Chai, Chunying Yu, Huanhuan Wei, Zhengjie Liu, Yanpeng Li, Hongjun Wang, Ruifeng Luo, Xianzhong Zhao
Main category: cs.CL
TL;DR: AECBench: A comprehensive benchmark for evaluating LLMs in Architecture, Engineering, and Construction with a 5-level cognition framework and 4,800-question dataset.
Details
Motivation: LLMs are increasingly used in safety-critical AEC domains, but their robustness and reliability need systematic evaluation to ensure safe integration into engineering practices.Method: Created AECBench with 5-level cognition framework (Knowledge Memorization, Understanding, Reasoning, Calculation, Application), 23 tasks from authentic AEC practice, 4,800-question dataset crafted by engineers, and LLM-as-a-Judge evaluation approach.
Result: Evaluation of 9 LLMs showed clear performance decline across cognitive levels - proficient in foundational tasks but significant deficits in interpreting tables, complex reasoning/calculation, and generating domain-specific documents.
Conclusion: The benchmark establishes groundwork for robust LLM integration into safety-critical engineering, revealing current limitations and guiding future development.
Abstract: Large language models (LLMs), as a novel information technology, are seeing increasing adoption in the Architecture, Engineering, and Construction (AEC) field. They have shown their potential to streamline processes throughout the building lifecycle. However, the robustness and reliability of LLMs in such a specialized and safety-critical domain remain to be evaluated. To address this challenge, this paper establishes AECBench, a comprehensive benchmark designed to quantify the strengths and limitations of current LLMs in the AEC domain. The benchmark features a five-level, cognition-oriented evaluation framework (i.e., Knowledge Memorization, Understanding, Reasoning, Calculation, and Application). Based on the framework, 23 representative evaluation tasks were defined. These tasks were derived from authentic AEC practice, with scope ranging from codes retrieval to specialized documents generation. Subsequently, a 4,800-question dataset encompassing diverse formats, including open-ended questions, was crafted primarily by engineers and validated through a two-round expert review. Furthermore, an “LLM-as-a-Judge” approach was introduced to provide a scalable and consistent methodology for evaluating complex, long-form responses leveraging expert-derived rubrics. Through the evaluation of nine LLMs, a clear performance decline across five cognitive levels was revealed. Despite demonstrating proficiency in foundational tasks at the Knowledge Memorization and Understanding levels, the models showed significant performance deficits, particularly in interpreting knowledge from tables in building codes, executing complex reasoning and calculation, and generating domain-specific documents. Consequently, this study lays the groundwork for future research and development aimed at the robust and reliable integration of LLMs into safety-critical engineering practices.
[109] d$^2$Cache: Accelerating Diffusion-Based LLMs via Dual Adaptive Caching
Yuchu Jiang, Yue Cai, Xiangzhong Luo, Jiale Fu, Jiarui Wang, Chonghan Liu, Xu Yang
Main category: cs.CL
TL;DR: Training-free KV cache framework (d²Cache) for diffusion-based LLMs that accelerates inference while improving generation quality through adaptive token selection and caching.
Details
Motivation: Diffusion-based LLMs suffer from inferior inference efficiency compared to autoregressive models because they rely on bidirectional attention and cannot directly benefit from standard KV cache mechanisms.Method: Proposes Dual adaptive Cache (d²Cache) with two-stage fine-grained selection strategy: identifies tokens to update KV states adaptively at each decoding step while caching remaining tokens’ KV states for reuse.
Result: Achieves substantial inference speedups on LLaDA and Dream models while yielding consistent improvements in generation quality, plus enables quasi left-to-right generation and mitigates premature overconfidence.
Conclusion: d²Cache provides effective training-free solution for accelerating diffusion-based LLM inference while enhancing generation quality through adaptive caching strategy.
Abstract: Diffusion-based large language models (dLLMs), despite their promising performance, still suffer from inferior inference efficiency. This is because dLLMs rely on bidirectional attention and cannot directly benefit from the standard key-value (KV) cache as autoregressive models (ARMs) do. To tackle this issue, we introduce \textit{Dual aDaptive Cache} (d$^2$Cache), which is a training-free approximate KV cache framework for accelerating dLLM inference. d$^2$Cache features a two-stage fine-grained selection strategy to identify tokens and adaptively update their KV states at each decoding step, while caching the KV states of the remaining tokens for reuse. Furthermore, d$^2$Cache naturally offers a more reliable decoding alternative, which can enable quasi left-to-right generation and mitigate premature overconfidence in tokens at the end of the sequence. Extensive experimental results on two representative dLLMs (\ie, LLaDA and Dream) demonstrate that d$^2$Cache not only achieves substantial inference speedups, but also yields consistent improvements in generation quality. The code is available at https://github.com/Kamichanw/d2Cache.
[110] Pragmatic Inference for Moral Reasoning Acquisition: Generalization via Metapragmatic Links
Guangliang Liu, Xi Chen, Bocheng Chen, Xitong Zhang, Kristen Johnson
Main category: cs.CL
TL;DR: A pragmatic-inference approach using metapragmatic links and moral foundations theory to improve LLMs’ moral reasoning generalization by bridging the gap between explicit statements and moral implications.
Details
Motivation: Moral reasoning in LLMs faces a critical generalization challenge due to the gap between what is explicitly stated and what is morally implied. Current approaches struggle with robust generalization across different moral situations.Method: Developed a pragmatic-inference approach based on metapragmatic links and moral foundations theory. The method helps LLMs acquire links between moral reasoning objectives and relevant social variables for given moral situations, adapted to three different moral reasoning tasks.
Result: Experimental results show the approach significantly enhances LLMs’ generalization in moral reasoning, demonstrating adaptability across different tasks.
Conclusion: The pragmatic-inference approach successfully bridges the gap in moral reasoning and paves the way for future research to utilize pragmatic inference in various moral reasoning tasks.
Abstract: While moral reasoning has emerged as a promising research direction for large language models (LLMs), achieving robust generalization remains a critical challenge. This challenge arises from the gap between what is said and what is morally implied. In this paper, we build on metapragmatic links and the moral foundations theory to close the gap. Specifically, we develop a pragmatic-inference approach that facilitates LLMs, for a given moral situation, to acquire the metapragmantic links between moral reasoning objectives and the social variables that affect them. This approach is adapted to three different moral reasoning tasks to demonstrate its adaptability and generalizability. Experimental results demonstrate that our approach significantly enhances LLMs’ generalization in moral reasoning, paving the road for future research to utilize pragmatic inference in various moral reasoning tasks.
[111] BiasFreeBench: a Benchmark for Mitigating Bias in Large Language Model Responses
Xin Xu, Xunzhi He, Churan Zhi, Ruizhe Chen, Julian McAuley, Zexue He
Main category: cs.CL
TL;DR: BiasFreeBench: A unified benchmark for evaluating bias mitigation methods in LLMs using response-level metrics and real-world interaction scenarios
Details
Motivation: Current bias mitigation evaluations are inconsistent due to diverse baselines and metrics, and focus on probability comparisons rather than real-world user interactions where people read model responses and expect fair outputsMethod: Introduces BiasFreeBench benchmark comparing 8 mainstream bias mitigation techniques (4 prompting-based, 4 training-based) on two test scenarios (multi-choice QA and open-ended multi-turn QA) by reorganizing existing datasets into unified query-response setting
Result: Systematic comparison of debiasing performance across key dimensions: prompting vs. training paradigm, model size, and generalization of different training strategies to unseen bias types
Conclusion: Provides a unified testbed for bias mitigation research with response-level evaluation metrics that better reflect real-world use cases
Abstract: Existing studies on bias mitigation methods for large language models (LLMs) use diverse baselines and metrics to evaluate debiasing performance, leading to inconsistent comparisons among them. Moreover, their evaluations are mostly based on the comparison between LLMs’ probabilities of biased and unbiased contexts, which ignores the gap between such evaluations and real-world use cases where users interact with LLMs by reading model responses and expect fair and safe outputs rather than LLMs’ probabilities. To enable consistent evaluation across debiasing methods and bridge this gap, we introduce BiasFreeBench, an empirical benchmark that comprehensively compares eight mainstream bias mitigation techniques (covering four prompting-based and four training-based methods) on two test scenarios (multi-choice QA and open-ended multi-turn QA) by reorganizing existing datasets into a unified query-response setting. We further introduce a response-level metric, Bias-Free Score, to measure the extent to which LLM responses are fair, safe, and anti-stereotypical. Debiasing performances are systematically compared and analyzed across key dimensions: the prompting vs. training paradigm, model size, and generalization of different training strategies to unseen bias types. We release our benchmark, aiming to establish a unified testbed for bias mitigation research.
[112] Finding Diamonds in Conversation Haystacks: A Benchmark for Conversational Data Retrieval
Yohan Lee, Yongwoo Song, Sangyeop Kim
Main category: cs.CL
TL;DR: CDR benchmark is the first comprehensive test set for evaluating conversational data retrieval systems, revealing significant performance gaps in existing models.
Details
Motivation: There's a lack of standardized benchmarks for evaluating conversational data retrieval systems, which are crucial for extracting product insights from conversation data but face unique challenges not addressed by traditional document retrieval.Method: Created a benchmark with 1.6k queries across five analytical tasks and 9.1k conversations, then evaluated 16 popular embedding models to measure conversational data retrieval performance.
Result: Even the best models only achieve around NDCG@10 of 0.51, showing a substantial gap between document and conversational data retrieval capabilities, with unique challenges identified in implicit state recognition, turn dynamics, and contextual references.
Conclusion: The CDR benchmark establishes a reliable standard for conversational data retrieval evaluation, reveals significant limitations in current models, and provides practical tools including query templates and error analysis for advancing the field.
Abstract: We present the Conversational Data Retrieval (CDR) benchmark, the first comprehensive test set for evaluating systems that retrieve conversation data for product insights. With 1.6k queries across five analytical tasks and 9.1k conversations, our benchmark provides a reliable standard for measuring conversational data retrieval performance. Our evaluation of 16 popular embedding models shows that even the best models reach only around NDCG@10 of 0.51, revealing a substantial gap between document and conversational data retrieval capabilities. Our work identifies unique challenges in conversational data retrieval (implicit state recognition, turn dynamics, contextual references) while providing practical query templates and detailed error analysis across different task categories. The benchmark dataset and code are available at https://github.com/l-yohai/CDR-Benchmark.
[113] SECA: Semantically Equivalent and Coherent Attacks for Eliciting LLM Hallucinations
Buyun Liang, Liangzu Peng, Jinqi Luo, Darshan Thaker, Kwan Ho Ryan Chan, René Vidal
Main category: cs.CL
TL;DR: SECA introduces realistic adversarial attacks on LLMs that preserve semantic meaning while eliciting hallucinations, addressing limitations of prior unrealistic attack methods.
Details
Motivation: Current adversarial attacks on LLMs often use unrealistic prompts (nonsensical tokens or altered semantics), limiting insights into real-world hallucination risks. The paper aims to develop realistic attacks that preserve prompt meaning while eliciting hallucinations.Method: Formulates realistic attacks as constrained optimization over input prompt space with semantic equivalence and coherence constraints. Introduces constraint-preserving zeroth-order method to search for adversarial yet feasible prompts.
Result: SECA achieves higher attack success rates while maintaining almost no semantic equivalence or coherence errors compared to existing methods on open-ended multiple-choice QA tasks.
Conclusion: SECA demonstrates LLM sensitivity to realistic prompt variations, highlighting hallucination risks in practical settings and providing a framework for more realistic adversarial evaluation.
Abstract: Large Language Models (LLMs) are increasingly deployed in high-risk domains. However, state-of-the-art LLMs often exhibit hallucinations, raising serious concerns about their reliability. Prior work has explored adversarial attacks to elicit hallucinations in LLMs, but these methods often rely on unrealistic prompts, either by inserting nonsensical tokens or by altering the original semantic intent. Consequently, such approaches provide limited insight into how hallucinations arise in real-world settings. In contrast, adversarial attacks in computer vision typically involve realistic modifications to input images. However, the problem of identifying realistic adversarial prompts for eliciting LLM hallucinations remains largely underexplored. To address this gap, we propose Semantically Equivalent and Coherent Attacks (SECA), which elicit hallucinations via realistic modifications to the prompt that preserve its meaning while maintaining semantic coherence. Our contributions are threefold: (i) we formulate finding realistic attacks for hallucination elicitation as a constrained optimization problem over the input prompt space under semantic equivalence and coherence constraints; (ii) we introduce a constraint-preserving zeroth-order method to effectively search for adversarial yet feasible prompts; and (iii) we demonstrate through experiments on open-ended multiple-choice question answering tasks that SECA achieves higher attack success rates while incurring almost no semantic equivalence or semantic coherence errors compared to existing methods. SECA highlights the sensitivity of both open-source and commercial gradient-inaccessible LLMs to realistic and plausible prompt variations. Code is available at https://github.com/Buyun-Liang/SECA.
[114] Peeking inside the Black-Box: Reinforcement Learning for Explainable and Accurate Relation Extraction
Xinyu Guo, Zhengliang Shi, Minglai Yang, Mahdi Rahimi, Mihai Surdeanu
Main category: cs.CL
TL;DR: CogRE is a novel relation extraction framework that combines cognitive science-inspired reasoning with reinforcement learning optimization to improve both accuracy and explainability through relation keyword generation.
Details
Motivation: Traditional relation extraction methods lack language-based explanations and struggle with poor attention focus and limited one-shot learning capability. The authors aim to enhance RE from both accuracy and explainability perspectives.Method: Two key components: (1) A cognitive science-inspired reasoning mechanism that formulates RE as a series of text-processing steps, and (2) An RL optimization process with a novel reward function that rewards generating relation keywords using an automatically constructed keywords dictionary.
Result: Achieves 24.65% F1 on One-shot NYT29 with Qwen2.5-15B-Instruct, surpassing prior reasoning-based designs. RL optimization provides +23.46% absolute improvement. Models trained on NYT29 achieve +16.9% F1 gain on out-of-distribution WIKIDATA. Human evaluation shows 54% relative increase in explanation quality ratings.
Conclusion: CogRE successfully improves both accuracy and explainability in relation extraction by combining cognitive-structured reasoning with RL optimization, addressing key failure patterns in one-shot RE and providing better language-based explanations.
Abstract: We introduce CogRE, a novel framework for relation extraction (RE), enhancing RE from both accuracy and explainability. The framework has two key components: (i) a reasoning mechanism that formulates relation extraction as a series of text-processing steps inspired by cognitive science, and (ii) an optimization process driven by a novel reinforcement learning (RL) reward function. Our framework introduces relation keywords and rewards generating such keywords using an automatically constructed keywords dictionary. This design addresses the lack of language-based explanations in traditional RE and provides supervision for explanation during RL training. Our experiments show that CogRE improves explanation quality by addressing two common failure patterns in one-shot RE: poor attention focus and limited one-shot learning capability. For example, our cognitive-structured reasoning with Qwen2.5-15B-Instruct on One-shot NYT29 achieves 24.65% F1, surpassing prior reasoning-based designs. Optimizing this approach with RL using our reward further improves performance by +23.46% (absolute). Further, models trained on NYT29 with our reward achieve a +16.9% F1 gain on out-of-distribution WIKIDATA. Finally, human evaluation shows that our best model generates relational keywords closely aligned with gold labels, increasing human explanation quality ratings by 54% (relative).
[115] AWM: Accurate Weight-Matrix Fingerprint for Large Language Models
Boyi Zeng, Lin Chen, Ziwei He, Xinbing Wang, Zhouhan Lin
Main category: cs.CL
TL;DR: A training-free fingerprinting method using weight matrices and Linear Assignment Problem with unbiased CKA similarity to verify LLM lineage despite intensive post-training modifications.
Details
Motivation: Protecting intellectual property of LLMs is crucial due to substantial training resources. Need to determine if suspect LLMs are trained from scratch or derived from existing models, but intensive post-training processes (fine-tuning, continued pretraining, RL, multimodal extension, pruning, upcycling) make reliable identification challenging.Method: Proposes training-free fingerprinting based on weight matrices. Uses Linear Assignment Problem (LAP) and unbiased Centered Kernel Alignment (CKA) similarity to neutralize effects of parameter manipulations, creating robust similarity metric.
Result: Tested on 60 positive and 90 negative model pairs, method shows exceptional robustness against all six post-training categories with near-zero false positive risk. Achieves perfect scores on all classification metrics, establishes reliable model lineage verification. Computation completes within 30s on NVIDIA 3090 GPU.
Conclusion: Method provides strong basis for reliable LLM lineage verification that is robust to various post-training modifications, addressing intellectual property protection needs for model owners and third parties.
Abstract: Protecting the intellectual property of large language models (LLMs) is crucial, given the substantial resources required for their training. Consequently, there is an urgent need for both model owners and third parties to determine whether a suspect LLM is trained from scratch or derived from an existing base model. However, the intensive post-training processes that models typically undergo-such as supervised fine-tuning, extensive continued pretraining, reinforcement learning, multi-modal extension, pruning, and upcycling-pose significant challenges to reliable identification. In this work, we propose a training-free fingerprinting method based on weight matrices. We leverage the Linear Assignment Problem (LAP) and an unbiased Centered Kernel Alignment (CKA) similarity to neutralize the effects of parameter manipulations, yielding a highly robust and high-fidelity similarity metric. On a comprehensive testbed of 60 positive and 90 negative model pairs, our method demonstrates exceptional robustness against all six aforementioned post-training categories while exhibiting a near-zero risk of false positives. By achieving perfect scores on all classification metrics, our approach establishes a strong basis for reliable model lineage verification. Moreover, the entire computation completes within 30s on an NVIDIA 3090 GPU. The code is available at https://github.com/LUMIA-Group/AWM.
[116] MemoTime: Memory-Augmented Temporal Knowledge Graph Enhanced Large Language Model Reasoning
Xingyu Tan, Xiaoyang Wang, Qing Liu, Xiwei Xu, Xin Yuan, Liming Zhu, Wenjie Zhang
Main category: cs.CL
TL;DR: MemoTime: A memory-augmented temporal knowledge graph framework that enhances LLM reasoning on temporal questions by decomposing them into hierarchical temporal trees, using operator-aware reasoning, and storing verified reasoning traces for reuse.
Details
Motivation: LLMs struggle with temporal understanding, especially for complex questions involving multiple entities, compound operators, and evolving event sequences. Existing TKG-based LLM reasoning methods face challenges with temporal faithfulness in multi-hop reasoning, multi-entity synchronization, adapting to diverse temporal operators, and reusing prior reasoning experience.Method: Proposes MemoTime framework with: 1) Hierarchical Tree of Time decomposition of complex temporal questions, 2) Operator-aware reasoning with monotonic timestamp enforcement and multi-entity constraints, 3) Dynamic evidence retrieval layer with operator-specific strategies, 4) Self-evolving experience memory storing verified reasoning traces, toolkit decisions, and sub-question embeddings for cross-type reuse.
Result: Achieves state-of-the-art results on multiple temporal QA benchmarks, outperforming strong baselines by up to 24.0%. Enables smaller models (e.g., Qwen3-4B) to achieve reasoning performance comparable to GPT-4-Turbo.
Conclusion: MemoTime effectively addresses temporal reasoning challenges in LLMs through structured grounding, recursive reasoning, and continual experience learning, demonstrating significant improvements in temporal question answering performance.
Abstract: Large Language Models (LLMs) have achieved impressive reasoning abilities, but struggle with temporal understanding, especially when questions involve multiple entities, compound operators, and evolving event sequences. Temporal Knowledge Graphs (TKGs), which capture vast amounts of temporal facts in a structured format, offer a reliable source for temporal reasoning. However, existing TKG-based LLM reasoning methods still struggle with four major challenges: maintaining temporal faithfulness in multi-hop reasoning, achieving multi-entity temporal synchronization, adapting retrieval to diverse temporal operators, and reusing prior reasoning experience for stability and efficiency. To address these issues, we propose MemoTime, a memory-augmented temporal knowledge graph framework that enhances LLM reasoning through structured grounding, recursive reasoning, and continual experience learning. MemoTime decomposes complex temporal questions into a hierarchical Tree of Time, enabling operator-aware reasoning that enforces monotonic timestamps and co-constrains multiple entities under unified temporal bounds. A dynamic evidence retrieval layer adaptively selects operator-specific retrieval strategies, while a self-evolving experience memory stores verified reasoning traces, toolkit decisions, and sub-question embeddings for cross-type reuse. Comprehensive experiments on multiple temporal QA benchmarks show that MemoTime achieves overall state-of-the-art results, outperforming the strong baseline by up to 24.0%. Furthermore, MemoTime enables smaller models (e.g., Qwen3-4B) to achieve reasoning performance comparable to that of GPT-4-Turbo.
[117] Batch Speculative Decoding Done Right
Ranran Haoran Zhang, Soumik Dey, Ashirbad Mishra, Hansi Wu, Binbin Li, Rui Zhang
Main category: cs.CL
TL;DR: First authentic batch speculative decoding framework that guarantees output equivalence by solving ragged tensor synchronization problems, achieving up to 3x throughput improvement while maintaining algorithmic correctness.
Details
Motivation: Existing batch speculative decoding implementations violate the fundamental requirement of output equivalence with standard autoregressive generation, producing corrupted outputs due to ragged tensor problems where sequences in the same batch accept different numbers of draft tokens, desynchronizing position IDs, attention masks, and KV-cache state.Method: Presents EQSPEC (first algorithm guaranteeing output equivalence) and EXSPEC (reduces alignment overhead through cross-batch scheduling). Formalizes synchronization invariants, analyzes cost structure showing alignment overhead grows superlinearly, and introduces dynamic grouping of same-length sequences.
Result: Achieves up to 3x throughput improvement at batch size 8 on SpecBench across Vicuna-7B/68M, Qwen3-8B/0.6B, and GLM-4-9B/0.6B pairs. Methods achieve 95% decoding-equivalence, with residual divergence attributable to floating-point non-determinism in GPU inference, not the synchronization failures that cause near-zero equivalence of prior methods.
Conclusion: Presents the first authentic batch speculative decoding framework that solves fundamental synchronization problems, guaranteeing output equivalence while achieving significant throughput improvements, addressing a critical correctness issue in existing implementations.
Abstract: Speculative decoding must produce outputs distribution identical to standard autoregressive generation-this output equivalence is not an optimization target but the defining criterion of valid speculative decoding. We demonstrate that all existing batch speculative decoding implementations violate this fundamental requirement, producing corrupted outputs ranging from repetitive tokens to gibberish. These failures stem from the ragged tensor problem: sequences in the same batch accept different numbers of draft tokens, desynchronizing position IDs, attention masks, and KV-cache state. We present the first authentic batch speculative decoding framework. We (1) formalize the synchronization invariants that valid batch speculative decoding must satisfy, (2) present EQSPEC, the first algorithm that guarantees output equivalence, and analyze its cost structure to show that alignment overhead grows superlinearly and consumes up to 40% of computation, and (3) introduce EXSPEC, which reduces this overhead through cross-batch scheduling that dynamically groups same-length sequences. On SpecBench across Vicuna-7B/68M, Qwen3-8B/0.6B, and GLM-4-9B/0.6B pairs, our methods achieve up to 3x throughput improvement at batch size 8 while maintaining algorithmic correctness. Our methods achieve 95% decoding-equivalence, with residual divergence attributable to floating-point non-determinism in GPU inference, not the synchronization failures that cause near-zero equivalence of prior methods. Our code is available at https://github.com/eBay/spec_dec.
[118] ParoQuant: Pairwise Rotation Quantization for Efficient Reasoning LLM Inference
Yesheng Liang, Haisheng Chen, Zihan Zhang, Song Han, Zhijian Liu
Main category: cs.CL
TL;DR: ParoQuant is a post-training quantization method that uses pairwise rotation and channel-wise scaling to address outlier issues in LLMs, improving accuracy for reasoning tasks with minimal inference overhead.
Details
Motivation: Existing PTQ methods struggle with outliers in weights and activations of LLMs, especially in reasoning models where errors accumulate across long chains of thought, leading to accuracy degradation. Current methods either insufficiently suppress outliers or introduce significant inference overhead.Method: Proposes Pairwise Rotation Quantization (ParoQuant) combining hardware-efficient independent Givens rotations with channel-wise scaling to even out magnitudes across channels and narrow dynamic range within each quantization group. Also includes co-designed inference kernels to exploit GPU parallelism while keeping rotations and scaling lightweight at runtime.
Result: Under weight-only quantization, ParoQuant achieves average 2.4% accuracy improvement over AWQ on reasoning tasks with less than 10% overhead. Also matches accuracy of state-of-the-art weight-activation quantization methods.
Conclusion: ParoQuant effectively addresses outlier issues in LLM quantization, enabling more efficient and accurate deployment of reasoning LLMs through hardware-efficient rotation-based quantization with minimal inference overhead.
Abstract: Post-training quantization (PTQ) compresses the weights and activations of large language models (LLMs) into low-precision representations to reduce memory footprint and accelerate inference. However, the presence of outliers in weights and activations often leads to large quantization errors and severe accuracy degradation, especially in recent reasoning LLMs where errors accumulate across long chains of thought. Existing PTQ methods either fail to sufficiently suppress outliers or introduce significant overhead during inference. In this paper, we propose Pairwise Rotation Quantization (ParoQuant), a PTQ method that combines hardware-efficient and optimizable independent Givens rotations with channel-wise scaling to even out the magnitudes across channels and narrow the dynamic range within each quantization group, effectively addressing the outlier issue. We further co-design the inference kernel to fully exploit GPU parallelism and keep the rotations and scaling lightweight at runtime. Under weight-only quantization, ParoQuant achieves an average 2.4% accuracy improvement over AWQ on reasoning tasks, with less than 10% overhead. ParoQuant also matches the accuracy of state-of-the-art weight-activation quantization methods. This paves the way for more efficient and accurate deployment of reasoning LLMs.
[119] Context-Emotion Aware Therapeutic Dialogue Generation: A Multi-component Reinforcement Learning Approach to Language Models for Mental Health Support
Eric Hua Qing Zhang, Julia Ive
Main category: cs.CL
TL;DR: This paper investigates using supervised fine-tuning and reinforcement learning to enhance GPT-2 for therapeutic dialogue generation, addressing gaps in existing mental health LLM approaches.
Details
Motivation: Mental health disorders create a global burden, and while LLMs offer 24/7 support, current models lack contextual coherence and emotional alignment for therapeutic dialogue. Existing methods have three critical gaps: SFT produces repetitive outputs, RL uses generic reward functions prioritizing lexical similarity over clinical appropriateness, and LLMs are resource-intensive with privacy concerns.Method: The study applies SFT and RL techniques to enhance GPT-2’s therapeutic dialogue capabilities. It restructures input formats to process contextual information and emotional states alongside user input, and employs a novel multi-component reward function that aligns outputs with professional therapeutic logic and annotated emotions rather than just lexical overlap.
Result: Results show substantial improvements through RL over baseline GPT-2 across multiple metrics: BLEU (0.0111), ROUGE-1 (0.1397), ROUGE-2 (0.0213), ROUGE-L (0.1317), and METEOR (0.0581). LLM evaluation confirmed high contextual relevance and professionalism, while RL achieved 99.34% emotion accuracy compared to 66.96% for baseline GPT-2.
Conclusion: The findings demonstrate RL’s effectiveness in developing therapeutic dialogue systems that can serve as valuable assistive tools for therapists while maintaining essential human clinical oversight. The approach addresses key limitations of existing methods for mental health applications.
Abstract: Mental health disorders impose a substantial global socioeconomic burden. While large language models (LLMs) offer 24/7, non-judgmental interactions to address this gap, pretrained models lack contextual coherence and emotional alignment for appropriate therapeutic dialogue. Existing methods suffer from three critical methodological gaps: 1) Supervised Fine-Tuning (SFT) produces repetitive, context-insensitive outputs that fail to balance clinical accuracy with genuine empathy; 2) Reinforcement Learning (RL)-based therapeutic systems rely on generic reward functions (e.g., BLEU, ROUGE) that prioritise lexical similarity over clinical-specific emotional appropriateness and contextual relevance; 3) LLMs are resource-intensive and pose data privacy risks, making local deployment in clinical settings infeasible. To address these gaps, this study investigates the application of SFT and RL techniques to enhance GPT-2’s capacity for therapeutic dialogue generation. The methodology restructured input formats to enable simultaneous processing of contextual information and emotional states alongside user input, employing a novel multi-component reward function that explicitly aligns model outputs with professional therapeutic logic (not just lexical overlap) and annotated emotions. Results demonstrated substantial improvements through RLs over baseline GPT-2 across multiple evaluation metrics: BLEU (0.0111), ROUGE-1 (0.1397), ROUGE-2 (0.0213), ROUGE-L (0.1317), and METEOR (0.0581). LLM evaluation confirmed high contextual relevance and professionalism, while RL achieved 99.34% emotion accuracy compared to 66.96% for baseline GPT-2. These findings demonstrate RL’s effectiveness in developing therapeutic dialogue systems that can serve as valuable assistive tools for therapists, while maintaining essential human clinical oversight.
[120] ArtistMus: A Globally Diverse, Artist-Centric Benchmark for Retrieval-Augmented Music Question Answering
Daeyong Kwon, SeungHeon Doh, Juhan Nam
Main category: cs.CL
TL;DR: MusWikiDB: A vector database of 3.2M music-related Wikipedia passages and ArtistMus benchmark for evaluating retrieval-augmented generation in music question answering, showing significant improvements in factual accuracy for LLMs.
Details
Motivation: LLMs have limited music knowledge due to sparse music data in pretraining, and existing music information retrieval resources don't adequately support factual and contextual music question answering grounded in artist metadata and historical context.Method: Created MusWikiDB (3.2M passages from 144K music Wikipedia pages) and ArtistMus benchmark (1,000 questions on 500 artists with metadata). Evaluated retrieval-augmented generation (RAG) for music question answering, comparing open-source and proprietary models, and conducted RAG-style fine-tuning.
Result: RAG significantly improves factual accuracy: open-source models gain up to +56.8 percentage points (Qwen3 8B improves from 35.0% to 91.8%). RAG-style fine-tuning boosts both factual recall and contextual reasoning. MusWikiDB achieves ~6% higher accuracy and 40% faster retrieval than general Wikipedia corpus.
Conclusion: MusWikiDB and ArtistMus advance music information retrieval and domain-specific QA, establishing foundation for retrieval-augmented reasoning in culturally rich domains like music, with RAG significantly enhancing LLM performance on music-related tasks.
Abstract: Recent advances in large language models (LLMs) have transformed open-domain question answering, yet their effectiveness in music-related reasoning remains limited due to sparse music knowledge in pretraining data. While music information retrieval and computational musicology have explored structured and multimodal understanding, few resources support factual and contextual music question answering (MQA) grounded in artist metadata or historical context. We introduce MusWikiDB, a vector database of 3.2M passages from 144K music-related Wikipedia pages, and ArtistMus, a benchmark of 1,000 questions on 500 diverse artists with metadata such as genre, debut year, and topic. These resources enable systematic evaluation of retrieval-augmented generation (RAG) for MQA. Experiments show that RAG markedly improves factual accuracy; open-source models gain up to +56.8 percentage points (for example, Qwen3 8B improves from 35.0 to 91.8), approaching proprietary model performance. RAG-style fine-tuning further boosts both factual recall and contextual reasoning, improving results on both in-domain and out-of-domain benchmarks. MusWikiDB also yields approximately 6 percentage points higher accuracy and 40% faster retrieval than a general-purpose Wikipedia corpus. We release MusWikiDB and ArtistMus to advance research in music information retrieval and domain-specific question answering, establishing a foundation for retrieval-augmented reasoning in culturally rich domains such as music.
[121] Multi-LLM Thematic Analysis with Dual Reliability Metrics: Combining Cohen’s Kappa and Semantic Similarity for Qualitative Research Validation
Nilesh Jain, Hyungil Suh, Seyi Adeyinka, Leor Roseman, Aza Allsop
Main category: cs.CL
TL;DR: A multi-perspective validation framework for LLM-based thematic analysis using ensemble validation with dual reliability metrics (Cohen’s Kappa and cosine similarity) to improve reliability in qualitative research.
Details
Motivation: Qualitative research faces reliability challenges with traditional inter-rater agreement methods being time-intensive and yielding moderate consistency. There's a need for more reliable AI-assisted qualitative analysis methods.Method: Multi-perspective validation framework combining ensemble validation with dual reliability metrics (Cohen’s Kappa for inter-rater agreement, cosine similarity for semantic consistency). Configurable analysis parameters (1-6 seeds, temperature 0.0-2.0), custom prompt structures with variable substitution, and consensus theme extraction across any JSON format.
Result: Evaluated three LLMs (Gemini 2.5 Pro, GPT-4o, Claude 3.5 Sonnet) on psychedelic art therapy interview transcript. Gemini achieved highest reliability (κ=0.907, cosine=95.3%), followed by GPT-4o (κ=0.853, cosine=92.6%) and Claude (κ=0.842, cosine=92.1%). All models achieved high agreement (κ>0.80). Gemini identified 6 consensus themes, GPT-4o 5 themes, Claude 4 themes.
Conclusion: The framework successfully validates multi-run ensemble approach for LLM-based thematic analysis, providing transparent reliability metrics, flexible configuration, and structure-agnostic consensus extraction, establishing methodological foundations for reliable AI-assisted qualitative research.
Abstract: Qualitative research faces a critical reliability challenge: traditional inter-rater agreement methods require multiple human coders, are time-intensive, and often yield moderate consistency. We present a multi-perspective validation framework for LLM-based thematic analysis that combines ensemble validation with dual reliability metrics: Cohen’s Kappa ($κ$) for inter-rater agreement and cosine similarity for semantic consistency. Our framework enables configurable analysis parameters (1-6 seeds, temperature 0.0-2.0), supports custom prompt structures with variable substitution, and provides consensus theme extraction across any JSON format. As proof-of-concept, we evaluate three leading LLMs (Gemini 2.5 Pro, GPT-4o, Claude 3.5 Sonnet) on a psychedelic art therapy interview transcript, conducting six independent runs per model. Results demonstrate Gemini achieves highest reliability ($κ= 0.907$, cosine=95.3%), followed by GPT-4o ($κ= 0.853$, cosine=92.6%) and Claude ($κ= 0.842$, cosine=92.1%). All three models achieve a high agreement ($κ> 0.80$), validating the multi-run ensemble approach. The framework successfully extracts consensus themes across runs, with Gemini identifying 6 consensus themes (50-83% consistency), GPT-4o identifying 5 themes, and Claude 4 themes. Our open-source implementation provides researchers with transparent reliability metrics, flexible configuration, and structure-agnostic consensus extraction, establishing methodological foundations for reliable AI-assisted qualitative research.
[122] Ara-HOPE: Human-Centric Post-Editing Evaluation for Dialectal Arabic to Modern Standard Arabic Translation
Abdullah Alabdullah, Lifeng Han, Chenghua Lin
Main category: cs.CL
TL;DR: Ara-HOPE is a human-centric post-editing evaluation framework for Dialectal Arabic to Modern Standard Arabic translation, featuring a five-category error taxonomy and decision-tree annotation protocol to address limitations of existing MT evaluation metrics.
Details
Motivation: Existing automatic evaluation metrics and general-purpose human evaluation frameworks fail to capture dialect-specific MT errors in DA-MSA translation, hindering progress in translation assessment and system improvement.Method: Developed Ara-HOPE framework with five-category error taxonomy and decision-tree annotation protocol; evaluated three MT systems (Jais, GPT-3.5, NLLB-200) using this framework to assess dialect-specific translation quality.
Result: Ara-HOPE effectively highlighted systematic performance differences between MT systems, revealing that dialect-specific terminology and semantic preservation remain the most persistent challenges in DA-MSA translation.
Conclusion: Ara-HOPE establishes a new framework for evaluating Dialectal Arabic MT quality and provides actionable guidance for improving dialect-aware MT systems, with annotation materials made publicly available for reproducibility.
Abstract: Dialectal Arabic to Modern Standard Arabic (DA-MSA) translation is a challenging task in Machine Translation (MT) due to significant lexical, syntactic, and semantic divergences between Arabic dialects and MSA. Existing automatic evaluation metrics and general-purpose human evaluation frameworks struggle to capture dialect-specific MT errors, hindering progress in translation assessment. This paper introduces Ara-HOPE, a human-centric post-editing evaluation framework designed to systematically address these challenges. The framework includes a five-category error taxonomy and a decision-tree annotation protocol. Through comparative evaluation of three MT systems (Arabic-centric Jais, general-purpose GPT-3.5, and baseline NLLB-200), Ara-HOPE effectively highlights systematic performance differences between these systems. Our results show that dialect-specific terminology and semantic preservation remain the most persistent challenges in DA-MSA translation. Ara-HOPE establishes a new framework for evaluating Dialectal Arabic MT quality and provides actionable guidance for improving dialect-aware MT systems. For reproducibility, we make the annotation files and related materials publicly available at https://github.com/abdullahalabdullah/Ara-HOPE
[123] EmoLoom-2B: Fast Base-Model Screening for Emotion Classification and VAD with Lexicon-Weak Supervision and KV-Off Evaluation
Zilin Li, Weiwei Xu, Xuanbo Lu, Zheda Liu
Main category: cs.CL
TL;DR: EmoLoom-2B is a lightweight pipeline that transforms small language models (<2B parameters) into efficient emotion analysis systems capable of joint emotion classification and Valence-Arousal-Dominance prediction with protocol-faithful evaluation.
Details
Motivation: The paper addresses the need for lightweight, reproducible emotion analysis systems that can perform both categorical emotion classification and dimensional VAD prediction. Current approaches often lack protocol consistency, introduce evaluation variance, and require heavy computational resources. The authors aim to create a budget-aware, auditable pipeline suitable as a screening pass before more intensive training or multimodal fusion.Method: The method uses Qwen-1.8B-Chat as base model with a unified JSON input-output contract for data loading, training, and inference. Key innovations include: 1) KV-off decoding to reduce variance, 2) Two orthogonal semantic regularizers (VAD-preserving constraint and lightweight external appraisal classifier), 3) Valence Flip augmentation for polarity sensitivity, and 4) A/B mixture sampling with entropy-aware temperature scheduling during supervised fine-tuning.
Result: EmoLoom-2B achieves strong performance on GoEmotions and EmpatheticDialogues datasets, and demonstrates robust cross-corpus generalization on DailyDialog. The pipeline is shown to be effective for emotion analysis while maintaining lightweight computational requirements.
Conclusion: The proposed EmoLoom-2B pipeline provides a dependable, budget-aware screening solution for emotion analysis tasks. It offers protocol-faithful evaluation, reduces avoidable variance, and serves as a practical foundation before heavier training or multimodal fusion approaches.
Abstract: We introduce EmoLoom-2B, a lightweight and reproducible pipeline that turns small language models under 2B parameters into fast screening candidates for joint emotion classification and Valence-Arousal-Dominance prediction. To ensure protocol-faithful and fair evaluation, we unify data loading, training, and inference under a single JSON input-output contract and remove avoidable variance by adopting KV-off decoding as the default setting. We incorporate two orthogonal semantic regularizers: a VAD-preserving constraint that aligns generated text with target VAD triples, and a lightweight external appraisal classifier that provides training-time guidance on goal attainment, controllability, certainty, and fairness without injecting long rationales. To improve polarity sensitivity, we introduce Valence Flip augmentation based on mirrored emotional pairs. During supervised fine-tuning, we apply A/B mixture sampling with entropy-aware temperature scheduling to balance coverage and convergence. Using Qwen-1.8B-Chat as the base model, EmoLoom-2B achieves strong performance on GoEmotions and EmpatheticDialogues, and demonstrates robust cross-corpus generalization on DailyDialog. The proposed recipe is budget-aware, auditable, and re-entrant, serving as a dependable screening pass before heavier training or multimodal fusion.
[124] FormationEval, an open multiple-choice benchmark for petroleum geoscience
Almaz Ermilov
Main category: cs.CL
TL;DR: FormationEval is a specialized benchmark for evaluating language models on petroleum geoscience knowledge, featuring 505 multiple-choice questions across 7 domains with 72 models tested.
Details
Motivation: To create a domain-specific evaluation benchmark for language models in petroleum geoscience and subsurface disciplines, addressing the lack of specialized assessment tools in this technical field.Method: Created a dataset of 505 questions across 7 domains (petrophysics, petroleum geology, reservoir engineering, etc.) using a reasoning model with detailed instructions and concept-based approach to avoid copyright issues. Questions include source metadata for traceability.
Result: Top performers achieved over 97% accuracy (Gemini 3 Pro Preview: 99.8%). Among open-weight models, GLM-4.7 led at 98.6%. Petrophysics was the most challenging domain. Performance gap between open-weight and closed models was narrower than expected, with several open-weight models exceeding 90% accuracy.
Conclusion: The benchmark provides valuable insights into language model performance on specialized technical domains, showing that open-weight models can compete with closed models in domain-specific knowledge tasks. The dataset and evaluation tools are publicly available.
Abstract: This paper presents FormationEval, an open multiple-choice question benchmark for evaluating language models on petroleum geoscience and subsurface disciplines. The dataset contains 505 questions across seven domains including petrophysics, petroleum geology and reservoir engineering, derived from three authoritative sources using a reasoning model with detailed instructions and a concept-based approach that avoids verbatim copying of copyrighted text. Each question includes source metadata to support traceability and audit. The evaluation covers 72 models from major providers including OpenAI, Anthropic, Google, Meta and open-weight alternatives. The top performers achieve over 97% accuracy, with Gemini 3 Pro Preview reaching 99.8%, while tier and domain gaps persist. Among open-weight models, GLM-4.7 leads at 98.6%, with several DeepSeek, Llama, Qwen and Mistral models also exceeding 93%. The performance gap between open-weight and closed models is narrower than expected, with several lower-cost open-weight models exceeding 90% accuracy. Petrophysics emerges as the most challenging domain across all models, while smaller models show wider performance variance. Residual length bias in the dataset (correct answers tend to be longer) is documented along with bias mitigation strategies applied during construction. The benchmark, evaluation code and results are publicly available.
[125] SYNAPSE: Empowering LLM Agents with Episodic-Semantic Memory via Spreading Activation
Hanqi Jiang, Junhao Chen, Yi Pan, Ling Chen, Weihang You, Yifan Zhou, Ruidong Zhang, Andrea Sikora, Lin Zhao, Yohannes Abate, Tianming Liu
Main category: cs.CL
TL;DR: Synapse introduces a unified memory architecture for LLMs that models memory as a dynamic graph with spreading activation, lateral inhibition, and temporal decay, outperforming state-of-the-art methods on complex temporal reasoning tasks.
Details
Motivation: Standard retrieval-augmented approaches fail to address the disconnected nature of long-term agentic memory in LLMs, particularly for complex temporal and multi-hop reasoning tasks where static vector similarity is insufficient.Method: Synapse models memory as a dynamic graph where relevance emerges from spreading activation rather than pre-computed links. It integrates lateral inhibition and temporal decay, and implements a Triple Hybrid Retrieval strategy that fuses geometric embeddings with activation-based graph traversal.
Result: Comprehensive evaluations on the LoCoMo benchmark show that Synapse significantly outperforms state-of-the-art methods in complex temporal and multi-hop reasoning tasks, effectively addressing the “Contextual Tunneling” problem.
Conclusion: Synapse offers a robust solution to memory limitations in LLMs by providing a unified memory architecture that transcends static vector similarity, enabling better long-term agentic memory and complex reasoning.
Abstract: While Large Language Models (LLMs) excel at generalized reasoning, standard retrieval-augmented approaches fail to address the disconnected nature of long-term agentic memory. To bridge this gap, we introduce Synapse (Synergistic Associative Processing Semantic Encoding), a unified memory architecture that transcends static vector similarity. Drawing from cognitive science, Synapse models memory as a dynamic graph where relevance emerges from spreading activation rather than pre-computed links. By integrating lateral inhibition and temporal decay, the system dynamically highlights relevant sub-graphs while filtering interference. We implement a Triple Hybrid Retrieval strategy that fuses geometric embeddings with activation-based graph traversal. Comprehensive evaluations on the LoCoMo benchmark show that Synapse significantly outperforms state-of-the-art methods in complex temporal and multi-hop reasoning tasks, offering a robust solution to the “Contextual Tunneling” problem. Our code and data will be made publicly available upon acceptance.
[126] Reward Modeling from Natural Language Human Feedback
Zongqi Wang, Rui Wang, Yuchuan Wu, Yiyao Yu, Pinyi Zhang, Shaoning Sun, Yujiu Yang, Yongbin Li
Main category: cs.CL
TL;DR: RM-NLHF improves reward modeling by using natural language feedback instead of binary preferences, addressing spurious success in GRMs through process-level supervision and a MetaRM for scalability.
Details
Motivation: Current RLVR approaches using binary preference labels allow GRMs to guess correct outcomes without proper reasoning, introducing noisy rewards that impair reinforcement learning effectiveness.Method: Proposes RM-NLHF using similarity between GRM-generated and human critiques as process reward, plus MetaRM to predict process reward from human-critique data and generalize to data without human critiques.
Result: Experiments on multiple benchmarks show consistent outperformance over state-of-the-art GRMs trained with outcome-only reward, confirming superiority of natural language feedback over binary supervision.
Conclusion: Natural language feedback provides more accurate reward signals than binary outcomes, and MetaRM enables scalable process-level supervision by learning from limited human critique data.
Abstract: Reinforcement Learning with Verifiable reward (RLVR) on preference data has become the mainstream approach for training Generative Reward Models (GRMs). Typically in pairwise rewarding tasks, GRMs generate reasoning chains ending with critiques and preference labels, and RLVR then relies on the correctness of the preference labels as the training reward. However, in this paper, we demonstrate that such binary classification tasks make GRMs susceptible to guessing correct outcomes without sound critiques. Consequently, these spurious successes introduce substantial noise into the reward signal, thereby impairing the effectiveness of reinforcement learning. To address this issue, we propose Reward Modeling from Natural Language Human Feedback (RM-NLHF), which leverages natural language feedback to obtain process reward signals, thereby mitigating the problem of limited solution space inherent in binary tasks. Specifically, we compute the similarity between GRM-generated and human critiques as the training reward, which provides more accurate reward signals than outcome-only supervision. Additionally, considering that human critiques are difficult to scale up, we introduce Meta Reward Model (MetaRM) which learns to predict process reward from datasets with human critiques and then generalizes to data without human critiques. Experiments on multiple benchmarks demonstrate that our method consistently outperforms state-of-the-art GRMs trained with outcome-only reward, confirming the superiority of integrating natural language over binary human feedback as supervision.
[127] Context Volume Drives Performance: Tackling Domain Shift in Extremely Low-Resource Translation via RAG
David Samuel Setiawan, Raphaël Merx, Jey Han Lau
Main category: cs.CL
TL;DR: Hybrid NMT+LLM framework improves low-resource language translation under domain shift by using NMT for initial draft and LLM with RAG for refinement.
Details
Motivation: NMT models for low-resource languages suffer significant performance degradation under domain shift, as demonstrated with Dhao language where translation quality drops sharply when moving from New Testament to Old Testament domains.Method: Hybrid framework: fine-tuned NMT model generates initial draft, then LLM with Retrieval-Augmented Generation (RAG) refines it. Analysis focuses on impact of retrieved examples quantity vs. retrieval algorithm choice.
Result: System achieves 35.21 chrF++ (+8.10 recovery), effectively matching original in-domain quality. Performance driven primarily by number of retrieved examples rather than retrieval algorithm. LLM acts as robust “safety net” for severe failures.
Conclusion: Hybrid NMT+LLM with RAG effectively addresses domain shift in low-resource translation, with LLM serving as critical safety net for zero-shot domains.
Abstract: Neural Machine Translation (NMT) models for low-resource languages suffer significant performance degradation under domain shift. We quantify this challenge using Dhao, an indigenous language of Eastern Indonesia with no digital footprint beyond the New Testament (NT). When applied to the unseen Old Testament (OT), a standard NMT model fine-tuned on the NT drops from an in-domain score of 36.17 chrF++ to 27.11 chrF++. To recover this loss, we introduce a hybrid framework where a fine-tuned NMT model generates an initial draft, which is then refined by a Large Language Model (LLM) using Retrieval-Augmented Generation (RAG). The final system achieves 35.21 chrF++ (+8.10 recovery), effectively matching the original in-domain quality. Our analysis reveals that this performance is driven primarily by the number of retrieved examples rather than the choice of retrieval algorithm. Qualitative analysis confirms the LLM acts as a robust “safety net,” repairing severe failures in zero-shot domains.
[128] Human Values in a Single Sentence: Moral Presence, Hierarchies, and Transformer Ensembles on the Schwartz Continuum
Víctor Yeste, Paolo Rosso
Main category: cs.CL
TL;DR: Sentence-level detection of 19 human values using DeBERTa models with lightweight feature augmentation, comparing direct vs hierarchical approaches under constrained compute budget.
Details
Motivation: To develop compute-efficient methods for detecting refined Schwartz human values at sentence level, comparing different architectural approaches under limited GPU resources.Method: Uses DeBERTa-base classifiers with calibrated thresholds, compares direct multi-label vs presence-gated hierarchical approaches, incorporates lightweight features (context, LIWC-22, moral lexica), and benchmarks instruction-tuned LLMs in zero/few-shot and QLoRA setups.
Result: Best supervised ensemble achieves macro-F1 = 0.332 on 19 values, improving over previous baseline (0.28). Moral presence detection achieves F1 = 0.74. LLMs lag behind supervised ensemble under same compute constraints.
Conclusion: Direct prediction with lightweight feature augmentation outperforms hierarchical approaches, and supervised methods are more compute-efficient than LLMs for value detection tasks under resource constraints.
Abstract: We study sentence-level detection of the 19 human values in the refined Schwartz continuum in about 74k English sentences from news and political manifestos (ValueEval'24 corpus). Each sentence is annotated with value presence, yielding a binary moral-presence label and a 19-way multi-label task under severe class imbalance. First, we show that moral presence is learnable from single sentences: a DeBERTa-base classifier attains positive-class F1 = 0.74 with calibrated thresholds. Second, we compare direct multi-label value detectors with presence-gated hierarchies in a setting where only a single consumer-grade GPU with 8 GB of VRAM is available, and we explicitly choose all training and inference configurations to fit within this budget. Presence gating does not improve over direct prediction, indicating that gate recall becomes a bottleneck. Third, we investigate lightweight auxiliary signals - short-range context, LIWC-22, and moral lexica - and small ensembles. Our best supervised configuration, a soft-voting ensemble of DeBERTa-based models enriched with such signals, reaches macro-F1 = 0.332 on the 19 values, improving over the best previous English-only baseline on this corpus, namely the best official ValueEval'24 English run (macro-F1 = 0.28 on the same 19-value test set). Methodologically, our study provides, to our knowledge, the first systematic comparison of direct versus presence-gated architectures, lightweight feature-augmented encoders, and medium-sized instruction-tuned Large Language Models (LLMs) for refined Schwartz values at sentence level. We additionally benchmark 7-9B instruction-tuned LLMs (Gemma 2 9B, Llama 3.1 8B, Mistral 8B, Qwen 2.5 7B) in zero-/few-shot and QLoRA setups, and find that they lag behind the supervised ensemble under the same compute budget. Overall, our results provide empirical guidance for building compute-efficient, value-aware NLP models.
[129] Do LLMs Truly Benefit from Longer Context in Automatic Post-Editing?
Ahrii Kim, Seong-heum Kim
Main category: cs.CL
TL;DR: LLMs show near-human APE quality with simple prompting but fail to effectively use document context, with proprietary models being robust but impractical due to cost/latency.
Details
Motivation: To systematically evaluate LLMs for automatic post-editing of machine translations, particularly understanding their effectiveness with document-level context, which remains insufficiently explored despite LLMs' strong translation capabilities.Method: Systematic comparison of proprietary and open-weight LLMs using naive document-level prompting setup, analyzing APE quality, contextual behavior, robustness to data poisoning attacks, and efficiency metrics.
Result: Proprietary LLMs achieve near human-level APE quality with simple one-shot prompting, regardless of document context. They show higher robustness to attacks than open-weight models but largely fail to exploit document-level context. Standard metrics don’t reflect qualitative improvements, and proprietary models have impractical cost/latency overheads.
Conclusion: LLM-based document-aware APE shows promise but has limitations: proprietary models are robust but impractical, and current approaches fail to effectively leverage document context, pointing to need for more efficient long-context modeling for translation refinement.
Abstract: Automatic post-editing (APE) aims to refine machine translations by correcting residual errors. Although recent large language models (LLMs) demonstrate strong translation capabilities, their effectiveness for APE–especially under document-level context–remains insufficiently understood. We present a systematic comparison of proprietary and open-weight LLMs under a naive document-level prompting setup, analyzing APE quality, contextual behavior, robustness, and efficiency. Our results show that proprietary LLMs achieve near human-level APE quality even with simple one-shot prompting, regardless of whether document context is provided. While these models exhibit higher robustness to data poisoning attacks than open-weight counterparts, this robustness also reveals a limitation: they largely fail to exploit document-level context for contextual error correction. Furthermore, standard automatic metrics do not reliably reflect these qualitative improvements, highlighting the continued necessity of human evaluation. Despite their strong performance, the substantial cost and latency overheads of proprietary LLMs render them impractical for real-world APE deployment. Overall, our findings elucidate both the promise and current limitations of LLM-based document-aware APE, and point toward the need for more efficient long-context modeling approaches for translation refinement.
[130] Accelerating Scientific Research with Gemini: Case Studies and Common Techniques
David P. Woodruff, Vincent Cohen-Addad, Lalit Jain, Jieming Mao, Song Zuo, MohammadHossein Bateni, Simina Branzei, Michael P. Brenner, Lin Chen, Ying Feng, Lance Fortnow, Gang Fu, Ziyi Guan, Zahra Hadizadeh, Mohammad T. Hajiaghayi, Mahdi JafariRaviz, Adel Javanmard, Karthik C. S., Ken-ichi Kawarabayashi, Ravi Kumar, Silvio Lattanzi, Euiwoong Lee, Yi Li, Ioannis Panageas, Dimitris Paparas, Benjamin Przybocki, Bernardo Subercaseaux, Ola Svensson, Shayan Taherijam, Xuan Wu, Eylon Yogev, Morteza Zadimoghaddam, Samson Zhou, Yossi Matias, James Manyika, Vahab Mirrokni
Main category: cs.CL
TL;DR: Researchers demonstrate successful human-AI collaboration using Google’s Gemini models to solve open problems, refute conjectures, and generate proofs across theoretical computer science, economics, optimization, and physics.
Details
Motivation: To explore how advanced AI models can contribute to novel, expert-level mathematical discovery and scientific research beyond routine task automation, examining their potential as genuine partners in creative scientific processes.Method: Collection of case studies using Google’s Gemini-based models (Gemini Deep Think and variants) with interactive conversational methodology, including iterative refinement, problem decomposition, cross-disciplinary knowledge transfer, adversarial review, and neuro-symbolic loops for code verification.
Result: Successful collaborations solving open problems, refuting conjectures, and generating new proofs across diverse theoretical fields, demonstrating AI’s ability to contribute to genuine scientific discovery beyond automation.
Conclusion: AI models can serve as versatile partners in scientific discovery when combined with effective human-AI collaboration techniques, pushing beyond standard chat interfaces to include adversarial review and neuro-symbolic verification systems.
Abstract: Recent advances in large language models (LLMs) have opened new avenues for accelerating scientific research. While models are increasingly capable of assisting with routine tasks, their ability to contribute to novel, expert-level mathematical discovery is less understood. We present a collection of case studies demonstrating how researchers have successfully collaborated with advanced AI models, specifically Google’s Gemini-based models (in particular Gemini Deep Think and its advanced variants), to solve open problems, refute conjectures, and generate new proofs across diverse areas in theoretical computer science, as well as other areas such as economics, optimization, and physics. Based on these experiences, we extract common techniques for effective human-AI collaboration in theoretical research, such as iterative refinement, problem decomposition, and cross-disciplinary knowledge transfer. While the majority of our results stem from this interactive, conversational methodology, we also highlight specific instances that push beyond standard chat interfaces. These include deploying the model as a rigorous adversarial reviewer to detect subtle flaws in existing proofs, and embedding it within a “neuro-symbolic” loop that autonomously writes and executes code to verify complex derivations. Together, these examples highlight the potential of AI not just as a tool for automation, but as a versatile, genuine partner in the creative process of scientific discovery.
[131] Tokenization and Morphological Fidelity in Uralic NLP: A Cross-Lingual Evaluation
Nuo Xu, Ahrii Kim
Main category: cs.CL
TL;DR: Systematic comparison of subword tokenization methods (BPE, OBPE, Unigram) for Uralic languages shows OBPE achieves better morphological alignment and POS tagging accuracy, especially for low-resource agglutinative languages.
Details
Motivation: Subword tokenization's impact on NLP performance in morphologically rich and low-resource language families is under-explored, particularly for Uralic languages with varying resource availability and typological diversity.Method: Compared three subword paradigms (Byte Pair Encoding, Overlap BPE, and Unigram Language Model) across six Uralic languages using part-of-speech tagging as a controlled downstream task to evaluate morphological alignment and transfer efficacy.
Result: OBPE consistently achieves stronger morphological alignment and higher tagging accuracy than conventional methods, particularly for Latin-script languages. Gains come from reduced fragmentation in open-class categories and better frequency spectrum balance.
Conclusion: Morphology-sensitive tokenization is a decisive factor for effective cross-lingual transfer in agglutinative, low-resource languages, not just a preprocessing choice.
Abstract: Subword tokenization critically affects Natural Language Processing (NLP) performance, yet its behavior in morphologically rich and low-resource language families remains under-explored. This study systematically compares three subword paradigms – Byte Pair Encoding (BPE), Overlap BPE (OBPE), and Unigram Language Model – across six Uralic languages with varying resource availability and typological diversity. Using part-of-speech (POS) tagging as a controlled downstream task, we show that OBPE consistently achieves stronger morphological alignment and higher tagging accuracy than conventional methods, particularly within the Latin-script group. These gains arise from reduced fragmentation in open-class categories and a better balance across the frequency spectrum. Transfer efficacy further depends on the downstream tagging architecture, interacting with both training volume and genealogical proximity. Taken together, these findings highlight that morphology-sensitive tokenization is not merely a preprocessing choice but a decisive factor in enabling effective cross-lingual transfer for agglutinative, low-resource languages.
[132] CoT is Not the Chain of Truth: An Empirical Internal Analysis of Reasoning LLMs for Fake News Generation
Zhao Tong, Chunlin Gong, Yiping Zhang, Haichao Shi, Qiang Liu, Xingcheng Xu, Shu Wu, Xiao-Yu Zhang
Main category: cs.CL
TL;DR: LLMs can internally propagate unsafe narratives in their Chain-of-Thought reasoning even when they refuse harmful requests, challenging the assumption that refusal implies safety throughout the reasoning process.
Details
Motivation: The paper challenges the common safety assumption that when LLMs refuse harmful requests, their entire reasoning process is safe. The authors investigate whether unsafe narratives can still propagate internally during Chain-of-Thought reasoning even when the final output is a refusal.Method: The authors introduce a unified safety-analysis framework that systematically deconstructs CoT generation across model layers and evaluates individual attention heads using Jacobian-based spectral metrics. They propose three interpretable measures: stability, geometry, and energy to quantify how specific attention heads respond to or embed deceptive reasoning patterns.
Result: Experiments on multiple reasoning-oriented LLMs show that generation risk rises significantly when thinking mode is activated, with critical routing decisions concentrated in only a few contiguous mid-depth layers. The framework successfully identifies specific attention heads responsible for divergence between safe outputs and unsafe internal reasoning.
Conclusion: The work challenges the assumption that refusal implies safety in LLMs and provides a new perspective for understanding and mitigating latent reasoning risks by identifying specific attention mechanisms that propagate unsafe narratives internally.
Abstract: From generating headlines to fabricating news, the Large Language Models (LLMs) are typically assessed by their final outputs, under the safety assumption that a refusal response signifies safe reasoning throughout the entire process. Challenging this assumption, our study reveals that during fake news generation, even when a model rejects a harmful request, its Chain-of-Thought (CoT) reasoning may still internally contain and propagate unsafe narratives. To analyze this phenomenon, we introduce a unified safety-analysis framework that systematically deconstructs CoT generation across model layers and evaluates the role of individual attention heads through Jacobian-based spectral metrics. Within this framework, we introduce three interpretable measures: stability, geometry, and energy to quantify how specific attention heads respond or embed deceptive reasoning patterns. Extensive experiments on multiple reasoning-oriented LLMs show that the generation risk rise significantly when the thinking mode is activated, where the critical routing decisions concentrated in only a few contiguous mid-depth layers. By precisely identifying the attention heads responsible for this divergence, our work challenges the assumption that refusal implies safety and provides a new understanding perspective for mitigating latent reasoning risks.
[133] CAST: Character-and-Scene Episodic Memory for Agents
Kexin Ma, Bojun Li, Yuhua Tang, Liting Sun, Ruochun Jin
Main category: cs.CL
TL;DR: CAST: A Character-and-Scene based episodic memory architecture for AI agents that organizes experiences into 3D scenes (time/place/topic) within character profiles, complemented by graph-based semantic memory.
Details
Motivation: Current agent memory systems focus mainly on semantic recall and treat experiences as simple structures (key-value, vector, graph), which struggle to represent and retrieve coherent episodic memories that capture the who, when, and where of events.Method: Inspired by dramatic theory, CAST constructs 3D scenes organized by time, place, and topic, and organizes them into character profiles that summarize events. This episodic memory is complemented by a graph-based semantic memory, creating a dual memory design.
Result: CAST improves performance by an average of 8.11% F1 and 10.21% J(LLM-as-a-Judge) compared to baselines across various datasets, with particularly strong improvements on open and time-sensitive conversational questions.
Conclusion: The CAST architecture successfully addresses the challenge of representing and retrieving coherent episodic memories in AI agents, demonstrating significant performance improvements over existing memory systems.
Abstract: Episodic memory is a central component of human memory, which refers to the ability to recall coherent events grounded in who, when, and where. However, most agent memory systems only emphasize semantic recall and treat experience as structures such as key-value, vector, or graph, which makes them struggle to represent and retrieve coherent events. To address this challenge, we propose a Character-and-Scene based memory architecture(CAST) inspired by dramatic theory. Specifically, CAST constructs 3D scenes (time/place/topic) and organizes them into character profiles that summarize the events of a character to represent episodic memory. Moreover, CAST complements this episodic memory with a graph-based semantic memory, which yields a robust dual memory design. Experiments demonstrate that CAST has averagely improved 8.11% F1 and 10.21% J(LLM-as-a-Judge) than baselines on various datasets, especially on open and time-sensitive conversational questions.
[134] Language Modeling and Understanding Through Paraphrase Generation and Detection
Jan Philip Wahle
Main category: cs.CL
TL;DR: Thesis proposes decomposing paraphrases into linguistic aspects (paraphrase types) for fine-grained semantic understanding, showing models trained on this approach outperform human baselines in plagiarism detection and duplicate question identification.
Details
Motivation: Current computational language models reduce paraphrasing to binary decisions or single rewrites, obscuring linguistic factors responsible for meaning preservation. The thesis argues that decomposing paraphrases into constituent linguistic aspects offers a more fine-grained and cognitively grounded view of semantic equivalence.Method: Proposes decomposing paraphrases into paraphrase types (constituent linguistic aspects) and training models explicitly on these types rather than binary paraphrase decisions. Demonstrates this approach through experiments in plagiarism detection and duplicate question identification.
Result: Models trained on paraphrase types achieve 89.6% accuracy vs 78.4% human baseline for Wikipedia plagiarism detection, and 66.5% vs 55.7% for arXiv scientific paper plagiarism. Also improves performance on Quora duplicate question identification over binary pair training.
Conclusion: Decomposing paraphrases into linguistic aspects provides better semantic understanding in computational models, enabling stronger performance on paraphrase-related tasks and downstream applications compared to traditional binary approaches.
Abstract: Language enables humans to share knowledge, reason about the world, and pass on strategies for survival and innovation across generations. At the heart of this process is not just the ability to communicate but also the remarkable flexibility in how we can express ourselves. We can express the same thoughts in virtually infinite ways using different words and structures - this ability to rephrase and reformulate expressions is known as paraphrase. Modeling paraphrases is a keystone to meaning in computational language models; being able to construct different variations of texts that convey the same meaning or not shows strong abilities of semantic understanding. If computational language models are to represent meaning, they must understand and control the different aspects that construct the same meaning as opposed to different meanings at a fine granularity. Yet most existing approaches reduce paraphrasing to a binary decision between two texts or to producing a single rewrite of a source, obscuring which linguistic factors are responsible for meaning preservation. In this thesis, I propose that decomposing paraphrases into their constituent linguistic aspects (paraphrase types) offers a more fine-grained and cognitively grounded view of semantic equivalence. I show that even advanced machine learning models struggle with this task. Yet, when explicitly trained on paraphrase types, models achieve stronger performance on related paraphrase tasks and downstream applications. For example, in plagiarism detection, language models trained on paraphrase types surpass human baselines: 89.6% accuracy compared to 78.4% for plagiarism cases from Wikipedia, and 66.5% compared to 55.7% for plagiarism of scientific papers from arXiv. In identifying duplicate questions on Quora, models trained with paraphrase types improve over models trained on binary pairs. Furthermore, I demonstrate that…
[135] Who is the richest club in the championship? Detecting and Rewriting Underspecified Questions Improve QA Performance
Yunchong Huang, Gianni Barlacchi, Sandro Pezzelle
Main category: cs.CL
TL;DR: LLMs struggle with underspecified questions in QA benchmarks, with 16-50% of questions being ambiguous; rewriting them to be fully specified significantly improves performance, showing underspecification is a major confound in evaluation.
Details
Motivation: Standard QA benchmarks remain unsolved despite LLM advances, possibly due to underspecified questions that lack unique interpretation without additional context, creating evaluation confounds.Method: Developed an LLM-based classifier to identify underspecified questions in QA datasets, then conducted controlled rewriting experiments to transform underspecified questions into fully specified variants while keeping gold answers fixed.
Result: Found 16% to over 50% of benchmark questions are underspecified; LLMs perform significantly worse on them; QA performance consistently improves when questions are rewritten to be fully specified.
Conclusion: Underspecification is a major confound in QA evaluation, and many apparent LLM failures stem from ambiguous questions rather than model limitations, highlighting need for clearer benchmark design.
Abstract: Large language models (LLMs) perform well on well-posed questions, yet standard question-answering (QA) benchmarks remain far from solved. We argue that this gap is partly due to underspecified questions - queries whose interpretation cannot be uniquely determined without additional context. To test this hypothesis, we introduce an LLM-based classifier to identify underspecified questions and apply it to several widely used QA datasets, finding that 16% to over 50% of benchmark questions are underspecified and that LLMs perform significantly worse on them. To isolate the effect of underspecification, we conduct a controlled rewriting experiment that serves as an upper-bound analysis, rewriting underspecified questions into fully specified variants while holding gold answers fixed. QA performance consistently improves under this setting, indicating that many apparent QA failures stem from question underspecification rather than model limitations. Our findings highlight underspecification as an important confound in QA evaluation and motivate greater attention to question clarity in benchmark design.
cs.CV
[136] Beyond Ground: Map-Free LiDAR Relocalization for UAVs
Hengyu Mu, Jianshi Wu, Yuxin Guo, XianLian Lin, Qingyong Hu, Chenglu Wen, Cheng Wang
Main category: cs.CV
TL;DR: MAILS: A map-free LiDAR relocalization framework for UAVs using locality-preserving attention and coordinate-independent features to handle sparse point clouds, yaw rotations, and altitude variations.
Details
Motivation: Existing LiDAR relocalization methods are designed for autonomous driving and perform poorly in UAV scenarios due to sparse point clouds, substantial yaw rotations, altitude variations, and irregular flight trajectories. There's also a lack of appropriate datasets capturing real UAV flight characteristics.Method: Proposes MAILS framework with: 1) Locality-Preserving Sliding Window Attention module for extracting locally discriminative geometric features from sparse point clouds; 2) Coordinate-independent feature initialization module; 3) Locally invariant positional encoding mechanism for robustness against yaw rotations and altitude variations; 4) A new large-scale LiDAR localization dataset for UAVs with four scenes and various flight trajectories.
Result: Extensive experiments show the method achieves satisfactory localization precision and consistently outperforms existing techniques by a significant margin. The framework handles UAV-specific challenges effectively.
Conclusion: MAILS provides an effective map-free LiDAR relocalization solution for UAVs, addressing key challenges like sparse point clouds, yaw rotations, and altitude variations. The new dataset enables better evaluation of UAV relocalization methods under realistic conditions.
Abstract: Localization is a fundamental capability in unmanned aerial vehicle (UAV) systems. Map-free LiDAR relocalization offers an effective solution for achieving high-precision positioning in environments with weak or unavailable GNSS signals. However, existing LiDAR relocalization methods are primarily tailored to autonomous driving, exhibiting significantly degraded accuracy in UAV scenarios. In this paper, we propose MAILS, a novel map-free LiDAR relocalization framework for UAVs. A Locality-Preserving Sliding Window Attention module is first introduced to extract locally discriminative geometric features from sparse point clouds. To handle substantial yaw rotations and altitude variations encountered during UAV flight, we then design a coordinate-independent feature initialization module and a locally invariant positional encoding mechanism, which together significantly enhance the robustness of feature extraction. Furthermore, existing LiDAR-based relocalization datasets fail to capture real-world UAV flight characteristics, such as irregular trajectories and varying altitudes. To address this gap, we construct a large-scale LiDAR localization dataset for UAVs, which comprises four scenes and various flight trajectories, designed to evaluate UAV relocalization performance under realistic conditions. Extensive experiments demonstrate that our method achieves satisfactory localization precision and consistently outperforms existing techniques by a significant margin. Our code and dataset will be released soon.
[137] Explanatory Interactive Machine Learning for Bias Mitigation in Visual Gender Classification
Nathanya Satriani, Djordje Slijepčević, Markus Schedl, Matthias Zeppelzauer
Main category: cs.CV
TL;DR: XIL methods (CAIPI, RRR, and hybrid) help reduce bias in visual classifiers by allowing users to provide feedback on model explanations, improving fairness in gender classification while maintaining or slightly improving accuracy.
Details
Motivation: To address bias and spurious correlations in visual classifiers, particularly in gender classification where data bias is common, by leveraging explanatory interactive learning to guide models toward relevant features and away from biased patterns.Method: Investigates three XIL strategies: CAIPI, Right for the Right Reasons (RRR), and a novel hybrid approach combining both. Uses Gradient-weighted Class Activation Mapping (GradCAM) and Bounded Logit Attention (BLA) for explanation generation and evaluation via segmentation mask comparisons.
Result: XIL methods effectively guide models to focus on relevant image features (especially CAIPI) and reduce model bias by balancing misclassification rates between male and female predictions. CAIPI shows potential to improve classification accuracy while other methods cause slight performance decreases.
Conclusion: XIL methods demonstrate significant potential for improving fairness and transparency in visual classifiers, with CAIPI being particularly promising for both bias reduction and maintaining/improving accuracy in gender classification tasks.
Abstract: Explanatory interactive learning (XIL) enables users to guide model training in machine learning (ML) by providing feedback on the model’s explanations, thereby helping it to focus on features that are relevant to the prediction from the user’s perspective. In this study, we explore the capability of this learning paradigm to mitigate bias and spurious correlations in visual classifiers, specifically in scenarios prone to data bias, such as gender classification. We investigate two methodologically different state-of-the-art XIL strategies, i.e., CAIPI and Right for the Right Reasons (RRR), as well as a novel hybrid approach that combines both strategies. The results are evaluated quantitatively by comparing segmentation masks with explanations generated using Gradient-weighted Class Activation Mapping (GradCAM) and Bounded Logit Attention (BLA). Experimental results demonstrate the effectiveness of these methods in (i) guiding ML models to focus on relevant image features, particularly when CAIPI is used, and (ii) reducing model bias (i.e., balancing the misclassification rates between male and female predictions). Our analysis further supports the potential of XIL methods to improve fairness in gender classifiers. Overall, the increased transparency and fairness obtained by XIL leads to slight performance decreases with an exception being CAIPI, which shows potential to even improve classification accuracy.
[138] COOPERTRIM: Adaptive Data Selection for Uncertainty-Aware Cooperative Perception
Shilpa Mukhopadhyay, Amit Roy-Chowdhury, Hang Qiu
Main category: cs.CV
TL;DR: COOPERTRIM: Adaptive feature selection framework for cooperative perception that reduces bandwidth by 80% while maintaining accuracy by exploiting temporal continuity to identify dynamic features.
Details
Motivation: Cooperative perception requires sharing sensor data between autonomous agents, but limited bandwidth conflicts with rich sensor information. Current approaches still stress wireless technologies, requiring a fundamental solution that exploits temporal continuity to reduce redundant transmissions.Method: Proposes COOPERTRIM framework with: 1) Conformal temporal uncertainty metric to gauge feature relevance by identifying dynamic vs static information, 2) Data-driven mechanism to dynamically determine sharing quantity based on environment complexity, 3) Temporal awareness to adapt sharing to environmental dynamics.
Result: Achieves up to 80.28% bandwidth reduction for semantic segmentation and 72.52% for 3D detection while maintaining comparable accuracy. Improves IoU by up to 45.54% with 72% less bandwidth. Combined with compression, reduces bandwidth to 1.46% without compromising IoU.
Conclusion: COOPERTRIM effectively addresses bandwidth limitations in cooperative perception by exploiting temporal continuity, demonstrating adaptability to environmental dynamics, localization error, and communication latency, enabling practical real-world deployment.
Abstract: Cooperative perception enables autonomous agents to share encoded representations over wireless communication to enhance each other’s live situational awareness. However, the tension between the limited communication bandwidth and the rich sensor information hinders its practical deployment. Recent studies have explored selection strategies that share only a subset of features per frame while striving to keep the performance on par. Nevertheless, the bandwidth requirement still stresses current wireless technologies. To fundamentally ease the tension, we take a proactive approach, exploiting the temporal continuity to identify features that capture environment dynamics, while avoiding repetitive and redundant transmission of static information. By incorporating temporal awareness, agents are empowered to dynamically adapt the sharing quantity according to environment complexity. We instantiate this intuition into an adaptive selection framework, COOPERTRIM, which introduces a novel conformal temporal uncertainty metric to gauge feature relevance, and a data-driven mechanism to dynamically determine the sharing quantity. To evaluate COOPERTRIM, we take semantic segmentation and 3D detection as example tasks. Across multiple open-source cooperative segmentation and detection models, COOPERTRIM achieves up to 80.28% and 72.52% bandwidth reduction respectively while maintaining a comparable accuracy. Relative to other selection strategies, COOPERTRIM also improves IoU by as much as 45.54% with up to 72% less bandwidth. Combined with compression strategies, COOPERTRIM can further reduce bandwidth usage to as low as 1.46% without compromising IoU performance. Qualitative results show COOPERTRIM gracefully adapts to environmental dynamics, localization error, and communication latency, demonstrating flexibility and paving the way for real-world deployment.
[139] GOT-JEPA: Generic Object Tracking with Model Adaptation and Occlusion Handling using Joint-Embedding Predictive Architecture
Shih-Fang Chen, Jun-Cheng Chen, I-Hong Jhuo, Yen-Yu Lin
Main category: cs.CV
TL;DR: GOT-JEPA extends JEPA framework from image feature prediction to tracking model prediction for better generalization and occlusion handling in object tracking, with OccuSolver enhancing occlusion perception through iterative visibility refinement.
Details
Motivation: Current object trackers lack robustness in unseen scenarios and have coarse occlusion reasoning, failing to match human visual system's ability to adapt to changes and reason about occlusion at fine granularity.Method: GOT-JEPA uses model-predictive pretraining where teacher generates pseudo-tracking models from clean frames and student learns to predict same models from corrupted frames. OccuSolver adds point-centric visibility estimation and iterative occlusion-pattern refinement using object priors.
Result: Extensive evaluations on seven benchmarks show the method effectively enhances tracker generalization and robustness, improving performance in dynamic environments with occlusions and distractors.
Conclusion: The proposed framework addresses limitations in generalization and occlusion perception for object tracking, bridging the gap between current trackers and human visual capabilities.
Abstract: The human visual system tracks objects by integrating current observations with previously observed information, adapting to target and scene changes, and reasoning about occlusion at fine granularity. In contrast, recent generic object trackers are often optimized for training targets, which limits robustness and generalization in unseen scenarios, and their occlusion reasoning remains coarse, lacking detailed modeling of occlusion patterns. To address these limitations in generalization and occlusion perception, we propose GOT-JEPA, a model-predictive pretraining framework that extends JEPA from predicting image features to predicting tracking models. Given identical historical information, a teacher predictor generates pseudo-tracking models from a clean current frame, and a student predictor learns to predict the same pseudo-tracking models from a corrupted version of the current frame. This design provides stable pseudo supervision and explicitly trains the predictor to produce reliable tracking models under occlusions, distractors, and other adverse observations, improving generalization to dynamic environments. Building on GOT-JEPA, we further propose OccuSolver to enhance occlusion perception for object tracking. OccuSolver adapts a point-centric point tracker for object-aware visibility estimation and detailed occlusion-pattern capture. Conditioned on object priors iteratively generated by the tracker, OccuSolver incrementally refines visibility states, strengthens occlusion handling, and produces higher-quality reference labels that progressively improve subsequent model predictions. Extensive evaluations on seven benchmarks show that our method effectively enhances tracker generalization and robustness.
[140] Evaluating the Impact of Post-Training Quantization on Reliable VQA with Multimodal LLMs
Paul Jonas Kurz, Tobias Jan Wieczorek, Mohamed A. Abdelsalam, Rahaf Aljundi, Marcus Rohrbach
Main category: cs.CV
TL;DR: Systematic study of how post-training quantization affects both accuracy and reliability in multimodal LLMs for VQA, showing data-aware methods and confidence estimators can mitigate reliability degradation.
Details
Motivation: MLLMs face dual challenges: overconfidence in incorrect answers and large size limiting edge deployment. Need to understand how compression (quantization) affects both accuracy and reliability in multimodal settings.Method: Evaluate Qwen2-VL-7B and Idefics3-8B with PTQ using data-free (HQQ) and data-aware (MBQ) methods across multiple bit widths. Adapt Selector confidence estimator for quantized multimodal settings and test robustness across quantization levels and OOD scenarios.
Result: PTQ degrades both accuracy and reliability. Data-aware methods soften the effect. Selector substantially mitigates reliability impact. Int4 MBQ + Selector achieves best efficiency-reliability trade-off, closing in on uncompressed performance at ~75% less memory.
Conclusion: First systematic study linking quantization and reliability in multimodal settings. Shows data-aware quantization methods combined with confidence estimators can maintain reliability while achieving significant compression for edge deployment.
Abstract: Multimodal Large Language Models (MLLM) are increasingly deployed in domains where both reliability and efficiency are critical. However, current models remain overconfident, producing highly certain but incorrect answers. At the same time, their large size limits deployment on edge devices, necessitating compression. We study the intersection of these two challenges by analyzing how Post-Training Quantization (PTQ) compression affects both accuracy and reliability in Visual Question Answering (VQA). We evaluate two MLLMs, Qwen2-VL-7B and Idefics3-8B, quantized with data-free (HQQ) and data-aware (MBQ) methods across multiple bit widths. To counteract the reduction in reliability caused by quantization, we adapt the Selector confidence estimator for quantized multimodal settings and test its robustness across various quantization levels and out-of-distribution (OOD) scenarios. We find that PTQ degrades both accuracy and reliability. Data-aware methods soften the effect thereof. The Selector substantially mitigates the reliability impact. The combination of int4 MBQ and the Selector achieves the best efficiency-reliability trade-off, closing in on uncompressed performance at approx. 75% less memory demand. Overall, we present the first systematic study linking quantization and reliability in multimodal settings.
[141] NutVLM: A Self-Adaptive Defense Framework against Full-Dimension Attacks for Vision Language Models in Autonomous Driving
Xiaoxu Peng, Dong Zhou, Jianwen Zhang, Guanghui Sun, Anh Tu Ngo, Anupam Chattopadhyay
Main category: cs.CV
TL;DR: NutVLM is a self-adaptive defense framework for Vision Language Models in autonomous driving that detects and mitigates adversarial threats through three-way classification and expert-guided prompt tuning.
Details
Motivation: Vision Language Models in autonomous driving are vulnerable to adversarial threats (local patches and global perturbations), but existing defense methods are limited and struggle to balance robustness with clean-sample performance.Method: Uses NutNet++ for three-way classification (benign, local patches, global perturbations). Local threats are purified via grayscale masking, while global perturbations trigger Expert-guided Adversarial Prompt Tuning (EAPT) that generates corrective driving prompts via gradient-based latent optimization instead of full-model fine-tuning.
Result: On the Dolphins benchmark, NutVLM achieves 4.89% improvement in overall metrics including Accuracy, Language Score, and GPT Score.
Conclusion: NutVLM provides a scalable security solution for intelligent transportation systems by securing the entire perception-decision lifecycle of VLMs against adversarial threats.
Abstract: Vision Language Models (VLMs) have advanced perception in autonomous driving (AD), but they remain vulnerable to adversarial threats. These risks range from localized physical patches to imperceptible global perturbations. Existing defense methods for VLMs remain limited and often fail to reconcile robustness with clean-sample performance. To bridge these gaps, we propose NutVLM, a comprehensive self-adaptive defense framework designed to secure the entire perception-decision lifecycle. Specifically, we first employ NutNet++ as a sentinel, which is a unified detection-purification mechanism. It identifies benign samples, local patches, and global perturbations through three-way classification. Subsequently, localized threats are purified via efficient grayscale masking, while global perturbations trigger Expert-guided Adversarial Prompt Tuning (EAPT). Instead of the costly parameter updates of full-model fine-tuning, EAPT generates “corrective driving prompts” via gradient-based latent optimization and discrete projection. These prompts refocus the VLM’s attention without requiring exhaustive full-model retraining. Evaluated on the Dolphins benchmark, our NutVLM yields a 4.89% improvement in overall metrics (e.g., Accuracy, Language Score, and GPT Score). These results validate NutVLM as a scalable security solution for intelligent transportation. Our code is available at https://github.com/PXX/NutVLM.
[142] VisPhyWorld: Probing Physical Reasoning via Code-Driven Video Reconstruction
Jiarong Liang, Max Ku, Ka-Hei Hui, Ping Nie, Wenhu Chen
Main category: cs.CV
TL;DR: VisPhyWorld is an execution-based framework that evaluates MLLMs’ physical reasoning by requiring them to generate executable simulator code from visual observations, making world representations inspectable and falsifiable.
Details
Motivation: Existing benchmarks for evaluating MLLMs' physical reasoning rely on recognition-style protocols (VQA, VoE) that can be answered without explicit physical hypotheses, making it hard to assess genuine physical understanding.Method: Proposes VisPhyWorld framework requiring models to generate executable simulator code from visual observations, and VisPhyBench with 209 scenes from 108 physical templates to evaluate appearance reconstruction and physically plausible motion reproduction.
Result: Pipeline produces valid reconstructed videos in 97.7% of cases. State-of-the-art MLLMs show strong semantic scene understanding but struggle with accurate physical parameter inference and consistent physical dynamics simulation.
Conclusion: Execution-based evaluation reveals limitations in current MLLMs’ physical reasoning capabilities, highlighting the need for frameworks that separate physical reasoning from rendering and make world representations testable.
Abstract: Evaluating whether Multimodal Large Language Models (MLLMs) genuinely reason about physical dynamics remains challenging. Most existing benchmarks rely on recognition-style protocols such as Visual Question Answering (VQA) and Violation of Expectation (VoE), which can often be answered without committing to an explicit, testable physical hypothesis. We propose VisPhyWorld, an execution-based framework that evaluates physical reasoning by requiring models to generate executable simulator code from visual observations. By producing runnable code, the inferred world representation is directly inspectable, editable, and falsifiable. This separates physical reasoning from rendering. Building on this framework, we introduce VisPhyBench, comprising 209 evaluation scenes derived from 108 physical templates and a systematic protocol that evaluates how well models reconstruct appearance and reproduce physically plausible motion. Our pipeline produces valid reconstructed videos in 97.7% on the benchmark. Experiments show that while state-of-the-art MLLMs achieve strong semantic scene understanding, they struggle to accurately infer physical parameters and to simulate consistent physical dynamics.
[143] MFN Decomposition and Related Metrics for High-Resolution Range Profiles Generative Models
Edwyn Brient, Santiago Velasco-Forero, Rami Kassab
Main category: cs.CV
TL;DR: Proposes two physics-based metrics for evaluating generated radar HRRP data by decomposing it into mask, features, and noise components, addressing limitations of black-box classification evaluation methods.
Details
Motivation: Current evaluation methods for generated HRRP (High-Resolution Range Profile) data rely on black-box classification models that lack explainability and multi-level evaluation capabilities. There's a need for more interpretable evaluation metrics that align with the physical properties of radar data.Method: Decomposes HRRP data into three components: mask, features, and noise. Based on this decomposition, proposes two physics-based metrics that leverage the physical interpretation of radar data. Evaluates these metrics using an expensive dataset on a challenging task.
Result: Demonstrates the discriminative ability of the proposed metrics, showing they can effectively evaluate generated HRRP data in ways that black-box classification models cannot.
Conclusion: The proposed physics-based decomposition and metrics provide more interpretable and multi-level evaluation of generated HRRP data compared to existing black-box classification approaches, offering better alignment with the physical properties of radar signals.
Abstract: High-resolution range profile (HRRP ) data are in vogue in radar automatic target recognition (RATR). With the interest in classifying models using HRRP, filling gaps in datasets using generative models has recently received promising contributions. Evaluating generated data is a challenging topic, even for explicit data like face images. However, the evaluation methods used in the state-ofthe-art of HRRP generation rely on classification models. Such models, called ‘‘black-box’’, do not allow either explainability on generated data or multi-level evaluation. This work focuses on decomposing HRRP data into three components: the mask, the features, and the noise. Using this decomposition, we propose two metrics based on the physical interpretation of those data. We take profit from an expensive dataset to evaluate our metrics on a challenging task and demonstrate the discriminative ability of those.
[144] NeRV360: Neural Representation for 360-Degree Videos with a Viewport Decoder
Daichi Arai, Kyohei Unno, Yasuko Sugito, Yuichi Kusakabe
Main category: cs.CV
TL;DR: NeRV360: An end-to-end framework for efficient 360-degree video compression using implicit neural representations that decodes only user-selected viewports instead of entire panoramic frames.
Details
Motivation: Current implicit neural representations for videos (NeRV) applied to high-resolution 360-degree videos suffer from high memory usage and slow decoding, making real-time applications impractical. There's a need for more efficient decoding of only the relevant portions of panoramic content.Method: NeRV360 integrates viewport extraction directly into the decoding process and introduces a spatial-temporal affine transform module for conditional decoding based on viewpoint and time. This allows selective decoding of only the user-selected viewport rather than reconstructing the entire panoramic frame.
Result: On 6K-resolution videos, NeRV360 achieves 7-fold reduction in memory consumption and 2.5-fold increase in decoding speed compared to HNeRV (a prior work), while delivering better image quality in objective metrics.
Conclusion: NeRV360 provides an efficient solution for 360-degree video compression using implicit neural representations by focusing decoding on relevant viewports, enabling practical real-time applications with significant improvements in memory and speed.
Abstract: Implicit neural representations for videos (NeRV) have shown strong potential for video compression. However, applying NeRV to high-resolution 360-degree videos causes high memory usage and slow decoding, making real-time applications impractical. We propose NeRV360, an end-to-end framework that decodes only the user-selected viewport instead of reconstructing the entire panoramic frame. Unlike conventional pipelines, NeRV360 integrates viewport extraction into decoding and introduces a spatial-temporal affine transform module for conditional decoding based on viewpoint and time. Experiments on 6K-resolution videos show that NeRV360 achieves a 7-fold reduction in memory consumption and a 2.5-fold increase in decoding speed compared to HNeRV, a representative prior work, while delivering better image quality in terms of objective metrics.
[145] Conditional Generative Models for High-Resolution Range Profiles: Capturing Geometry-Driven Trends in a Large-Scale Maritime Dataset
Edwyn Brient, Santiago Velasco-Forero, Rami Kassab
Main category: cs.CV
TL;DR: This paper studies high-resolution range profile (HRRP) generation for radar automatic target recognition, focusing on maritime targets and conditioning on geometric variables like ship dimensions and aspect angle.
Details
Motivation: HRRPs are useful for fast onboard radar processing but suffer from sensitivity to acquisition conditions, limiting robustness across operational scenarios. Prior work has been constrained by small, specific datasets, so the authors study HRRP synthesis on a large-scale maritime database to address coastal surveillance variability.Method: The authors analyze HRRP synthesis on a large-scale maritime database and identify that the fundamental scenario drivers are geometric: ship dimensions and desired aspect angle. They condition generative models on these variables and train them to synthesize radar signatures.
Result: The synthesized signatures reproduce the expected line-of-sight geometric trend observed in real data, demonstrating that acquisition geometry plays a central role in robust HRRP generation.
Conclusion: The study highlights the importance of acquisition geometry for robust HRRP generation and shows that conditioning on geometric variables (ship dimensions and aspect angle) enables effective synthesis of radar signatures that match real-world trends.
Abstract: High-resolution range profiles (HRRPs) enable fast onboard processing for radar automatic target recognition, but their strong sensitivity to acquisition conditions limits robustness across operational scenarios. Conditional HRRP generation can mitigate this issue, yet prior studies are constrained by small, highly specific datasets. We study HRRP synthesis on a largescale maritime database representative of coastal surveillance variability. Our analysis indicates that the fundamental scenario drivers are geometric: ship dimensions and the desired aspect angle. Conditioning on these variables, we train generative models and show that the synthesized signatures reproduce the expected line-of-sight geometric trend observed in real data. These results highlight the central role of acquisition geometry for robust HRRP generation.
[146] Effect of Convolutional Depth on Image Recognition Performance: VGG vs. ResNet vs. GoogLeNet
Manfred M. Fischer, Joshua Pitts
Main category: cs.CV
TL;DR: Depth alone doesn’t guarantee better performance; effective depth (enabled by architectural mechanisms like residuals/inception) matters more than nominal depth for accuracy, stability, and efficiency.
Details
Motivation: To understand why deeper convolutional networks don't always yield better performance, and to isolate how depth influences classification accuracy, convergence behavior, and computational efficiency through a controlled study of different architectures.Method: Conducted a controlled comparative study of three canonical CNN architectures (VGG, ResNet, GoogLeNet) with standardized training protocols, distinguishing between nominal depth (number of layers) and effective depth (actual depth manifested during training).
Result: Plain deep networks (like VGG) show early accuracy saturation and optimization instability, while residual (ResNet) and inception-based (GoogLeNet) architectures consistently translate additional depth into improved accuracy at lower effective depth with better accuracy-compute trade-offs.
Conclusion: Effective depth, not nominal depth, is the key factor governing depth’s role as a productive scaling dimension in convolutional networks, enabled by architectural mechanisms that constrain how depth manifests during training.
Abstract: Increasing convolutional depth has been central to advances in image recognition, yet deeper networks do not uniformly yield higher accuracy, stable optimization, or efficient computation. We present a controlled comparative study of three canonical convolutional neural network architectures - VGG, ResNet, and GoogLeNet - to isolate how depth influences classification performance, convergence behavior, and computational efficiency. By standardizing training protocols and explicitly distinguishing between nominal and effective depth, we show that the benefits of depth depend critically on architectural mechanisms that constrain its effective manifestation during training rather than on nominal depth alone. Although plain deep networks exhibit early accuracy saturation and optimization instability, residual and inception-based architectures consistently translate additional depth into improved accuracy at lower effective depth and favorable accuracy-compute trade-offs. These findings demonstrate that effective depth, not nominal depth, is the operative quantity governing depth’s role as a productive scaling dimension in convolutional networks.
[147] KidMesh: Computational Mesh Reconstruction for Pediatric Congenital Hydronephrosis Using Deep Neural Networks
Haoran Sun, Zhanpeng Zhu, Anguo Zhang, Bo Liu, Zhaohua Lin, Liqin Huang, Mingjing Yang, Lei Liu, Shan Lin, Wangbin Ding
Main category: cs.CV
TL;DR: KidMesh: An end-to-end deep learning method for automatically reconstructing congenital hydronephrosis meshes directly from MRU images without requiring post-processing steps.
Details
Motivation: Existing voxel-based segmentation methods for pediatric congenital hydronephrosis focus only on morphological features and require complex post-processing for functional assessments like urodynamic simulations. There's a need for direct mesh reconstruction from MRU images to enable functional analysis.Method: KidMesh extracts feature maps from MRU images, converts them to feature vertices through grid sampling, and deforms a template mesh according to these feature vertices. The method includes a novel training schema that doesn’t require accurate mesh-level annotations, which are difficult to obtain due to sparsely sampled MRU slices.
Result: KidMesh reconstructs CH meshes in 0.4 seconds on average, achieving comparable performance to conventional methods without post-processing. Reconstructed meshes have no self-intersections, with only 3.7% and 0.2% of vertices having error distances exceeding 3.2mm and 6.4mm respectively. After rasterization, Dice score of 0.86 against manual masks. Meshes can be used for renal urine flow simulations.
Conclusion: KidMesh provides an efficient end-to-end solution for direct mesh reconstruction from MRU images, enabling functional urodynamic assessments without complex post-processing, with potential clinical value for congenital hydronephrosis diagnosis and treatment planning.
Abstract: Pediatric congenital hydronephrosis (CH) is a common urinary tract disorder, primarily caused by obstruction at the renal pelvis-ureter junction. Magnetic resonance urography (MRU) can visualize hydronephrosis, including renal pelvis and calyces, by utilizing the natural contrast provided by water. Existing voxel-based segmentation approaches can extract CH regions from MRU, facilitating disease diagnosis and prognosis. However, these segmentation methods predominantly focus on morphological features, such as size, shape, and structure. To enable functional assessments, such as urodynamic simulations, external complex post-processing steps are required to convert these results into mesh-level representations. To address this limitation, we propose an end-to-end method based on deep neural networks, namely KidMesh, which could automatically reconstruct CH meshes directly from MRU. Generally, KidMesh extracts feature maps from MRU images and converts them into feature vertices through grid sampling. It then deforms a template mesh according to these feature vertices to generate the specific CH meshes of MRU images. Meanwhile, we develop a novel schema to train KidMesh without relying on accurate mesh-level annotations, which are difficult to obtain due to the sparsely sampled MRU slices. Experimental results show that KidMesh could reconstruct CH meshes in an average of 0.4 seconds, and achieve comparable performance to conventional methods without requiring post-processing. The reconstructed meshes exhibited no self-intersections, with only 3.7% and 0.2% of the vertices having error distances exceeding 3.2mm and 6.4mm, respectively. After rasterization, these meshes achieved a Dice score of 0.86 against manually delineated CH masks. Furthermore, these meshes could be used in renal urine flow simulations, providing valuable urodynamic information for clinical practice.
[148] DriveMamba: Task-Centric Scalable State Space Model for Efficient End-to-End Autonomous Driving
Haisheng Su, Wei Wu, Feixiang Song, Junjie Zhang, Zhenjie Yang, Junchi Yan
Main category: cs.CV
TL;DR: DriveMamba: A Task-Centric Scalable paradigm for End-to-End Autonomous Driving using Mamba architecture for efficient long-context sequential token modeling with linear complexity.
Details
Motivation: Current E2E-AD systems use sequential modular designs (perception-prediction-planning) with separable Transformer decoders, causing information loss, cumulative errors, and inefficient quadratic-complexity attention. Need more flexible relation modeling and efficient spatiotemporal processing.Method: Proposes DriveMamba with a single-stage Unified Mamba decoder that integrates dynamic task relation modeling, implicit view correspondence learning, and long-term temporal fusion. Converts image features and task outputs to token-level sparse representations sorted by 3D positions, uses linear-complexity Mamba for efficient long-context modeling, and employs bidirectional trajectory-guided “local-to-global” scan for spatial locality preservation.
Result: Extensive experiments on nuScenes and Bench2Drive datasets demonstrate superiority, generalizability, and great efficiency of DriveMamba compared to existing approaches.
Conclusion: DriveMamba provides an efficient, scalable paradigm for E2E-AD that overcomes limitations of sequential modular designs through unified Mamba-based architecture with linear complexity and better task relation modeling.
Abstract: Recent advances towards End-to-End Autonomous Driving (E2E-AD) have been often devoted on integrating modular designs into a unified framework for joint optimization e.g. UniAD, which follow a sequential paradigm (i.e., perception-prediction-planning) based on separable Transformer decoders and rely on dense BEV features to encode scene representations. However, such manual ordering design can inevitably cause information loss and cumulative errors, lacking flexible and diverse relation modeling among different modules and sensors. Meanwhile, insufficient training of image backbone and quadratic-complexity of attention mechanism also hinder the scalability and efficiency of E2E-AD system to handle spatiotemporal input. To this end, we propose DriveMamba, a Task-Centric Scalable paradigm for efficient E2E-AD, which integrates dynamic task relation modeling, implicit view correspondence learning and long-term temporal fusion into a single-stage Unified Mamba decoder. Specifically, both extracted image features and expected task outputs are converted into token-level sparse representations in advance, which are then sorted by their instantiated positions in 3D space. The linear-complexity operator enables efficient long-context sequential token modeling to capture task-related inter-dependencies simultaneously. Additionally, a bidirectional trajectory-guided “local-to-global” scan method is designed to preserve spatial locality from ego-perspective, thus facilitating the ego-planning. Extensive experiments conducted on nuScenes and Bench2Drive datasets demonstrate the superiority, generalizability and great efficiency of DriveMamba.
[149] Spectral Collapse in Diffusion Inversion
Nicolas Bourriez, Alexandre Verine, Auguste Genovesio
Main category: cs.CV
TL;DR: OVG (Orthogonal Variance Guidance) addresses spectral collapse in diffusion inversion for image-to-image translation by correcting ODE dynamics to restore Gaussian noise variance in the null-space of structural gradients.
Details
Motivation: Standard deterministic diffusion inversion (DDIM) fails when source domain is spectrally sparse compared to target domain (e.g., super-resolution, sketch-to-image), causing spectral collapse where recovered latents don't follow expected Gaussian distribution, leading to oversmoothed, texture-poor generations.Method: Proposes Orthogonal Variance Guidance (OVG), an inference-time method that corrects ODE dynamics to enforce theoretical Gaussian noise magnitude within the null-space of the structural gradient, resolving the structure-texture trade-off.
Result: Extensive experiments on microscopy super-resolution (BBBC021) and sketch-to-image (Edges2Shoes) demonstrate OVG effectively restores photorealistic textures while preserving structural fidelity.
Conclusion: OVG successfully addresses spectral collapse in conditional diffusion inversion for cross-domain image translation, enabling high-quality texture generation while maintaining structural accuracy.
Abstract: Conditional diffusion inversion provides a powerful framework for unpaired image-to-image translation. However, we demonstrate through an extensive analysis that standard deterministic inversion (e.g. DDIM) fails when the source domain is spectrally sparse compared to the target domain (e.g., super-resolution, sketch-to-image). In these contexts, the recovered latent from the input does not follow the expected isotropic Gaussian distribution. Instead it exhibits a signal with lower frequencies, locking target sampling to oversmoothed and texture-poor generations. We term this phenomenon spectral collapse. We observe that stochastic alternatives attempting to restore the noise variance tend to break the semantic link to the input, leading to structural drift. To resolve this structure-texture trade-off, we propose Orthogonal Variance Guidance (OVG), an inference-time method that corrects the ODE dynamics to enforce the theoretical Gaussian noise magnitude within the null-space of the structural gradient. Extensive experiments on microscopy super-resolution (BBBC021) and sketch-to-image (Edges2Shoes) demonstrate that OVG effectively restores photorealistic textures while preserving structural fidelity.
[150] Progressive Contrast Registration for High-Fidelity Bidirectional Photoacoustic Microscopy Alignment
Jiahao Qin
Main category: cs.CV
TL;DR: PCReg-Net is a deep learning framework for aligning forward and backward scan lines in bidirectional OR-PAM imaging, using progressive contrast-guided registration with four lightweight modules to achieve high-fidelity alignment and real-time performance.
Details
Motivation: Bidirectional raster scanning in high-speed OR-PAM doubles imaging speed but introduces coupled domain shift and geometric misalignment between forward and backward scan lines. Existing methods based on brightness constancy assumptions achieve limited alignment quality (NCC ≤ 0.96), necessitating better registration techniques.Method: PCReg-Net uses a progressive contrast-guided registration framework with four modules: (1) registration U-Net for coarse alignment, (2) reference feature extractor capturing multi-scale structural cues, (3) contrast module identifying residual misalignment by comparing coarse-registered and reference features, and (4) refinement U-Net with feature injection for high-fidelity output. Also proposes Temporal NCC (TNCC) and Temporal NCC Gap (TNCG) for reference-free evaluation.
Result: On OR-PAM-Reg-4K dataset (432 test samples), PCReg-Net achieves NCC of 0.983, SSIM of 0.982, and PSNR of 46.96 dB, surpassing state-of-the-art by over 14 dB at real-time speed.
Conclusion: PCReg-Net effectively addresses alignment challenges in bidirectional OR-PAM imaging through a progressive contrast-guided framework, achieving superior registration quality and temporal consistency while maintaining real-time performance.
Abstract: High-speed optical-resolution photoacoustic microscopy (OR-PAM) with bidirectional raster scanning doubles imaging speed but introduces coupled domain shift and geometric misalignment between forward and backward scan lines. Existing methods, constrained by brightness constancy assumptions, achieve limited alignment quality (NCC~$\leq 0.96$). We propose PCReg-Net, a progressive contrast-guided registration framework that performs coarse-to-fine alignment through four lightweight modules: (1)~a registration U-Net for coarse alignment, (2)~a reference feature extractor capturing multi-scale structural cues, (3)~a contrast module that identifies residual misalignment by comparing coarse-registered and reference features, and (4)~a refinement U-Net with feature injection for high-fidelity output. We further propose the Temporal NCC (TNCC) and Temporal NCC Gap (TNCG) for reference-free evaluation of inter-frame temporal consistency. On OR-PAM-Reg-4K (432 test samples), PCReg-Net achieves NCC of 0.983, SSIM of 0.982, and PSNR of 46.96 dB, surpassing the state-of-the-art by over 14 dB at real-time speed. Code is available at https://github.com/JiahaoQin/PCReg-Net
[151] FireRed-Image-Edit-1.0 Techinical Report
Super Intelligence Team, Changhao Qiao, Chao Hui, Chen Li, Cunzheng Wang, Dejia Song, Jiale Zhang, Jing Li, Qiang Xiang, Runqi Wang, Shuang Sun, Wei Zhu, Xu Tang, Yao Hu, Yibo Chen, Yuhao Huang, Yuxuan Duan, Zhiyi Chen, Ziyuan Guo
Main category: cs.CV
TL;DR: FireRed-Image-Edit is a diffusion transformer for instruction-based image editing that achieves SOTA performance through optimized data curation, multi-stage training, and novel techniques for data efficiency and controllability.
Details
Motivation: To advance instruction-based image editing by addressing limitations in existing systems through systematic optimization of data quality, training methodology, and evaluation design, while improving data efficiency and controllability.Method: Constructs a 1.6B-sample training corpus with rigorous cleaning/stratification; uses multi-stage training (pre-training, supervised fine-tuning, reinforcement learning); introduces Multi-Condition Aware Bucket Sampler for variable-resolution batching, Stochastic Instruction Alignment, Asymmetric Gradient Optimization for DPO, DiffusionNFT with layout-aware OCR rewards, and differentiable Consistency Loss for identity preservation.
Result: Achieves competitive or superior performance on REDEdit-Bench (15 editing categories) and public benchmarks (ImgEdit and GEdit) against both open-source and proprietary systems.
Conclusion: FireRed-Image-Edit demonstrates state-of-the-art instruction-based image editing through systematic optimization, with released code, models, and benchmark suite to support future research.
Abstract: We present FireRed-Image-Edit, a diffusion transformer for instruction-based image editing that achieves state-of-the-art performance through systematic optimization of data curation, training methodology, and evaluation design. We construct a 1.6B-sample training corpus, comprising 900M text-to-image and 700M image editing pairs from diverse sources. After rigorous cleaning, stratification, auto-labeling, and two-stage filtering, we retain over 100M high-quality samples balanced between generation and editing, ensuring strong semantic coverage and instruction alignment. Our multi-stage training pipeline progressively builds editing capability via pre-training, supervised fine-tuning, and reinforcement learning. To improve data efficiency, we introduce a Multi-Condition Aware Bucket Sampler for variable-resolution batching and Stochastic Instruction Alignment with dynamic prompt re-indexing. To stabilize optimization and enhance controllability, we propose Asymmetric Gradient Optimization for DPO, DiffusionNFT with layout-aware OCR rewards for text editing, and a differentiable Consistency Loss for identity preservation. We further establish REDEdit-Bench, a comprehensive benchmark spanning 15 editing categories, including newly introduced beautification and low-level enhancement tasks. Extensive experiments on REDEdit-Bench and public benchmarks (ImgEdit and GEdit) demonstrate competitive or superior performance against both open-source and proprietary systems. We release code, models, and the benchmark suite to support future research.
[152] WildfireVLM: AI-powered Analysis for Early Wildfire Detection and Risk Assessment Using Satellite Imagery
Aydin Ayanzadeh, Prakhar Dixit, Sadia Kamal, Milton Halem
Main category: cs.CV
TL;DR: WildfireVLM: AI framework combining satellite imagery wildfire detection with language-driven risk assessment using YOLOv12 for detection and MLLMs for contextual risk analysis.
Details
Motivation: Wildfires are increasing due to climate change and human activities, but early detection remains challenging due to faint smoke signals, dynamic weather conditions, and the need for real-time analysis over large areas using satellite imagery.Method: Construct labeled wildfire/smoke dataset from Landsat-8/9, GOES-16, and other Earth observation sources; use YOLOv12 for fire zone and smoke plume detection; integrate Multimodal Large Language Models (MLLMs) to convert detection outputs into contextualized risk assessments; validate with LLM-as-judge evaluation; deploy with service-oriented architecture.
Result: System demonstrates value of combining computer vision with language-based reasoning for scalable wildfire monitoring, supporting real-time processing, visual risk dashboards, and long-term wildfire tracking.
Conclusion: WildfireVLM shows the effectiveness of integrating satellite imagery detection with language-driven risk assessment for improved wildfire monitoring and disaster management.
Abstract: Wildfires are a growing threat to ecosystems, human lives, and infrastructure, with their frequency and intensity rising due to climate change and human activities. Early detection is critical, yet satellite-based monitoring remains challenging due to faint smoke signals, dynamic weather conditions, and the need for real-time analysis over large areas. We introduce WildfireVLM, an AI framework that combines satellite imagery wildfire detection with language-driven risk assessment. We construct a labeled wildfire and smoke dataset using imagery from Landsat-8/9, GOES-16, and other publicly available Earth observation sources, including harmonized products with aligned spectral bands. WildfireVLM employs YOLOv12 to detect fire zones and smoke plumes, leveraging its ability to detect small, complex patterns in satellite imagery. We integrate Multimodal Large Language Models (MLLMs) that convert detection outputs into contextualized risk assessments and prioritized response recommendations for disaster management. We validate the quality of risk reasoning using an LLM-as-judge evaluation with a shared rubric. The system is deployed using a service-oriented architecture that supports real-time processing, visual risk dashboards, and long-term wildfire tracking, demonstrating the value of combining computer vision with language-based reasoning for scalable wildfire monitoring.
[153] Fine-Tuning a Large Vision-Language Model for Artwork’s Scoring and Critique
Zhehan Zhang, Meihua Qian, Li Luo, Siyu Huang, Chaoyi Zhou, Ripon Saha, Xinxin Song
Main category: cs.CV
TL;DR: Fine-tuning Qwen2-VL-7B vision-language model for automated creativity assessment of human paintings with multi-task learning to predict numerical scores and generate rubric-aligned feedback.
Details
Motivation: Manual creativity assessment (like Torrance Tests) is labor-intensive at scale. Existing machine learning approaches for visual creativity scoring rely mainly on image features and provide limited explanatory feedback.Method: Fine-tune Qwen2-VL-7B vision-language model with multi-task learning on 1000 human-created paintings scored 1-100 with expert rubrics (originality, color, texture, composition, content). Add lightweight regression head on visual encoder to predict scores and generate rubric-aligned feedback in single forward pass, using structured rubric and artwork descriptions in system prompt.
Result: Strong accuracy with Pearson r > 0.97 and MAE ~3.95 on 100-point scale. Generated feedback shows high semantic similarity to expert critiques (average SBERT cosine similarity = 0.798).
Conclusion: The approach bridges computer vision and art assessment, offering a scalable tool for creativity research and classroom feedback through automated scoring with explanatory feedback.
Abstract: Assessing artistic creativity is foundational to creativity research and arts education, yet manual scoring (e.g., Torrance Tests of Creative Thinking) is labor-intensive at scale. Prior machine-learning approaches show promise for visual creativity scoring, but many rely mainly on image features and provide limited or no explanatory feedback. We propose a framework for automated creativity assessment of human paintings by fine-tuning the vision-language model Qwen2-VL-7B with multi-task learning. Our dataset contains 1000 human-created paintings scored on a 1-100 scale and paired with a short human-written description (content or artist explanation). Two expert raters evaluated each work using a five-dimension rubric (originality, color, texture, composition, content) and provided written critiques; we use an 80/20 train-test split. We add a lightweight regression head on the visual encoder output so the model can predict a numerical score and generate rubric-aligned feedback in a single forward pass. By embedding the structured rubric and the artwork description in the system prompt, we constrain the generated text to match the quantitative prediction. Experiments show strong accuracy, achieving Pearson r > 0.97 and MAE about 3.95 on the 100-point scale. Qualitative evaluation indicates the generated feedback is semantically close to expert critiques (average SBERT cosine similarity = 0.798). The proposed approach bridges computer vision and art assessment and offers a scalable tool for creativity research and classroom feedback.
[154] Visual Para-Thinker: Divide-and-Conquer Reasoning for Visual Comprehension
Haoran Xu, Hongyu Wang, Jiaze Li, Shunpeng Chen, Zizhao Tong, Jianzhong Ju, Zhenbo Luo, Jian Luan
Main category: cs.CV
TL;DR: Visual Para-Thinker introduces the first parallel reasoning framework for MLLMs, shifting from depth-based to parallel thinking in visual domains using visual partitioning strategies, Pa-Attention, and LPRoPE to maintain path independence and reasoning diversity.
Details
Motivation: Existing LLM scaling laws focus on extended reasoning length (vertical scaling) which leads to exploration plateaus as models get locked into specific thinking patterns. While parallel thinking helps in text domains, extending this to visual reasoning remains unexplored.Method: Proposes Visual Para-Thinker framework with two visual partitioning strategies for parallelized reasoning. Integrates Pa-Attention and LPRoPE to maintain path independence and promote reasoning diversity. Built on vLLM framework for efficient multimodal parallel processing.
Result: Empirical results on V*, CountBench, RefCOCO, and HallusionBench benchmarks confirm successful extension of parallel reasoning benefits to visual domain.
Conclusion: Visual Para-Thinker successfully extends parallel reasoning paradigm to multimodal settings, addressing exploration limitations of depth-based scaling in visual understanding tasks.
Abstract: Existing LLM test-time scaling laws emphasize the emergence of self-reflective behaviors through extended reasoning length. Nevertheless, this vertical scaling strategy often encounters plateaus in exploration as the model becomes locked into specific thinking pattern. By shifting from depth to parallelism, parallel thinking mitigates the narrowing of exploration. However, the extension of this paradigm to visual domain remains an open research question. In this paper, we first examine the role of visual partitioning in parallelized reasoning and subsequently propose two distinct strategies. Based on the above, we introduce Visual Para-Thinker, representing the inaugural parallel reasoning framework for MLLMs. To maintain path independence and promote diversity in reasoning, our approach integrates Pa-Attention alongside LPRoPE. Leveraging the vLLM framework, we have developed a native multimodal implementation that facilitates high-efficiency parallel processing. Empirical results on benchmark datasets such as V*, CountBench, RefCOCO, and HallusionBench confirm that Visual Para-Thinker successfully extends the benefits of parallel reasoning to the visual domain.
[155] Agentic Spatio-Temporal Grounding via Collaborative Reasoning
Heng Zhao, Yew-Soon Ong, Joey Tianyi Zhou
Main category: cs.CV
TL;DR: ASTG is a training-free framework using MLLM agents for spatio-temporal video grounding that outperforms weakly-supervised methods and matches some fully-supervised approaches.
Details
Motivation: Existing STVG methods have redundant computation, heavy supervision requirements, and limited generalization. Weakly-supervised variants reduce annotation costs but still follow dataset-level train-and-fit paradigms with inferior performance.Method: Proposes Agentic Spatio-Temporal Grounder (ASTG) with two specialized MLLM agents: Spatial Reasoning Agent (SRA) and Temporal Reasoning Agent (TRA). They work collaboratively in a propose-and-evaluation paradigm with visual memory and dialogue context to automate tube extraction, verification, and temporal localization.
Result: Outperforms existing weakly-supervised and zero-shot approaches by a margin and is comparable to some fully-supervised methods on popular benchmarks.
Conclusion: ASTG enables open-world, training-free spatio-temporal video grounding through collaborative MLLM agents, addressing limitations of existing supervised approaches while maintaining competitive performance.
Abstract: Spatio-Temporal Video Grounding (STVG) aims to retrieve the spatio-temporal tube of a target object or person in a video given a text query. Most existing approaches perform frame-wise spatial localization within a predicted temporal span, resulting in redundant computation, heavy supervision requirements, and limited generalization. Weakly-supervised variants mitigate annotation costs but remain constrained by the dataset-level train-and-fit paradigm with an inferior performance. To address these challenges, we propose the Agentic Spatio-Temporal Grounder (ASTG) framework for the task of STVG towards an open-world and training-free scenario. Specifically, two specialized agents SRA (Spatial Reasoning Agent) and TRA (Temporal Reasoning Agent) constructed leveraging on modern Multimoal Large Language Models (MLLMs) work collaboratively to retrieve the target tube in an autonomous and self-guided manner. Following a propose-and-evaluation paradigm, ASTG duly decouples spatio-temporal reasoning and automates the tube extraction, verification and temporal localization processes. With a dedicate visual memory and dialogue context, the retrieval efficiency is significantly enhanced. Experiments on popular benchmarks demonstrate the superiority of the proposed approach where it outperforms existing weakly-supervised and zero-shot approaches by a margin and is comparable to some of the fully-supervised methods.
[156] Sim2Radar: Toward Bridging the Radar Sim-to-Real Gap with VLM-Guided Scene Reconstruction
Emily Bejerano, Federico Tondolo, Aayan Qayyum, Xiaofan Yu, Xiaofan Jiang
Main category: cs.CV
TL;DR: Sim2Radar: Framework for synthesizing mmWave radar training data from single-view RGB images using physics-based simulation to address radar dataset scarcity.
Details
Motivation: Learning-based radar perception is limited by the scarcity and high cost of collecting and annotating large-scale radar datasets, especially for mmWave radar which provides reliable perception in visually degraded indoor environments.Method: End-to-end framework that reconstructs material-aware 3D scenes from single RGB images using monocular depth estimation, segmentation, and vision-language reasoning to infer object materials, then simulates mmWave propagation with configurable physics-based ray tracer using Fresnel reflection models parameterized by ITU-R electromagnetic properties.
Result: Sim2Radar improves downstream 3D radar perception via transfer learning: pre-training a radar point-cloud object detection model on synthetic data and fine-tuning on real radar yields up to +3.7 3D AP (IoU 0.3), with gains primarily from improved spatial localization.
Conclusion: Physics-based, vision-driven radar simulation can provide effective geometric priors for radar learning and measurably improve performance under limited real-data supervision.
Abstract: Millimeter-wave (mmWave) radar provides reliable perception in visually degraded indoor environments (e.g., smoke, dust, and low light), but learning-based radar perception is bottlenecked by the scarcity and cost of collecting and annotating large-scale radar datasets. We present Sim2Radar, an end-to-end framework that synthesizes training radar data directly from single-view RGB images, enabling scalable data generation without manual scene modeling. Sim2Radar reconstructs a material-aware 3D scene by combining monocular depth estimation, segmentation, and vision-language reasoning to infer object materials, then simulates mmWave propagation with a configurable physics-based ray tracer using Fresnel reflection models parameterized by ITU-R electromagnetic properties. Evaluated on real-world indoor scenes, Sim2Radar improves downstream 3D radar perception via transfer learning: pre-training a radar point-cloud object detection model on synthetic data and fine-tuning on real radar yields up to +3.7 3D AP (IoU 0.3), with gains driven primarily by improved spatial localization. These results suggest that physics-based, vision-driven radar simulation can provide effective geometric priors for radar learning and measurably improve performance under limited real-data supervision.
[157] IDPruner: Harmonizing Importance and Diversity in Visual Token Pruning for MLLMs
Yifan Tan, Yifu Sun, Shirui Huang, Hong Liu, Guanghua Yu, Jianchen Zhu, Yangdong Deng
Main category: cs.CV
TL;DR: IDPruner is a visual token pruning method for MLLMs that balances token importance and diversity using Maximal Marginal Relevance, achieving efficient inference without attention maps.
Details
Motivation: MLLMs face computational bottlenecks from massive visual tokens. Existing pruning methods lack principled frameworks for optimally integrating token importance and diversity trade-offs.Method: Systematic analysis of importance-diversity trade-off, then proposes IDPruner using Maximal Marginal Relevance algorithm for Pareto-optimal balance, operating without attention maps for FlashAttention compatibility.
Result: State-of-the-art performance across various architectures and benchmarks. On Qwen2.5-VL-7B-Instruct: retains 95.18% performance with 75% pruning, 86.40% with 90% pruning.
Conclusion: IDPruner provides principled framework for visual token pruning that balances importance and diversity, enabling efficient MLLM inference with strong performance retention.
Abstract: Multimodal Large Language Models (MLLMs) have demonstrated impressive capabilities, yet they encounter significant computational bottlenecks due to the massive volume of visual tokens. Consequently, visual token pruning, which substantially reduces the token count, has emerged as a critical technique for accelerating MLLM inference. Existing approaches focus on token importance, diversity, or an intuitive combination of both, without a principled framework for their optimal integration. To address this issue, we first conduct a systematic analysis to characterize the trade-off between token importance and semantic diversity. Guided by this analysis, we propose the \textbf{I}mportance and \textbf{D}iversity Pruner (\textbf{IDPruner}), which leverages the Maximal Marginal Relevance (MMR) algorithm to achieve a Pareto-optimal balance between these two objectives. Crucially, our method operates without requiring attention maps, ensuring full compatibility with FlashAttention and efficient deployment via one-shot pruning. We conduct extensive experiments across various model architectures and multimodal benchmarks, demonstrating that IDPruner achieves state-of-the-art performance and superior generalization across diverse architectures and tasks. Notably, on Qwen2.5-VL-7B-Instruct, IDPruner retains 95.18% of baseline performance when pruning 75% of the tokens, and still maintains 86.40% even under an extreme 90% pruning ratio. Our code is available at https://github.com/Tencent/AngelSlim.
[158] Diagnostic Benchmarks for Invariant Learning Dynamics: Empirical Validation of the Eidos Architecture
Datorien L. Anderson
Main category: cs.CV
TL;DR: PSI dataset isolates topological invariance from texture correlations; Eidos architecture achieves high accuracy validating “Form-First” hypothesis that generalization comes from geometric integrity, not statistical scale.
Details
Motivation: Standard vision benchmarks are dominated by textural correlations, making it difficult to isolate and study topological invariance - the ability to maintain structural identity across affine transformations. The paper aims to create diagnostic tools to test whether generalization in vision models comes from geometric integrity rather than statistical scale.Method: Created PolyShapes-Ideal (PSI) dataset with three diagnostic probes: 1) polygon classification under noise, 2) zero-shot font transfer from MNIST, and 3) geometric collapse mapping under progressive deformation. Tested Eidos architecture on these benchmarks.
Result: Eidos architecture achieved >99% accuracy on PSI dataset and 81.67% zero-shot transfer across 30 unseen typefaces without pre-training. These results demonstrate strong topological invariance capabilities.
Conclusion: Results validate the “Form-First” hypothesis: generalization in structurally constrained architectures is a property of geometric integrity, not statistical scale. The PSI dataset provides diagnostic tools for evaluating topological invariance separate from texture-based learning.
Abstract: We present the PolyShapes-Ideal (PSI) dataset, a suite of diagnostic benchmarks designed to isolate topological invariance – the ability to maintain structural identity across affine transformations – from the textural correlations that dominate standard vision benchmarks. Through three diagnostic probes (polygon classification under noise, zero-shot font transfer from MNIST, and geometric collapse mapping under progressive deformation), we demonstrate that the Eidos architecture achieves >99% accuracy on PSI and 81.67% zero-shot transfer across 30 unseen typefaces without pre-training. These results validate the “Form-First” hypothesis: generalization in structurally constrained architectures is a property of geometric integrity, not statistical scale.
[159] Synthesizing the Kill Chain: A Zero-Shot Framework for Target Verification and Tactical Reasoning on the Edge
Jesse Barkley, Abraham George, Amir Barati Farimani
Main category: cs.CV
TL;DR: Hierarchical zero-shot framework combines lightweight object detection with compact VLMs for autonomous edge robotics in military environments, achieving high accuracy in tasks like false-positive filtering, damage assessment, and vehicle classification.
Details
Motivation: Autonomous edge robotics in dynamic military environments face constraints from scarce domain-specific training data and computational limits of edge hardware, requiring efficient zero-shot solutions.Method: Cascaded pipeline using Grounding DINO as text-promptable region proposer, then passing high-confidence frames to compact VLMs (Qwen and Gemma families, 4B-12B parameters) for semantic verification. Extended to agentic Scout-Commander workflow with novel “Controlled Input” methodology.
Result: Achieved up to 100% accuracy in false-positive filtering, 97.5% in damage assessment, 55-90% in vehicle classification. Agentic workflow achieved 100% correct asset deployment with 9.8/10 reasoning score and sub-75-second latency.
Conclusion: Hierarchical zero-shot architectures are validated for edge autonomy, with diagnostic framework for certifying VLM suitability in safety-critical applications. Different VLMs show distinct failure patterns in perception vs reasoning.
Abstract: Deploying autonomous edge robotics in dynamic military environments is constrained by both scarce domain-specific training data and the computational limits of edge hardware. This paper introduces a hierarchical, zero-shot framework that cascades lightweight object detection with compact Vision-Language Models (VLMs) from the Qwen and Gemma families (4B-12B parameters). Grounding DINO serves as a high-recall, text-promptable region proposer, and frames with high detection confidence are passed to edge-class VLMs for semantic verification. We evaluate this pipeline on 55 high-fidelity synthetic videos from Battlefield 6 across three tasks: false-positive filtering (up to 100% accuracy), damage assessment (up to 97.5%), and fine-grained vehicle classification (55-90%). We further extend the pipeline into an agentic Scout-Commander workflow, achieving 100% correct asset deployment and a 9.8/10 reasoning score (graded by GPT-4o) with sub-75-second latency. A novel “Controlled Input” methodology decouples perception from reasoning, revealing distinct failure phenotypes: Gemma3-12B excels at tactical logic but fails in visual perception, while Gemma3-4B exhibits reasoning collapse even with accurate inputs. These findings validate hierarchical zero-shot architectures for edge autonomy and provide a diagnostic framework for certifying VLM suitability in safety-critical applications.
[160] MotionWeaver: Holistic 4D-Anchored Framework for Multi-Humanoid Image Animation
Xirui Hu, Yanbo Ding, Jiahao Wang, Tingting Shi, Yali Wang, Guo Zhi Zhi, Weizhan Zhang
Main category: cs.CV
TL;DR: MotionWeaver is a framework for multi-humanoid image animation that generalizes across diverse humanoid forms and handles complex interactions/occlusions using unified motion representations and a 4D-anchored paradigm.
Details
Motivation: Existing character animation methods are limited to single-human settings and struggle with multi-humanoid scenarios involving diverse forms, complex interactions, and frequent occlusions.Method: Two key innovations: 1) Unified motion representations that extract identity-agnostic motions and bind them to corresponding characters, 2) Holistic 4D-anchored paradigm that constructs shared 4D space to fuse motion representations with video latents, reinforced with hierarchical 4D-level supervision.
Result: MotionWeaver achieves state-of-the-art results on a new 300-video benchmark and generalizes effectively across diverse humanoid forms, complex interactions, and challenging multi-humanoid scenarios.
Conclusion: The paper presents a novel framework for multi-humanoid image animation that addresses limitations of existing single-human methods through unified motion representations and 4D-anchored supervision.
Abstract: Character image animation, which synthesizes videos of reference characters driven by pose sequences, has advanced rapidly but remains largely limited to single-human settings. Existing methods struggle to generalize to multi-humanoid scenarios, which involve diverse humanoid forms, complex interactions, and frequent occlusions. We address this gap with two key innovations. First, we introduce unified motion representations that extract identity-agnostic motions and explicitly bind them to corresponding characters, enabling generalization across diverse humanoid forms and seamless extension to multi-humanoid scenarios. Second, we propose a holistic 4D-anchored paradigm that constructs a shared 4D space to fuse motion representations with video latents, and further reinforces this process with hierarchical 4D-level supervision to better handle interactions and occlusions. We instantiate these ideas in MotionWeaver, an end-to-end framework for multi-humanoid image animation. To support this setting, we curate a 46-hour dataset of multi-human videos with rich interactions, and construct a 300-video benchmark featuring paired humanoid characters. Quantitative and qualitative experiments demonstrate that MotionWeaver not only achieves state-of-the-art results on our benchmark but also generalizes effectively across diverse humanoid forms, complex interactions, and challenging multi-humanoid scenarios.
[161] HiST-VLA: A Hierarchical Spatio-Temporal Vision-Language-Action Model for End-to-End Autonomous Driving
Yiru Wang, Zichong Gu, Yu Gao, Anqing Jiang, Zhigang Sun, Shuo Wang, Yuwen Heng, Hao Sun
Main category: cs.CV
TL;DR: HiST-VLA is a hierarchical spatio-temporal VLA model for autonomous driving that improves 3D spatial reasoning, integrates geometric awareness, uses dynamic token sparsification for efficiency, and refines waypoints into trajectories with strict spatial grounding.
Details
Motivation: Current Vision-Language-Action models for autonomous driving have limitations including imprecise numerical reasoning, weak 3D spatial awareness, and high sensitivity to context, which constrain their use in safety-critical scenarios.Method: Proposes HiST-VLA with: 1) integration of geometric awareness with fine-grained driving commands and state history prompting, 2) dynamic token sparsification that fuses redundant tokens instead of filtering, 3) hierarchical transformer-based planner that refines coarse VLA waypoints into fine-grained trajectories, and 4) dynamic latent regularization to incorporate language commands for spatial grounding.
Result: Achieves state-of-the-art performance on NAVSIM v2 benchmark with EPDMS of 88.6 on Navtest and 50.9 on pseudo closed-loop Navhard benchmark.
Conclusion: HiST-VLA addresses key limitations of VLA models for autonomous driving through enhanced 3D spatial-temporal reasoning, computational efficiency via token fusion, and hierarchical trajectory refinement with strict spatial grounding.
Abstract: Vision-Language-Action (VLA) models offer promising capabilities for autonomous driving through multimodal understanding. However, their utilization in safety-critical scenarios is constrained by inherent limitations, including imprecise numerical reasoning, weak 3D spatial awareness, and high sensitivity to context. To address these challenges, we propose HiST-VLA, a novel Hierarchical Spatio-Temporal VLA model designed for reliable trajectory generation. Our framework enhances 3D spatial and temporal reasoning by integrating geometric awareness with fine-grained driving commands and state history prompting. To ensure computational efficiency, we integrate dynamic token sparsification into the VLA architecture. This approach fuses redundant tokens rather than filtering them, effectively reducing redundancy without sacrificing model performance. Furthermore, we employ a hierarchical transformer-based planner to progressively refine coarse VLA waypoints into fine-grained trajectories. Crucially, the planner utilizes dynamic latent regularization to incorporate language commands, ensuring strict spatial grounding and temporal coherence. Extensive evaluation on the NAVSIM v2 benchmark demonstrates state-of-the-art performance on Navtest, achieving an EPDMS of 88.6, and EPDMS of 50.9 on pseudo closed-loop Navhard benchmark.
[162] Zwitscherkasten – DIY Audiovisual bird monitoring
Dominik Blum, Elias Häring, Fabian Jirges, Martin Schäffer, David Schick, Florian Schulenberg, Torsten Schön
Main category: cs.CV
TL;DR: Zwitscherkasten: A DIY multimodal system for bird species monitoring using audio and visual data on edge devices with deep learning models for bioacoustic and image classification.
Details
Motivation: To create a scalable, non-invasive bird species monitoring system that can operate on resource-constrained edge devices for biodiversity monitoring and citizen science applications.Method: Deploys deep learning models for bioacoustic classification and image-based recognition on edge hardware, with acoustic activity detection to reduce energy consumption and fine-grained visual detection pipelines.
Result: Demonstrates that accurate bird species identification is feasible on embedded platforms, enabling real-time monitoring with reduced energy consumption.
Conclusion: The system supports scalable biodiversity monitoring through multimodal edge computing, making bird species monitoring more accessible for conservation and citizen science.
Abstract: This paper presents Zwitscherkasten, a DiY, multimodal system for bird species monitoring using audio and visual data on edge devices. Deep learning models for bioacoustic and image-based classification are deployed on resource-constrained hardware, enabling real-time, non-invasive monitoring. An acoustic activity detector reduces energy consumption, while visual recognition is performed using fine-grained detection and classification pipelines. Results show that accurate bird species identification is feasible on embedded platforms, supporting scalable biodiversity monitoring and citizen science applications.
[163] MedScope: Incentivizing “Think with Videos” for Clinical Reasoning via Coarse-to-Fine Tool Calling
Wenjie Li, Yujie Zhang, Haoran Sun, Xingqi He, Hongcheng Gao, Chenglong Ma, Ming Hu, Guankun Wang, Shiyi Yao, Renhao Yang, Hongliang Ren, Lei Wang, Junjun He, Yankai Jiang
Main category: cs.CV
TL;DR: MedScope is a tool-using clinical video reasoning model that performs coarse-to-fine evidence seeking over long-form medical videos, using targeted tool calls and verification to produce accurate predictions grounded in temporally localized visual evidence.
Details
Motivation: Current multimodal LLMs process videos with passive sampling or weakly grounded inspection, limiting their ability to iteratively locate, verify, and justify predictions with temporally targeted evidence in clinical settings like surgical robotics.Method: Proposes MedScope with coarse-to-fine evidence seeking, interleaving reasoning with targeted tool calls and verification. Uses Grounding-Aware Group Relative Policy Optimization (GA-GRPO) with grounding-aligned rewards and evidence-weighted advantages. Built ClinVideoSuite for supervision.
Result: Achieves state-of-the-art performance on both in-domain and out-of-domain video understanding benchmarks, with more accurate and trustworthy predictions explicitly grounded in temporally localized visual evidence.
Conclusion: MedScope illuminates a path toward medical AI agents that can genuinely “think with videos” through tool-integrated reasoning, advancing clinical video understanding with explicit evidence grounding.
Abstract: Long-form clinical videos are central to visual evidence-based decision-making, with growing importance for applications such as surgical robotics and related settings. However, current multimodal large language models typically process videos with passive sampling or weakly grounded inspection, which limits their ability to iteratively locate, verify, and justify predictions with temporally targeted evidence. To close this gap, we propose MedScope, a tool-using clinical video reasoning model that performs coarse-to-fine evidence seeking over long-form procedures. By interleaving intermediate reasoning with targeted tool calls and verification on retrieved observations, MedScope produces more accurate and trustworthy predictions that are explicitly grounded in temporally localized visual evidence. To address the lack of high-fidelity supervision, we build ClinVideoSuite, an evidence-centric, fine-grained clinical video suite. We then optimize MedScope with Grounding-Aware Group Relative Policy Optimization (GA-GRPO), which directly reinforces tool use with grounding-aligned rewards and evidence-weighted advantages. On full and fine-grained video understanding benchmarks, MedScope achieves state-of-the-art performance in both in-domain and out-of-domain evaluations. Our approach illuminates a path toward medical AI agents that can genuinely “think with videos” through tool-integrated reasoning. We will release our code, models, and data.
[164] Ask the Expert: Collaborative Inference for Vision Transformers with Near-Edge Accelerators
Hao Liu, Suhaib A. Fahmy
Main category: cs.CV
TL;DR: A collaborative inference framework for Vision Transformers that uses edge device with lightweight ViT and near-edge accelerator with expert ViTs, with dynamic routing based on confidence and progressive specialist training to improve accuracy and reduce latency/energy.
Details
Motivation: Vision Transformers are computationally expensive for edge devices, while full cloud offloading introduces high latency. Need efficient deployment strategy that balances edge and near-edge computation.Method: Collaborative framework with lightweight generalist ViT on edge device and multiple medium-sized expert ViTs on near-edge accelerator. Novel routing mechanism uses edge model’s Top-k predictions to dynamically select relevant expert for low-confidence samples. Progressive specialist training strategy enhances expert accuracy on dataset subsets.
Result: On CIFAR-100 with real-world testbed: training strategy improves expert specialization accuracy by 4.12% on target subsets and overall accuracy by 2.76% over static experts. Reduces latency by up to 45% compared to edge execution, and energy consumption by up to 46% compared to near-edge offload.
Conclusion: Proposed collaborative inference framework effectively balances edge and near-edge computation for Vision Transformers, achieving significant improvements in accuracy, latency, and energy efficiency.
Abstract: Deploying Vision Transformers on edge devices is challenging due to their high computational complexity, while full offloading to cloud resources presents significant latency overheads. We propose a novel collaborative inference framework, which orchestrates a lightweight generalist ViT on an edge device and multiple medium-sized expert ViTs on a near-edge accelerator. A novel routing mechanism uses the edge model’s Top-$\mathit{k}$ predictions to dynamically select the most relevant expert for samples with low confidence. We further design a progressive specialist training strategy to enhance expert accuracy on dataset subsets. Extensive experiments on the CIFAR-100 dataset using a real-world edge and near-edge testbed demonstrate the superiority of our framework. Specifically, the proposed training strategy improves expert specialization accuracy by 4.12% on target subsets and enhances overall accuracy by 2.76% over static experts. Moreover, our method reduces latency by up to 45% compared to edge execution, and energy consumption by up to 46% compared to just near-edge offload.
[165] Meningioma Analysis and Diagnosis using Limited Labeled Samples
Jiamiao Lu, Wei Wu, Ke Gao, Ping Mao, Weichuan Zhang, Tuo Wang, Lingkun Ma, Jiapan Guo, Zanyi Wu, Yuqing Hu, Changming Sun
Main category: cs.CV
TL;DR: Proposes AMSF-Net, an adaptive multi-scale feature fusion network for few-shot meningioma classification using MRI, combining spatial and frequency domain features with adaptive weighting.
Details
Motivation: Accurate meningioma grading is crucial for treatment planning and prognosis, but current methods may not optimally leverage both spatial and frequency domain information, especially in few-shot learning scenarios.Method: Uses discrete wavelet transform to extract frequency band features, proposes adaptive weighting mechanism to fuse spatial and frequency domain information, and applies this to few-shot meningioma classification on MRI data.
Result: Outperforms state-of-the-art methods on three datasets, including a newly introduced MRI meningioma dataset, demonstrating effectiveness of adaptive feature fusion.
Conclusion: Adaptive fusion of spatial-frequency domain features significantly improves meningioma classification, especially in few-shot learning scenarios, with potential clinical applications.
Abstract: The biological behavior and treatment response of meningiomas depend on their grade, making an accurate diagnosis essential for treatment planning and prognosis assessment. We observed that the weighted fusion of spatial-frequency domain features significantly influences meningioma classification performance. Notably, the contribution of specific frequency bands obtained by discrete wavelet transform varies considerably across different images. A feature fusion architecture with adaptive weights of different frequency band information and spatial domain information is proposed for few-shot meningioma learning. To verify the effectiveness of the proposed method, a new MRI dataset of meningiomas is introduced. The experimental results demonstrate the superiority of the proposed method compared with existing state-of-the-art methods in three datasets. The code will be available at: https://github.com/ICL-SUST/AMSF-Net
[166] An Integrated Causal Inference Framework for Traffic Safety Modeling with Semantic Street-View Visual Features
Lishan Sun, Yujia Cheng, Pengfei Cui, Lei Han, Mohamed Abdel-Aty, Yunhan Zheng, Xingchen Zhang
Main category: cs.CV
TL;DR: Using street view imagery and causal inference methods, this study finds that urban greenery has a significant negative causal effect on traffic crashes, with spatial heterogeneity in its protective benefits.
Details
Motivation: Current macroscopic traffic safety models rely on static sociodemographic and infrastructure metrics but overlook drivers' visual perception of the driving environment. While visual features impact crashes, existing evidence is observational and lacks robust causality for policy evaluation.Method: Applied semantic segmentation on Google Street View images to extract visual environmental features, used Double Machine Learning framework to quantify causal effects on regional crashes, utilized SHAP values for nonlinear influence analysis, and applied causal forests for conditional average treatment effects.
Result: Greenery proportion has significant negative causal effect on traffic crashes (ATE = -6.38, p = 0.005). Protective effect is strongest in densely populated, socially vulnerable urban cores. Greenery mitigates angle and rear-end crashes but has limited protective benefit for vulnerable road users.
Conclusion: Provides causal evidence for greening as a safety intervention, suggests prioritizing hazardous visual environments, and highlights need for distinct design optimizations to protect vulnerable road users.
Abstract: Macroscopic traffic safety modeling aims to identify critical risk factors for regional crashes, thereby informing targeted policy interventions for safety improvement. However, current approaches rely heavily on static sociodemographic and infrastructure metrics, frequently overlooking the impacts from drivers’ visual perception of driving environment. Although visual environment features have been found to impact driving and traffic crashes, existing evidence remains largely observational, failing to establish the robust causality for traffic policy evaluation under complex spatial environment. To fill these gaps, we applied semantic segmentation on Google Street View imageries to extract visual environmental features and proposed a Double Machine Learning framework to quantify their causal effects on regional crashes. Meanwhile, we utilized SHAP values to characterize the nonlinear influence mechanisms of confounding variables in the models and applied causal forests to estimate conditional average treatment effects. Leveraging crash records from the Miami metropolitan area, Florida, and 220,000 street view images, evidence shows that greenery proportion exerts a significant and robust negative causal effect on traffic crashes (Average Treatment Effect = -6.38, p = 0.005). This protective effect exhibits spatial heterogeneity, being most pronounced in densely populated and socially vulnerable urban cores. While greenery significantly mitigates angle and rear-end crashes, its protective benefit for vulnerable road users (VRUs) remains limited. Our findings provide causal evidence for greening as a potential safety intervention, prioritizing hazardous visual environments while highlighting the need for distinct design optimizations to protect VRUs.
[167] Visual Foresight for Robotic Stow: A Diffusion-Based World Model from Sparse Snapshots
Lijun Zhang, Nikhil Chacko, Petter Nilsson, Ruinian Xu, Shantanu Thakar, Bai Lou, Harpreet Sawhney, Zhebin Zhang, Mudit Agrawal, Bhavana Chandrashekhar, Aaron Parness
Main category: cs.CV
TL;DR: FOREST: A stow-intent-conditioned world model for predicting post-stow bin configurations in automated warehouses using latent diffusion transformers.
Details
Motivation: Automated warehouses need to anticipate how storage bins will look after stow operations before actual execution for better planning and decision-making.Method: Represents bin states as item-aligned instance masks and uses a latent diffusion transformer to predict post-stow configurations from observed context and planned stow behavior.
Result: FOREST substantially improves geometric agreement between predicted and true post-stow layouts compared to heuristic baselines, with only modest performance loss in downstream tasks when using predictions.
Conclusion: FOREST provides useful foresight signals for warehouse planning by accurately predicting post-stow configurations, enabling better load-quality assessment and multi-stow reasoning.
Abstract: Automated warehouses execute millions of stow operations, where robots place objects into storage bins. For these systems it is valuable to anticipate how a bin will look from the current observations and the planned stow behavior before real execution. We propose FOREST, a stow-intent-conditioned world model that represents bin states as item-aligned instance masks and uses a latent diffusion transformer to predict the post-stow configuration from the observed context. Our evaluation shows that FOREST substantially improves the geometric agreement between predicted and true post-stow layouts compared with heuristic baselines. We further evaluate the predicted post-stow layouts in two downstream tasks, in which replacing the real post-stow masks with FOREST predictions causes only modest performance loss in load-quality assessment and multi-stow reasoning, indicating that our model can provide useful foresight signals for warehouse planning.
[168] From Prompt to Production:Automating Brand-Safe Marketing Imagery with Text-to-Image Models
Parmida Atighehchian, Henry Wang, Andrei Kapustin, Boris Lerner, Tiancheng Jiang, Taylor Jensen, Negin Sokhandan
Main category: cs.CV
TL;DR: A scalable production pipeline for text-to-image models that balances automation with human oversight to generate marketing images of commercial products with improved fidelity and human preference.
Details
Motivation: While text-to-image models produce impressive results, deploying them in production at scale while maintaining quality and creative alignment remains challenging. Current approaches struggle to balance automation efficiency with necessary human feedback for quality control.Method: Proposes a new pipeline offering fully automated, scalable solution for generating marketing images of commercial products using text-to-image models. The system maintains image quality and fidelity while introducing creative variation that adheres to marketing guidelines.
Result: Achieves 30.77% increase in marketing object fidelity using DINOV2 and 52.00% increase in human preference over generated outcomes compared to baseline approaches.
Conclusion: The proposed pipeline successfully balances automation and human oversight, enabling scalable deployment of text-to-image models in production while maintaining quality standards and creative alignment for marketing applications.
Abstract: Text-to-image models have made significant strides, producing impressive results in generating images from textual descriptions. However, creating a scalable pipeline for deploying these models in production remains a challenge. Achieving the right balance between automation and human feedback is critical to maintain both scale and quality. While automation can handle large volumes, human oversight is still an essential component to ensure that the generated images meet the desired standards and are aligned with the creative vision. This paper presents a new pipeline that offers a fully automated, scalable solution for generating marketing images of commercial products using text-to-image models. The proposed system maintains the quality and fidelity of images, while also introducing sufficient creative variation to adhere to marketing guidelines. By streamlining this process, we ensure a seamless blend of efficiency and human oversight, achieving a $30.77%$ increase in marketing object fidelity using DINOV2 and a $52.00%$ increase in human preference over the generated outcome.
[169] Detecting Brick Kiln Infrastructure at Scale: Graph, Foundation, and Remote Sensing Models for Satellite Imagery Data
Usman Nazir, Xidong Chen, Hafiz Muhammad Abubakar, Hadia Abu Bakar, Raahim Arbaz, Fezan Rasool, Bin Chen, Sara Khalid
Main category: cs.CV
TL;DR: Brick kiln detection from satellite imagery using graph-based models and foundation models for environmental monitoring
Details
Motivation: Brick kilns are major sources of air pollution and forced labor in South Asia, but large-scale monitoring is limited by sparse ground data. There's a need for scalable detection methods using satellite imagery.Method: Created a high-resolution (0.149m/pixel) dataset of 1.3M image tiles across 5 regions. Proposed ClimateGraph, a region-adaptive graph-based model that captures spatial and directional structure in kiln layouts. Compared against graph learning baselines, remote sensing pipelines, and recent foundation models for satellite imagery.
Result: Results show complementary strengths across graph-based, foundation model, and remote sensing approaches. Provides practical guidance for scalable brick kiln monitoring from satellite imagery.
Conclusion: Multiple approaches (graph-based, foundation models, remote sensing) offer complementary strengths for scalable brick kiln detection from satellite imagery, enabling better environmental and labor monitoring.
Abstract: Brick kilns are a major source of air pollution and forced labor in South Asia, yet large-scale monitoring remains limited by sparse and outdated ground data. We study brick kiln detection at scale using high-resolution satellite imagery and curate a multi city zoom-20 (0.149 meters per pixel) resolution dataset comprising over 1.3 million image tiles across five regions in South and Central Asia. We propose ClimateGraph, a region-adaptive graph-based model that captures spatial and directional structure in kiln layouts, and evaluate it against established graph learning baselines. In parallel, we assess a remote sensing based detection pipeline and benchmark it against recent foundation models for satellite imagery. Our results highlight complementary strengths across graph, foundation, and remote sensing approaches, providing practical guidance for scalable brick kiln monitoring from satellite imagery.
[170] Using Deep Learning to Generate Semantically Correct Hindi Captions
Wasim Akram Khan, Anil Kumar Vuppala
Main category: cs.CV
TL;DR: This paper presents a multimodal image captioning system that generates Hindi captions by combining visual features from pre-trained CNNs with attention-based bidirectional LSTM text encoding, achieving BLEU scores of 0.59 (BLEU-1) and 0.19 (BLEU-4).
Details
Motivation: Most image captioning research focuses on English, leaving a gap for other popular languages like Hindi. The paper aims to develop a multimodal system for generating semantically accurate Hindi image captions by leveraging computer vision and natural language processing techniques.Method: Uses Google Cloud Translator to generate Hindi captions from Flickr8k dataset. Extracts visual features using pre-trained CNNs (VGG16, ResNet50, Inception V3). Employs uni-directional and bi-directional LSTM with attention mechanisms for text encoding. Attention layer generates weight vectors to combine image features at each time step into sentence-level feature vectors.
Result: Attention-based bidirectional LSTM with VGG16 produced the best results: BLEU-1 score of 0.59 and BLEU-4 score of 0.19. The system demonstrates ability to produce relevant, semantically accurate Hindi image captions.
Conclusion: The research successfully develops a multimodal architecture for Hindi image captioning, achieving good performance metrics. The model can guide future research in multilingual multimodal captioning systems.
Abstract: Automated image captioning using the content from the image is very appealing when done by harnessing the capability of computer vision and natural language processing. Extensive research has been done in the field with a major focus on the English language which gives the scope for further developments in the same with consideration of popular foreign languages. This research utilizes distinct models for translating the image caption into Hindi, the fourth most popular language across the world. Exploring the multi-modal architectures this research comprises local visual features, global visual features, attention mechanisms, and pre-trained models. Using google cloud translator on the image dataset from Flickr8k, Hindi image descriptions have been generated. Pre-trained CNNs like VGG16, ResNet50, and Inception V3 helped in retrieving image characteristics, while the uni-directional and bi-directional techniques of text encoding are used for the text encoding process. An additional Attention layer helps to generate a weight vector and, by multiplying it, combine image characteristics from each time step into a sentence-level feature vector. Bilingual evaluation understudy scores are used to compare the research outcome. Many experiments that serve as a baseline are done for the comparative analysis of the research. An image with a score of BLEU-1 is considered sufficient, whereas one with a score of BLEU-4 is considered to have fluid image captioning. For both BLEU scores, the attention-based bidirectional LSTM with VGG16 produced the best results of 0.59 and 0.19 respectively. The experiments conclude that researchs ability to produce relevant, semantically accurate image captions in Hindi. The research accomplishes the goals and future research can be guided by this research model.
[171] AdaCorrection: Adaptive Offset Cache Correction for Accurate Diffusion Transformers
Dong Liu, Yanxuan Yu, Ben Lengerich, Ying Nian Wu
Main category: cs.CV
TL;DR: AdaCorrection is an adaptive cache correction framework for Diffusion Transformers that improves inference efficiency while maintaining generation quality by dynamically blending cached and fresh activations based on spatio-temporal signals.
Details
Motivation: Diffusion Transformers (DiTs) achieve state-of-the-art image/video generation but suffer from expensive iterative inference. Existing cache acceleration methods use static reuse schedules or heuristics that cause temporal drift and cache misalignment, degrading generation quality.Method: AdaCorrection uses lightweight spatio-temporal signals to estimate cache validity at each timestep, then adaptively blends cached and fresh activations. This on-the-fly correction requires no additional supervision or retraining.
Result: Maintains near-original FID scores while providing moderate acceleration. Experiments on image and video diffusion benchmarks show consistent generation performance improvements.
Conclusion: AdaCorrection enables efficient cache reuse in Diffusion Transformers while preserving high generation fidelity through adaptive correction, offering a practical solution to accelerate diffusion inference without quality degradation.
Abstract: Diffusion Transformers (DiTs) achieve state-of-the-art performance in high-fidelity image and video generation but suffer from expensive inference due to their iterative denoising structure. While prior methods accelerate sampling by caching intermediate features, they rely on static reuse schedules or coarse-grained heuristics, which often lead to temporal drift and cache misalignment that significantly degrade generation quality. We introduce \textbf{AdaCorrection}, an adaptive offset cache correction framework that maintains high generation fidelity while enabling efficient cache reuse across Transformer layers during diffusion inference. At each timestep, AdaCorrection estimates cache validity with lightweight spatio-temporal signals and adaptively blends cached and fresh activations. This correction is computed on-the-fly without additional supervision or retraining. Our approach achieves strong generation quality with minimal computational overhead, maintaining near-original FID while providing moderate acceleration. Experiments on image and video diffusion benchmarks show that AdaCorrection consistently improves generation performance.
[172] The Diffusion Duet: Harmonizing Dual Channels with Wavelet Suppression for Image Separation
Jingwei Li, Wei Pu
Main category: cs.CV
TL;DR: DCDSM introduces diffusion models to blind image separation, using dual-channel diffusion with wavelet suppression for better rain/snow removal and complex mixture separation.
Details
Motivation: Traditional blind image separation methods struggle with complex real-world scenes, suffering from estimation bias, texture distortion, and artifacts under strong noise and nonlinear mixing. There's a need for better approaches that can handle these challenges.Method: Proposes Dual-Channel Diffusion Separation Model (DCDSM) that uses diffusion models to learn source image feature distributions. Includes a Wavelet Suppression Module (WSM) in the dual-branch reverse denoising process to form an interactive separation network that exploits mutual coupling noise between source images.
Result: Achieves state-of-the-art performance: 1) For rain/snow removal: PSNR/SSIM of 35.0023 dB/0.9549 (rain) and 29.8108 dB/0.9243 (snow), outperforming Histoformer and LDRCNet; 2) For complex mixture separation: average PSNR/SSIM of 25.0049 dB/0.7997, surpassing comparative methods by 4.1249 dB/0.0926.
Conclusion: DCDSM demonstrates superiority in addressing rain/snow residue removal and detail preservation challenges through both subjective and objective evaluations, showing the effectiveness of diffusion models for blind image separation tasks.
Abstract: Blind image separation (BIS) refers to the inverse problem of simultaneously estimating and restoring multiple independent source images from a single observation image under conditions of unknown mixing mode and without prior knowledge of the source images. Traditional methods relying on statistical independence assumptions or CNN/GAN variants struggle to characterize complex feature distributions in real scenes, leading to estimation bias, texture distortion, and artifact residue under strong noise and nonlinear mixing. This paper innovatively introduces diffusion models into dual-channel BIS, proposing an efficient Dual-Channel Diffusion Separation Model (DCDSM). DCDSM leverages diffusion models’ powerful generative capability to learn source image feature distributions and reconstruct feature structures effectively. A novel Wavelet Suppression Module (WSM) is designed within the dual-branch reverse denoising process, forming an interactive separation network that enhances detail separation by exploiting the mutual coupling noise characteristic between source images. Extensive experiments on synthetic datasets containing rain/snow and complex mixtures demonstrate that DCDSM achieves state-of-the-art performance: 1) In image restoration tasks, it obtains PSNR/SSIM values of 35.0023 dB/0.9549 and 29.8108 dB/0.9243 for rain and snow removal respectively, outperforming Histoformer and LDRCNet by 1.2570 dB/0.9272 dB (PSNR) and 0.0262/0.0289 (SSIM) on average; 2) For complex mixture separation, the restored dual-source images achieve average PSNR and SSIM of 25.0049 dB and 0.7997, surpassing comparative methods by 4.1249 dB and 0.0926. Both subjective and objective evaluations confirm DCDSM’s superiority in addressing rain/snow residue removal and detail preservation challenges.
[173] An Online Reference-Free Evaluation Framework for Flowchart Image-to-Code Generation
Giang Son Nguyen, Zi Pong Lim, Sarthak Ketanbhai Modi, Yon Shin Teo, Wenya Wang
Main category: cs.CV
TL;DR: A reference-free evaluation framework for flowchart image-to-code generation that uses OCR-based recall and visual entailment-based precision metrics to assess output quality without ground truth.
Details
Motivation: VLMs are increasingly used to convert flowchart images to structured code in production, but lack of ground truth makes quality assessment difficult. Need for automated quality monitoring without reference data.Method: Proposes two automated metrics: Recall_OCR (extracts text via OCR from input image as proxy reference to estimate content coverage) and Precision_VE (uses visual entailment against original image to detect hallucinated elements). Their harmonic mean F1_OCR-VE provides unified quality score.
Result: Validation on FlowVQA dataset shows strong agreement with ground-truth metrics: average Pearson’s r = 0.97, 0.91, and 0.94 for Recall, Precision, and F1 respectively, confirming reliability as practical reference-free alternative.
Conclusion: The framework provides reliable, reference-free quality monitoring for flowchart image-to-code generation in production settings, enabling continuous assessment without ground truth data.
Abstract: Vision-Language Models (VLMs) are increasingly used in document processing pipelines to convert flowchart images into structured code (e.g., Mermaid). In production, these systems process arbitrary inputs for which no ground-truth code exists, making output quality difficult to assess. We propose a reference-free evaluation framework that monitors flowchart image-to-code generation quality at inference time, using only the input image and the generated output. The framework introduces two automated metrics: $\text{Recall}{\text{OCR}}$, which estimates content coverage by extracting text from the input image via OCR as a proxy reference, and $\text{Precision}{\text{VE}}$, which detects hallucinated elements through Visual Entailment against the original image. Their harmonic mean, $\text{F1}{\text{OCR-VE}}$, provides a unified quality score. Validation on the FlowVQA dataset shows strong agreement with ground-truth metrics (average Pearson’s $r = 0.97$, $0.91$, and $0.94$ for Recall, Precision, and F1, respectively), confirming the framework’s reliability as a practical, reference-free alternative for continuous quality monitoring in production settings.
[174] LAF-YOLOv10 with Partial Convolution Backbone, Attention-Guided Feature Pyramid, Auxiliary P2 Head, and Wise-IoU Loss for Small Object Detection in Drone Aerial Imagery
Sohail Ali Farooqui, Zuhair Ahmed Khan Taha, Mohammed Mudassir Uddin, Shahnawaz Alam
Main category: cs.CV
TL;DR: LAF-YOLOv10 improves small object detection in drone imagery by integrating four complementary techniques into YOLOv10n framework, achieving better performance with fewer parameters for embedded UAV deployment.
Details
Motivation: Current object detectors struggle with UAV-specific challenges including tiny targets (few pixels), cluttered backgrounds, heavy occlusion, and strict onboard computational constraints for real-time deployment on drones.Method: Four integrated techniques: 1) Partial Convolution C2f (PC-C2f) reduces redundant computation, 2) Attention-Guided Feature Pyramid Network (AG-FPN) improves multi-scale fusion, 3) Auxiliary P2 detection head for sub-8×8 pixel objects, 4) Wise-IoU v3 for stable regression under noisy annotations.
Result: Achieves 35.1±0.3% mAP@0.5 on VisDrone-DET2019 (3.3 points over YOLOv10n) with 2.3M parameters, 35.8±0.4% mAP@0.5 on UAVDT, and 24.3 FPS on NVIDIA Jetson Orin Nano at FP16 precision.
Conclusion: The joint integration of complementary techniques within YOLOv10 framework effectively addresses UAV-specific detection challenges while maintaining computational efficiency suitable for embedded deployment.
Abstract: Unmanned aerial vehicles serve as primary sensing platforms for surveillance, traffic monitoring, and disaster response, making aerial object detection a central problem in applied computer vision. Current detectors struggle with UAV-specific challenges: targets spanning only a few pixels, cluttered backgrounds, heavy occlusion, and strict onboard computational budgets. This study introduces LAF-YOLOv10, built on YOLOv10n, integrating four complementary techniques to improve small-object detection in drone imagery. A Partial Convolution C2f (PC-C2f) module restricts spatial convolution to one quarter of backbone channels, reducing redundant computation while preserving discriminative capacity. An Attention-Guided Feature Pyramid Network (AG-FPN) inserts Squeeze-and-Excitation channel gates before multi-scale fusion and replaces nearest-neighbor upsampling with DySample for content-aware interpolation. An auxiliary P2 detection head at 160$\times$160 resolution extends localization to objects below 8$\times$8 pixels, while the P5 head is removed to redistribute parameters. Wise-IoU v3 replaces CIoU for bounding box regression, attenuating gradients from noisy annotations in crowded aerial scenes. The four modules address non-overlapping bottlenecks: PC-C2f compresses backbone computation, AG-FPN refines cross-scale fusion, the P2 head recovers spatial resolution, and Wise-IoU stabilizes regression under label noise. No individual component is novel; the contribution is the joint integration within a single YOLOv10 framework. Across three training runs (seeds 42, 123, 256), LAF-YOLOv10 achieves 35.1$\pm$0.3% mAP@0.5 on VisDrone-DET2019 with 2.3,M parameters, exceeding YOLOv10n by 3.3 points. Cross-dataset evaluation on UAVDT yields 35.8$\pm$0.4% mAP@0.5. Benchmarks on NVIDIA Jetson Orin Nano confirm 24.3 FPS at FP16, demonstrating viability for embedded UAV deployment.
[175] Handling Supervision Scarcity in Chest X-ray Classification: Long-Tailed and Zero-Shot Learning
Ha-Hieu Pham, Hai-Dang Nguyen, Thanh-Huy Nguyen, Min Xu, Ulas Bagci, Trung-Nghia Le, Huy-Hieu Pham
Main category: cs.CV
TL;DR: The paper presents solutions for chest X-ray classification addressing long-tailed multi-label distributions and zero-shot out-of-distribution recognition, achieving top performance on the CXR-LT 2026 challenge.
Details
Motivation: Clinical chest X-ray classification faces two key challenges: (1) extreme long-tailed multi-label disease distributions where rare diseases have few examples, and (2) missing annotations for rare or previously unseen findings that require zero-shot recognition.Method: For Task 1 (long-tailed multi-label classification), they use an imbalance-aware multi-label learning strategy to improve tail class recognition while maintaining performance on frequent findings. For Task 2 (zero-shot OOD recognition), they propose a prediction approach that scores unseen disease categories without using any supervised labels or examples from OOD classes during training.
Result: The method achieves strong performance on both tasks, ranking first on the public leaderboard of the development phase of the CXR-LT 2026 challenge, as measured by macro-averaged mean Average Precision (mAP).
Conclusion: The proposed task-specific solutions effectively address the challenges of imperfect supervision in clinical CXR classification, demonstrating robust performance on both long-tailed multi-label learning and zero-shot out-of-distribution recognition.
Abstract: Chest X-ray (CXR) classification in clinical practice is often limited by imperfect supervision, arising from (i) extreme long-tailed multi-label disease distributions and (ii) missing annotations for rare or previously unseen findings. The CXR-LT 2026 challenge addresses these issues on a PadChest-based benchmark with a 36-class label space split into 30 in-distribution classes for training and 6 out-of-distribution (OOD) classes for zero-shot evaluation. We present task-specific solutions tailored to the distinct supervision regimes. For Task 1 (long-tailed multi-label classification), we adopt an imbalance-aware multi-label learning strategy to improve recognition of tail classes while maintaining stable performance on frequent findings. For Task 2 (zero-shot OOD recognition), we propose a prediction approach that produces scores for unseen disease categories without using any supervised labels or examples from the OOD classes during training. Evaluated with macro-averaged mean Average Precision (mAP), our method achieves strong performance on both tasks, ranking first on the public leaderboard of the development phase. Code and pre-trained models are available at https://github.com/hieuphamha19/CXR_LT.
[176] Learning on the Fly: Replay-Based Continual Object Perception for Indoor Drones
Sebastian-Ion Nae, Mihai-Eugen Barbu, Sebastian Mocanu, Marius Leordeanu
Main category: cs.CV
TL;DR: Indoor UAV dataset for class-incremental learning with replay-based methods under tight memory constraints, showing Forgetting-Aware Replay performs best.
Details
Motivation: Autonomous indoor drones need to learn new object classes in real-time while avoiding catastrophic forgetting, but existing UAV datasets focus on outdoor scenes and lack temporally coherent indoor videos.Method: Created indoor dataset of 14,400 frames with inter-drone and ground vehicle footage using semi-automatic annotation. Benchmarked 3 replay-based CIL strategies (ER, MIR, FAR) using YOLOv11-nano detector under tight memory budgets (5-10% replay). Used Grad-CAM for attention analysis.
Result: FAR performed best under tight memory budgets, achieving 82.96% average accuracy with 5% replay. Grad-CAM analysis showed attention shifts across classes in mixed scenes, associated with reduced localization quality for drones.
Conclusion: Replay-based continual learning can be effectively applied to edge aerial systems. The work contributes an indoor UAV video dataset with temporal coherence and evaluation of replay-based CIL under limited replay budgets.
Abstract: Autonomous agents such as indoor drones must learn new object classes in real-time while limiting catastrophic forgetting, motivating Class-Incremental Learning (CIL). However, most unmanned aerial vehicle (UAV) datasets focus on outdoor scenes and offer limited temporally coherent indoor videos. We introduce an indoor dataset of $14,400$ frames capturing inter-drone and ground vehicle footage, annotated via a semi-automatic workflow with a $98.6%$ first-pass labeling agreement before final manual verification. Using this dataset, we benchmark 3 replay-based CIL strategies: Experience Replay (ER), Maximally Interfered Retrieval (MIR), and Forgetting-Aware Replay (FAR), using YOLOv11-nano as a resource-efficient detector for deployment-constrained UAV platforms. Under tight memory budgets ($5-10%$ replay), FAR performs better than the rest, achieving an average accuracy (ACC, $mAP_{50-95}$ across increments) of $82.96%$ with $5%$ replay. Gradient-weighted class activation mapping (Grad-CAM) analysis shows attention shifts across classes in mixed scenes, which is associated with reduced localization quality for drones. The experiments further demonstrate that replay-based continual learning can be effectively applied to edge aerial systems. Overall, this work contributes an indoor UAV video dataset with preserved temporal coherence and an evaluation of replay-based CIL under limited replay budgets. Project page: https://spacetime-vision-robotics-laboratory.github.io/learning-on-the-fly-cl
[177] GLIMPSE : Real-Time Text Recognition and Contextual Understanding for VQA in Wearables
Akhil Ramachandran, Ankit Arun, Ashish Shenoy, Abhay Harpale, Srihari Jayakumar, Debojeet Chatterjee, Mohsen Moslehpour, Pierce Chuang, Yichao Lu, Vikas Bhardwaj, Peyman Heidari
Main category: cs.CV
TL;DR: Hybrid architecture for wearable Text VQA that uses selective high-resolution OCR on-device while streaming low-resolution video to balance power consumption and text recognition accuracy.
Details
Motivation: Deploying Text VQA on wearable devices faces tension between high-resolution requirements for text recognition and power/thermal constraints. Existing models struggle with coherent temporal context in real-time streams.Method: Exploits asymmetric resolution requirements: OCR needs fine detail while scene understanding tolerates coarse features. Uses hybrid architecture with selective high-resolution OCR on-device and low-resolution video streaming for visual context.
Result: Achieves 72% accuracy at 0.49x power consumption of full-resolution streaming on benchmark of text-based VQA samples across five task categories.
Conclusion: Enables sustained VQA sessions on resource-constrained wearables without sacrificing text understanding quality by balancing resolution requirements.
Abstract: Video Large Language Models (Video LLMs) have shown remarkable progress in understanding and reasoning about visual content, particularly in tasks involving text recognition and text-based visual question answering (Text VQA). However, deploying Text VQA on wearable devices faces a fundamental tension: text recognition requires high-resolution video, but streaming high-quality video drains battery and causes thermal throttling. Moreover, existing models struggle to maintain coherent temporal context when processing text across multiple frames in real-time streams. We observe that text recognition and visual reasoning have asymmetric resolution requirements - OCR needs fine detail while scene understanding tolerates coarse features. We exploit this asymmetry with a hybrid architecture that performs selective high-resolution OCR on-device while streaming low-resolution video for visual context. On a benchmark of text-based VQA samples across five task categories, our system achieves 72% accuracy at 0.49x the power consumption of full-resolution streaming, enabling sustained VQA sessions on resource-constrained wearables without sacrificing text understanding quality.
[178] Benchmarking Video Foundation Models for Remote Parkinson’s Disease Screening
Md Saiful Islam, Ekram Hossain, Abdelrahman Abdelkader, Tariq Adnan, Fazla Rabbi Mashrur, Sooyong Park, Praveen Kumar, Qasim Sudais, Natalia Chunga, Nami Shah, Jan Freyberg, Christopher Kanan, Ruth Schneider, Ehsan Hoque
Main category: cs.CV
TL;DR: Systematic evaluation of 7 video foundation models for Parkinson’s disease screening using 32,847 videos from 1,888 participants across 16 clinical tasks, finding model-dependent task saliency with AUCs of 76.4-85.3%.
Details
Motivation: Remote video-based assessments offer scalable Parkinson's disease screening, but the comparative effectiveness of different video foundation model architectures across diverse clinical tasks remains poorly understood despite recent advances.Method: Large-scale study using novel dataset from 1,888 participants (727 with PD) with 32,847 videos across 16 standardized clinical tasks. Evaluated 7 state-of-the-art VFMs (VideoPrism, V-JEPA, ViViT, VideoMAE, TimeSformer) using frozen embeddings with linear classification head to assess model robustness in clinical screening.
Result: Task saliency is highly model-dependent: VideoPrism excels in visual speech kinematics and facial expressivity, V-JEPA superior for upper-limb motor tasks, TimeSformer competitive for rhythmic tasks. Achieved AUCs of 76.4-85.3% and accuracies of 71.5-80.6%. High specificity (up to 90.3%) but lower sensitivity (43.2-57.3%).
Conclusion: Establishes rigorous baseline for VFM-based PD screening and provides roadmap for selecting suitable tasks and architectures in remote neurological monitoring, highlighting need for task-aware calibration and multi-task/multi-modal integration.
Abstract: Remote, video-based assessments offer a scalable pathway for Parkinson’s disease (PD) screening. While traditional approaches rely on handcrafted features mimicking clinical scales, recent advances in video foundation models (VFMs) enable representation learning without task-specific customization. However, the comparative effectiveness of different VFM architectures across diverse clinical tasks remains poorly understood. We present a large-scale systematic study using a novel video dataset from 1,888 participants (727 with PD), comprising 32,847 videos across 16 standardized clinical tasks. We evaluate seven state-of-the-art VFMs – including VideoPrism, V-JEPA, ViViT, and VideoMAE – to determine their robustness in clinical screening. By evaluating frozen embeddings with a linear classification head, we demonstrate that task saliency is highly model-dependent: VideoPrism excels in capturing visual speech kinematics (no audio) and facial expressivity, while V-JEPA proves superior for upper-limb motor tasks. Notably, TimeSformer remains highly competitive for rhythmic tasks like finger tapping. Our experiments yield AUCs of 76.4-85.3% and accuracies of 71.5-80.6%. While high specificity (up to 90.3%) suggests strong potential for ruling out healthy individuals, the lower sensitivity (43.2-57.3%) highlights the need for task-aware calibration and integration of multiple tasks and modalities. Overall, this work establishes a rigorous baseline for VFM-based PD screening and provides a roadmap for selecting suitable tasks and architectures in remote neurological monitoring. Code and anonymized structured data are publicly available: https://anonymous.4open.science/r/parkinson_video_benchmarking-A2C5
[179] SpargeAttention2: Trainable Sparse Attention via Hybrid Top-k+Top-p Masking and Distillation Fine-Tuning
Jintao Zhang, Kai Jiang, Chendong Xiang, Weiqi Feng, Yuezhou Hu, Haocheng Xi, Jianfei Chen, Jun Zhu
Main category: cs.CV
TL;DR: SpargeAttention2: A trainable sparse attention method for diffusion models that achieves 95% sparsity with 16.2x speedup while maintaining generation quality through hybrid masking rules and distillation-inspired fine-tuning.
Details
Motivation: To address limitations of existing sparse attention methods for diffusion models, specifically: (1) failures of common masking rules (Top-k/Top-p) at high sparsity, (2) understanding why trainable methods achieve higher sparsity than training-free ones, and (3) limitations of fine-tuning sparse attention with diffusion loss.Method: Proposes SpargeAttention2 with three components: (i) hybrid masking rule combining Top-k and Top-p for robust high-sparsity masking, (ii) efficient trainable sparse attention implementation, and (iii) distillation-inspired fine-tuning objective to preserve generation quality during sparse attention fine-tuning.
Result: Achieves 95% attention sparsity and 16.2x attention speedup on video diffusion models while maintaining generation quality, consistently outperforming prior sparse attention methods.
Conclusion: SpargeAttention2 successfully addresses key limitations of sparse attention methods, enabling extremely high sparsity (95%) with significant speedups while preserving generation quality in diffusion models.
Abstract: Many training-free sparse attention methods are effective for accelerating diffusion models. Recently, several works suggest that making sparse attention trainable can further increase sparsity while preserving generation quality. We study three key questions: (1) when do the two common masking rules, i.e., Top-k and Top-p, fail, and how can we avoid these failures? (2) why can trainable sparse attention reach higher sparsity than training-free methods? (3) what are the limitations of fine-tuning sparse attention using the diffusion loss, and how can we address them? Based on this analysis, we propose SpargeAttention2, a trainable sparse attention method that achieves high sparsity without degrading generation quality. SpargeAttention2 includes (i) a hybrid masking rule that combines Top-k and Top-p for more robust masking at high sparsity, (ii) an efficient trainable sparse attention implementation, and (iii) a distillation-inspired fine-tuning objective to better preserve generation quality during fine-tuning using sparse attention. Experiments on video diffusion models show that SpargeAttention2 reaches 95% attention sparsity and a 16.2x attention speedup while maintaining generation quality, consistently outperforming prior sparse attention methods.
[180] Nighttime Autonomous Driving Scene Reconstruction with Physically-Based Gaussian Splatting
Tae-Kyeong Kim, Xingxin Chen, Guile Wu, Chengjie Huang, Dongfeng Bai, Bingbing Liu
Main category: cs.CV
TL;DR: A novel approach integrating physically based rendering into 3D Gaussian Splatting for enhanced nighttime scene reconstruction in autonomous driving, outperforming state-of-the-art methods.
Details
Motivation: Existing NeRF and 3DGS methods achieve photorealistic modeling for autonomous driving scene reconstruction but primarily focus on normal-light conditions. Nighttime scenes present complex lighting and appearance challenges that degrade performance of current methods.Method: Integrates physically based rendering into 3D Gaussian Splatting with composite scene Gaussian representations. Jointly optimizes BRDF-based material properties, explicitly models diffuse components through global illumination and specular components via anisotropic spherical Gaussians.
Result: Outperforms state-of-the-art methods quantitatively and qualitatively across diverse nighttime scenarios on nuScenes and Waymo datasets. Maintains real-time rendering while improving reconstruction quality for outdoor nighttime driving scenes.
Conclusion: The approach successfully addresses nighttime scene reconstruction challenges in autonomous driving by integrating physically based rendering with 3DGS, enabling better handling of complex lighting conditions while preserving real-time performance.
Abstract: This paper focuses on scene reconstruction under nighttime conditions in autonomous driving simulation. Recent methods based on Neural Radiance Fields (NeRFs) and 3D Gaussian Splatting (3DGS) have achieved photorealistic modeling in autonomous driving scene reconstruction, but they primarily focus on normal-light conditions. Low-light driving scenes are more challenging to model due to their complex lighting and appearance conditions, which often causes performance degradation of existing methods. To address this problem, this work presents a novel approach that integrates physically based rendering into 3DGS to enhance nighttime scene reconstruction for autonomous driving. Specifically, our approach integrates physically based rendering into composite scene Gaussian representations and jointly optimizes Bidirectional Reflectance Distribution Function (BRDF) based material properties. We explicitly model diffuse components through a global illumination module and specular components by anisotropic spherical Gaussians. As a result, our approach improves reconstruction quality for outdoor nighttime driving scenes, while maintaining real-time rendering. Extensive experiments across diverse nighttime scenarios on two real-world autonomous driving datasets, including nuScenes and Waymo, demonstrate that our approach outperforms the state-of-the-art methods both quantitatively and qualitatively.
[181] Privacy-Concealing Cooperative Perception for BEV Scene Segmentation
Song Wang, Lingling Li, Marcus Santos, Guanghui Wang
Main category: cs.CV
TL;DR: Privacy-preserving cooperative perception framework for autonomous vehicles that protects visual privacy while maintaining BEV segmentation performance through adversarial learning.
Details
Motivation: Cooperative perception systems improve autonomous driving performance but create privacy risks by sharing sensitive visual data that could be reconstructed. Need to protect privacy while maintaining perception capabilities.Method: Proposes Privacy-Concealing Cooperation (PCC) framework with three components: hiding network to conceal visual clues in BEV features, reconstruction network that tries to recover images, and perception network for segmentation. Uses adversarial learning where hiding and reconstruction networks compete, while perception network is integrated and optimized end-to-end.
Result: Effectively degrades reconstructed image quality with minimal impact on segmentation performance, providing privacy protection for cooperating vehicles.
Conclusion: PCC framework successfully balances privacy protection and perception performance in cooperative autonomous driving systems through adversarial feature concealment.
Abstract: Cooperative perception systems for autonomous driving aim to overcome the limited perception range of a single vehicle by communicating with adjacent agents to share sensing information. While this improves perception performance, these systems also face a significant privacy-leakage issue, as sensitive visual content can potentially be reconstructed from the shared data. In this paper, we propose a novel Privacy-Concealing Cooperation (PCC) framework for Bird’s Eye View (BEV) semantic segmentation. Based on commonly shared BEV features, we design a hiding network to prevent an image reconstruction network from recovering the input images from the shared features. An adversarial learning mechanism is employed to train the network, where the hiding network works to conceal the visual clues in the BEV features while the reconstruction network attempts to uncover these clues. To maintain segmentation performance, the perception network is integrated with the hiding network and optimized end-to-end. The experimental results demonstrate that the proposed PCC framework effectively degrades the quality of the reconstructed images with minimal impact on segmentation performance, providing privacy protection for cooperating vehicles. The source code will be made publicly available upon publication.
[182] Diff-Aid: Inference-time Adaptive Interaction Denoising for Rectified Text-to-Image Generation
Binglei Li, Mengping Yang, Zhiyu Tan, Junping Zhang, Hao Li
Main category: cs.CV
TL;DR: Diff-Aid: A lightweight inference-time method that adaptively adjusts per-token text-image interactions across transformer blocks and denoising timesteps in text-to-image diffusion models to improve prompt adherence and semantic alignment.
Details
Motivation: Current text-to-image diffusion models struggle with faithfully following complex textual descriptions due to insufficient interactions between textual and visual features. Prior approaches lack flexibility and overlook dynamic interactions across different blocks and denoising stages.Method: Proposes Diff-Aid, a plug-and-play inference-time method that adaptively modulates per-token text and image interactions across transformer blocks and denoising timesteps. It provides interpretable modulation patterns and can be integrated into downstream applications like style LoRAs, controllable generation, and zero-shot editing.
Result: Experiments on SD 3.5 and FLUX baselines show consistent improvements in prompt adherence, visual quality, and human preference across various metrics. The method yields interpretable patterns revealing how different components contribute to semantic alignment.
Conclusion: Diff-Aid provides a flexible and efficient solution for enhancing text-image interactions in diffusion models, improving generation quality while offering interpretability and seamless integration into downstream applications.
Abstract: Recent text-to-image (T2I) diffusion models have achieved remarkable advancement, yet faithfully following complex textual descriptions remains challenging due to insufficient interactions between textual and visual features. Prior approaches enhance such interactions via architectural design or handcrafted textual condition weighting, but lack flexibility and overlook the dynamic interactions across different blocks and denoising stages. To provide a more flexible and efficient solution to this problem, we propose Diff-Aid, a lightweight inference-time method that adaptively adjusts per-token text and image interactions across transformer blocks and denoising timesteps. Beyond improving generation quality, Diff-Aid yields interpretable modulation patterns that reveal how different blocks, timesteps, and textual tokens contribute to semantic alignment during denoising. As a plug-and-play module, Diff-Aid can be seamlessly integrated into downstream applications for further improvement, including style LoRAs, controllable generation, and zero-shot editing. Experiments on strong baselines (SD 3.5 and FLUX) demonstrate consistent improvements in prompt adherence, visual quality, and human preference across various metrics. Our code and models will be released.
[183] Two-Stream Interactive Joint Learning of Scene Parsing and Geometric Vision Tasks
Guanfeng Tang, Hongbo Zhao, Ziwei Long, Jiayao Li, Bohong Xiao, Wei Ye, Hanli Wang, Rui Fan
Main category: cs.CV
TL;DR: TwInS is a bio-inspired joint learning framework that simultaneously performs scene parsing and geometric vision tasks using two interactive streams, with cross-task feature fusion and semi-supervised training.
Details
Motivation: Inspired by the human visual system's parallel streams for contextual and spatial understanding, the paper aims to create a unified framework that jointly addresses scene parsing (contextual understanding) and geometric vision tasks (spatial understanding) through interactive feature sharing.Method: TwInS uses two interactive streams: scene parsing stream for contextual features and geometric vision stream for spatial features. Features are bidirectionally fused using a cross-task adapter, with geometric features projected into contextual space. A semi-supervised training strategy eliminates need for human-annotated correspondence ground truth.
Result: Extensive experiments on three public datasets show TwInS outperforms state-of-the-art approaches, validating the effectiveness of its core components including the interactive streams and cross-task adapter.
Conclusion: TwInS successfully demonstrates that joint learning of scene parsing and geometric vision through interactive streams leads to superior performance, with the semi-supervised approach enabling large-scale training without costly annotations.
Abstract: Inspired by the human visual system, which operates on two parallel yet interactive streams for contextual and spatial understanding, this article presents Two Interactive Streams (TwInS), a novel bio-inspired joint learning framework capable of simultaneously performing scene parsing and geometric vision tasks. TwInS adopts a unified, general-purpose architecture in which multi-level contextual features from the scene parsing stream are infused into the geometric vision stream to guide its iterative refinement. In the reverse direction, decoded geometric features are projected into the contextual feature space for selective heterogeneous feature fusion via a novel cross-task adapter, which leverages rich cross-view geometric cues to enhance scene parsing. To eliminate the dependence on costly human-annotated correspondence ground truth, TwInS is further equipped with a tailored semi-supervised training strategy, which unleashes the potential of large-scale multi-view data and enables continuous self-evolution without requiring ground-truth correspondences. Extensive experiments conducted on three public datasets validate the effectiveness of TwInS’s core components and demonstrate its superior performance over existing state-of-the-art approaches. The source code will be made publicly available upon publication.
[184] AdaVBoost: Mitigating Hallucinations in LVLMs via Token-Level Adaptive Visual Attention Boosting
Jiacheng Zhang, Feng Liu, Chao Du, Tianyu Pang
Main category: cs.CV
TL;DR: AdaVBoost: A token-level adaptive visual attention boosting framework that uses Visual Grounding Entropy to estimate hallucination risk and dynamically adjusts attention scaling during generation to mitigate hallucinations in Large Vision-Language Models.
Details
Motivation: Existing visual attention boosting methods use predefined scaling factors that create a fundamental trade-off: too weak scaling leaves hallucinations unresolved, while too strong scaling introduces new hallucinations. The paper identifies the need for adaptive, token-level intervention based on real-time hallucination risk assessment.Method: Proposes AdaVBoost with Visual Grounding Entropy (VGE) to estimate hallucination risk by leveraging visual grounding as a complementary signal to capture evidence mismatches. The framework applies stronger visual attention boosting to high-risk tokens and weaker boosting to low-risk tokens, enabling token-level adaptive intervention at each generation step.
Result: Extensive experiments show AdaVBoost significantly outperforms baseline methods across multiple LVLMs and hallucination benchmarks, demonstrating effectiveness in mitigating hallucinations through adaptive attention boosting.
Conclusion: AdaVBoost provides an effective solution to the fundamental trade-off in visual attention boosting by enabling adaptive, token-level intervention based on real-time hallucination risk assessment, significantly improving hallucination mitigation in LVLMs.
Abstract: Visual attention boosting has emerged as a promising direction for mitigating hallucinations in Large Vision-Language Models (LVLMs), where existing methods primarily focus on where to boost by applying a predefined scaling to the attention of method-specific visual tokens during autoregressive generation. In this paper, we identify a fundamental trade-off in these methods: a predefined scaling factor can be too weak at some generation steps, leaving hallucinations unresolved, yet too strong at others, leading to new hallucinations. Motivated by this finding, we propose AdaVBoost, a token-level visual attention boosting framework that adaptively determines how much attention to boost at each generation step. Specifically, we introduce Visual Grounding Entropy (VGE) to estimate hallucination risk, which leverages visual grounding as a complementary signal to capture evidence mismatches beyond entropy. Guided by VGE, AdaVBoost applies stronger visual attention boosting to high-risk tokens and weaker boosting to low-risk tokens, enabling token-level adaptive intervention at each generation step. Extensive experiments show that AdaVBoost significantly outperforms baseline methods across multiple LVLMs and hallucination benchmarks.
[185] Towards Sparse Video Understanding and Reasoning
Chenwei Xu, Zhen Ye, Shang Wu, Weijian Li, Zihan Wang, Zhuofan Xia, Lie Lu, Pranav Maneriker, Fan Du, Manling Li, Han Liu
Main category: cs.CV
TL;DR: REVISE is a multi-round agent for video QA that selects sparse informative frames, maintains summary-as-state across rounds, and stops early when confident, reducing computational costs while improving accuracy.
Details
Motivation: Current video QA methods often process all frames uniformly, which is computationally expensive and inefficient. There's a need for sparse reasoning that focuses only on relevant video content while maintaining accuracy.Method: REVISE uses multi-round sparse frame selection, maintains a summary-as-state across rounds, and supports both proprietary VLMs (plug-and-play) and open-source models (via EAGER reinforcement fine-tuning). EAGER uses three reward terms: confidence gain, summary sufficiency, and correct-and-early stop.
Result: Across multiple VQA benchmarks, REVISE improves accuracy while significantly reducing frames processed, rounds needed, and prompt tokens used, demonstrating practical sparse video reasoning.
Conclusion: REVISE enables efficient video QA through sparse reasoning, making video understanding more practical by reducing computational costs while maintaining or improving performance.
Abstract: We present \revise (\underline{Re}asoning with \underline{Vi}deo \underline{S}parsity), a multi-round agent for video question answering (VQA). Instead of uniformly sampling frames, \revise selects a small set of informative frames, maintains a summary-as-state across rounds, and stops early when confident. It supports proprietary vision-language models (VLMs) in a ``plug-and-play’’ setting and enables reinforcement fine-tuning for open-source models. For fine-tuning, we introduce EAGER (Evidence-Adjusted Gain for Efficient Reasoning), an annotation-free reward with three terms: (1) Confidence gain: after new frames are added, we reward the increase in the log-odds gap between the correct option and the strongest alternative; (2) Summary sufficiency: at answer time we re-ask using only the last committed summary and reward success; (3) Correct-and-early stop: answering correctly within a small turn budget is rewarded. Across multiple VQA benchmarks, \revise improves accuracy while reducing frames, rounds, and prompt tokens, demonstrating practical sparse video reasoning.
[186] A generalizable foundation model for intraoperative understanding across surgical procedures
Kanggil Park, Yongjun Jeon, Soyoung Lim, Seonmin Park, Jongmin Shin, Jung Yong Kim, Sehyeon An, Jinsoo Rhu, Jongman Kim, Gyu-Seong Choi, Namkee Oh, Kyu-Hwan Jung
Main category: cs.CV
TL;DR: ZEN is a generalizable foundation model for surgical video understanding trained on 4M+ frames from 21 procedures using self-supervised multi-teacher distillation, showing strong cross-procedure generalization across 20 downstream tasks.
Details
Motivation: Current surgical AI models are task-specific and don't generalize across procedures or institutions, limiting consistent assessment, training, and development of reliable AI systems due to variability in intraoperative perception among surgeons.Method: Developed ZEN using self-supervised multi-teacher distillation framework trained on a large diverse dataset of over 4 million frames from 21 different surgical procedures, with systematic evaluation of multiple representation learning strategies.
Result: ZEN consistently outperforms existing surgical foundation models across 20 downstream tasks in full fine-tuning, frozen-backbone, few-shot and zero-shot settings, demonstrating robust cross-procedure generalization.
Conclusion: ZEN represents a step toward unified representations for surgical scene understanding and supports future applications in intraoperative assistance and surgical training assessment.
Abstract: In minimally invasive surgery, clinical decisions depend on real-time visual interpretation, yet intraoperative perception varies substantially across surgeons and procedures. This variability limits consistent assessment, training, and the development of reliable artificial intelligence systems, as most surgical AI models are designed for narrowly defined tasks and do not generalize across procedures or institutions. Here we introduce ZEN, a generalizable foundation model for intraoperative surgical video understanding trained on more than 4 million frames from over 21 procedures using a self-supervised multi-teacher distillation framework. We curated a large and diverse dataset and systematically evaluated multiple representation learning strategies within a unified benchmark. Across 20 downstream tasks and full fine-tuning, frozen-backbone, few-shot and zero-shot settings, ZEN consistently outperforms existing surgical foundation models and demonstrates robust cross-procedure generalization. These results suggest a step toward unified representations for surgical scene understanding and support future applications in intraoperative assistance and surgical training assessment.
[187] Layer-Guided UAV Tracking: Enhancing Efficiency and Occlusion Robustness
Yang Zhou, Derui Ding, Ran Sun, Ying Sun, Haohua Zhang
Main category: cs.CV
TL;DR: LGTrack is a lightweight UAV tracking framework that balances accuracy and efficiency through dynamic layer selection, efficient feature enhancement, and robust representation learning for occlusions.
Details
Motivation: Addressing the trade-off between accuracy and efficiency in visual object tracking for UAV applications, particularly under challenging conditions like unpredictable occlusion.Method: Introduces a unified UAV tracking framework with: 1) Global-Grouped Coordinate Attention (GGCA) module for capturing long-range dependencies and global contexts with minimal computational overhead, 2) Similarity-Guided Layer Adaptation (SGLA) module to replace knowledge distillation for optimal balance between precision and efficiency, and 3) dynamic layer selection and efficient feature enhancement.
Result: Achieves state-of-the-art real-time speed (258.7 FPS on UAVDT) while maintaining competitive tracking accuracy (82.8% precision) across three datasets.
Conclusion: LGTrack provides an effective solution for UAV tracking that successfully balances accuracy and efficiency, making it suitable for real-time applications.
Abstract: Visual object tracking (VOT) plays a pivotal role in unmanned aerial vehicle (UAV) applications. Addressing the trade-off between accuracy and efficiency, especially under challenging conditions like unpredictable occlusion, remains a significant challenge. This paper introduces LGTrack, a unified UAV tracking framework that integrates dynamic layer selection, efficient feature enhancement, and robust representation learning for occlusions. By employing a novel lightweight Global-Grouped Coordinate Attention (GGCA) module, LGTrack captures long-range dependencies and global contexts, enhancing feature discriminability with minimal computational overhead. Additionally, a lightweight Similarity-Guided Layer Adaptation (SGLA) module replaces knowledge distillation, achieving an optimal balance between tracking precision and inference efficiency. Experiments on three datasets demonstrate LGTrack’s state-of-the-art real-time speed (258.7 FPS on UAVDT) while maintaining competitive tracking accuracy (82.8% precision). Code is available at https://github.com/XiaoMoc/LGTrack
[188] DCDM: Divide-and-Conquer Diffusion Models for Consistency-Preserving Video Generation
Haoyu Zhao, Yuang Zhang, Junqi Cheng, Jiaxi Gu, Zenghui Lu, Peng Shu, Zuxuan Wu, Yu-Gang Jiang
Main category: cs.CV
TL;DR: DCDM is a system-level framework for video generation that addresses three consistency challenges: intra-clip world knowledge, inter-clip camera motion, and inter-shot element consistency through dedicated components sharing a unified backbone.
Details
Motivation: Current video generative models show impressive visual fidelity but struggle with semantic, geometric, and identity consistency across different temporal scales, limiting their practical applications.Method: DCDM decomposes video consistency into three components: 1) LLM-parsed semantic representations for intra-clip consistency, 2) temporal camera representation in noise space for inter-clip camera control, and 3) holistic scene generation with windowed cross-attention and sparse inter-shot self-attention for long-range coherence.
Result: Validated on the CVM Competition at AAAI'26 test set, demonstrating that the proposed strategies effectively address the three consistency challenges in video generation.
Conclusion: DCDM provides a comprehensive framework for improving video generation consistency across multiple temporal scales through specialized components while maintaining computational efficiency.
Abstract: Recent video generative models have demonstrated impressive visual fidelity, yet they often struggle with semantic, geometric, and identity consistency. In this paper, we propose a system-level framework, termed the Divide-and-Conquer Diffusion Model (DCDM), to address three key challenges: (1) intra-clip world knowledge consistency, (2) inter-clip camera consistency, and (3) inter-shot element consistency. DCDM decomposes video consistency modeling under these scenarios into three dedicated components while sharing a unified video generation backbone. For intra-clip consistency, DCDM leverages a large language model to parse input prompts into structured semantic representations, which are subsequently translated into coherent video content by a diffusion transformer. For inter-clip camera consistency, we propose a temporal camera representation in the noise space that enables precise and stable camera motion control, along with a text-to-image initialization mechanism to further enhance controllability. For inter-shot consistency, DCDM adopts a holistic scene generation paradigm with windowed cross-attention and sparse inter-shot self-attention, ensuring long-range narrative coherence while maintaining computational efficiency. We validate our framework on the test set of the CVM Competition at AAAI'26, and the results demonstrate that the proposed strategies effectively address these challenges.
[189] KorMedMCQA-V: A Multimodal Benchmark for Evaluating Vision-Language Models on the Korean Medical Licensing Examination
Byungjin Choi, Seongsu Bae, Sunjun Kweon, Edward Choi
Main category: cs.CV
TL;DR: KorMedMCQA-V: A Korean medical multimodal benchmark with 1,534 questions and 2,043 medical images for evaluating vision-language models on medical reasoning tasks.
Details
Motivation: To create a comprehensive multimodal benchmark for evaluating vision-language models on Korean medical licensing exam questions, complementing existing text-only benchmarks and addressing the need for medical multimodal evaluation in Korean language contexts.Method: Created a dataset of 1,534 questions with 2,043 associated medical images from Korean Medical Licensing Examinations (2012-2023), with about 30% containing multiple images requiring cross-image integration. Images cover various clinical modalities including X-ray, CT, ECG, ultrasound, and endoscopy. Evaluated over 50 VLMs across proprietary, open-source, general-purpose, medical-specialized, and Korean-specialized categories using zero-shot evaluation.
Result: Best proprietary model (Gemini-3.0-Pro) achieved 96.9% accuracy, best open-source model (Qwen3-VL-32B-Thinking) 83.7%, and best Korean-specialized model (VARCO-VISION-2.0-14B) only 43.2%. Key findings: reasoning-oriented models gain up to +20% over instruction-tuned counterparts, medical domain specialization yields inconsistent gains, all models degrade on multi-image questions, and performance varies across imaging modalities.
Conclusion: KorMedMCQA-V provides a valuable multimodal benchmark for Korean medical reasoning, revealing significant gaps in current VLMs’ capabilities, especially for Korean-specialized models and multi-image reasoning tasks. The benchmark complements text-only evaluation and highlights areas for improvement in medical multimodal understanding.
Abstract: We introduce KorMedMCQA-V, a Korean medical licensing-exam-style multimodal multiple-choice question answering benchmark for evaluating vision-language models (VLMs). The dataset consists of 1,534 questions with 2,043 associated images from Korean Medical Licensing Examinations (2012-2023), with about 30% containing multiple images requiring cross-image evidence integration. Images cover clinical modalities including X-ray, computed tomography (CT), electrocardiography (ECG), ultrasound, endoscopy, and other medical visuals. We benchmark over 50 VLMs across proprietary and open-source categories-spanning general-purpose, medical-specialized, and Korean-specialized families-under a unified zero-shot evaluation protocol. The best proprietary model (Gemini-3.0-Pro) achieves 96.9% accuracy, the best open-source model (Qwen3-VL-32B-Thinking) 83.7%, and the best Korean-specialized model (VARCO-VISION-2.0-14B) only 43.2%. We further find that reasoning-oriented model variants gain up to +20 percentage points over instruction-tuned counterparts, medical domain specialization yields inconsistent gains over strong general-purpose baselines, all models degrade on multi-image questions, and performance varies notably across imaging modalities. By complementing the text-only KorMedMCQA benchmark, KorMedMCQA-V forms a unified evaluation suite for Korean medical reasoning across text-only and multimodal conditions. The dataset is available via Hugging Face Datasets: https://huggingface.co/datasets/seongsubae/KorMedMCQA-V.
[190] Optimizing Point-of-Care Ultrasound Video Acquisition for Probabilistic Multi-Task Heart Failure Detection
Armin Saadat, Nima Hashemi, Bahar Khodabakhshian, Michael Y. Tsang, Christina Luong, Teresa S. M. Tsang, Purang Abolmaesumi
Main category: cs.CV
TL;DR: RL agent for personalized echocardiography view selection balances diagnostic accuracy with acquisition cost for heart failure assessment
Details
Motivation: Point-of-care ultrasound (POCUS) requires efficient bedside decision-making under time and operator constraints. Current approaches lack personalized, cost-aware acquisition strategies that balance diagnostic performance with practical workflow needs.Method: Model POCUS as sequential acquisition problem with RL agent selecting next view or terminating acquisition. Upon termination, multi-view transformer performs multi-task inference (ordinal AS classification and LVEF regression) with Gaussian predictive distributions. Reward function balances expected diagnostic benefit against acquisition cost.
Result: On 1,820 test studies, method matches full-study performance using 32% fewer videos, achieving 77.2% mean balanced accuracy across AS severity classification and LVEF estimation. Demonstrates robust multi-task performance under acquisition budgets.
Conclusion: Patient-tailored, cost-aware acquisition can streamline POCUS workflows while preserving decision quality, producing interpretable scan pathways suited to bedside use. Framework extensible to additional cardiac endpoints.
Abstract: Purpose: Echocardiography with point-of-care ultrasound (POCUS) must support clinical decision-making under tight bedside time and operator-effort constraints. We introduce a personalized data acquisition strategy in which an RL agent, given a partially observed multi-view study, selects the next view to acquire or terminates acquisition to support heart-failure (HF) assessment. Upon termination, a diagnostic model jointly predicts aortic stenosis (AS) severity and left ventricular ejection fraction (LVEF), two key HF biomarkers, and outputs uncertainty, enabling an explicit trade-off between diagnostic performance and acquisition cost. Methods: We model POCUS as a sequential acquisition problem: at each step, a video selector (RL agent) chooses the next view to acquire or terminates acquisition. Upon termination, a shared multi-view transformer performs multi-task inference with two heads, ordinal AS classification, and LVEF regression, and outputs Gaussian predictive distributions yielding ordinal probabilities over AS classes and EF thresholds. These probabilities drive a reward that balances expected diagnostic benefit against acquisition cost, producing patient-specific acquisition pathways. Results: The dataset comprises 12,180 patient-level studies, split into training/validation/test sets (75/15/15). On the 1,820 test studies, our method matches full-study performance while using 32% fewer videos, achieving 77.2% mean balanced accuracy (bACC) across AS severity classification and LVEF estimation, demonstrating robust multi-task performance under acquisition budgets. Conclusion: Patient-tailored, cost-aware acquisition can streamline POCUS workflows while preserving decision quality, producing interpretable scan pathways suited to bedside use. The framework is extensible to additional cardiac endpoints and merits prospective evaluation for clinical integration.
[191] LeafNet: A Large-Scale Dataset and Comprehensive Benchmark for Foundational Vision-Language Understanding of Plant Diseases
Khang Nguyen Quoc, Phuong D. Dao, Luyl-Da Quach
Main category: cs.CV
TL;DR: LeafNet dataset and LeafBench benchmark for evaluating VLMs on plant disease understanding, showing VLMs outperform vision-only models but struggle with fine-grained identification.
Details
Motivation: Current VLMs lack application in domain-specific agricultural tasks like plant pathology due to missing large-scale multimodal datasets and benchmarks.Method: Created LeafNet dataset with 186K leaf images across 97 disease classes and LeafBench benchmark with 13,950 QA pairs across 6 agricultural tasks to evaluate 12 state-of-the-art VLMs.
Result: VLMs outperform vision-only models, with >90% accuracy on binary classification but <65% on fine-grained pathogen/species identification, showing significant performance gaps across tasks.
Conclusion: Multimodal architectures enhance diagnostic precision over vision-only models, but current VLMs have critical gaps in plant pathology, highlighting need for specialized benchmarks like LeafBench.
Abstract: Foundation models and vision-language pre-training have significantly advanced Vision-Language Models (VLMs), enabling multimodal processing of visual and linguistic data. However, their application in domain-specific agricultural tasks, such as plant pathology, remains limited due to the lack of large-scale, comprehensive multimodal image–text datasets and benchmarks. To address this gap, we introduce LeafNet, a comprehensive multimodal dataset, and LeafBench, a visual question-answering benchmark developed to systematically evaluate the capabilities of VLMs in understanding plant diseases. The dataset comprises 186,000 leaf digital images spanning 97 disease classes, paired with metadata, generating 13,950 question-answer pairs spanning six critical agricultural tasks. The questions assess various aspects of plant pathology understanding, including visual symptom recognition, taxonomic relationships, and diagnostic reasoning. Benchmarking 12 state-of-the-art VLMs on our LeafBench dataset, we reveal substantial disparity in their disease understanding capabilities. Our study shows performance varies markedly across tasks: binary healthy–diseased classification exceeds 90% accuracy, while fine-grained pathogen and species identification remains below 65%. Direct comparison between vision-only models and VLMs demonstrates the critical advantage of multimodal architectures: fine-tuned VLMs outperform traditional vision models, confirming that integrating linguistic representations significantly enhances diagnostic precision. These findings highlight critical gaps in current VLMs for plant pathology applications and underscore the need for LeafBench as a rigorous framework for methodological advancement and progress evaluation toward reliable AI-assisted plant disease diagnosis. Code is available at https://github.com/EnalisUs/LeafBench.
[192] EchoTorrent: Towards Swift, Sustained, and Streaming Multi-Modal Video Generation
Rang Meng, Weipeng Wu, Yingjie Yin, Yuming Li, Chenguang Ma
Main category: cs.CV
TL;DR: EchoTorrent is a novel schema for real-time multi-modal video generation that addresses latency, temporal stability, and multimodal degradation issues through multi-teacher training, adaptive CFG calibration, hybrid long tail forcing, and VAE decoder refinement.
Details
Motivation: Current multi-modal video generation models suffer from prohibitive latency, limited temporal stability, and multimodal degradation (spatial blurring, temporal drift, lip desynchronization) in streaming inference, creating an unresolved efficiency-performance trade-off.Method: Fourfold design: (1) Multi-Teacher Training with domain experts transferring knowledge to a student model; (2) Adaptive CFG Calibration (ACC-DMD) for audio CFG error calibration; (3) Hybrid Long Tail Forcing for alignment on tail frames; (4) VAE Decoder Refiner for pixel-domain optimization.
Result: Achieves few-pass autoregressive generation with substantially extended temporal consistency, identity preservation, and audio-lip synchronization, demonstrating improved performance in streaming mode.
Conclusion: EchoTorrent effectively addresses the efficiency-performance trade-off in multi-modal video generation, enabling real-time deployment with enhanced temporal stability and multimodal alignment.
Abstract: Recent multi-modal video generation models have achieved high visual quality, but their prohibitive latency and limited temporal stability hinder real-time deployment. Streaming inference exacerbates these issues, leading to pronounced multimodal degradation, such as spatial blurring, temporal drift, and lip desynchronization, which creates an unresolved efficiency-performance trade-off. To this end, we propose EchoTorrent, a novel schema with a fourfold design: (1) Multi-Teacher Training fine-tunes a pre-trained model on distinct preference domains to obtain specialized domain experts, which sequentially transfer domain-specific knowledge to a student model; (2) Adaptive CFG Calibration (ACC-DMD), which calibrates the audio CFG augmentation errors in DMD via a phased spatiotemporal schedule, eliminating redundant CFG computations and enabling single-pass inference per step; (3) Hybrid Long Tail Forcing, which enforces alignment exclusively on tail frames during long-horizon self-rollout training via a causal-bidirectional hybrid architecture, effectively mitigates spatiotemporal degradation in streaming mode while enhancing fidelity to reference frames; and (4) VAE Decoder Refiner through pixel-domain optimization of the VAE decoder to recover high-frequency details while circumventing latent-space ambiguities. Extensive experiments and analysis demonstrate that EchoTorrent achieves few-pass autoregressive generation with substantially extended temporal consistency, identity preservation, and audio-lip synchronization.
[193] An Ensemble Learning Approach towards Waste Segmentation in Cluttered Environment
Maimoona Jafar, Syed Imran Ali, Ahsan Saadat, Muhammad Bilal, Shah Khalid
Main category: cs.CV
TL;DR: An ensemble learning approach combining U-Net and FPN segmentation models improves waste segregation accuracy for robotic recycling systems.
Details
Motivation: Waste segregation is crucial for recycling but challenging due to complex real-world environments with deformed, overlapping objects. Computer vision can help robots accurately localize and pick waste items from conveyor belts.Method: Proposes an ensemble learning approach combining U-Net (good at capturing fine details) and FPN (effective at handling scale variation) using weighted average method. Uses dataset mimicking real-life waste scenarios with preprocessing to enhance feature learning.
Result: Ensemble model (EL-4) achieved IoU of 0.8306 (improved from U-Net’s 0.8065) and reduced Dice loss to 0.09019 (from FPN’s 0.1183).
Conclusion: The ensemble approach improves segmentation accuracy for waste segregation, potentially enhancing efficiency at Material Recovery Facilities with minimal human intervention.
Abstract: Environmental pollution is a critical global issue, with recycling emerging as one of the most viable solutions. This study focuses on waste segregation, a crucial step in recycling processes to obtain raw material. Recent advancements in computer vision have significantly contributed to waste classification and recognition. In waste segregation, segmentation masks are essential for robots to accurately localize and pick objects from conveyor belts. The complexity of real-world waste environments, characterized by deformed items without specific patterns and overlapping objects, further complicates waste segmentation tasks. This paper proposes an Ensemble Learning approach to improve segmentation accuracy by combining high performing segmentation models, U-Net and FPN, using a weighted average method. U-Net excels in capturing fine details and boundaries in segmentation tasks, while FPN effectively handles scale variation and context in complex environments, and their combined masks result in more precise predictions. The dataset used closely mimics real-life waste scenarios, and preprocessing techniques were applied to enhance feature learning for deep learning segmentation models. The ensemble model, referred to as EL-4, achieved an IoU value of 0.8306, an improvement over U-Net’s 0.8065, and reduced Dice loss to 0.09019 from FPN’s 0.1183. This study could contribute to the efficiency of waste sorting at Material Recovery Facility, facilitating better raw material acquisition for recycling with minimal human intervention and enhancing the overall throughput.
[194] A WDLoRA-Based Multimodal Generative Framework for Clinically Guided Corneal Confocal Microscopy Image Synthesis in Diabetic Neuropathy
Xin Zhang, Liangxiu Han, Yue Shi, Yalin Zheng, Uazman Alam, Maryam Ferdousi, Rayaz Malik
Main category: cs.CV
TL;DR: A multimodal generative framework using Weight-Decomposed LoRA for synthesizing corneal confocal microscopy images with clinical guidance, improving diagnostic model training for diabetic peripheral neuropathy.
Details
Motivation: Limited labelled data and fine-grained variability in corneal nerve morphology hinder development of robust deep learning diagnostic models for diabetic peripheral neuropathy using corneal confocal microscopy. Existing foundation generative models struggle with medical imaging due to lack of domain-specific training and anatomical fidelity.Method: Proposes Weight-Decomposed Low-Rank Adaptation (WDLoRA)-based multimodal generative framework that decouples magnitude and directional weight updates, enabling foundation models to independently learn nerve topology (orientation) and stromal contrast (intensity). Jointly conditions on nerve segmentation masks and disease-specific clinical prompts to synthesize anatomically coherent images across DPN spectrum.
Result: Achieves state-of-the-art visual fidelity (FID: 5.18) and structural integrity (SSIM: 0.630), significantly outperforming GAN and standard diffusion baselines. Synthetic images preserve clinical biomarkers and are statistically equivalent to real patient data. Improves downstream diagnostic accuracy by 2.1% and segmentation performance by 2.2% when used for training.
Conclusion: The proposed framework effectively addresses data bottlenecks in medical AI by generating clinically valid synthetic images that enhance diagnostic model performance for diabetic peripheral neuropathy assessment.
Abstract: Corneal Confocal Microscopy (CCM) is a sensitive tool for assessing small-fiber damage in Diabetic Peripheral Neuropathy (DPN), yet the development of robust, automated deep learning-based diagnostic models is limited by scarce labelled data and fine-grained variability in corneal nerve morphology. Although Artificial Intelligence (AI)-driven foundation generative models excel at natural image synthesis, they often struggle in medical imaging due to limited domain-specific training, compromising the anatomical fidelity required for clinical analysis. To overcome these limitations, we propose a Weight-Decomposed Low-Rank Adaptation (WDLoRA)-based multimodal generative framework for clinically guided CCM image synthesis. WDLoRA is a parameter-efficient fine-tuning (PEFT) mechanism that decouples magnitude and directional weight updates, enabling foundation generative models to independently learn the orientation (nerve topology) and intensity (stromal contrast) required for medical realism. By jointly conditioning on nerve segmentation masks and disease-specific clinical prompts, the model synthesises anatomically coherent images across the DPN spectrum (Control, T1NoDPN, T1DPN). A comprehensive three-pillar evaluation demonstrates that the proposed framework achieves state-of-the-art visual fidelity (Fréchet Inception Distance (FID): 5.18) and structural integrity (Structural Similarity Index Measure (SSIM): 0.630), significantly outperforming GAN and standard diffusion baselines. Crucially, the synthetic images preserve gold-standard clinical biomarkers and are statistically equivalent to real patient data. When used to train automated diagnostic models, the synthetic dataset improves downstream diagnostic accuracy by 2.1% and segmentation performance by 2.2%, validating the framework’s potential to alleviate data bottlenecks in medical AI.
[195] Fine-tuned Vision Language Model for Localization of Parasitic Eggs in Microscopic Images
Chan Hao Sien, Hezerul Abdul Karim, Nouar AlDahoul
Main category: cs.CV
TL;DR: Vision language model fine-tuned for parasitic egg localization in microscopic images outperforms traditional object detection methods
Details
Motivation: Soil-transmitted helminth infections affect large populations in tropical regions where diagnostic expertise is limited. Manual microscopic diagnosis is labor-intensive, time-consuming, and error-prone, necessitating automated solutions.Method: Fine-tuned Microsoft Florence vision language model (VLM) to localize parasitic eggs within microscopic images, comparing performance against traditional object detection methods like EfficientDet.
Result: The localization VLM achieved mIOU of 0.94, performing comparatively better than other object detection methods, demonstrating strong potential for automated parasitological diagnosis.
Conclusion: The proposed VLM shows promise as a core component of an automated framework for intelligent parasitological diagnosis, offering a scalable engineering solution for regions with limited diagnostic expertise.
Abstract: Soil-transmitted helminth (STH) infections continuously affect a large proportion of the global population, particularly in tropical and sub-tropical regions, where access to specialized diagnostic expertise is limited. Although manual microscopic diagnosis of parasitic eggs remains the diagnostic gold standard, the approach can be labour-intensive, time-consuming, and prone to human error. This paper aims to utilize a vision language model (VLM) such as Microsoft Florence that was fine-tuned to localize all parasitic eggs within microscopic images. The preliminary results show that our localization VLM performs comparatively better than the other object detection methods, such as EfficientDet, with an mIOU of 0.94. This finding demonstrates the potential of the proposed VLM to serve as a core component of an automated framework, offering a scalable engineering solution for intelligent parasitological diagnosis.
[196] RGA-Net: A Vision Enhancement Framework for Robotic Surgical Systems Using Reciprocal Attention Mechanisms
Quanjun Li, Weixuan Li, Han Xia, Junhua Zhou, Chi-Man Pun, Xuhang Chen
Main category: cs.CV
TL;DR: RGA-Net is a deep learning framework for surgical smoke removal in robotic surgery using hierarchical encoder-decoder with dual-stream hybrid attention and axis-decomposed attention modules.
Details
Motivation: Surgical smoke from energy-based devices degrades endoscopic video feeds in robotic surgery, compromising visual feedback for teleoperation and surgical outcomes, necessitating effective smoke removal solutions.Method: Hierarchical encoder-decoder architecture with two key innovations: Dual-Stream Hybrid Attention (DHA) combining shifted window attention with frequency-domain processing, and Axis-Decomposed Attention (ADA) with factorized attention mechanisms, connected via reciprocal cross-gating blocks.
Result: Superior performance on DesmokeData and LSD3K surgical datasets, effectively restoring visual clarity suitable for robotic surgery integration and enhancing surgeon-robot interface.
Conclusion: RGA-Net represents a significant step toward more reliable and safer robotic surgical systems through computational vision enhancement, with potential to reduce cognitive burden and iatrogenic injury risks.
Abstract: Robotic surgical systems rely heavily on high-quality visual feedback for precise teleoperation; yet, surgical smoke from energy-based devices significantly degrades endoscopic video feeds, compromising the human-robot interface and surgical outcomes. This paper presents RGA-Net (Reciprocal Gating and Attention-fusion Network), a novel deep learning framework specifically designed for smoke removal in robotic surgery workflows. Our approach addresses the unique challenges of surgical smoke-including dense, non-homogeneous distribution and complex light scattering-through a hierarchical encoder-decoder architecture featuring two key innovations: (1) a Dual-Stream Hybrid Attention (DHA) module that combines shifted window attention with frequency-domain processing to capture both local surgical details and global illumination changes, and (2) an Axis-Decomposed Attention (ADA) module that efficiently processes multi-scale features through factorized attention mechanisms. These components are connected via reciprocal cross-gating blocks that enable bidirectional feature modulation between encoder and decoder pathways. Extensive experiments on the DesmokeData and LSD3K surgical datasets demonstrate that RGA-Net achieves superior performance in restoring visual clarity suitable for robotic surgery integration. Our method enhances the surgeon-robot interface by providing consistently clear visualization, laying a technical foundation for alleviating surgeons’ cognitive burden, optimizing operation workflows, and reducing iatrogenic injury risks in minimally invasive procedures. These practical benefits could be further validated through future clinical trials involving surgeon usability assessments. The proposed framework represents a significant step toward more reliable and safer robotic surgical systems through computational vision enhancement.
[197] Explore Intrinsic Geometry for Query-based Tiny and Oriented Object Detector with Momentum-based Bipartite Matching
Junpeng Zhang, Zewei Yang, Jie Feng, Yuhui Zheng, Ronghua Shang, Mengxuan Zhang
Main category: cs.CV
TL;DR: IGOFormer: A query-based oriented object detector that integrates intrinsic geometry into feature decoding and uses momentum-based bipartite matching for better detection of oriented objects, especially tiny ones in aerial imagery.
Details
Motivation: Current query-based detectors struggle with oriented objects, particularly tiny objects with limited texture information, due to underutilization of intrinsic geometry during feature decoding and inter-stage matching inconsistency from stage-wise bipartite matching.Method: Proposes IGOFormer with two key components: 1) Intrinsic Geometry-aware Decoder that enhances object-related features by injecting geometric embeddings to capture object orientation, and 2) Momentum-based Bipartite Matching that aggregates historical matching costs using exponential moving average with query-specific smoothing factors to stabilize inter-stage matching.
Result: Achieves state-of-the-art performance for aerial oriented object detection, with AP₅₀ score of 78.00% on DOTA-V1.0 using Swin-T backbone under single-scale setting.
Conclusion: IGOFormer effectively addresses limitations in oriented object detection by integrating intrinsic geometry into feature decoding and stabilizing inter-stage matching, demonstrating superior performance especially for challenging tiny objects in aerial imagery.
Abstract: Recent query-based detectors have achieved remarkable progress, yet their performance remains constrained when handling objects with arbitrary orientations, especially for tiny objects capturing limited texture information. This limitation primarily stems from the underutilization of intrinsic geometry during pixel-based feature decoding and the occurrence of inter-stage matching inconsistency caused by stage-wise bipartite matching. To tackle these challenges, we present IGOFormer, a novel query-based oriented object detector that explicitly integrates intrinsic geometry into feature decoding and enhances inter-stage matching stability. Specifically, we design an Intrinsic Geometry-aware Decoder, which enhances the object-related features conditioned on an object query by injecting complementary geometric embeddings extrapolated from their correlations to capture the geometric layout of the object, thereby offering a critical geometric insight into its orientation. Meanwhile, a Momentum-based Bipartite Matching scheme is developed to adaptively aggregate historical matching costs by formulating an exponential moving average with query-specific smoothing factors, effectively preventing conflicting supervisory signals arising from inter-stage matching inconsistency. Extensive experiments and ablation studies demonstrate the superiority of our IGOFormer for aerial oriented object detection, achieving an AP$_{50}$ score of 78.00% on DOTA-V1.0 using Swin-T backbone under the single-scale setting. The code will be made publicly available.
[198] Generative Latent Representations of 3D Brain MRI for Multi-Task Downstream Analysis in Down Syndrome
Jordi Malé, Juan Fortea, Mateus Rozalem-Aranha, Neus Martínez-Abadías, Xavier Sevillano
Main category: cs.CV
TL;DR: VAEs encode 3D brain MRI scans into latent representations for generative and predictive applications, evaluated through reconstruction quality, latent space visualization, and classification of Down syndrome vs. euploid individuals.
Details
Motivation: Latent representations in medical imaging generative models are underexplored despite their potential for neuroimaging research and clinical applications. Understanding their structure and information content is crucial for advancing generative models in medical decision-making.Method: Develop multiple variational autoencoders (VAEs) to encode 3D brain MRI scans into compact latent representations. Systematically evaluate through: (1) quantitative/qualitative MRI reconstruction assessment, (2) latent space visualization using Principal Component Analysis, and (3) downstream classification tasks on proprietary dataset of euploid and Down syndrome brain MRI scans.
Result: VAE successfully captures essential brain features with high reconstruction fidelity. Latent space exhibits clear clustering patterns, particularly distinguishing individuals with Down syndrome from euploid controls.
Conclusion: VAE-based latent representations effectively encode meaningful information from 3D brain MRI scans, demonstrating utility for both generative reconstruction and predictive classification tasks in neuroimaging.
Abstract: Generative models have emerged as powerful tools in medical imaging, enabling tasks such as segmentation, anomaly detection, and high-quality synthetic data generation. These models typically rely on learning meaningful latent representations, which are particularly valuable given the high-dimensional nature of 3D medical images like brain magnetic resonance imaging (MRI) scans. Despite their potential, latent representations remain underexplored in terms of their structure, information content, and applicability to downstream clinical tasks. Investigating these representations is crucial for advancing the use of generative models in neuroimaging research and clinical decision-making. In this work, we develop multiple variational autoencoders (VAEs) to encode 3D brain MRI scans into compact latent space representations for generative and predictive applications. We systematically evaluate the effectiveness of the learned representations through three key analyses: (i) a quantitative and qualitative assessment of MRI reconstruction quality, (ii) a visualisation of the latent space structure using Principal Component Analysis, and (iii) downstream classification tasks on a proprietary dataset of euploid and Down syndrome individuals brain MRI scans. Our results demonstrate that the VAE successfully captures essential brain features while maintaining high reconstruction fidelity. The latent space exhibits clear clustering patterns, particularly in distinguishing individuals with Down syndrome from euploid controls.
[199] T2MBench: A Benchmark for Out-of-Distribution Text-to-Motion Generation
Bin Yang, Rong Ou, Weisheng Xu, Jiaqi Xiong, Xintao Li, Taowen Wang, Luyu Zhu, Xu Jiang, Jing Tan, Renjing Xu
Main category: cs.CV
TL;DR: Proposes a benchmark for out-of-distribution text-to-motion evaluation with 1,025 OOD prompts, evaluating 14 baseline models using LLM-based, multi-factor motion, and fine-grained accuracy metrics.
Details
Motivation: Existing text-to-motion evaluations focus on in-distribution inputs and limited criteria, failing to assess model generalization under complex OOD textual conditions.Method: Constructs OOD prompt dataset (1,025 textual descriptions), introduces unified evaluation framework with LLM-based Evaluation, Multi-factor Motion evaluation, and Fine-grained Accuracy Evaluation, tests 14 baseline models.
Result: Models show strengths in semantic alignment, motion generalizability, and physical quality, but struggle with fine-grained accuracy in OOD scenarios.
Conclusion: Highlights limitations of existing methods in OOD scenarios and provides practical guidance for future production-level text-to-motion model design and evaluation.
Abstract: Most existing evaluations of text-to-motion generation focus on in-distribution textual inputs and a limited set of evaluation criteria, which restricts their ability to systematically assess model generalization and motion generation capabilities under complex out-of-distribution (OOD) textual conditions. To address this limitation, we propose a benchmark specifically designed for OOD text-to-motion evaluation, which includes a comprehensive analysis of 14 representative baseline models and the two datasets derived from evaluation results. Specifically, we construct an OOD prompt dataset consisting of 1,025 textual descriptions. Based on this prompt dataset, we introduce a unified evaluation framework that integrates LLM-based Evaluation, Multi-factor Motion evaluation, and Fine-grained Accuracy Evaluation. Our experimental results reveal that while different baseline models demonstrate strengths in areas such as text-to-motion semantic alignment, motion generalizability, and physical quality, most models struggle to achieve strong performance with Fine-grained Accuracy Evaluation. These findings highlight the limitations of existing methods in OOD scenarios and offer practical guidance for the design and evaluation of future production-level text-to-motion models.
[200] OmniScience: A Large-scale Multi-modal Dataset for Scientific Image Understanding
Haoyi Tao, Chaozheng Huang, Nan Wang, Han Lyu, Linfeng Zhang, Guolin Ke, Xi Fang
Main category: cs.CV
TL;DR: OmniScience: A large-scale scientific multimodal dataset with 1.5M figure-caption-context triplets across 10+ disciplines, featuring high-fidelity annotations via dynamic model-routing re-captioning pipeline, significantly improving scientific image understanding in MLLMs.
Details
Motivation: Current MLLMs perform well on natural images but struggle with scientific images (diagrams, charts, characterizations) due to limited domain coverage, coarse annotations, and weak semantic grounding in existing datasets.Method: Created OmniScience dataset with 1.5M scientific figure-caption-context triplets. Developed dynamic model-routing re-captioning pipeline that synthesizes visual features, original captions, and in-text references using SOTA MLLMs, with quality filtering and human alignment.
Result: Pipeline boosted image-text similarity from 0.769 to 0.956. Qwen2.5-VL-3B finetuned on OmniScience achieved gains of 0.378 on MM-MT-Bench and 0.140 on MMMU benchmarks, showing substantial improvement in scientific visual understanding.
Conclusion: OmniScience addresses the scientific image understanding gap in MLLMs through high-quality dataset creation and re-captioning pipeline, enabling significant performance improvements on scientific visual reasoning tasks.
Abstract: Multimodal Large Language Models demonstrate strong performance on natural image understanding, yet exhibit limited capability in interpreting scientific images, including but not limited to schematic diagrams, experimental characterizations, and analytical charts. This limitation is particularly pronounced in open-source MLLMs. The gap largely stems from existing datasets with limited domain coverage, coarse structural annotations, and weak semantic grounding. We introduce OmniScience, a large-scale, high-fidelity multi-modal dataset comprising 1.5 million figure-caption-context triplets, spanning more than 10 major scientific disciplines. To obtain image caption data with higher information density and accuracy for multi-modal large-model training, we develop a dynamic model-routing re-captioning pipeline that leverages state-of-the-art multi-modal large language models to generate dense, self-contained descriptions by jointly synthesizing visual features, original figure captions, and corresponding in-text references authored by human scientists. The pipeline is further reinforced with rigorous quality filtering and alignment with human expert judgments, ensuring both factual accuracy and semantic completeness, and boosts the image-text multi-modal similarity score from 0.769 to 0.956. We further propose a caption QA protocol as a proxy task for evaluating visual understanding. Under this setting, Qwen2.5-VL-3B model finetuned on OmniScience show substantial gains over baselines, achieving a gain of 0.378 on MM-MT-Bench and a gain of 0.140 on MMMU.
[201] SAM4Dcap: Training-free Biomechanical Twin System from Monocular Video
Li Wang, HaoYu Wang, Xi Chen, ZeKun Jiang, Kang Li, Jian Li
Main category: cs.CV
TL;DR: SAM4Dcap is an open-source framework that estimates biomechanical metrics from monocular video by combining 4D human mesh recovery with biomechanical simulation, enabling motion analysis outside laboratory settings.
Details
Motivation: Current biomechanical analysis is limited to expensive laboratory setups with optical motion capture systems. While multi-view video approaches have reduced costs, they remain impractical for home-based scenarios that require monocular capture. There's a need for accessible, non-laboratory motion analysis tools.Method: SAM4Dcap integrates SAM-Body4D for temporally consistent 4D human mesh recovery from monocular video with the OpenSim biomechanical solver. The pipeline converts reconstructed meshes into trajectory files compatible with musculoskeletal models, includes automated prompting strategies, and provides a Linux-native build for processing.
Result: Preliminary evaluations on walking and drop-jump tasks show SAM4Dcap can achieve knee kinematic predictions comparable to multi-view systems, though some discrepancies in hip flexion and residual jitter remain. The framework provides a flexible foundation for non-laboratory motion analysis.
Conclusion: SAM4Dcap bridges advanced computer vision with established biomechanical simulation to create an accessible, open-source framework for estimating biomechanical metrics from monocular video, enabling motion analysis outside traditional laboratory settings.
Abstract: Quantitative biomechanical analysis is essential for clinical diagnosis and injury prevention but is often restricted to laboratories due to the high cost of optical motion capture systems. While multi-view video approaches have lowered barriers, they remain impractical for home-based scenarios requiring monocular capture. This paper presents SAM4Dcap, an open-source, end-to-end framework for estimating biomechanical metrics from monocular video without additional training. SAM4Dcap integrates the temporally consistent 4D human mesh recovery of SAM-Body4D with the OpenSim biomechanical solver. The pipeline converts reconstructed meshes into trajectory files compatible with diverse musculoskeletal models. We introduce automated prompting strategies and a Linux-native build for processing. Preliminary evaluations on walking and drop-jump tasks indicate that SAM4Dcap has the potential to achieve knee kinematic predictions comparable to multi-view systems, although some discrepancies in hip flexion and residual jitter remain. By bridging advanced computer vision with established biomechanical simulation, SAM4Dcap provides a flexible, accessible foundation for non-laboratory motion analysis.
[202] Offline-Poly: A Polyhedral Framework For Offline 3D Multi-Object Tracking
Xiaoyu Li, Yitao Wu, Xian Wu, Haolin Zhuo, Lijun Zhao, Lining Sun
Main category: cs.CV
TL;DR: Offline-Poly: A general offline 3D multi-object tracking method using Tracking-by-Tracking paradigm that refines coarse tracking outputs through hierarchical matching and fusion, achieving state-of-the-art performance on nuScenes and KITTI datasets.
Details
Motivation: Existing offline 3D MOT methods are direct extensions of online frameworks and fail to fully exploit offline advantages like resource unconstrainedness and future observability. They also depend on fixed upstream architectures, limiting adaptability.Method: Proposes Tracking-by-Tracking (TBT) paradigm that operates on arbitrary tracking outputs. Offline-Poly processes coarse tracking results through pre-processing, hierarchical matching/fusion, and tracklet refinement modules, leveraging offline properties for global optimization and full temporal reasoning.
Result: Achieves SOTA performance with 77.6% AMOTA on nuScenes and leading results with 83.00% HOTA on KITTI. Comprehensive experiments validate flexibility, generalizability, and modular effectiveness.
Conclusion: Offline-Poly provides a general, flexible offline 3D MOT framework that decouples from specific upstream detectors/trackers and effectively exploits offline advantages for superior tracking performance.
Abstract: Offline 3D multi-object tracking (MOT) is a critical component of the 4D auto-labeling (4DAL) process. It enhances pseudo-labels generated by high-performance detectors through the incorporation of temporal context. However, existing offline 3D MOT approaches are direct extensions of online frameworks and fail to fully exploit the advantages of offline setting. Moreover, these methods often depend on fixed upstream and customized architectures, limiting their adaptability. To address these limitations, we propose Offline-Poly, a general offline 3D MOT method based on a tracking-centric design. We introduce a standardized paradigm termed Tracking-by-Tracking (TBT), which operates exclusively on arbitrary off-the-shelf tracking outputs and produces offline-refined tracklets. This formulation decouples offline tracker from specific upstream detectors or trackers. Under the TBT paradigm, Offline-Poly accepts one or multiple coarse tracking results and processes them through a structured pipeline comprising pre-processing, hierarchical matching and fusion, and tracklet refinement. Each module is designed to capitalize on the two fundamental properties of offline tracking: resource unconstrainedness, which permits global optimization beyond real-time limits, and future observability, which enables tracklet reasoning over the full temporal horizon. Offline-Poly first eliminates short-term ghost tracklets and re-identifies fragmented segments using global scene context. It then constructs scene-level similarity to associate tracklets across multiple input sources. Finally, Offline-Poly refines tracklets by jointly leveraging local and global motion patterns. On nuScenes, we achieve SOTA performance with 77.6% AMOTA. On KITTI, it achieves leading results with 83.00% HOTA. Comprehensive experiments further validate the flexibility, generalizability, and modular effectiveness of Offline-Poly.
[203] Skeleton2Stage: Reward-Guided Fine-Tuning for Physically Plausible Dance Generation
Jidong Jia, Youjian Zhang, Huan Fu, Dacheng Tao
Main category: cs.CV
TL;DR: RLFT with physics-based rewards improves dance generation by addressing skeleton-to-mesh gap, reducing self-penetration and foot-ground contact issues while preserving motion dynamics.
Details
Motivation: Current dance generation methods work in skeletal domain but ignore mesh-level physical constraints, leading to body self-penetration and foot-ground contact anomalies when visualized with human body mesh, reducing aesthetic appeal and limiting real-world applications.Method: Uses Reinforcement Learning Fine-Tuning (RLFT) with physics-based rewards derived from body mesh: imitation reward measuring motion plausibility via physical simulator, Foot-Ground Deviation reward with test-time guidance, and anti-freezing reward to preserve motion dynamics.
Result: Experiments on multiple dance datasets show significant improvement in physical plausibility of generated motions, yielding more realistic and aesthetically pleasing dances while maintaining motion dynamics.
Conclusion: The proposed RLFT approach effectively bridges the skeleton-to-mesh gap in dance generation, producing physically plausible motions suitable for mesh visualization while preserving the dynamic qualities of dance.
Abstract: Despite advances in dance generation, most methods are trained in the skeletal domain and ignore mesh-level physical constraints. As a result, motions that look plausible as joint trajectories often exhibit body self-penetration and Foot-Ground Contact (FGC) anomalies when visualized with a human body mesh, reducing the aesthetic appeal of generated dances and limiting their real-world applications. We address this skeleton-to-mesh gap by deriving physics-based rewards from the body mesh and applying Reinforcement Learning Fine-Tuning (RLFT) to steer the diffusion model toward physically plausible motion synthesis under mesh visualization. Our reward design combines (i) an imitation reward that measures a motion’s general plausibility by its imitability in a physical simulator (penalizing penetration and foot skating), and (ii) a Foot-Ground Deviation (FGD) reward with test-time FGD guidance to better capture the dynamic foot-ground interaction in dance. However, we find that the physics-based rewards tend to push the model to generate freezing motions for fewer physical anomalies and better imitability. To mitigate it, we propose an anti-freezing reward to preserve motion dynamics while maintaining physical plausibility. Experiments on multiple dance datasets consistently demonstrate that our method can significantly improve the physical plausibility of generated motions, yielding more realistic and aesthetically pleasing dances. The project page is available at: https://jjd1123.github.io/Skeleton2Stage/
[204] Foundation Model-Driven Semantic Change Detection in Remote Sensing Imagery
Hengtong Shen, Li Yan, Hong Xie, Yaxuan Wei, Xinhao Li, Wenfei Shen, Peixian Lv, Fei Tan
Main category: cs.CV
TL;DR: PerASCD is a semantic change detection method for remote sensing imagery that uses foundation model PerA with a Cascaded Gated Decoder and Soft Semantic Consistency Loss to improve multi-scale semantic understanding and performance.
Details
Motivation: Existing semantic change detection methods for remote sensing face challenges due to limited semantic understanding capability of models and inherent task complexity, leading to performance limitations and paradigm complexity.Method: Proposes PerASCD using RS foundation model PerA with a modular Cascaded Gated Decoder to simplify complex SCD decoding pipelines and promote multi-level feature interaction, plus Soft Semantic Consistency Loss to mitigate training instability.
Result: Achieves state-of-the-art performance on two public benchmark datasets, effectively simplifies SCD paradigm, and shows seamless adaptation across various vision encoders.
Conclusion: PerASCD enhances multi-scale semantic understanding and overall performance in semantic change detection for remote sensing imagery through foundation model integration and simplified decoding architecture.
Abstract: Remote sensing (RS) change detection methods can extract critical information on surface dynamics and are an essential means for humans to understand changes in the earth’s surface and environment. Among these methods, semantic change detection (SCD) can more effectively interpret the multi-class information contained in bi-temporal RS imagery, providing semantic-level predictions that support dynamic change monitoring. However, due to the limited semantic understanding capability of the model and the inherent complexity of the SCD tasks, existing SCD methods face significant challenges in both performance and paradigm complexity. In this paper, we propose PerASCD, a SCD method driven by RS foundation model PerA, designed to enhance the multi-scale semantic understanding and overall performance. We introduce a modular Cascaded Gated Decoder (CG-Decoder) that simplifies complex SCD decoding pipelines while promoting effective multi-level feature interaction and fusion. In addition, we propose a Soft Semantic Consistency Loss (SSCLoss) to mitigate the numerical instability commonly encountered during SCD training. We further explore the applicability of multiple existing RS foundation models on the SCD task when equipped with the proposed decoder. Experimental results demonstrate that our decoder not only effectively simplifies the paradigm of SCD, but also achieves seamless adaptation across various vision encoders. Our method achieves state-of-the-art (SOTA) performance on two public benchmark datasets, validating its effectiveness. The code is available at https://github.com/SathShen/PerASCD.git.
[205] Joint Orientation and Weight Optimization for Robust Watertight Surface Reconstruction via Dirichlet-Regularized Winding Fields
Jiaze Li, Daisheng Jin, Fei Hou, Junhui Hou, Zheng Liu, Shiqing Xin, Wenping Wang, Ying He
Main category: cs.CV
TL;DR: DiWR is a robust method for reconstructing watertight surfaces from unoriented point clouds with non-uniform sampling, noise, and outliers by jointly optimizing orientations, area weights, and confidence coefficients using generalized winding number field optimization.
Details
Motivation: Existing methods struggle with reconstructing watertight surfaces from challenging point clouds with non-uniform sampling, noise, and outliers, especially from sources like 3D Gaussian Splatting and corrupted graphics data. Traditional multi-stage pipelines and recent joint methods have limitations in handling these real-world imperfections.Method: Uses generalized winding number (GWN) field as implicit representation and jointly optimizes point orientations, per-point area weights, and confidence coefficients in a single pipeline. Minimizes Dirichlet energy of the induced winding field with additional GWN-based constraints to compensate for non-uniform sampling, reduce noise impact, and downweight outliers without separate preprocessing.
Result: DiWR produces plausible watertight surfaces on challenging inputs including point clouds from 3D Gaussian Splatting and corrupted graphics benchmarks, outperforming both traditional multi-stage pipelines and recent joint orientation-reconstruction methods.
Conclusion: DiWR provides a robust solution for watertight surface reconstruction from imperfect point clouds, effectively handling non-uniform sampling, noise, and outliers through joint optimization in a single pipeline without preprocessing dependencies.
Abstract: We propose Dirichlet Winding Reconstruction (DiWR), a robust method for reconstructing watertight surfaces from unoriented point clouds with non-uniform sampling, noise, and outliers. Our method uses the generalized winding number (GWN) field as the target implicit representation and jointly optimizes point orientations, per-point area weights, and confidence coefficients in a single pipeline. The optimization minimizes the Dirichlet energy of the induced winding field together with additional GWN-based constraints, allowing DiWR to compensate for non-uniform sampling, reduce the impact of noise, and downweight outliers during reconstruction, with no reliance on separate preprocessing. We evaluate DiWR on point clouds from 3D Gaussian Splatting, a computer-vision pipeline, and corrupted graphics benchmarks. Experiments show that DiWR produces plausible watertight surfaces on these challenging inputs and outperforms both traditional multi-stage pipelines and recent joint orientation-reconstruction methods.
[206] Gaussian Sequences with Multi-Scale Dynamics for 4D Reconstruction from Monocular Casual Videos
Can Li, Jie Gu, Jingmin Chen, Fangzhou Qiu, Lei Sun
Main category: cs.CV
TL;DR: Multi-scale dynamics factorization enables accurate 4D reconstruction from monocular videos using Gaussian sequences with layered motion representation and multi-modal priors.
Details
Motivation: Dynamic scene understanding from casual monocular videos is critical for scalable robot learning, but 4D reconstruction under strictly monocular settings remains highly ill-posed due to ambiguity in motion estimation.Method: Proposes multi-scale dynamics mechanism that factorizes complex motion fields, Gaussian sequences with multi-scale dynamics representation derived through compositions of multi-level motion, and incorporates multi-modal priors from vision foundation models for complementary supervision.
Result: Achieves accurate and globally consistent 4D reconstruction from monocular casual videos, with considerable improvements over existing methods in dynamic novel-view synthesis on benchmark and real-world manipulation datasets.
Conclusion: Multi-scale dynamics factorization with layered Gaussian representation and multi-modal priors effectively addresses the ambiguity in monocular 4D reconstruction, enabling physically plausible dynamic scene understanding.
Abstract: Understanding dynamic scenes from casual videos is critical for scalable robot learning, yet four-dimensional (4D) reconstruction under strictly monocular settings remains highly ill-posed. To address this challenge, our key insight is that real-world dynamics exhibits a multi-scale regularity from object to particle level. To this end, we design the multi-scale dynamics mechanism that factorizes complex motion fields. Within this formulation, we propose Gaussian sequences with multi-scale dynamics, a novel representation for dynamic 3D Gaussians derived through compositions of multi-level motion. This layered structure substantially alleviates ambiguity of reconstruction and promotes physically plausible dynamics. We further incorporate multi-modal priors from vision foundation models to establish complementary supervision, constraining the solution space and improving the reconstruction fidelity. Our approach enables accurate and globally consistent 4D reconstruction from monocular casual videos. Experiments of dynamic novel-view synthesis (NVS) on benchmark and real-world manipulation datasets demonstrate considerable improvements over existing methods.
[207] VAR-3D: View-aware Auto-Regressive Model for Text-to-3D Generation via a 3D Tokenizer
Zongcheng Han, Dongyan Cao, Haoran Sun, Yu Hong
Main category: cs.CV
TL;DR: VAR-3D is a novel text-to-3D generation framework that uses view-aware 3D VQ-VAE and rendering-supervised training to improve geometric coherence and text-3D alignment.
Details
Motivation: Text-to-3D generation faces challenges due to information loss during 3D representation encoding and representational distortion before quantization, which is amplified by vector quantization, degrading geometric coherence. The conventional two-stage training also causes objective mismatch between reconstruction and text-conditioned generation.Method: Proposes VAR-3D with: 1) View-aware 3D VQ-VAE to convert complex 3D geometric structures into discrete tokens, 2) Rendering-supervised training strategy that couples discrete token prediction with visual reconstruction to preserve visual fidelity and structural consistency relative to input text.
Result: VAR-3D significantly outperforms existing methods in both generation quality and text-3D alignment.
Conclusion: The proposed approach effectively addresses bottlenecks in learning discrete 3D representations and improves text-to-3D generation through better geometric coherence and alignment with textual descriptions.
Abstract: Recent advances in auto-regressive transformers have achieved remarkable success in generative modeling. However, text-to-3D generation remains challenging, primarily due to bottlenecks in learning discrete 3D representations. Specifically, existing approaches often suffer from information loss during encoding, causing representational distortion before the quantization process. This effect is further amplified by vector quantization, ultimately degrading the geometric coherence of text-conditioned 3D shapes. Moreover, the conventional two-stage training paradigm induces an objective mismatch between reconstruction and text-conditioned auto-regressive generation. To address these issues, we propose View-aware Auto-Regressive 3D (VAR-3D), which intergrates a view-aware 3D Vector Quantized-Variational AutoEncoder (VQ-VAE) to convert the complex geometric structure of 3D models into discrete tokens. Additionally, we introduce a rendering-supervised training strategy that couples discrete token prediction with visual reconstruction, encouraging the generative process to better preserve visual fidelity and structural consistency relative to the input text. Experiments demonstrate that VAR-3D significantly outperforms existing methods in both generation quality and text-3D alignment.
[208] Embed-RL: Reinforcement Learning for Reasoning-Driven Multimodal Embeddings
Haonan Jiang, Yuji Wang, Yongjie Zhu, Xin Lu, Wenyu Qin, Meng Wang, Pengfei Wan, Yansong Tang
Main category: cs.CV
TL;DR: A reasoning-driven Universal Multimodal Embeddings framework using Embedder-Guided Reinforcement Learning to optimize traceable Chain-of-Thought reasoning for improved cross-modal retrieval.
Details
Motivation: Existing generative embedding methods using Chain-of-Thought reasoning are limited to textual analysis of queries and irrelevant to target retrieval, lacking multimodal evidence and retrieval alignment.Method: Proposes Embedder-Guided Reinforcement Learning (EG-RL) framework where the Embedder supervises the Reasoner to produce evidential Traceability CoT (T-CoT) that extracts critical multimodal cues and provides retrieval-relevant inputs for the Embedder.
Result: Outperforms pioneering embedding models on MMEB-V2 and UVRB benchmarks with limited computational resources, demonstrating improved cross-modal semantic consistency and fine-grained matching capability.
Conclusion: Targeted reasoning optimization significantly improves multimodal embedding quality, providing a practical and efficient solution for reasoning-driven Universal Multimodal Embeddings development.
Abstract: Leveraging Multimodal Large Language Models (MLLMs) has become pivotal for advancing Universal Multimodal Embeddings (UME) in addressing diverse cross-modal tasks. Recent studies demonstrate that incorporating generative Chain-of-Thought (CoT) reasoning can substantially enhance task-specific representations compared to discriminative methods. However, the generated reasoning CoTs of existing generative embedding methods are limited to the textual analysis of queries and are irrelevant to the retrieval of the targets. To address these limitations, we propose a reasoning-driven UME framework that integrates Embedder-Guided Reinforcement Learning (EG-RL) to optimize the Reasoner to produce evidential Traceability CoT (T-CoT). Our key contributions are threefold: (1) We design an EG-RL framework where the Embedder provides explicit supervision to the Reasoner, ensuring the generated CoT traces are aligned with embedding tasks. (2) We introduce T-CoT, which extracts critical multimodal cues to focus on retrieval-relevant elements and provides multimodal inputs for the Embedder. (3) With limited computational resources, our framework outperforms the pioneering embedding model on both MMEB-V2 and UVRB benchmarks. The integration of multimodal evidence in structured reasoning, paired with retrieval-oriented alignment, effectively strengthens cross-modal semantic consistency and boosts the fine-grained matching capability of the model as well as the generalization across complex scenarios. Our work demonstrates that targeted reasoning optimization can significantly improve multimodal embedding quality, providing a practical and efficient solution for reasoning-driven UME development.
[209] Prior-guided Hierarchical Instance-pixel Contrastive Learning for Ultrasound Speckle Noise Suppression
Zhenyu Bu, Yuanxin Xie, Guang-Quan Zhou
Main category: cs.CV
TL;DR: A hierarchical contrastive learning model for ultrasound denoising that combines pixel-level and instance-level contrastive learning with a hybrid Transformer-CNN architecture to suppress speckle noise while preserving anatomical structures.
Details
Motivation: Ultrasound denoising is crucial for improving image quality and diagnostic reliability, but speckle patterns contain both noise and important anatomical details, making it challenging to suppress noise while preserving structural fidelity.Method: Proposes a prior-guided hierarchical instance-pixel contrastive learning model with: 1) statistics-guided pixel-level contrastive learning to enhance distributional discrepancies between noisy and clean pixels, 2) memory bank-based instance-level contrastive learning in feature space, and 3) hybrid Transformer-CNN architecture combining Transformer encoder for global context with CNN decoder for fine-grained structure restoration.
Result: Extensive evaluations on two publicly available ultrasound datasets demonstrate that the proposed model consistently outperforms existing methods, confirming its effectiveness and superiority.
Conclusion: The proposed hierarchical contrastive learning approach effectively addresses the ultrasound denoising challenge by promoting noise-invariant and structure-aware feature representations through complementary pixel and instance-level learning strategies combined with a hybrid architecture.
Abstract: Ultrasound denoising is essential for mitigating speckle-induced degradations, thereby enhancing image quality and improving diagnostic reliability. Nevertheless, because speckle patterns inherently encode both texture and fine anatomical details, effectively suppressing noise while preserving structural fidelity remains a significant challenge. In this study, we propose a prior-guided hierarchical instance-pixel contrastive learning model for ultrasound denoising, designed to promote noise-invariant and structure-aware feature representations by maximizing the separability between noisy and clean samples at both pixel and instance levels. Specifically, a statistics-guided pixel-level contrastive learning strategy is introduced to enhance distributional discrepancies between noisy and clean pixels, thereby improving local structural consistency. Concurrently, a memory bank is employed to facilitate instance-level contrastive learning in the feature space, encouraging representations that more faithfully approximate the underlying data distribution. Furthermore, a hybrid Transformer-CNN architecture is adopted, coupling a Transformer-based encoder for global context modeling with a CNN-based decoder optimized for fine-grained anatomical structure restoration, thus enabling complementary exploitation of long-range dependencies and local texture details. Extensive evaluations on two publicly available ultrasound datasets demonstrate that the proposed model consistently outperforms existing methods, confirming its effectiveness and superiority.
[210] High-Fidelity Causal Video Diffusion Models for Real-Time Ultra-Low-Bitrate Semantic Communication
Cem Eteke, Batuhan Tosun, Alexander Griessel, Wolfgang Kellerer, Eckehard Steinbach
Main category: cs.CV
TL;DR: A video diffusion model for high-fidelity, real-time video generation under ultra-low-bitrate semantic communication constraints, using semantic scene structure and compressed low-res frames with modular adapters and temporal distillation.
Details
Motivation: To enable high-quality video generation in ultra-low-bitrate semantic communication scenarios where traditional compression methods fail, requiring efficient transmission of semantic information while preserving visual fidelity and temporal consistency.Method: Uses lossy semantic video coding to transmit semantic scene structure plus compressed low-resolution frames for texture. Features modular video diffusion model with Semantic Control, Restoration Adapter, and Temporal Adapter modules. Includes efficient temporal distillation for real-time causal synthesis, reducing parameters by 300x and training time by 2x.
Result: Achieves strong perceptual quality, semantic fidelity, and temporal consistency at ultra-low bitrates (< 0.0003 bpp), outperforming classical, neural, and generative baselines in quantitative, qualitative, and subjective evaluations across diverse datasets.
Conclusion: The framework successfully enables high-fidelity, real-time video generation under extreme communication constraints through semantic coding and efficient diffusion modeling, advancing video generation for bandwidth-limited applications.
Abstract: We introduce a video diffusion model for high-fidelity, causal, and real-time video generation under ultra-low-bitrate semantic communication constraints. Our approach utilizes lossy semantic video coding to transmit the semantic scene structure, complemented by a stream of highly compressed, low-resolution frames that provide sufficient texture information to preserve fidelity. Building on these inputs, we introduce a modular video diffusion model that contains Semantic Control, Restoration Adapter, and Temporal Adapter. We further introduce an efficient temporal distillation procedure that enables extension to real-time and causal synthesis, reducing trainable parameters by 300x and training time by 2x, while adhering to communication constraints. Evaluated across diverse datasets, the framework achieves strong perceptual quality, semantic fidelity, and temporal consistency at ultra-low bitrates (< 0.0003 bpp), outperforming classical, neural, and generative baselines in extensive quantitative, qualitative, and subjective evaluations.
[211] Automated Prediction of Paravalvular Regurgitation before Transcatheter Aortic Valve Implantation
Michele Cannito, Riccardo Renzulli, Adson Duarte, Farzad Nikfam, Carlo Alberto Barbano, Enrico Chiesa, Francesco Bruno, Federico Giacobbe, Wojciech Wanha, Arturo Giordano, Marco Grangetto, Fabrizio D’Ascenzo
Main category: cs.CV
TL;DR: Deep learning approach using 3D CNNs on preoperative cardiac CT scans to predict paravalvular aortic regurgitation (PVR) risk after TAVI procedures.
Details
Motivation: Paravalvular aortic regurgitation (PVR) is a frequent and serious complication after Transcatheter Aortic Valve Implantation (TAVI) that impacts long-term prognosis, creating a need for better preoperative risk assessment methods.Method: Used 3D convolutional neural networks trained on isotropic CT volumes from preoperative TAVI patients to predict PVR occurrence from anatomical features in cardiac CT scans.
Result: The deep learning approach demonstrated potential to capture subtle anatomical features from pre-TAVI imaging, suggesting feasibility for personalized risk assessment and procedural optimization.
Conclusion: Volumetric deep learning shows promise for predicting PVR complications from preoperative CT, opening new perspectives for improving TAVI outcomes through personalized risk assessment.
Abstract: Severe aortic stenosis is a common and life-threatening condition in elderly patients, often treated with Transcatheter Aortic Valve Implantation (TAVI). Despite procedural advances, paravalvular aortic regurgitation (PVR) remains one of the most frequent post-TAVI complications, with a proven impact on long-term prognosis. In this work, we investigate the potential of deep learning to predict the occurrence of PVR from preoperative cardiac CT. To this end, a dataset of preoperative TAVI patients was collected, and 3D convolutional neural networks were trained on isotropic CT volumes. The results achieved suggest that volumetric deep learning can capture subtle anatomical features from pre-TAVI imaging, opening new perspectives for personalized risk assessment and procedural optimization. Source code is available at https://github.com/EIDOSLAB/tavi.
[212] Synthetic Dataset Generation and Validation for Robotic Surgery Instrument Segmentation
Giorgio Chiesa, Rossella Borra, Vittorio Lauro, Sabrina De Cillis, Daniele Amparore, Cristian Fiori, Riccardo Renzulli, Marco Grangetto
Main category: cs.CV
TL;DR: Synthetic dataset generation pipeline for robotic surgery instrument segmentation using 3D-reconstructed Da Vinci arms with photorealistic rendering, motion patterns, and synthetic blood textures to improve segmentation model generalization.
Details
Motivation: Need for large-scale, accurately labeled datasets for surgical instrument segmentation in robotic surgery, where real data collection is expensive and privacy-sensitive, requiring synthetic alternatives that maintain realism.Method: 3D reconstruction of Da Vinci robotic arms animated in Autodesk Maya via automated Python pipeline producing photorealistic labeled videos with randomized motion, lighting variations, and synthetic blood textures while preserving pixel-accurate ground truth masks.
Result: Balanced composition of real and synthetic data significantly improves segmentation model generalization compared to real-only training, while excessive synthetic data causes domain shift; framework provides reproducible tool for surgical computer vision.
Conclusion: Synthetic data generation pipeline effectively augments real surgical data for instrument segmentation, offering scalable solution for data augmentation, domain adaptation, and simulation-based pretraining in robotic-assisted surgery.
Abstract: This paper presents a comprehensive workflow for generating and validating a synthetic dataset designed for robotic surgery instrument segmentation. A 3D reconstruction of the Da Vinci robotic arms was refined and animated in Autodesk Maya through a fully automated Python-based pipeline capable of producing photorealistic, labeled video sequences. Each scene integrates randomized motion patterns, lighting variations, and synthetic blood textures to mimic intraoperative variability while preserving pixel-accurate ground truth masks. To validate the realism and effectiveness of the generated data, several segmentation models were trained under controlled ratios of real and synthetic data. Results demonstrate that a balanced composition of real and synthetic samples significantly improves model generalization compared to training on real data only, while excessive reliance on synthetic data introduces a measurable domain shift. The proposed framework provides a reproducible and scalable tool for surgical computer vision, supporting future research in data augmentation, domain adaptation, and simulation-based pretraining for robotic-assisted surgery. Data and code are available at https://github.com/EIDOSLAB/Sintetic-dataset-DaVinci.
[213] Cardiac Output Prediction from Echocardiograms: Self-Supervised Learning with Limited Data
Adson Duarte, Davide Vitturini, Emanuele Milillo, Andrea Bragagnolo, Carlo Alberto Barbano, Riccardo Renzulli, Michele Cannito, Federico Giacobbe, Francesco Bruno, Ovidio de Filippo, Fabrizio D’Ascenzo, Marco Grangetto
Main category: cs.CV
TL;DR: Self-supervised learning pretraining using SimCLR improves cardiac output prediction from echocardiographic videos, outperforming larger supervised models despite limited data.
Details
Motivation: Accurate cardiac output measurement requires invasive right-heart catheterization, motivating non-invasive alternatives using echocardiography. However, limited labeled data for training models presents a challenge.Method: Proposes a self-supervised learning pretraining strategy based on SimCLR (contrastive learning) using the same limited dataset available for downstream cardiac output prediction from apical four-chamber echocardiographic videos.
Result: SSL pretraining mitigates overfitting and improves representation learning, achieving an average Pearson correlation of 0.41 on the test set, outperforming PanEcho (a model trained on over one million echocardiographic exams).
Conclusion: Self-supervised learning demonstrates potential for improving medical imaging analysis even under data scarcity, showing that SSL can outperform larger supervised models on specific tasks.
Abstract: Cardiac Output (CO) is a key parameter in the diagnosis and management of cardiovascular diseases. However, its accurate measurement requires right-heart catheterization, an invasive and time-consuming procedure, motivating the development of reliable non-invasive alternatives using echocardiography. In this work, we propose a self-supervised learning (SSL) pretraining strategy based on SimCLR to improve CO prediction from apical four-chamber echocardiographic videos. The pretraining is performed using the same limited dataset available for the downstream task, demonstrating the potential of SSL even under data scarcity. Our results show that SSL mitigates overfitting and improves representation learning, achieving an average Pearson correlation of 0.41 on the test set and outperforming PanEcho, a model trained on over one million echocardiographic exams. Source code is available at https://github.com/EIDOSLAB/cardiac-output.
[214] Low-Pass Filtering Improves Behavioral Alignment of Vision Models
Max Wolff, Thomas Klein, Evgenia Rusak, Felix Wichmann, Wieland Brendel
Main category: cs.CV
TL;DR: Removing high-frequency spatial information from discriminative models (like CLIP) through blurring or low-pass filtering drastically improves their alignment with human visual behavior, achieving state-of-the-art performance on model-vs-human benchmarks.
Details
Motivation: Deep Neural Networks (DNNs) still fall short of adequately modeling human visual behavior despite impressive benchmark performance. Recent work suggested generative classifiers improve behavioral alignment, but this paper investigates whether simpler frequency-based explanations exist.Method: Conducted controlled experiments removing high-frequency spatial information from discriminative models like CLIP using blurring operations. Tested both training on blurred images and test-time blurring. Directly optimized filters for alignment and computed Pareto-optimal solutions frontier for the benchmark.
Result: Test-time blurring achieves new state-of-the-art on model-vs-human benchmark, halving the alignment gap between DNNs and human observers. Low-pass filters are likely optimal, and optimal Gaussian filters match the frequency spectrum of human visual system’s band-pass filters.
Conclusion: The increased alignment of generative models can be largely explained by low-pass filtering effects rather than generative modeling itself. Human visual behavior alignment in DNNs can be dramatically improved by simple frequency filtering that matches human visual system characteristics.
Abstract: Despite their impressive performance on computer vision benchmarks, Deep Neural Networks (DNNs) still fall short of adequately modeling human visual behavior, as measured by error consistency and shape bias. Recent work hypothesized that behavioral alignment can be drastically improved through \emph{generative} – rather than \emph{discriminative} – classifiers, with far-reaching implications for models of human vision. Here, we instead show that the increased alignment of generative models can be largely explained by a seemingly innocuous resizing operation in the generative model which effectively acts as a low-pass filter. In a series of controlled experiments, we show that removing high-frequency spatial information from discriminative models like CLIP drastically increases their behavioral alignment. Simply blurring images at test-time – rather than training on blurred images – achieves a new state-of-the-art score on the model-vs-human benchmark, halving the current alignment gap between DNNs and human observers. Furthermore, low-pass filters are likely optimal, which we demonstrate by directly optimizing filters for alignment. To contextualize the performance of optimal filters, we compute the frontier of all possible pareto-optimal solutions to the benchmark, which was formerly unknown. We explain our findings by observing that the frequency spectrum of optimal Gaussian filters roughly matches the spectrum of band-pass filters implemented by the human visual system. We show that the contrast sensitivity function, describing the inverse of the contrast threshold required for humans to detect a sinusoidal grating as a function of spatiotemporal frequency, is approximated well by Gaussian filters of the specific width that also maximizes error consistency.
[215] Human-Aligned Evaluation of a Pixel-wise DNN Color Constancy Model
Hamed Heidari-Gorji, Raquel Gil Rodriguez, Karl R. Gegenfurtner
Main category: cs.CV
TL;DR: A study comparing human color constancy performance with a DNN model in VR environments, showing strong correspondence between model predictions and human behavior across different visual cue conditions.
Details
Motivation: To compare and study color constancy mechanisms between humans and deep neural networks in virtual reality environments, specifically examining how both perform when different visual cues (local surround, maximum flux, spatial mean) are manipulated.Method: Used a ResNet-based U-Net pre-trained on rendered images to predict surface reflectance, then fine-tuned only the decoder on VR baseline condition images. The model was evaluated using the same achromatic object selection task as human experiments across various cue manipulation conditions.
Result: The model showed strong correspondence with human behavior - both achieved high color constancy under baseline conditions and exhibited similar, condition-dependent performance declines when local surround or spatial mean color cues were removed.
Conclusion: Deep neural networks can effectively model human color constancy mechanisms in VR environments, demonstrating similar sensitivity to visual cues and suggesting potential for using DNNs to study human visual perception.
Abstract: We previously investigated color constancy in photorealistic virtual reality (VR) and developed a Deep Neural Network (DNN) that predicts reflectance from rendered images. Here, we combine both approaches to compare and study a model and human performance with respect to established color constancy mechanisms: local surround, maximum flux and spatial mean. Rather than evaluating the model against physical ground truth, model performance was assessed using the same achromatic object selection task employed in the human experiments. The model, a ResNet based U-Net from our previous work, was pre-trained on rendered images to predict surface reflectance. We then applied transfer learning, fine-tuning only the network’s decoder on images from the baseline VR condition. To parallel the human experiment, the model’s output was used to perform the same achromatic object selection task across all conditions. Results show a strong correspondence between the model and human behavior. Both achieved high constancy under baseline conditions and showed similar, condition-dependent performance declines when the local surround or spatial mean color cues were removed.
[216] Parameter-Efficient Fine-Tuning of DINOv2 for Large-Scale Font Classification
Daniel Chen, Zaria Zinn, Marcus Lowe
Main category: cs.CV
TL;DR: Fine-tuned DINOv2 Vision Transformer with LoRA achieves 86% accuracy on 394 font family classification using synthetic Google Fonts data with diverse augmentations.
Details
Motivation: To create an accurate font classification system that can identify a large number of font families from rendered text images, addressing the challenge of limited real-world training data through synthetic generation.Method: Fine-tunes DINOv2 Vision Transformer using Low-Rank Adaptation (LoRA) with only 1% of parameters trained. Uses synthetic dataset generation pipeline rendering Google Fonts with diverse augmentations (random colors, alignment, line wrapping, Gaussian noise). Includes built-in preprocessing for consistency.
Result: Achieves approximately 86% top-1 accuracy on 394 font families while training fewer than 1% of the model’s 87.2M parameters. Model, dataset, and training pipeline released as open-source.
Conclusion: LoRA fine-tuning on synthetic data enables efficient and accurate font classification, demonstrating the effectiveness of parameter-efficient adaptation for vision tasks with limited real data.
Abstract: We present a font classification system capable of identifying 394 font families from rendered text images. Our approach fine-tunes a DINOv2 Vision Transformer using Low-Rank Adaptation (LoRA), achieving approximately 86% top-1 accuracy while training fewer than 1% of the model’s 87.2M parameters. We introduce a synthetic dataset generation pipeline that renders Google Fonts at scale with diverse augmentations including randomized colors, alignment, line wrapping, and Gaussian noise, producing training images that generalize to real-world typographic samples. The model incorporates built-in preprocessing to ensure consistency between training and inference, and is deployed as a HuggingFace Inference Endpoint. We release the model, dataset, and full training pipeline as open-source resources.
[217] RPGD: RANSAC-P3P Gradient Descent for Extrinsic Calibration in 3D Human Pose Estimation
Zhanyu Tuo
Main category: cs.CV
TL;DR: RPGD: A robust human-pose-driven extrinsic calibration framework that aligns MoCap 3D skeletal data with RGB cameras using natural human motion through RANSAC-P3P and gradient descent refinement.
Details
Motivation: The paper addresses the need for reliable extrinsic calibration in 3D human pose estimation datasets, particularly for aligning motion capture (MoCap) data with RGB camera views. Current methods often require manual intervention or specialized calibration patterns, which are impractical for large-scale, in-the-wild data collection.Method: RPGD formulates extrinsic calibration as a coarse-to-fine problem using human poses. It combines RANSAC-P3P for global robustness in initial alignment with gradient-descent-based refinement for precise optimization. The approach works with both monocular and multi-view RGB cameras using only natural human motion without requiring calibration patterns.
Result: The method achieves sub-pixel MPJPE (Mean Per Joint Position Error) reprojection error on three large-scale public 3D HPE datasets and a self-collected in-the-wild dataset. It recovers extrinsic parameters with accuracy comparable to ground truth, even in challenging, noisy settings.
Conclusion: RPGD provides a practical and automatic solution for reliable extrinsic calibration in large-scale 3D human pose estimation dataset collection, enabling more efficient and accurate alignment of MoCap data with RGB camera systems using only natural human motion.
Abstract: In this paper, we propose RPGD (RANSAC-P3P Gradient Descent), a human-pose-driven extrinsic calibration framework that robustly aligns MoCap-based 3D skeletal data with monocular or multi-view RGB cameras using only natural human motion. RPGD formulates extrinsic calibration as a coarse-to-fine problem tailored to human poses, combining the global robustness of RANSAC-P3P with Gradient-Descent-based refinement. We evaluate RPGD on three large-scale public 3D HPE datasets as well as on a self-collected in-the-wild dataset. Experimental results demonstrate that RPGD consistently recovers extrinsic parameters with accuracy comparable to the provided ground truth, achieving sub-pixel MPJPE reprojection error even in challenging, noisy settings. These results indicate that RPGD provides a practical and automatic solution for reliable extrinsic calibration of large-scale 3D HPE dataset collection.
[218] MamaDino: A Hybrid Vision Model for Breast Cancer 3-Year Risk Prediction
Ruggiero Santeramo, Igor Zubarev, Florian Jug
Main category: cs.CV
TL;DR: MamaDino combines convolutional and transformer architectures with explicit contralateral asymmetry modeling to achieve state-of-the-art 3-year breast cancer risk prediction using substantially lower-resolution mammograms than previous methods.
Details
Motivation: Current deep learning systems for breast cancer risk prediction use high-resolution inputs and simple multi-view fusion with limited explicit modeling of contralateral asymmetry. The authors hypothesize that combining complementary inductive biases with explicit asymmetry modeling can achieve state-of-the-art performance with lower-resolution images.Method: MamaDino fuses frozen self-supervised DINOv3 ViT-S features with a trainable CNN encoder at 512x512 resolution, and aggregates bilateral breast information via a BilateralMixer to output 3-year breast cancer risk scores. The model is trained on 53,883 women from OPTIMAM (UK) and evaluated on internal and external test sets.
Result: MamaDino matches Mirai’s performance on both internal and external tests while using ~13x fewer input pixels. Adding the BilateralMixer improves discrimination to AUC 0.736 (vs 0.713) in-distribution and 0.677 (vs 0.666) out-of-distribution, with consistent performance across various demographic and clinical factors.
Conclusion: Explicit contralateral modeling and complementary inductive biases enable predictions that match state-of-the-art methods despite operating on substantially lower-resolution mammograms, demonstrating that using less detailed images in a more structured way can recover state-of-the-art accuracy.
Abstract: Breast cancer screening programmes increasingly seek to move from one-size-fits-all interval to risk-adapted and personalized strategies. Deep learning (DL) has enabled image-based risk models with stronger 1- to 5-year prediction than traditional clinical models, but leading systems (e.g., Mirai) typically use convolutional backbones, very high-resolution inputs (>1M pixels) and simple multi-view fusion, with limited explicit modelling of contralateral asymmetry. We hypothesised that combining complementary inductive biases (convolutional and transformer-based) with explicit contralateral asymmetry modelling would allow us to match state-of-the-art 3-year risk prediction performance even when operating on substantially lower-resolution mammograms, indicating that using less detailed images in a more structured way can recover state-of-the-art accuracy. We present MamaDino, a mammography-aware multi-view attentional DINO model. MamaDino fuses frozen self-supervised DINOv3 ViT-S features with a trainable CNN encoder at 512x512 resolution, and aggregates bilateral breast information via a BilateralMixer to output a 3-year breast cancer risk score. We train on 53,883 women from OPTIMAM (UK) and evaluate on matched 3-year case-control cohorts: an in-distribution test set from four screening sites and an external out-of-distribution cohort from an unseen site. At breast-level, MamaDino matches Mirai on both internal and external tests while using ~13x fewer input pixels. Adding the BilateralMixer improves discrimination to AUC 0.736 (vs 0.713) in-distribution and 0.677 (vs 0.666) out-of-distribution, with consistent performance across age, ethnicity, scanner, tumour type and grade. These findings demonstrate that explicit contralateral modelling and complementary inductive biases enable predictions that match Mirai, despite operating on substantially lower-resolution mammograms.
[219] Fusing Pixels and Genes: Spatially-Aware Learning in Computational Pathology
Minghao Han, Dingkang Yang, Linhao Qu, Zizhi Chen, Gang Li, Han Wang, Jiacong Wang, Lihua Zhang
Main category: cs.CV
TL;DR: STAMP integrates spatial transcriptomics with pathology images for molecule-guided multimodal representation learning in computational pathology, achieving strong performance across multiple tasks.
Details
Motivation: Existing multimodal pathology models rely on vision and language, but language lacks molecular specificity and provides limited pathological supervision, creating representational bottlenecks. Spatial transcriptomics offers molecular-level information that can enhance pathology image understanding.Method: Proposes STAMP framework that integrates spatially-resolved gene expression profiles with pathology images using self-supervised, gene-guided training. Uses hierarchical multi-scale contrastive alignment and cross-scale patch localization to align spatial transcriptomics with pathology images. Built SpaVis-6M, the largest Visium-based spatial transcriptomics dataset, and trained a spatially-aware gene encoder.
Result: Validated across six datasets and four downstream tasks, consistently achieving strong performance. Demonstrates that gene-guided training provides robust, task-agnostic signals for learning pathology image representations.
Conclusion: Integrating spatially resolved molecular supervision is valuable and necessary for advancing multimodal learning in computational pathology, overcoming limitations of language-only supervision.
Abstract: Recent years have witnessed remarkable progress in multimodal learning within computational pathology. Existing models primarily rely on vision and language modalities; however, language alone lacks molecular specificity and offers limited pathological supervision, leading to representational bottlenecks. In this paper, we propose STAMP, a Spatial Transcriptomics-Augmented Multimodal Pathology representation learning framework that integrates spatially-resolved gene expression profiles to enable molecule-guided joint embedding of pathology images and transcriptomic data. Our study shows that self-supervised, gene-guided training provides a robust and task-agnostic signal for learning pathology image representations. Incorporating spatial context and multi-scale information further enhances model performance and generalizability. To support this, we constructed SpaVis-6M, the largest Visium-based spatial transcriptomics dataset to date, and trained a spatially-aware gene encoder on this resource. Leveraging hierarchical multi-scale contrastive alignment and cross-scale patch localization mechanisms, STAMP effectively aligns spatial transcriptomics with pathology images, capturing spatial structure and molecular variation. We validate STAMP across six datasets and four downstream tasks, where it consistently achieves strong performance. These results highlight the value and necessity of integrating spatially resolved molecular supervision for advancing multimodal learning in computational pathology. The code is included in the supplementary materials. The pretrained weights and SpaVis-6M are available at: https://github.com/Hanminghao/STAMP.
[220] MarsRetrieval: Benchmarking Vision-Language Models for Planetary-Scale Geospatial Retrieval on Mars
Shuoyuan Wang, Yiran Wang, Hongxin Wei
Main category: cs.CV
TL;DR: MarsRetrieval is a vision-language benchmark for Martian geospatial discovery with three retrieval tasks: image-text retrieval, landform retrieval, and global geo-localization, showing that even strong foundation models struggle with domain-specific geomorphic distinctions.
Details
Motivation: Most existing planetary science benchmarks focus on closed-set supervised visual tasks and lack text-guided retrieval capabilities needed for geospatial discovery on Mars. There's a need for comprehensive evaluation of vision-language models in planetary exploration contexts.Method: Introduces MarsRetrieval benchmark with three tasks covering multiple spatial scales and geomorphic origins. Proposes unified retrieval-centric protocol to evaluate multimodal embedding architectures including contrastive dual-tower encoders and generative vision-language models.
Result: MarsRetrieval is challenging - even strong foundation models often fail to capture domain-specific geomorphic distinctions. Domain-specific fine-tuning is critical for generalizable geospatial discovery in planetary settings.
Conclusion: MarsRetrieval provides a comprehensive benchmark for evaluating vision-language models in planetary science, highlighting the importance of domain adaptation for effective geospatial discovery on Mars.
Abstract: Data-driven approaches like deep learning are rapidly advancing planetary science, particularly in Mars exploration. Despite recent progress, most existing benchmarks remain confined to closed-set supervised visual tasks and do not support text-guided retrieval for geospatial discovery. We introduce MarsRetrieval, a retrieval benchmark for evaluating vision-language models for Martian geospatial discovery. MarsRetrieval includes three tasks: (1) paired image-text retrieval, (2) landform retrieval, and (3) global geo-localization, covering multiple spatial scales and diverse geomorphic origins. We propose a unified retrieval-centric protocol to benchmark multimodal embedding architectures, including contrastive dual-tower encoders and generative vision-language models. Our evaluation shows MarsRetrieval is challenging: even strong foundation models often fail to capture domain-specific geomorphic distinctions. We further show that domain-specific fine-tuning is critical for generalizable geospatial discovery in planetary settings. Our code is available at https://github.com/ml-stat-Sustech/MarsRetrieval
[221] Elastic Diffusion Transformer
Jiangshan Wang, Zeqiang Lai, Jiarui Chen, Jiayi Guo, Hang Guo, Xiu Li, Xiangyu Yue, Chunchao Guo
Main category: cs.CV
TL;DR: E-DiT: An adaptive acceleration framework for Diffusion Transformers that dynamically skips blocks and reduces MLP widths based on sample-dependent sparsity, achieving 2× speedup with minimal quality loss.
Details
Motivation: Diffusion Transformers (DiT) have excellent generative capabilities but are computationally expensive. Existing acceleration methods use fixed computational capacity, leading to insufficient speedup and degraded quality. The authors observed that DiT generative processes exhibit substantial and variable sparsity across samples.Method: E-DiT adds lightweight routers to each DiT block that dynamically identify sample-dependent sparsity from input latents. Each router decides whether to skip its block and, if not skipped, predicts optimal MLP width reduction ratio. A block-level feature caching mechanism eliminates redundant computations during inference.
Result: Extensive experiments on 2D image (Qwen-Image, FLUX) and 3D asset (Hunyuan3D-3.0) generation show E-DiT achieves up to ~2× speedup with negligible loss in generation quality.
Conclusion: E-DiT provides an effective adaptive acceleration framework for DiT that maintains generation quality while significantly improving efficiency through dynamic sparsity-aware computation.
Abstract: Diffusion Transformers (DiT) have demonstrated remarkable generative capabilities but remain highly computationally expensive. Previous acceleration methods, such as pruning and distillation, typically rely on a fixed computational capacity, leading to insufficient acceleration and degraded generation quality. To address this limitation, we propose \textbf{Elastic Diffusion Transformer (E-DiT)}, an adaptive acceleration framework for DiT that effectively improves efficiency while maintaining generation quality. Specifically, we observe that the generative process of DiT exhibits substantial sparsity (i.e., some computations can be skipped with minimal impact on quality), and this sparsity varies significantly across samples. Motivated by this observation, E-DiT equips each DiT block with a lightweight router that dynamically identifies sample-dependent sparsity from the input latent. Each router adaptively determines whether the corresponding block can be skipped. If the block is not skipped, the router then predicts the optimal MLP width reduction ratio within the block. During inference, we further introduce a block-level feature caching mechanism that leverages router predictions to eliminate redundant computations in a training-free manner. Extensive experiments across 2D image (Qwen-Image and FLUX) and 3D asset (Hunyuan3D-3.0) demonstrate the effectiveness of E-DiT, achieving up to $\sim$2$\times$ speedup with negligible loss in generation quality. Code will be available at https://github.com/wangjiangshan0725/Elastic-DiT.
[222] Inject Where It Matters: Training-Free Spatially-Adaptive Identity Preservation for Text-to-Image Personalization
Guandong Li, Mengxia Ye
Main category: cs.CV
TL;DR: SpatialID: A training-free spatially-adaptive identity modulation framework for personalized text-to-image generation that prevents identity features from contaminating non-facial regions while maintaining text adherence.
Details
Motivation: Existing tuning-free methods for personalized text-to-image generation use spatially uniform visual injection, which causes identity features to contaminate non-facial regions (backgrounds, lighting) and degrades text adherence. There's a need for a solution that doesn't require expensive fine-tuning.Method: Proposes SpatialID, a training-free spatially-adaptive identity modulation framework that decouples identity injection into face-relevant and context-free regions using a Spatial Mask Extractor derived from cross-attention responses. Also introduces Temporal-Spatial Scheduling that dynamically adjusts spatial constraints (transitioning from Gaussian priors to attention-based masks with adaptive relaxation) to align with diffusion generation dynamics.
Result: Extensive experiments on IBench show SpatialID achieves state-of-the-art performance in text adherence (CLIP-T: 0.281), visual consistency (CLIP-I: 0.827), and image quality (IQ: 0.523), significantly eliminating background contamination while maintaining robust identity preservation.
Conclusion: SpatialID provides an effective training-free solution for personalized text-to-image generation that spatially adapts identity injection to prevent contamination of non-facial regions while preserving identity and text adherence.
Abstract: Personalized text-to-image generation aims to integrate specific identities into arbitrary contexts. However, existing tuning-free methods typically employ Spatially Uniform Visual Injection, causing identity features to contaminate non-facial regions (e.g., backgrounds and lighting) and degrading text adherence. To address this without expensive fine-tuning, we propose SpatialID, a training-free spatially-adaptive identity modulation framework. SpatialID fundamentally decouples identity injection into face-relevant and context-free regions using a Spatial Mask Extractor derived from cross-attention responses. Furthermore, we introduce a Temporal-Spatial Scheduling strategy that dynamically adjusts spatial constraints - transitioning from Gaussian priors to attention-based masks and adaptive relaxation - to align with the diffusion generation dynamics. Extensive experiments on IBench demonstrate that SpatialID achieves state-of-the-art performance in text adherence (CLIP-T: 0.281), visual consistency (CLIP-I: 0.827), and image quality (IQ: 0.523), significantly eliminating background contamination while maintaining robust identity preservation.
[223] A Deployment-Friendly Foundational Framework for Efficient Computational Pathology
Yu Cai, Cheng Jin, Jiabo Ma, Fengtao Zhou, Yingxue Xu, Zhengrui Guo, Yihui Wang, Zhengyu Zhang, Ling Liang, Yonghao Tan, Pingcheng Dong, Du Cai, On Ki Tang, Chenglong Zhao, Xi Wang, Can Yang, Yali Xu, Jing Cui, Zhenhui Li, Ronald Cheong Kin Chan, Yueping Liu, Feng Gao, Xiuming Zhang, Li Liang, Hao Chen, Kwang-Ting Cheng
Main category: cs.CV
TL;DR: LitePath is a deployment-friendly pathology foundation model framework that reduces computational costs by 28x parameters and 403.5x FLOPs while maintaining 99.71% of Virchow2’s AUC, enabling efficient deployment on edge hardware.
Details
Motivation: Pathology foundation models (PFMs) have strong generalization capabilities but suffer from high computational costs that limit clinical accessibility and scalability, especially for gigapixel whole slide images.Method: LitePath integrates LiteFM (a compact model distilled from three large PFMs using 190M patches) and Adaptive Patch Selector (APS) for task-specific patch selection, reducing model over-parameterization and patch redundancy.
Result: Achieves 28x parameter reduction, 403.5x FLOP reduction, processes 208 slides/hour on NVIDIA Jetson Orin Nano (104.5x faster than Virchow2), consumes 0.36 kWh/3,000 slides (171x lower), and ranks 2nd among 19 models while maintaining 99.71% of Virchow2’s AUC.
Conclusion: LitePath enables rapid, cost-effective, energy-efficient pathology image analysis on accessible hardware while maintaining accuracy comparable to state-of-the-art PFMs and reducing AI deployment carbon footprint.
Abstract: Pathology foundation models (PFMs) have enabled robust generalization in computational pathology through large-scale datasets and expansive architectures, but their substantial computational cost, particularly for gigapixel whole slide images, limits clinical accessibility and scalability. Here, we present LitePath, a deployment-friendly foundational framework designed to mitigate model over-parameterization and patch level redundancy. LitePath integrates LiteFM, a compact model distilled from three large PFMs (Virchow2, H-Optimus-1 and UNI2) using 190 million patches, and the Adaptive Patch Selector (APS), a lightweight component for task-specific patch selection. The framework reduces model parameters by 28x and lowers FLOPs by 403.5x relative to Virchow2, enabling deployment on low-power edge hardware such as the NVIDIA Jetson Orin Nano Super. On this device, LitePath processes 208 slides per hour, 104.5x faster than Virchow2, and consumes 0.36 kWh per 3,000 slides, 171x lower than Virchow2 on an RTX3090 GPU. We validated accuracy using 37 cohorts across four organs and 26 tasks (26 internal, 9 external, and 2 prospective), comprising 15,672 slides from 9,808 patients disjoint from the pretraining data. LitePath ranks second among 19 evaluated models and outperforms larger models including H-Optimus-1, mSTAR, UNI2 and GPFM, while retaining 99.71% of the AUC of Virchow2 on average. To quantify the balance between accuracy and efficiency, we propose the Deployability Score (D-Score), defined as the weighted geometric mean of normalized AUC and normalized FLOP, where LitePath achieves the highest value, surpassing Virchow2 by 10.64%. These results demonstrate that LitePath enables rapid, cost-effective and energy-efficient pathology image analysis on accessible hardware while maintaining accuracy comparable to state-of-the-art PFMs and reducing the carbon footprint of AI deployment.
[224] Flow4R: Unifying 4D Reconstruction and Tracking with Scene Flow
Shenhan Qian, Ganlin Zhang, Shangzhe Wu, Daniel Cremers
Main category: cs.CV
TL;DR: Flow4R: A unified framework using camera-space scene flow as central representation for 3D dynamic scene reconstruction and tracking from two-view inputs.
Details
Motivation: Existing approaches decouple geometry from motion - static reconstruction methods assume static scenes, while dynamic tracking relies on explicit camera pose estimation or separate motion models. There's a need for a unified approach that links 3D structure, object motion, and camera motion.Method: Flow4R treats camera-space scene flow as central representation, predicting per-pixel properties (3D point position, scene flow, pose weight, confidence) from two-view inputs using Vision Transformer. Uses flow-centric formulation allowing local geometry and bidirectional motion inference with shared decoder in single forward pass, without explicit pose regressors or bundle adjustment.
Result: Achieves state-of-the-art performance on 4D reconstruction and tracking tasks. Demonstrates effectiveness of flow-central representation for spatiotemporal scene understanding.
Conclusion: Flow4R provides a unified framework for dynamic 3D scene understanding through flow-centric representation, successfully linking geometry and motion without explicit pose estimation.
Abstract: Reconstructing and tracking dynamic 3D scenes remains a fundamental challenge in computer vision. Existing approaches often decouple geometry from motion: multi-view reconstruction methods assume static scenes, while dynamic tracking frameworks rely on explicit camera pose estimation or separate motion models. We propose Flow4R, a unified framework that treats camera-space scene flow as the central representation linking 3D structure, object motion, and camera motion. Flow4R predicts a minimal per-pixel property set-3D point position, scene flow, pose weight, and confidence-from two-view inputs using a Vision Transformer. This flow-centric formulation allows local geometry and bidirectional motion to be inferred symmetrically with a shared decoder in a single forward pass, without requiring explicit pose regressors or bundle adjustment. Trained jointly on static and dynamic datasets, Flow4R achieves state-of-the-art performance on 4D reconstruction and tracking tasks, demonstrating the effectiveness of the flow-central representation for spatiotemporal scene understanding.
[225] Train Short, Inference Long: Training-free Horizon Extension for Autoregressive Video Generation
Jia Li, Xiaomeng Fu, Xurui Peng, Weifeng Chen, Youwei Zheng, Tianyu Zhao, Jiexi Wang, Fangmin Chen, Xing Wang, Hayden Kwok-Hay So
Main category: cs.CV
TL;DR: FLEX is a training-free inference framework for autoregressive video diffusion models that addresses extrapolation failure in long video generation through frequency-aware RoPE modulation and antiphase noise sampling.
Details
Motivation: Autoregressive video diffusion models suffer from severe extrapolation failure when generating videos beyond training horizons due to spectral bias in 3D positional embeddings and lack of dynamic priors in noise sampling, leading to temporal degradation.Method: Proposes FLEX with three components: 1) Frequency-aware RoPE Modulation to adaptively interpolate under-trained low-frequency components while extrapolating high-frequency ones, 2) Antiphase Noise Sampling (ANS) to inject high-frequency dynamic priors, and 3) Inference-only Attention Sink to anchor global structure.
Result: FLEX significantly outperforms SOTA models at 6× extrapolation (30s duration) and matches long-video fine-tuned baselines at 12× scale (60s duration). It enables consistent video synthesis at 4-minute scale with existing models like LongLive.
Conclusion: FLEX is an effective plug-and-play inference-time framework that bridges the gap between short-term training and long-term inference for video diffusion models, enabling scalable long video generation without retraining.
Abstract: Autoregressive video diffusion models have emerged as a scalable paradigm for long video generation. However, they often suffer from severe extrapolation failure, where rapid error accumulation leads to significant temporal degradation when extending beyond training horizons. We identify that this failure primarily stems from the \textit{spectral bias} of 3D positional embeddings and the lack of \textit{dynamic priors} in noise sampling. To address these issues, we propose \textbf{FLEX} (\textbf{F}requency-aware \textbf{L}ength \textbf{EX}tension), a training-free inference-time framework that bridges the gap between short-term training and long-term inference. FLEX introduces Frequency-aware RoPE Modulation to adaptively interpolate under-trained low-frequency components while extrapolating high-frequency ones to preserve multi-scale temporal discriminability. This is integrated with Antiphase Noise Sampling (ANS) to inject high-frequency dynamic priors and Inference-only Attention Sink to anchor global structure. Extensive evaluations on VBench demonstrate that FLEX significantly outperforms state-of-the-art models at $6\times$ extrapolation (30s duration) and matches the performance of long-video fine-tuned baselines at $12\times$ scale (60s duration). As a plug-and-play augmentation, FLEX seamlessly integrates into existing inference pipelines for horizon extension. It effectively pushes the generation limits of models such as LongLive, supporting consistent and dynamic video synthesis at a 4-minute scale. Project page is available at \href{https://ga-lee.github.io/FLEX_demo}{https://ga-lee.github.io/FLEX}.
[226] Explainability-Inspired Layer-Wise Pruning of Deep Neural Networks for Efficient Object Detection
Abhinav Shukla, Nachiket Tapas
Main category: cs.CV
TL;DR: Explainability-inspired layer-wise pruning framework for object detection that uses gradient-activation attribution to assess layer importance, achieving better accuracy-efficiency trade-offs than magnitude-based pruning.
Details
Motivation: Traditional magnitude-based pruning methods for deep neural networks don't align with true functional contribution to task performance, especially for object detection on resource-constrained platforms.Method: Proposes a SHAP-inspired gradient-activation attribution method to estimate layer importance as a data-driven proxy for functional contribution, enabling layer-wise pruning tailored for object detection architectures.
Result: The method consistently identifies different least important layers than L1-norm pruning, achieving better accuracy-efficiency trade-offs: 10% speed increase for ShuffleNetV2 vs 13.7% degradation with L1, and preserved mAP for RetinaNet with negligible speed impact.
Conclusion: Explainability-inspired compression offers a principled direction for deploying DNNs on edge platforms while preserving both performance and interpretability, highlighting the importance of data-driven layer importance assessment.
Abstract: Deep neural networks (DNNs) have achieved remarkable success in object detection tasks, but their increasing complexity poses significant challenges for deployment on resource-constrained platforms. While model compression techniques such as pruning have emerged as essential tools, traditional magnitude-based pruning methods do not necessarily align with the true functional contribution of network components to task-specific performance. In this work, we present an explainability-inspired, layer-wise pruning framework tailored for efficient object detection. Our approach leverages a SHAP-inspired gradient–activation attribution to estimate layer importance, providing a data-driven proxy for functional contribution rather than relying solely on static weight magnitudes. We conduct comprehensive experiments across diverse object detection architectures, including ResNet-50, MobileNetV2, ShuffleNetV2, Faster R-CNN, RetinaNet, and YOLOv8, evaluating performance on the Microsoft COCO 2017 validation set. The results show that the proposed attribution-inspired pruning consistently identifies different layers as least important compared to L1-norm-based methods, leading to improved accuracy–efficiency trade-offs. Notably, for ShuffleNetV2, our method yields a 10% empirical increase in inference speed, whereas L1-pruning degrades performance by 13.7%. For RetinaNet, the proposed approach preserves the baseline mAP (0.151) with negligible impact on inference speed, while L1-pruning incurs a 1.3% mAP drop for a 6.2% speed increase. These findings highlight the importance of data-driven layer importance assessment and demonstrate that explainability-inspired compression offers a principled direction for deploying deep neural networks on edge and resource-constrained platforms while preserving both performance and interpretability.
[227] BitDance: Scaling Autoregressive Generative Models with Binary Tokens
Yuang Ai, Jiaming Han, Shaobin Zhuang, Weijia Mao, Xuefeng Hu, Ziyan Yang, Zhenheng Yang, Huaibo Huang, Xiangyu Yue, Hao Chen
Main category: cs.CV
TL;DR: BitDance is an autoregressive image generator that predicts binary visual tokens instead of codebook indices, using binary diffusion heads and next-patch diffusion for efficient high-quality image generation.
Details
Motivation: To create a scalable autoregressive image generator that overcomes limitations of standard classification-based token prediction by using high-entropy binary latents for more expressive discrete representations.Method: Uses binary visual tokens (up to 2^256 states per token) with binary diffusion heads instead of softmax classification. Introduces next-patch diffusion for parallel token prediction, enabling faster inference while maintaining accuracy.
Result: Achieves FID of 1.24 on ImageNet 256x256 (best among AR models). With 260M parameters, beats 1.4B parameter parallel AR models with 8.7x speedup. For text-to-image, generates high-resolution photorealistic images with over 30x speedup for 1024x1024 images.
Conclusion: BitDance demonstrates that binary tokens with diffusion heads enable highly efficient and scalable autoregressive image generation with state-of-the-art quality and significant speed improvements.
Abstract: We present BitDance, a scalable autoregressive (AR) image generator that predicts binary visual tokens instead of codebook indices. With high-entropy binary latents, BitDance lets each token represent up to $2^{256}$ states, yielding a compact yet highly expressive discrete representation. Sampling from such a huge token space is difficult with standard classification. To resolve this, BitDance uses a binary diffusion head: instead of predicting an index with softmax, it employs continuous-space diffusion to generate the binary tokens. Furthermore, we propose next-patch diffusion, a new decoding method that predicts multiple tokens in parallel with high accuracy, greatly speeding up inference. On ImageNet 256x256, BitDance achieves an FID of 1.24, the best among AR models. With next-patch diffusion, BitDance beats state-of-the-art parallel AR models that use 1.4B parameters, while using 5.4x fewer parameters (260M) and achieving 8.7x speedup. For text-to-image generation, BitDance trains on large-scale multimodal tokens and generates high-resolution, photorealistic images efficiently, showing strong performance and favorable scaling. When generating 1024x1024 images, BitDance achieves a speedup of over 30x compared to prior AR models. We release code and models to facilitate further research on AR foundation models. Code and models are available at: https://github.com/shallowdream204/BitDance.
[228] Restoration Adaptation for Semantic Segmentation on Low Quality Images
Kai Guan, Rongyuan Wu, Shuai Li, Wentao Zhu, Wenjun Zeng, Lei Zhang
Main category: cs.CV
TL;DR: RASS integrates semantic image restoration with segmentation to improve performance on low-quality images by injecting segmentation priors into restoration and transferring knowledge to segmentation models.
Details
Motivation: Semantic segmentation performance degrades on low-quality images lacking clear structures and details. Existing restoration models focus on pixel fidelity but fail to recover semantic cues, while segmentation models lack robustness to real-world degradations.Method: Proposes Restoration Adaptation for Semantic Segmentation (RASS) with two components: 1) Semantic-Constrained Restoration (SCR) that aligns restoration cross-attention maps with segmentation masks to inject semantic priors, and 2) LoRA-based module merging and fine-tuning to transfer restoration knowledge to segmentation models.
Result: Significantly outperforms state-of-the-art methods on both synthetic and real-world low-quality benchmarks. Constructed a real-world LQ image segmentation dataset with high-quality annotations for validation.
Conclusion: RASS effectively integrates semantic restoration into segmentation, enabling robust high-quality segmentation on low-quality images by bridging the gap between restoration fidelity and semantic understanding.
Abstract: In real-world scenarios, the performance of semantic segmentation often deteriorates when processing low-quality (LQ) images, which may lack clear semantic structures and high-frequency details. Although image restoration techniques offer a promising direction for enhancing degraded visual content, conventional real-world image restoration (Real-IR) models primarily focus on pixel-level fidelity and often fail to recover task-relevant semantic cues, limiting their effectiveness when directly applied to downstream vision tasks. Conversely, existing segmentation models trained on high-quality data lack robustness under real-world degradations. In this paper, we propose Restoration Adaptation for Semantic Segmentation (RASS), which effectively integrates semantic image restoration into the segmentation process, enabling high-quality semantic segmentation on the LQ images directly. Specifically, we first propose a Semantic-Constrained Restoration (SCR) model, which injects segmentation priors into the restoration model by aligning its cross-attention maps with segmentation masks, encouraging semantically faithful image reconstruction. Then, RASS transfers semantic restoration knowledge into segmentation through LoRA-based module merging and task-specific fine-tuning, thereby enhancing the model’s robustness to LQ images. To validate the effectiveness of our framework, we construct a real-world LQ image segmentation dataset with high-quality annotations, and conduct extensive experiments on both synthetic and real-world LQ benchmarks. The results show that SCR and RASS significantly outperform state-of-the-art methods in segmentation and restoration tasks. Code, models, and datasets will be available at https://github.com/Ka1Guan/RASS.git.
[229] CoCoEdit: Content-Consistent Image Editing via Region Regularized Reinforcement Learning
Yuhui Wu, Chenxi Xie, Ruibin Li, Liyi Chen, Qiaosi Yi, Lei Zhang
Main category: cs.CV
TL;DR: CoCoEdit is a post-training framework for content-consistent image editing using region-regularized reinforcement learning to prevent unwanted changes in unintended regions while maintaining editing quality.
Details
Motivation: Current image editing models often produce unwanted changes in unintended regions when editing specific objects/areas. There's a need for methods that maintain content consistency in non-edited regions while achieving high-quality editing effects.Method: 1) Curated 40K high-quality training samples with refined instructions and masks; 2) Introduced pixel-level similarity reward to complement MLLM-based rewards; 3) Proposed region-based regularizer to preserve non-edited regions for high-reward samples while encouraging editing for low-reward samples; 4) Applied framework to Qwen-Image-Edit and FLUX-Kontext models.
Result: Achieved competitive editing scores with state-of-the-art models while significantly improving content consistency, measured by PSNR/SSIM metrics and human subjective ratings. Annotated editing masks for GEdit-Bench and ImgEdit-Bench with pixel-level similarity metrics.
Conclusion: CoCoEdit effectively addresses the content consistency problem in image editing through region-regularized reinforcement learning, achieving both high editing quality and preservation of non-edited regions.
Abstract: Image editing has achieved impressive results with the development of large-scale generative models. However, existing models mainly focus on the editing effects of intended objects and regions, often leading to unwanted changes in unintended regions. We present a post-training framework for Content-Consistent Editing (CoCoEdit) via region regularized reinforcement learning. We first augment existing editing datasets with refined instructions and masks, from which 40K diverse and high quality samples are curated as training set. We then introduce a pixel-level similarity reward to complement MLLM-based rewards, enabling models to ensure both editing quality and content consistency during the editing process. To overcome the spatial-agnostic nature of the rewards, we propose a region-based regularizer, aiming to preserve non-edited regions for high-reward samples while encouraging editing effects for low-reward samples. For evaluation, we annotate editing masks for GEdit-Bench and ImgEdit-Bench, introducing pixel-level similarity metrics to measure content consistency and editing quality. Applying CoCoEdit to Qwen-Image-Edit and FLUX-Kontext, we achieve not only competitive editing scores with state-of-the-art models, but also significantly better content consistency, measured by PSNR/SSIM metrics and human subjective ratings.
[230] ForgeryVCR: Visual-Centric Reasoning via Efficient Forensic Tools in MLLMs for Image Forgery Detection and Localization
Youqi Wang, Shen Chen, Haowei Wang, Rongxuan Peng, Taiping Yao, Shunquan Tan, Changsheng Chen, Bin Li, Shouhong Ding
Main category: cs.CV
TL;DR: ForgeryVCR is a multimodal LLM framework for image forgery detection that uses visual-centric reasoning with forensic tools instead of text-centric approaches to avoid hallucinations when detecting pixel-level tampering traces.
Details
Motivation: Current MLLMs for image forgery detection rely on text-centric Chain-of-Thought approaches, which struggle to capture imperceptible low-level tampering traces and lead to hallucinations since linguistic modalities can't adequately represent fine-grained pixel-level inconsistencies.Method: Proposes ForgeryVCR with: 1) Forensic toolbox to materialize imperceptible traces into explicit visual intermediates via Visual-Centric Reasoning, 2) Strategic Tool Learning post-training with gain-driven trajectory construction for SFT and RL optimization guided by tool utility reward, enabling MLLM to act as proactive decision-maker invoking multi-view reasoning paths.
Result: Achieves state-of-the-art performance in both detection and localization tasks, demonstrating superior generalization and robustness with minimal tool redundancy.
Conclusion: Visual-centric reasoning with forensic tools overcomes limitations of text-centric approaches for image forgery detection, enabling more accurate capture of pixel-level inconsistencies through strategic tool learning.
Abstract: Existing Multimodal Large Language Models (MLLMs) for image forgery detection and localization predominantly operate under a text-centric Chain-of-Thought (CoT) paradigm. However, forcing these models to textually characterize imperceptible low-level tampering traces inevitably leads to hallucinations, as linguistic modalities are insufficient to capture such fine-grained pixel-level inconsistencies. To overcome this, we propose ForgeryVCR, a framework that incorporates a forensic toolbox to materialize imperceptible traces into explicit visual intermediates via Visual-Centric Reasoning. To enable efficient tool utilization, we introduce a Strategic Tool Learning post-training paradigm, encompassing gain-driven trajectory construction for Supervised Fine-Tuning (SFT) and subsequent Reinforcement Learning (RL) optimization guided by a tool utility reward. This paradigm empowers the MLLM to act as a proactive decision-maker, learning to spontaneously invoke multi-view reasoning paths including local zoom-in for fine-grained inspection and the analysis of invisible inconsistencies in compression history, noise residuals, and frequency domains. Extensive experiments reveal that ForgeryVCR achieves state-of-the-art (SOTA) performance in both detection and localization tasks, demonstrating superior generalization and robustness with minimal tool redundancy. The project page is available at https://youqiwong.github.io/projects/ForgeryVCR/.
[231] GeoFusionLRM: Geometry-Aware Self-Correction for Consistent 3D Reconstruction
Ahmet Burak Yildirim, Tuna Saygin, Duygu Ceylan, Aysegul Dundar
Main category: cs.CV
TL;DR: GeoFusionLRM: A geometry-aware self-correction framework for single-image 3D reconstruction that uses the model’s own normal and depth predictions to refine structural accuracy through feedback loops.
Details
Motivation: Current large reconstruction models (LRMs) for single-image 3D reconstruction often produce geometric inconsistencies and misaligned details that limit reconstruction fidelity, despite rapid advancements in the field.Method: Introduces a geometry-aware self-correction framework that leverages the model’s own normal and depth predictions to refine structural accuracy. Uses a dedicated transformer and fusion module to feed back geometric cues, enabling error correction and consistency enforcement with the conditioning image, without additional supervision or external signals.
Result: Extensive experiments show GeoFusionLRM achieves sharper geometry, more consistent normals, and higher fidelity than state-of-the-art LRM baselines.
Conclusion: GeoFusionLRM effectively improves 3D reconstruction quality by using self-generated geometric cues for refinement, addressing limitations of current LRMs without requiring additional supervision.
Abstract: Single-image 3D reconstruction with large reconstruction models (LRMs) has advanced rapidly, yet reconstructions often exhibit geometric inconsistencies and misaligned details that limit fidelity. We introduce GeoFusionLRM, a geometry-aware self-correction framework that leverages the model’s own normal and depth predictions to refine structural accuracy. Unlike prior approaches that rely solely on features extracted from the input image, GeoFusionLRM feeds back geometric cues through a dedicated transformer and fusion module, enabling the model to correct errors and enforce consistency with the conditioning image. This design improves the alignment between the reconstructed mesh and the input views without additional supervision or external signals. Extensive experiments demonstrate that GeoFusionLRM achieves sharper geometry, more consistent normals, and higher fidelity than state-of-the-art LRM baselines.
[232] EgoSound: Benchmarking Sound Understanding in Egocentric Videos
Bingwen Zhu, Yuqian Fu, Qiaole Dong, Guolei Sun, Tianwen Qian, Yuzheng Wu, Danda Pani Paudel, Xiangyang Xue, Yanwei Fu
Main category: cs.CV
TL;DR: EgoSound is the first benchmark for evaluating egocentric sound understanding in MLLMs, combining Ego4D and EgoBlind data with 7315 QA pairs across 7 tasks covering sound perception, spatial localization, causal inference, and cross-modal reasoning.
Details
Motivation: Current MLLMs focus on vision-language understanding, but human perception integrates multiple modalities including sound, which provides crucial spatial, off-screen, and causal information, especially in egocentric settings where audio and visual signals are tightly coupled.Method: Created EgoSound benchmark by unifying data from Ego4D and EgoBlind datasets, defining a seven-task taxonomy, and constructing 7315 validated QA pairs across 900 videos through a multi-stage auto-generative pipeline.
Result: Comprehensive evaluation of 9 state-of-the-art MLLMs shows emerging auditory reasoning abilities but limitations in fine-grained spatial and causal understanding, establishing EgoSound as a challenging foundation for multisensory egocentric intelligence.
Conclusion: EgoSound bridges the gap between seeing and truly hearing the world, providing a systematic benchmark to advance multisensory egocentric intelligence in MLLMs beyond current vision-language capabilities.
Abstract: Multimodal Large Language Models (MLLMs) have recently achieved remarkable progress in vision-language understanding. Yet, human perception is inherently multisensory, integrating sight, sound, and motion to reason about the world. Among these modalities, sound provides indispensable cues about spatial layout, off-screen events, and causal interactions, particularly in egocentric settings where auditory and visual signals are tightly coupled. To this end, we introduce EgoSound, the first benchmark designed to systematically evaluate egocentric sound understanding in MLLMs. EgoSound unifies data from Ego4D and EgoBlind, encompassing both sighted and sound-dependent experiences. It defines a seven-task taxonomy spanning intrinsic sound perception, spatial localization, causal inference, and cross-modal reasoning. Constructed through a multi-stage auto-generative pipeline, EgoSound contains 7315 validated QA pairs across 900 videos. Comprehensive experiments on nine state-of-the-art MLLMs reveal that current models exhibit emerging auditory reasoning abilities but remain limited in fine-grained spatial and causal understanding. EgoSound establishes a challenging foundation for advancing multisensory egocentric intelligence, bridging the gap between seeing and truly hearing the world.
[233] DenseMLLM: Standard Multimodal LLMs are Intrinsic Dense Predictors
Yi Li, Hongze Shen, Lexiang Tang, Xin Li, Xinpeng Ding, Yinsong Liu, Deqiang Jiang, Xing Sun, Xiaomeng Li
Main category: cs.CV
TL;DR: DenseMLLM enables standard multimodal LLMs to perform dense prediction tasks like semantic segmentation and depth estimation without task-specific decoders, using vision token supervision for multiple labels/tasks.
Details
Motivation: Current MLLMs excel at high-level visual understanding but require complex, task-specific decoders for fine-grained dense prediction tasks, which increases model complexity and deviates from generalist MLLM design principles.Method: Proposes DenseMLLM with a novel vision token supervision strategy for multiple labels and tasks, enabling standard MLLM architecture to handle dense predictions without additional decoders or architectural specialization.
Result: Achieves highly competitive performance across a wide range of dense prediction and vision-language benchmarks despite minimalist design.
Conclusion: Standard, general-purpose MLLMs can effectively support dense perception tasks without architectural specialization, challenging the paradigm that requires task-specific decoders for fine-grained predictions.
Abstract: Multimodal Large Language Models (MLLMs) have demonstrated exceptional capabilities in high-level visual understanding. However, extending these models to fine-grained dense prediction tasks, such as semantic segmentation and depth estimation, typically necessitates the incorporation of complex, task-specific decoders and other customizations. This architectural fragmentation increases model complexity and deviates from the generalist design of MLLMs, ultimately limiting their practicality. In this work, we challenge this paradigm by accommodating standard MLLMs to perform dense predictions without requiring additional task-specific decoders. The proposed model is called DenseMLLM, grounded in the standard architecture with a novel vision token supervision strategy for multiple labels and tasks. Despite its minimalist design, our model achieves highly competitive performance across a wide range of dense prediction and vision-language benchmarks, demonstrating that a standard, general-purpose MLLM can effectively support dense perception without architectural specialization.
[234] Detection of On-Ground Chestnuts Using Artificial Intelligence Toward Automated Picking
Kaixuan Fang, Yuzhen Lu, Xinyang Mu
Main category: cs.CV
TL;DR: Systematic evaluation of 29 real-time object detectors (YOLO and RT-DETR variants) for chestnut detection in orchard environments, with YOLOv12m achieving best mAP@0.5 of 95.1%
Details
Motivation: Traditional mechanized chestnut harvesting is costly and damaging. Accurate chestnut detection on orchard floors is needed for low-cost, vision-guided automated harvesting, but faces challenges from complex environments with shading, varying light, and interference from weeds/leaves/stones.Method: Collected 319 images with 6524 annotated chestnuts. Evaluated 29 state-of-the-art real-time object detectors (14 YOLO v11-13 variants, 15 RT-DETR v1-v4 variants) through replicated modeling experiments for chestnut detection.
Result: YOLOv12m achieved best mAP@0.5 of 95.1%, RT-DETRv2-R101 was best among RT-DETR models with 91.1% mAP@0.5. YOLOv11x achieved best mAP@[0.5:0.95] of 80.1%. YOLO models outperformed RT-DETR in both accuracy and inference speed, better suited for on-board deployment.
Conclusion: All models show significant potential for real-time chestnut detection. YOLO models are superior for on-board deployment. Dataset and software made publicly available for further research.
Abstract: Traditional mechanized chestnut harvesting is too costly for small producers, non-selective, and prone to damaging nuts. Accurate, reliable detection of chestnuts on the orchard floor is crucial for developing low-cost, vision-guided automated harvesting technology. However, developing a reliable chestnut detection system faces challenges in complex environments with shading, varying natural light conditions, and interference from weeds, fallen leaves, stones, and other foreign on-ground objects, which have remained unaddressed. This study collected 319 images of chestnuts on the orchard floor, containing 6524 annotated chestnuts. A comprehensive set of 29 state-of-the-art real-time object detectors, including 14 in the YOLO (v11-13) and 15 in the RT-DETR (v1-v4) families at varied model scales, was systematically evaluated through replicated modeling experiments for chestnut detection. Experimental results show that the YOLOv12m model achieves the best mAP@0.5 of 95.1% among all the evaluated models, while the RT-DETRv2-R101 was the most accurate variant among RT-DETR models, with mAP@0.5 of 91.1%. In terms of mAP@[0.5:0.95], the YOLOv11x model achieved the best accuracy of 80.1%. All models demonstrate significant potential for real-time chestnut detection, and YOLO models outperformed RT-DETR models in terms of both detection accuracy and inference, making them better suited for on-board deployment. Both the dataset and software programs in this study have been made publicly available at https://github.com/AgFood-Sensing-and-Intelligence-Lab/ChestnutDetection.
[235] LaViDa-R1: Advancing Reasoning for Unified Multimodal Diffusion Language Models
Shufan Li, Yuchen Zhu, Jiuxiang Gu, Kangning Liu, Zhe Lin, Yongxin Chen, Molei Tao, Aditya Grover, Jason Kuen
Main category: cs.CV
TL;DR: LaViDa-R1 is a multimodal reasoning diffusion language model that unifies diverse understanding and generation tasks through a novel post-training framework combining supervised finetuning and multi-task reinforcement learning.
Details
Motivation: To create a general-purpose multimodal reasoning diffusion language model that can handle diverse understanding and generation tasks in a unified manner, addressing limitations of existing approaches that use task-specific reinforcement learning.Method: Proposes LaViDa-R1 with a unified post-training framework integrating supervised finetuning (SFT) and multi-task reinforcement learning (RL). Uses novel techniques including answer-forcing, tree search, and complementary likelihood estimation for enhanced effectiveness and scalability.
Result: Extensive experiments demonstrate strong performance on a wide range of multimodal tasks including visual math reasoning, reason-intensive grounding, and image editing.
Conclusion: LaViDa-R1 represents an effective multimodal reasoning diffusion language model that successfully unifies diverse understanding and generation capabilities through its novel training framework.
Abstract: Diffusion language models (dLLMs) recently emerged as a promising alternative to auto-regressive LLMs. The latest works further extended it to multimodal understanding and generation tasks. In this work, we propose LaViDa-R1, a multimodal, general-purpose reasoning dLLM. Unlike existing works that build reasoning dLLMs through task-specific reinforcement learning, LaViDa-R1 incorporates diverse multimodal understanding and generation tasks in a unified manner. In particular, LaViDa-R1 is built with a novel unified post-training framework that seamlessly integrates supervised finetuning (SFT) and multi-task reinforcement learning (RL). It employs several novel training techniques, including answer-forcing, tree search, and complementary likelihood estimation, to enhance effectiveness and scalability. Extensive experiments demonstrate LaViDa-R1’s strong performance on a wide range of multimodal tasks, including visual math reasoning, reason-intensive grounding, and image editing.
[236] ARport: An Augmented Reality System for Markerless Image-Guided Port Placement in Robotic Surgery
Zheng Han, Zixin Yang, Yonghao Long, Lin Zhang, Peter Kazanzides, Qi Dou
Main category: cs.CV
TL;DR: ARport is an augmented reality system that automatically maps pre-planned trocar layouts onto patient’s body surface for robot-assisted surgery using markerless registration on optical see-through head-mounted displays.
Details
Motivation: To bridge the gap between preoperative planning and intraoperative execution in robot-assisted surgery by providing intuitive spatial guidance for precise port placement, which influences both visual access and instrument maneuverability.Method: Implemented on optical see-through head-mounted display (OST-HMD) without external sensors or markers. Reconstructs operative scene from RGB, depth, and pose data, extracts patient’s body surface using foundation model, performs surface-based markerless registration to align preoperative anatomical models, enabling in-situ visualization of planned trocar layouts.
Result: In full-scale human-phantom experiments, ARport accurately overlaid pre-planned trocar sites onto physical phantom, achieving consistent spatial correspondence between virtual plans and real anatomy.
Conclusion: ARport provides a fully marker-free and hardware-minimal solution for visualizing preoperative trocar plans directly on patient’s body surface, facilitating efficient intraoperative setup with potential for seamless integration into clinical workflows.
Abstract: Purpose: Precise port placement is a critical step in robot-assisted surgery, where port configuration influences both visual access to the operative field and instrument maneuverability. To bridge the gap between preoperative planning and intraoperative execution, we present ARport, an augmented reality (AR) system that automatically maps pre-planned trocar layouts onto the patient’s body surface, providing intuitive spatial guidance during surgical preparation. Methods: ARport, implemented on an optical see-through head-mounted display (OST-HMD), operates without any external sensors or markers, simplifying setup and enhancing workflow integration. It reconstructs the operative scene from RGB, depth, and pose data captured by the OST-HMD, extracts the patient’s body surface using a foundation model, and performs surface-based markerless registration to align preoperative anatomical models to the extracted patient’s body surface, enabling in-situ visualization of planned trocar layouts. A demonstration video illustrating the overall workflow is available online. Results: In full-scale human-phantom experiments, ARport accurately overlaid pre-planned trocar sites onto the physical phantom, achieving consistent spatial correspondence between virtual plans and real anatomy. Conclusion: ARport provides a fully marker-free and hardware-minimal solution for visualizing preoperative trocar plans directly on the patient’s body surface. The system facilitates efficient intraoperative setup and demonstrates potential for seamless integration into routine clinical workflows.
[237] When Test-Time Guidance Is Enough: Fast Image and Video Editing with Diffusion Guidance
Ahmed Ghorbel, Badr Moufad, Navid Bagheri Shouraki, Alain Oliviero Durmus, Thomas Hirtz, Eric Moulines, Jimmy Olsson, Yazid Janati
Main category: cs.CV
TL;DR: VJP-free test-time guidance for diffusion/flow models achieves competitive image/video editing performance without costly vector-Jacobian product computations
Details
Motivation: Text-driven image and video editing as inpainting problems requires efficient test-time guidance methods that avoid computationally expensive VJP approximations used in existing approachesMethod: Builds on Moufad et al. (2025) VJP-free approximation for test-time guidance in diffusion and flow models, with extensive empirical evaluation on large-scale image and video editing benchmarks
Result: Test-time guidance alone achieves performance comparable to or surpassing training-based methods for image and video editing tasks
Conclusion: VJP-free test-time guidance provides a practical and effective approach for text-driven image and video editing without the computational overhead of VJP computations
Abstract: Text-driven image and video editing can be naturally cast as inpainting problems, where masked regions are reconstructed to remain consistent with both the observed content and the editing prompt. Recent advances in test-time guidance for diffusion and flow models provide a principled framework for this task; however, existing methods rely on costly vector–Jacobian product (VJP) computations to approximate the intractable guidance term, limiting their practical applicability. Building upon the recent work of Moufad et al. (2025), we provide theoretical insights into their VJP-free approximation and substantially extend their empirical evaluation to large-scale image and video editing benchmarks. Our results demonstrate that test-time guidance alone can achieve performance comparable to, and in some cases surpass, training-based methods.
[238] Towards Spatial Transcriptomics-driven Pathology Foundation Models
Konstantin Hemker, Andrew H. Song, Cristina Almagro-Pérez, Guillaume Jaume, Sophia J. Wagner, Anurag Vaidya, Nikola Simidjievski, Mateja Jamnik, Faisal Mahmood
Main category: cs.CV
TL;DR: SEAL is a parameter-efficient vision-omics finetuning framework that infuses localized molecular information from spatial transcriptomics into pathology vision encoders, improving performance on various downstream tasks.
Details
Motivation: To leverage morphomolecular coupling between local gene expression and tissue morphology to enhance pathology foundation models, moving beyond vision-only approaches by incorporating spatial transcriptomics data.Method: SEAL uses self-supervised learning with over 700,000 paired gene expression spot-tissue region examples from 14 organs. It’s designed as a parameter-efficient finetuning method that can be flexibly applied to existing pathology foundation models rather than training new encoders from scratch.
Result: SEAL consistently outperforms vision-only and spatial transcriptomics prediction baselines across 38 slide-level and 15 patch-level tasks, including molecular status, pathway activity, treatment response prediction, and gene expression prediction. It also shows robust domain generalization and enables new cross-modal capabilities like gene-to-image retrieval.
Conclusion: Augmenting existing pathology foundation models with localized molecular supervision from spatial transcriptomics is an effective and practical approach for improving visual representations and expanding cross-modal utility in computational pathology.
Abstract: Spatial transcriptomics (ST) provides spatially resolved measurements of gene expression, enabling characterization of the molecular landscape of human tissue beyond histological assessment as well as localized readouts that can be aligned with morphology. Concurrently, the success of multimodal foundation models that integrate vision with complementary modalities suggests that morphomolecular coupling between local expression and morphology can be systematically used to improve histological representations themselves. We introduce Spatial Expression-Aligned Learning (SEAL), a vision-omics self-supervised learning framework that infuses localized molecular information into pathology vision encoders. Rather than training new encoders from scratch, SEAL is designed as a parameter-efficient vision-omics finetuning method that can be flexibly applied to widely used pathology foundation models. We instantiate SEAL by training on over 700,000 paired gene expression spot-tissue region examples spanning tumor and normal samples from 14 organs. Tested across 38 slide-level and 15 patch-level downstream tasks, SEAL provides a drop-in replacement for pathology foundation models that consistently improves performance over widely used vision-only and ST prediction baselines on slide-level molecular status, pathway activity, and treatment response prediction, as well as patch-level gene expression prediction tasks. Additionally, SEAL encoders exhibit robust domain generalization on out-of-distribution evaluations and enable new cross-modal capabilities such as gene-to-image retrieval. Our work proposes a general framework for ST-guided finetuning of pathology foundation models, showing that augmenting existing models with localized molecular supervision is an effective and practical step for improving visual representations and expanding their cross-modal utility.
[239] UniWeTok: An Unified Binary Tokenizer with Codebook Size $\mathit{2^{128}}$ for Unified Multimodal Large Language Model
Shaobin Zhuang, Yuang Ai, Jiaming Han, Weijia Mao, Xiaohui Li, Fangyikang Wang, Xiao Wang, Yan Li, Shanchuan Lin, Kun Xu, Zhenheng Yang, Huaibo Huang, Xiangyu Yue, Hao Chen, Yali Wang
Main category: cs.CV
TL;DR: UniWeTok: A unified multimodal tokenizer using massive binary codebook with novel training framework and architecture for high-fidelity reconstruction, semantic extraction, and generative suitability.
Details
Motivation: Existing visual tokenizers struggle to simultaneously support high-fidelity reconstruction, complex semantic extraction, and generative suitability within a single framework for unified multimodal LLMs.Method: Introduces UniWeTok with massive binary codebook (2^128), Pre-Post Distillation, Generative-Aware Prior, convolution-attention hybrid architecture with SigLu activation, and three-stage training framework for cross-resolution adaptability.
Result: Achieves SOTA image generation (FID: 1.38 vs REPA 1.42) with low training compute (33B vs 262B tokens), competitive multimodal understanding, generation (DPG: 86.63 vs FLUX.1 83.84), and editing (GEdit: 5.09 vs OmniGen 5.06).
Conclusion: UniWeTok provides a unified visual tokenizer that effectively balances reconstruction, semantic extraction, and generative capabilities, advancing multimodal LLM development with efficient training.
Abstract: Unified Multimodal Large Language Models (MLLMs) require a visual representation that simultaneously supports high-fidelity reconstruction, complex semantic extraction, and generative suitability. However, existing visual tokenizers typically struggle to satisfy these conflicting objectives within a single framework. In this paper, we introduce UniWeTok, a unified discrete tokenizer designed to bridge this gap using a massive binary codebook ($\mathit{2^{128}}$). For training framework, we introduce Pre-Post Distillation and a Generative-Aware Prior to enhance the semantic extraction and generative prior of the discrete tokens. In terms of model architecture, we propose a convolution-attention hybrid architecture with the SigLu activation function. SigLu activation not only bounds the encoder output and stabilizes the semantic distillation process but also effectively addresses the optimization conflict between token entropy loss and commitment loss. We further propose a three-stage training framework designed to enhance UniWeTok’s adaptability cross various image resolutions and perception-sensitive scenarios, such as those involving human faces and textual content. On ImageNet, UniWeTok achieves state-of-the-art image generation performance (FID: UniWeTok 1.38 vs. REPA 1.42) while requiring a remarkably low training compute (Training Tokens: UniWeTok 33B vs. REPA 262B). On general-domain, UniWeTok demonstrates highly competitive capabilities across a broad range of tasks, including multimodal understanding, image generation (DPG Score: UniWeTok 86.63 vs. FLUX.1 [Dev] 83.84), and editing (GEdit Overall Score: UniWeTok 5.09 vs. OmniGen 5.06). We release code and models to facilitate community exploration of unified tokenizer and MLLM.
[240] UniRef-Image-Edit: Towards Scalable and Consistent Multi-Reference Image Editing
Hongyang Wei, Bin Wen, Yancheng Long, Yankai Yang, Yuhang Hu, Tianke Zhang, Wei Chen, Haonan Fan, Kaiyu Jiang, Jiankang Chen, Changyi Liu, Kaiyu Tang, Haojie Ding, Xiao Yang, Jia Sun, Huaiqing Wang, Zhenyu Yang, Xinyu Wei, Xianglong He, Yangguang Li, Fan Yang, Tingting Gao, Lei Zhang, Guorui Zhou, Han Li
Main category: cs.CV
TL;DR: UniRef-Image-Edit: A unified framework for single-image editing and multi-image composition using Sequence-Extended Latent Fusion (SELF) and two-stage training with supervised fine-tuning and reinforcement learning.
Details
Motivation: Existing diffusion-based editing methods struggle with consistency across multiple conditions due to limited interaction between reference inputs. The paper aims to create a unified system that handles both single-image editing and multi-image composition effectively.Method: Introduces Sequence-Extended Latent Fusion (SELF) to serialize multiple reference images into coherent latent sequences. Uses two-stage training: 1) Supervised fine-tuning with progressive sequence length training (1024² to 2048² pixel budget), 2) Multi-Source GRPO reinforcement learning framework to optimize compositional consistency.
Result: Developed a high-performance multi-modal generation system that unifies single-image editing and multi-image composition. The system maintains consistency across multiple reference images and improves visual fidelity through progressive training.
Conclusion: UniRef-Image-Edit successfully addresses consistency issues in multi-reference image generation through unified latent representation and two-stage training, offering a comprehensive solution for both single and multi-image editing tasks.
Abstract: We present UniRef-Image-Edit, a high-performance multi-modal generation system that unifies single-image editing and multi-image composition within a single framework. Existing diffusion-based editing methods often struggle to maintain consistency across multiple conditions due to limited interaction between reference inputs. To address this, we introduce Sequence-Extended Latent Fusion (SELF), a unified input representation that dynamically serializes multiple reference images into a coherent latent sequence. During a dedicated training stage, all reference images are jointly constrained to fit within a fixed-length sequence under a global pixel-budget constraint. Building upon SELF, we propose a two-stage training framework comprising supervised fine-tuning (SFT) and reinforcement learning (RL). In the SFT stage, we jointly train on single-image editing and multi-image composition tasks to establish a robust generative prior. We adopt a progressive sequence length training strategy, in which all input images are initially resized to a total pixel budget of $1024^2$, and are then gradually increased to $1536^2$ and $2048^2$ to improve visual fidelity and cross-reference consistency. This gradual relaxation of compression enables the model to incrementally capture finer visual details while maintaining stable alignment across references. For the RL stage, we introduce Multi-Source GRPO (MSGRPO), to our knowledge the first reinforcement learning framework tailored for multi-reference image generation. MSGRPO optimizes the model to reconcile conflicting visual constraints, significantly enhancing compositional consistency. We will open-source the code, models, training data, and reward data for community research purposes.
[241] GeoEyes: On-Demand Visual Focusing for Evidence-Grounded Understanding of Ultra-High-Resolution Remote Sensing Imagery
Fengxiang Wang, Mingshuo Chen, Yueying Li, Yajie Yang, Yifan Zhang, Long Lan, Xue Yang, Hongda Sun, Yulin Wang, Di Wang, Jun Song, Jing Zhang, Bo Du
Main category: cs.CV
TL;DR: GeoEyes is a training framework for multimodal LLMs that addresses tool usage homogenization in zoom-enabled models for ultra-high-resolution remote sensing visual question answering.
Details
Motivation: Existing zoom-enabled MLLMs suffer from "Tool Usage Homogenization" where tool calls collapse into task-agnostic patterns, limiting effective evidence acquisition for ultra-high-resolution remote sensing VQA where relevant cues are sparse and tiny.Method: Two-stage framework: (1) Cold-start SFT dataset UHR Chain-of-Zoom (UHR-CoZ) covering diverse zooming regimes, and (2) Agentic reinforcement learning method AdaZoom-GRPO that explicitly rewards evidence gain and answer improvement during zoom interactions.
Result: The model learns on-demand zooming with proper stopping behavior and achieves substantial improvements on UHR remote sensing benchmarks, with 54.23% accuracy on XLRS-Bench.
Conclusion: GeoEyes effectively addresses tool usage homogenization in zoom-enabled MLLMs for remote sensing VQA through staged training combining specialized datasets and reinforcement learning with explicit evidence-based rewards.
Abstract: The “thinking-with-images” paradigm enables multimodal large language models (MLLMs) to actively explore visual scenes via zoom-in tools. This is essential for ultra-high-resolution (UHR) remote sensing VQA, where task-relevant cues are sparse and tiny. However, we observe a consistent failure mode in existing zoom-enabled MLLMs: Tool Usage Homogenization, where tool calls collapse into task-agnostic patterns, limiting effective evidence acquisition. To address this, we propose GeoEyes, a staged training framework consisting of (1) a cold-start SFT dataset, UHR Chain-of-Zoom (UHR-CoZ), which covers diverse zooming regimes, and (2) an agentic reinforcement learning method, AdaZoom-GRPO, that explicitly rewards evidence gain and answer improvement during zoom interactions. The resulting model learns on-demand zooming with proper stopping behavior and achieves substantial improvements on UHR remote sensing benchmarks, with 54.23% accuracy on XLRS-Bench.
[242] HiVid: LLM-Guided Video Saliency For Content-Aware VOD And Live Streaming
Jiahui Chen, Bo Peng, Lianchen Jia, Zeyu Zhang, Tianchi Huang, Lifeng Sun
Main category: cs.CV
TL;DR: HiVid uses LLMs as scalable human proxies to generate chunk-level importance weights for video streaming QoE optimization, addressing modality limitations, rating inconsistency, and live streaming latency challenges.
Details
Motivation: Content-aware streaming needs dynamic chunk-level importance weights for optimal QoE, but human annotation is expensive and vision-saliency models generalize poorly. LLMs offer a scalable alternative but face modality, token limit, and latency challenges.Method: Three-module framework: (1) Perception module uses local context windows for coherent video understanding; (2) Ranking module with LLM-guided merge-sort for global re-ranking in VOD; (3) Prediction module with multi-modal time series model for low-latency live streaming.
Result: HiVid improves weight prediction accuracy by up to 11.5% for VOD and 26% for live streaming over SOTA baselines. Real-world user study shows 14.7% boost in streaming QoE correlation.
Conclusion: HiVid successfully leverages LLMs as scalable human proxies for generating high-fidelity importance weights in both VOD and live streaming, significantly improving QoE optimization.
Abstract: Content-aware streaming requires dynamic, chunk-level importance weights to optimize subjective quality of experience (QoE). However, direct human annotation is prohibitively expensive while vision-saliency models generalize poorly. We introduce HiVid, the first framework to leverage Large Language Models (LLMs) as a scalable human proxy to generate high-fidelity weights for both Video-on-Demand (VOD) and live streaming. We address 3 non-trivial challenges: (1) To extend LLMs’ limited modality and circumvent token limits, we propose a perception module to assess frames in a local context window, autoregressively building a coherent understanding of the video. (2) For VOD with rating inconsistency across local windows, we propose a ranking module to perform global re-ranking with a novel LLM-guided merge-sort algorithm. (3) For live streaming which requires low-latency, online inference without future knowledge, we propose a prediction module to predict future weights with a multi-modal time series model, which comprises a content-aware attention and adaptive horizon to accommodate asynchronous LLM inference. Extensive experiments show HiVid improves weight prediction accuracy by up to 11.5% for VOD and 26% for live streaming over SOTA baselines. Real-world user study validates HiVid boosts streaming QoE correlation by 14.7%.
[243] Freq-DP Net: A Dual-Branch Network for Fence Removal using Dual-Pixel and Fourier Priors
Kunal Swami, Sudha Velusamy, Chandra Sekhar Seelamantula
Main category: cs.CV
TL;DR: Freq-DP Net: A dual-branch network using dual-pixel sensors for single-image fence removal, combining geometric defocus disparity and structural Fourier pattern priors.
Details
Motivation: Fence occlusions degrade visual quality and limit computer vision applications. Existing methods fail on static scenes or require multiple frames, so the authors propose leveraging dual-pixel sensors for single-image fence removal.Method: Proposes Freq-DP Net with two complementary priors: 1) geometric prior from defocus disparity using explicit cost volume, and 2) structural prior of fence’s global pattern via Fast Fourier Convolution. Uses attention mechanism to fuse these cues for accurate fence segmentation.
Result: Method significantly outperforms strong general-purpose baselines, establishing new state-of-the-art for single-image, DP-based fence removal. Authors built and released a diverse benchmark with different fence varieties.
Conclusion: First framework to leverage dual-pixel sensors for fence removal, combining geometric and structural priors effectively. Provides practical solution for single-image fence occlusion removal.
Abstract: Removing fence occlusions from single images is a challenging task that degrades visual quality and limits downstream computer vision applications. Existing methods often fail on static scenes or require motion cues from multiple frames. To overcome these limitations, we introduce the first framework to leverage dual-pixel (DP) sensors for this problem. We propose Freq-DP Net, a novel dual-branch network that fuses two complementary priors: a geometric prior from defocus disparity, modeled using an explicit cost volume, and a structural prior of the fence’s global pattern, learned via Fast Fourier Convolution (FFC). An attention mechanism intelligently merges these cues for highly accurate fence segmentation. To validate our approach, we build and release a diverse benchmark with different fence varieties. Experiments demonstrate that our method significantly outperforms strong general-purpose baselines, establishing a new state-of-the-art for single-image, DP-based fence removal.
[244] Learning Significant Persistent Homology Features for 3D Shape Understanding
Prachi Kudeshia, Jiju Poovvancheri
Main category: cs.CV
TL;DR: Paper introduces topologically-enriched 3D shape datasets (ModelNet40/ShapeNet with persistent homology features) and TopoGAT, a deep learning method for selecting significant topological features, improving point cloud classification and segmentation.
Details
Motivation: Existing 3D shape datasets focus on geometric information but neglect topological structure. There's a need for unified geometry-topology learning benchmarks and better methods to integrate topological features into deep learning workflows for 3D point cloud analysis.Method: 1) Created topologically-enriched versions of ModelNet40 and ShapeNet by augmenting point clouds with persistent homology features. 2) Proposed TopoGAT, a deep learning-based method that learns to identify the most informative topological features directly from input data and topological signatures, avoiding hand-crafted statistical selection criteria.
Result: TopoGAT outperforms traditional statistical approaches in stability and discriminative power. Integrating selected persistent points into standard point cloud classification and part-segmentation pipelines improves both classification accuracy and segmentation metrics.
Conclusion: The topologically-enriched datasets and learnable feature selection approach enable broader integration of persistent homology into practical deep learning workflows for 3D point cloud analysis, bridging the gap between geometric and topological shape descriptors.
Abstract: Geometry and topology constitute complementary descriptors of three-dimensional shape, yet existing benchmark datasets primarily capture geometric information while neglecting topological structure. This work addresses this limitation by introducing topologically-enriched versions of ModelNet40 and ShapeNet, where each point cloud is augmented with its corresponding persistent homology features. These benchmarks with the topological signatures establish a foundation for unified geometry-topology learning and enable systematic evaluation of topology-aware deep learning architectures for 3D shape analysis. Building on this foundation, we propose a deep learning-based significant persistent point selection method, \textit{TopoGAT}, that learns to identify the most informative topological features directly from input data and the corresponding topological signatures, circumventing the limitations of hand-crafted statistical selection criteria. A comparative study verifies the superiority of the proposed method over traditional statistical approaches in terms of stability and discriminative power. Integrating the selected significant persistent points into standard point cloud classification and part-segmentation pipelines yields improvements in both classification accuracy and segmentation metrics. The presented topologically-enriched datasets, coupled with our learnable significant feature selection approach, enable the broader integration of persistent homology into the practical deep learning workflows for 3D point cloud analysis.
[245] Dual-Signal Adaptive KV-Cache Optimization for Long-Form Video Understanding in Vision-Language Models
Vishnu Sai, Dheeraj Sai, Srinath B, Girish Varma, Priyesh Shukla
Main category: cs.CV
TL;DR: Sali-Cache is a novel memory optimization framework for Vision-Language Models that uses dual-signal adaptive caching to reduce KV cache memory bottleneck in long-form video processing.
Details
Motivation: VLMs face critical memory bottlenecks when processing long-form video due to linear growth of KV cache with sequence length. Existing reactive eviction strategies waste computation by computing full attention matrices before discarding tokens.Method: Proactive memory management framework with dual-signal adaptive caching: 1) temporal filter using optical flow analysis to detect inter-frame redundancy, and 2) spatial filter using saliency detection to identify visually significant regions. Manages memory allocation before expensive attention operations.
Result: Achieves 2.20x compression ratio in effective memory usage while maintaining 100% accuracy across BLEU, ROUGE-L, and Exact Match metrics. Preserves context-rich features over extended temporal durations without degrading performance under identical memory budget constraints.
Conclusion: Sali-Cache enables efficient processing of long-form video content on consumer-grade hardware by intelligently managing memory allocation through proactive optimization, addressing the critical memory bottleneck in VLMs.
Abstract: Vision-Language Models (VLMs) face a critical memory bottleneck when processing long-form video content due to the linear growth of the Key-Value (KV) cache with sequence length. Existing solutions predominantly employ reactive eviction strategies that compute full attention matrices before discarding tokens, resulting in substantial computational waste. We propose Sali-Cache, a novel a priori optimization framework that implements dual-signal adaptive caching through proactive memory management. By integrating a temporal filter based on optical flow analysis for detecting inter-frame redundancy and a spatial filter leveraging saliency detection for identifying visually significant regions, Sali-Cache intelligently manages memory allocation before entering computationally expensive attention operations. Experimental evaluation on the LLaVA 1.6 architecture demonstrates that our method achieves a 2.20x compression ratio in effective memory usage while maintaining 100% accuracy across BLEU, ROUGE-L, and Exact Match metrics. Furthermore, under identical memory budget constraints, Sali-Cache preserves context-rich features over extended temporal durations without degrading model performance, enabling efficient processing of long-form video content on consumer-grade hardware.
[246] AbracADDbra: Touch-Guided Object Addition by Decoupling Placement and Editing Subtasks
Kunal Swami, Raghu Chittersu, Yuvraj Rathore, Rajeev Irny, Shashavali Doodekula, Alok Shukla
Main category: cs.CV
TL;DR: AbracADDbra: A user-friendly framework for object addition using intuitive touch inputs instead of ambiguous text prompts or tedious masks, featuring touch-guided placement and diffusion-based generation with instance masks for high-fidelity blending.
Details
Motivation: Instruction-based object addition faces usability issues with ambiguous text-only prompts or tedious mask-based inputs. The paper aims to address this gap by creating a more intuitive, user-friendly approach to object placement in images.Method: Proposes a decoupled architecture: 1) a vision-language transformer for touch-guided spatial placement based on intuitive touch priors, and 2) a diffusion model that jointly generates the object and an instance mask for high-fidelity blending. Also introduces the Touch2Add benchmark for standardized evaluation.
Result: The placement model significantly outperforms both random placement and general-purpose VLM baselines. The framework produces high-fidelity edits, and analysis shows strong correlation between initial placement accuracy and final edit quality, validating the decoupled approach.
Conclusion: The work paves the way for more accessible and efficient creative tools by introducing intuitive touch-based interaction for object addition, with a decoupled architecture that enables precise placement and high-fidelity blending.
Abstract: Instruction-based object addition is often hindered by the ambiguity of text-only prompts or the tedious nature of mask-based inputs. To address this usability gap, we introduce AbracADDbra, a user-friendly framework that leverages intuitive touch priors to spatially ground succinct instructions for precise placement. Our efficient, decoupled architecture uses a vision-language transformer for touch-guided placement, followed by a diffusion model that jointly generates the object and an instance mask for high-fidelity blending. To facilitate standardized evaluation, we contribute the Touch2Add benchmark for this interactive task. Our extensive evaluations, where our placement model significantly outperforms both random placement and general-purpose VLM baselines, confirm the framework’s ability to produce high-fidelity edits. Furthermore, our analysis reveals a strong correlation between initial placement accuracy and final edit quality, validating our decoupled approach. This work thus paves the way for more accessible and efficient creative tools.
[247] Moving Beyond Sparse Grounding with Complete Screen Parsing Supervision
A. Said Gurbuz, Sunghwan Hong, Ahmed Nassar, Marc Pollefeys, Peter Staar
Main category: cs.CV
TL;DR: ScreenParse: A large-scale dataset for complete screen parsing with dense UI element annotations, used to train ScreenVLM, a compact vision-language model for structured screen understanding.
Details
Motivation: Current computer-use agents need structured screen understanding but existing datasets provide sparse supervision with limited coverage and diversity. Practical deployment also requires efficient, low-latency models for on-device use.Method: Created ScreenParse dataset with 771K web screenshots (21M elements) using Webshot pipeline for automated rendering, annotation extraction, and VLM-based relabeling. Trained ScreenVLM (316M parameters) with compact ScreenTag markup representation and structure-aware loss.
Result: ScreenVLM substantially outperforms larger foundation VLMs on dense parsing (0.592 vs 0.294 PageIoU) and shows strong transfer to public benchmarks. Finetuning foundation VLMs on ScreenParse consistently improves grounding performance.
Conclusion: Dense screen supervision provides transferable structural priors for UI understanding. ScreenParse enables training compact, efficient models for complete screen parsing with strong generalization capabilities.
Abstract: Modern computer-use agents (CUA) must perceive a screen as a structured state, what elements are visible, where they are, and what text they contain, before they can reliably ground instructions and act. Yet, most available grounding datasets provide sparse supervision, with insufficient and low-diversity labels that annotate only a small subset of task-relevant elements per screen, which limits both coverage and generalization; moreover, practical deployment requires efficiency to enable low-latency, on-device use. We introduce ScreenParse, a large-scale dataset for complete screen parsing, with dense annotations of all visible UI elements (boxes, 55-class types, and text) across 771K web screenshots (21M elements). ScreenParse is generated by Webshot, an automated, scalable pipeline that renders diverse urls, extracts annotations and applies VLM-based relabeling and quality filtering. Using ScreenParse, we train ScreenVLM, a compact, 316M-parameter vision language model (VLM) that decodes a compact ScreenTag markup representation with a structure-aware loss that upweights structure-critical tokens. ScreenVLM substantially outperforms much larger foundation VLMs on dense parsing (e.g., 0.592 vs. 0.294 PageIoU on ScreenParse) and shows strong transfer to public benchmarks. Moreover, finetuning foundation VLMs on ScreenParse consistently improves their grounding performance, suggesting that dense screen supervision provides transferable structural priors for UI understanding. Project page: https://saidgurbuz.github.io/screenparse/.
[248] Differential pose optimization in descriptor space – Combining Geometric and Photometric Methods for Motion Estimation
Andreas L. Teigen, Annette Stahl, Rudolf Mester
Main category: cs.CV
TL;DR: Proposes combining photometric and geometric feature paradigms by using dense geometric descriptors instead of photometric error for relative pose optimization, but finds it doesn’t outperform traditional reprojection error methods despite using more information.
Details
Motivation: Addresses the trade-off between photometric error (accuracy) and reprojection error (robustness, loop closing) in two-frame relative pose optimization by creating a unified approach that combines strengths of both paradigms.Method: Uses densely sampled geometric feature descriptors to replace photometric error with descriptor residuals, enabling sub-pixel accuracy from differential photometric methods while leveraging geometric descriptor expressiveness.
Result: Experiments show the proposed strategy results in accurate tracking but ultimately does not outperform pose optimization strategies based on reprojection error despite utilizing more information.
Conclusion: The descriptor similarity metric is too slowly varying and doesn’t necessarily correspond strictly to keypoint placement accuracy, explaining why the unified approach doesn’t outperform traditional methods.
Abstract: One of the fundamental problems in computer vision is the two-frame relative pose optimization problem. Primarily, two different kinds of error values are used: photometric error and re-projection error. The selection of error value is usually directly dependent on the selection of feature paradigm, photometric features, or geometric features. It is a trade-off between accuracy, robustness, and the possibility of loop closing. We investigate a third method that combines the strengths of both paradigms into a unified approach. Using densely sampled geometric feature descriptors, we replace the photometric error with a descriptor residual from a dense set of descriptors, thereby enabling the employment of sub-pixel accuracy in differential photometric methods, along with the expressiveness of the geometric feature descriptor. Experiments show that although the proposed strategy is an interesting approach that results in accurate tracking, it ultimately does not outperform pose optimization strategies based on re-projection error despite utilizing more information. We proceed to analyze the underlying reason for this discrepancy and present the hypothesis that the descriptor similarity metric is too slowly varying and does not necessarily correspond strictly to keypoint placement accuracy.
[249] A Generative AI Approach for Reducing Skin Tone Bias in Skin Cancer Classification
Areez Muhammed Shabu, Mohammad Samar Ansari, Asra Aslam
Main category: cs.CV
TL;DR: Using Stable Diffusion with LoRA fine-tuning to generate synthetic dermoscopic images for darker skin tones to address dataset imbalance in skin cancer detection, improving segmentation and classification performance.
Details
Motivation: Current AI diagnostic tools for skin cancer are biased toward lighter skin tones due to dataset imbalances (ISIC dataset has >70% light skin images, <8% dark skin), creating fairness issues and reduced accuracy for people with darker skin, highlighting the need for methods that address demographic diversity in medical imaging.Method: A generative augmentation pipeline that fine-tunes a pre-trained Stable Diffusion model using Low-Rank Adaptation (LoRA) on the dark-skin subset of the ISIC dataset, generating synthetic dermoscopic images conditioned on lesion type and skin tone for data augmentation.
Result: Models trained on augmented data showed consistent improvements in segmentation metrics (IoU, Dice coefficient, boundary accuracy) and achieved 92.14% accuracy for binary classification using EfficientNet-B0, demonstrating reduced bias and increased fairness.
Conclusion: Synthetic data augmentation with Generative AI can substantially reduce bias and increase fairness in dermatological diagnostics, addressing skin tone imbalance in medical imaging datasets.
Abstract: Skin cancer is one of the most common cancers worldwide and early detection is critical for effective treatment. However, current AI diagnostic tools are often trained on datasets dominated by lighter skin tones, leading to reduced accuracy and fairness for people with darker skin. The International Skin Imaging Collaboration (ISIC) dataset, one of the most widely used benchmarks, contains over 70% light skin images while dark skins fewer than 8%. This imbalance poses a significant barrier to equitable healthcare delivery and highlights the urgent need for methods that address demographic diversity in medical imaging. This paper addresses this challenge of skin tone imbalance in automated skin cancer detection using dermoscopic images. To overcome this, we present a generative augmentation pipeline that fine-tunes a pre-trained Stable Diffusion model using Low-Rank Adaptation (LoRA) on the image dark-skin subset of the ISIC dataset and generates synthetic dermoscopic images conditioned on lesion type and skin tone. In this study, we investigated the utility of these images on two downstream tasks: lesion segmentation and binary classification. For segmentation, models trained on the augmented dataset and evaluated on held-out real images show consistent improvements in IoU, Dice coefficient, and boundary accuracy. These evalutions provides the verification of Generated dataset. For classification, an EfficientNet-B0 model trained on the augmented dataset achieved 92.14% accuracy. This paper demonstrates that synthetic data augmentation with Generative AI integration can substantially reduce bias with increase fairness in conventional dermatological diagnostics and open challenges for future directions.
[250] Image-based Joint-level Detection for Inflammation in Rheumatoid Arthritis from Small and Imbalanced Data
Shun Kato, Yasushi Kondo, Shuntaro Saito, Yoshimitsu Aoki, Mariko Isogawa
Main category: cs.CV
TL;DR: A framework for detecting rheumatoid arthritis inflammation from RGB hand images using self-supervised pretraining on healthy hand data and imbalance-aware training to address medical imaging challenges.
Details
Motivation: Early diagnosis and monitoring of rheumatoid arthritis is crucial but specialist care access is limited. There's a need for accessible systems using RGB hand images captured at home to detect joint inflammation, addressing challenges like data scarcity, imbalance, and task difficulty.Method: Proposes an inflammation detection framework with global-local encoder that combines self-supervised pretraining on large-scale healthy hand images with imbalance-aware training to detect RA-related joint inflammation from RGB hand images.
Result: The proposed approach improves F1-score by 0.2 points and Gmean by 0.25 points compared with the baseline model, demonstrating effectiveness in addressing the challenges of RA inflammation detection.
Conclusion: The framework successfully addresses key challenges in medical imaging for RA detection, providing a promising approach for accessible inflammation detection using RGB hand images with improved performance metrics.
Abstract: Rheumatoid arthritis (RA) is an autoimmune disease characterized by systemic joint inflammation. Early diagnosis and tight follow-up are essential to the management of RA, as ongoing inflammation can cause irreversible joint damage. The detection of arthritis is important for diagnosis and assessment of disease activity; however, it often takes a long time for patients to receive appropriate specialist care. Therefore, there is a strong need to develop systems that can detect joint inflammation easily using RGB images captured at home. Consequently, we tackle the task of RA inflammation detection from RGB hand images. This task is highly challenging due to general issues in medical imaging, such as the scarcity of positive samples, data imbalance, and the inherent difficulty of the task itself. However, to the best of our knowledge, no existing work has explicitly addressed these challenges in RGB-based RA inflammation detection. This paper quantitatively demonstrates the difficulty of visually detecting inflammation by constructing a dedicated dataset, and we propose a inflammation detection framework with global local encoder that combines self-supervised pretraining on large-scale healthy hand images with imbalance-aware training to detect RA-related joint inflammation from RGB hand images. Our experiments demonstrated that the proposed approach improves F1-score by 0.2 points and Gmean by 0.25 points compared with the baseline model.
[251] Event-based Visual Deformation Measurement
Yuliang Wu, Wei Zhai, Yuxin Cui, Tiesong Zhao, Yang Cao, Zheng-Jun Zha
Main category: cs.CV
TL;DR: Event-frame fusion framework for visual deformation measurement using events for dense temporal motion cues and frames for spatial precision, with Affine Invariant Simplicial modeling and neighborhood-greedy optimization.
Details
Motivation: Traditional visual deformation measurement methods rely on minimal inter-frame motion, limiting applicability to highly dynamic scenes or requiring high-speed cameras with prohibitive storage/computational costs. Need for temporally dense motion tracking without these limitations.Method: Proposes event-frame fusion framework combining events (temporally dense motion cues) and frames (spatially dense precise estimation). Uses Affine Invariant Simplicial (AIS) framework partitioning deformation field into linearized sub-regions with low-parametric representation. Introduces neighborhood-greedy optimization strategy for faster parameter search and reduced error accumulation.
Result: Outperforms state-of-the-art baseline by 1.6% in survival rate. Achieves this using only 18.9% of data storage and processing resources compared to high-speed video methods. Established benchmark dataset with over 120 sequences spanning diverse deformation scenarios.
Conclusion: Event-frame fusion with AIS modeling and neighborhood-greedy optimization enables efficient, accurate dense deformation tracking in dynamic scenes with significantly reduced computational and storage requirements compared to high-speed video approaches.
Abstract: Visual Deformation Measurement (VDM) aims to recover dense deformation fields by tracking surface motion from camera observations. Traditional image-based methods rely on minimal inter-frame motion to constrain the correspondence search space, which limits their applicability to highly dynamic scenes or necessitates high-speed cameras at the cost of prohibitive storage and computational overhead. We propose an event-frame fusion framework that exploits events for temporally dense motion cues and frames for spatially dense precise estimation. Revisiting the solid elastic modeling prior, we propose an Affine Invariant Simplicial (AIS) framework. It partitions the deformation field into linearized sub-regions with low-parametric representation, effectively mitigating motion ambiguities arising from sparse and noisy events. To speed up parameter searching and reduce error accumulation, a neighborhood-greedy optimization strategy is introduced, enabling well-converged sub-regions to guide their poorly-converged neighbors, effectively suppress local error accumulation in long-term dense tracking. To evaluate the proposed method, a benchmark dataset with temporally aligned event streams and frames is established, encompassing over 120 sequences spanning diverse deformation scenarios. Experimental results show that our method outperforms the state-of-the-art baseline by 1.6% in survival rate. Remarkably, it achieves this using only 18.9% of the data storage and processing resources of high-speed video methods.
[252] Adapting VACE for Real-Time Autoregressive Video Diffusion
Ryan Fosdick
Main category: cs.CV
TL;DR: Adaptation of VACE for real-time autoregressive video generation by moving reference frames to parallel conditioning pathway, enabling streaming pipelines with fixed chunk sizes and causal attention while reusing pretrained weights.
Details
Motivation: VACE provides unified video control but uses bidirectional attention over full sequences, making it incompatible with streaming pipelines that require fixed chunk sizes and causal attention for real-time autoregressive generation.Method: Key modification moves reference frames from diffusion latent space into a parallel conditioning pathway, preserving fixed chunk sizes and KV caching required by autoregressive models. Reuses existing pretrained VACE weights without additional training.
Result: VACE adds 20-30% latency overhead for structural control and inpainting across 1.3B and 14B model scales with negligible VRAM cost. However, reference-to-video fidelity is severely degraded compared to batch VACE due to causal attention constraints.
Conclusion: The adaptation enables real-time autoregressive video generation with VACE’s unified control capabilities, though with trade-offs in fidelity due to causal attention requirements of streaming pipelines.
Abstract: We describe an adaptation of VACE (Video All-in-one Creation and Editing) for real-time autoregressive video generation. VACE provides unified video control (reference guidance, structural conditioning, inpainting, and temporal extension) but assumes bidirectional attention over full sequences, making it incompatible with streaming pipelines that require fixed chunk sizes and causal attention. The key modification moves reference frames from the diffusion latent space into a parallel conditioning pathway, preserving the fixed chunk sizes and KV caching that autoregressive models require. This adaptation reuses existing pretrained VACE weights without additional training. Across 1.3B and 14B model scales, VACE adds 20-30% latency overhead for structural control and inpainting, with negligible VRAM cost relative to the base model. Reference-to-video fidelity is severely degraded compared to batch VACE due to causal attention constraints. A reference implementation is available at https://github.com/daydreamlive/scope.
[253] Multi-Turn Adaptive Prompting Attack on Large Vision-Language Models
In Chong Choi, Jiacheng Zhang, Feng Liu, Yiliao Song
Main category: cs.CV
TL;DR: MAPA is a multi-turn adaptive prompting attack method for large vision-language models that alternates text-vision attack actions and refines trajectories across turns to bypass safety defenses.
Details
Motivation: Existing multi-turn jailbreak attacks work on text-only LLMs but fail on LVLMs because overly malicious visual inputs trigger safety mechanisms, making responses more conservative. There's a need for adaptive attacks that can bypass multimodal safety alignment.Method: Two-level design: 1) At each turn, alternates between text and vision attack actions to elicit the most malicious response; 2) Across turns, adjusts attack trajectory through iterative back-and-forth refinement to gradually amplify response maliciousness.
Result: MAPA consistently outperforms state-of-the-art methods, improving attack success rates by 11-35% on benchmarks against LLaVA-V1.6-Mistral-7B, Qwen2.5-VL-7B-Instruct, Llama-3.2-Vision-11B-Instruct and GPT-4o-mini.
Conclusion: The proposed adaptive multi-turn attack method effectively bypasses safety mechanisms in vision-language models by strategically alternating between text and visual inputs and refining attack trajectories.
Abstract: Multi-turn jailbreak attacks are effective against text-only large language models (LLMs) by gradually introducing malicious content across turns. When extended to large vision-language models (LVLMs), we find that naively adding visual inputs can cause existing multi-turn jailbreaks to be easily defended. For example, overly malicious visual input will easily trigger the defense mechanism of safety-aligned LVLMs, making the response more conservative. To address this, we propose MAPA: a multi-turn adaptive prompting attack that 1) at each turn, alternates text-vision attack actions to elicit the most malicious response; and 2) across turns, adjusts the attack trajectory through iterative back-and-forth refinement to gradually amplify response maliciousness. This two-level design enables MAPA to consistently outperform state-of-the-art methods, improving attack success rates by 11-35% on recent benchmarks against LLaVA-V1.6-Mistral-7B, Qwen2.5-VL-7B-Instruct, Llama-3.2-Vision-11B-Instruct and GPT-4o-mini.
[254] pFedNavi: Structure-Aware Personalized Federated Vision-Language Navigation for Embodied AI
Qingqian Yang, Hao Wang, Sai Qian Zhang, Jian Li, Yang Hua, Miao Pan, Tao Song, Zhengwei Qi, Haibing Guan
Main category: cs.CV
TL;DR: pFedNavi: A personalized federated learning framework for Vision-Language Navigation that addresses privacy concerns and client heterogeneity through adaptive layer-wise personalization and fine-grained parameter fusion.
Details
Motivation: VLN requires large-scale trajectory instruction data from private indoor environments, raising privacy concerns. Federated Learning helps keep data on-device but struggles with extreme cross-client heterogeneity in environments and instruction styles, making a single global model suboptimal.Method: Proposes pFedNavi, a structure-aware and dynamically adaptive personalized federated learning framework that adaptively identifies client-specific layers via layer-wise mixing coefficients and performs fine-grained parameter fusion on selected components (encoder-decoder projection and environment-sensitive decoder layers).
Result: Outperforms FedAvg-based VLN baseline on R2R and RxR benchmarks using both ResNet and CLIP visual representations, achieving up to 7.5% improvement in navigation success rate, up to 7.8% gain in trajectory fidelity, and converging 1.38x faster under non-IID conditions.
Conclusion: pFedNavi effectively balances global knowledge sharing with local specialization for VLN tasks, addressing privacy concerns while handling extreme client heterogeneity through personalized federated learning.
Abstract: Vision-Language Navigation VLN requires large-scale trajectory instruction data from private indoor environments, raising significant privacy concerns. Federated Learning FL mitigates this by keeping data on-device, but vanilla FL struggles under VLNs’ extreme cross-client heterogeneity in environments and instruction styles, making a single global model suboptimal. This paper proposes pFedNavi, a structure-aware and dynamically adaptive personalized federated learning framework tailored for VLN. Our key idea is to personalize where it matters: pFedNavi adaptively identifies client-specific layers via layer-wise mixing coefficients, and performs fine-grained parameter fusion on the selected components (e.g., the encoder-decoder projection and environment-sensitive decoder layers) to balance global knowledge sharing with local specialization. We evaluate pFedNavi on two standard VLN benchmarks, R2R and RxR, using both ResNet and CLIP visual representations. Across all metrics, pFedNavi consistently outperforms the FedAvg-based VLN baseline, achieving up to 7.5% improvement in navigation success rate and up to 7.8% gain in trajectory fidelity, while converging 1.38x faster under non-IID conditions.
[255] Feature Recalibration Based Olfactory-Visual Multimodal Model for Fine-Grained Rice Deterioration Detection
Rongqiang Zhao, Hengrui Hu, Yijing Wang, Mingchun Sun, Jie Liu
Main category: cs.CV
TL;DR: A multimodal olfactory-visual model for fine-grained rice deterioration detection using feature recalibration to enhance sensitivity to subtle surface abnormalities.
Details
Motivation: Current multimodal methods for rice deterioration detection have limited capability in representing fine-grained abnormal features and rely on expensive devices like hyperspectral cameras and mass spectrometers, increasing costs and data acquisition time.Method: Proposes a feature recalibration based olfactory-visual multimodal model with two components: 1) Fine-grained deterioration embedding constructor (FDEC) to reconstruct labeled multimodal embedded-feature dataset, and 2) Fine-grained deterioration recalibration attention network (FDRA-Net) to emphasize signal variations and increase sensitivity to fine-grained deterioration on rice surface.
Result: Achieves 99.89% classification accuracy, outperforming state-of-the-art methods while simplifying the detection procedure. Field detection demonstrates advantages in accuracy and operational simplicity.
Conclusion: The proposed method effectively addresses limitations of existing multimodal approaches for rice deterioration detection and can be extended to other agrifood applications in agriculture and food industry.
Abstract: Multimodal methods are widely used in rice deterioration detection, which exhibit limited capability in representing and extracting fine-grained abnormal features. Moreover, these methods rely on devices, such as hyperspectral cameras and mass spectrometers, increasing detection costs and prolonging data acquisition time. To address these issues, we propose a feature recalibration based olfactory-visual multimodal model for fine-grained rice deterioration detection. The fine-grained deterioration embedding constructor (FDEC) is proposed to reconstruct the labeled multimodal embedded-feature dataset, enhancing sample representation. The fine-grained deterioration recalibration attention network (FDRA-Net) is proposed to emphasize signal variations and increase sensitivity to fine-grained deterioration on the rice surface. Experiments show that the proposed method achieves a classification accuracy of 99.89%. Compared with state-of-the-art methods, the detection accuracy is improved and the procedure is simplified. Furthermore, field detection demonstrates the advantages of accuracy and operational simplicity. The proposed method can also be extended to other agrifood in agriculture and food industry.
[256] Learning Proposes, Geometry Disposes: A Modular Framework for Efficient Spatial Reasoning
Haichao Zhu, Zhaorui Yang, Qian Zhang
Main category: cs.CV
TL;DR: Learning-based methods augment but don’t replace geometric algorithms in spatial perception; a modular framework with learning proposing geometric hypotheses and geometry algorithms making final decisions shows best results for camera pose estimation.
Details
Motivation: To investigate whether learning components should directly replace geometric estimation or serve as intermediate modules in spatial perception pipelines, addressing the open question of how to effectively integrate learning-based methods with traditional geometric approaches.Method: Proposes an end-to-end modular framework where learning proposes geometric hypotheses (pose and depth) and geometric algorithms make final estimation decisions. Evaluated using VGGT as learning model for pose/depth proposals on RGB-D sequences, followed by classical point-to-plane RGB-D ICP as geometric backend on TUM RGB-D benchmark.
Result: Three key findings: (1) learning-based pose proposals alone are unreliable; (2) learning-proposed geometry can degrade performance when improperly aligned with camera intrinsics; (3) when learning-proposed depth is geometrically aligned and followed by geometric disposal stage, consistent improvements emerge in moderately challenging rigid settings.
Conclusion: Geometry is not merely a refinement component but an essential arbiter that validates and absorbs learning-based geometric observations. Modular, geometry-aware system design is crucial for robust spatial perception.
Abstract: Spatial perception aims to estimate camera motion and scene structure from visual observations, a problem traditionally addressed through geometric modeling and physical consistency constraints. Recent learning-based methods have demonstrated strong representational capacity for geometric perception and are increasingly used to augment classical geometry-centric systems in practice. However, whether learning components should directly replace geometric estimation or instead serve as intermediate modules within such pipelines remains an open question. In this work, we address this gap and investigate an end-to-end modular framework for effective spatial reasoning, where learning proposes geometric hypotheses, while geometric algorithms dispose estimation decisions. In particular, we study this principle in the context of relative camera pose estimation on RGB-D sequences. Using VGGT as a representative learning model, we evaluate learning-based pose and depth proposals under varying motion magnitudes and scene dynamics, followed by a classical point-to-plane RGB-D ICP as the geometric backend. Our experiments on the TUM RGB-D benchmark reveal three consistent findings: (1) learning-based pose proposals alone are unreliable; (2) learning-proposed geometry, when improperly aligned with camera intrinsics, can degrade performance; and (3) when learning-proposed depth is geometrically aligned and followed by a geometric disposal stage, consistent improvements emerge in moderately challenging rigid settings. These results demonstrate that geometry is not merely a refinement component, but an essential arbiter that validates and absorbs learning-based geometric observations. Our study highlights the importance of modular, geometry-aware system design for robust spatial perception.
[257] Understanding Sensor Vulnerabilities in Industrial XR Tracking
Sourya Saha, Md. Nurul Absur
Main category: cs.CV
TL;DR: VIO systems in XR show asymmetric fault tolerance: visual degradation causes centimeter-level errors, while inertial degradation can cause massive trajectory deviations up to kilometers.
Details
Motivation: XR systems in industrial settings rely on VIO for pose tracking, but real-world conditions often involve sensor degradation that deviates from ideal assumptions. Most VIO evaluations focus on nominal conditions, leaving insufficient understanding of sustained sensor degradation effects in operational environments.Method: Controlled empirical study using systematic fault injection to examine faults affecting visual and inertial modalities across various operating regimes, with quantitative evaluation of VIO behavior under degraded sensing conditions.
Result: Pronounced asymmetry in fault impact: visual sensing degradations lead to bounded pose errors (~centimeters), while inertial sensing degradations cause substantially larger trajectory deviations (hundreds to thousands of meters).
Conclusion: Greater emphasis on inertial reliability is needed in the evaluation and design of XR systems for real-life industrial settings, as inertial sensor degradation poses much greater risk than visual degradation.
Abstract: Extended Reality (XR) systems deployed in industrial and operational settings rely on Visual–Inertial Odometry (VIO) for continuous six-degree-of-freedom pose tracking, yet these environments often involve sensing conditions that deviate from ideal assumptions. Despite this, most VIO evaluations emphasize nominal sensor behavior, leaving the effects of sustained sensor degradation under operational conditions insufficiently understood. This paper presents a controlled empirical study of VIO behavior under degraded sensing, examining faults affecting visual and inertial modalities across a range of operating regimes. Through systematic fault injection and quantitative evaluation, we observe a pronounced asymmetry in fault impact where degradations affecting visual sensing typically lead to bounded pose errors on the order of centimeters, whereas degradations affecting inertial sensing can induce substantially larger trajectory deviations, in some cases reaching hundreds to thousands of meters. These observations motivate greater emphasis on inertial reliability in the evaluation and design of XR systems for real-life industrial settings.
[258] Hierarchical Vision-Language Interaction for Facial Action Unit Detection
Yong Li, Yi Ren, Yizhe Zhang, Wenhua Zhang, Tianyi Zhang, Muyun Jiang, Guo-Sen Xie, Cuntai Guan
Main category: cs.CV
TL;DR: HiVA: Hierarchical Vision-language Interaction method for Facial Action Unit detection using textual AU descriptions as semantic priors to guide AU representation learning.
Details
Motivation: AU detection faces challenges in learning discriminative and generalizable representations with limited annotated data. The paper aims to leverage textual AU descriptions as semantic priors to enhance AU detection capabilities.Method: Uses LLM to generate diverse AU descriptions, introduces AU-aware dynamic graph module for AU-specific visual representations, and hierarchical cross-modal attention with Disentangled Dual Cross-Attention (fine-grained AU-specific interactions) and Contextual Dual Cross-Attention (global inter-AU dependencies).
Result: HiVA consistently surpasses state-of-the-art approaches in AU detection. Qualitative analyses show semantically meaningful activation patterns and robust cross-modal correspondences.
Conclusion: HiVA effectively leverages vision-language interaction for comprehensive facial behavior analysis, producing robust and interpretable AU detection with enhanced semantic understanding.
Abstract: Facial Action Unit (AU) detection seeks to recognize subtle facial muscle activations as defined by the Facial Action Coding System (FACS). A primary challenge w.r.t AU detection is the effective learning of discriminative and generalizable AU representations under conditions of limited annotated data. To address this, we propose a Hierarchical Vision-language Interaction for AU Understanding (HiVA) method, which leverages textual AU descriptions as semantic priors to guide and enhance AU detection. Specifically, HiVA employs a large language model to generate diverse and contextually rich AU descriptions to strengthen language-based representation learning. To capture both fine-grained and holistic vision-language associations, HiVA introduces an AU-aware dynamic graph module that facilitates the learning of AU-specific visual representations. These features are further integrated within a hierarchical cross-modal attention architecture comprising two complementary mechanisms: Disentangled Dual Cross-Attention (DDCA), which establishes fine-grained, AU-specific interactions between visual and textual features, and Contextual Dual Cross-Attention (CDCA), which models global inter-AU dependencies. This collaborative, cross-modal learning paradigm enables HiVA to leverage multi-grained vision-based AU features in conjunction with refined language-based AU details, culminating in robust and semantically enriched AU detection capabilities. Extensive experiments show that HiVA consistently surpasses state-of-the-art approaches. Besides, qualitative analyses reveal that HiVA produces semantically meaningful activation patterns, highlighting its efficacy in learning robust and interpretable cross-modal correspondences for comprehensive facial behavior analysis.
[259] D-SECURE: Dual-Source Evidence Combination for Unified Reasoning in Misinformation Detection
Gagandeep Singh, Samudi Amarasinghe, Priyanka Singh
Main category: cs.CV
TL;DR: D-SECURE is a multimodal misinformation detection framework that combines internal manipulation detection (HAMMER) with external evidence-based reasoning (DEFAME) to identify both fine-grained visual/textual edits and factual inaccuracies in news-style posts.
Details
Motivation: Current misinformation detection systems have limitations: content-based detectors only find local inconsistencies but can't determine global factual truth, while retrieval-based fact-checkers treat inputs as coarse claims and miss subtle visual/textual manipulations. This separation allows internally consistent fabrications to bypass manipulation detectors and fact-checkers to verify claims with pixel-level or token-level corruption.Method: D-SECURE integrates HAMMER (manipulation detector) with DEFAME (retrieval pipeline). DEFAME performs broad verification first, then HAMMER analyzes residual or uncertain cases that may contain fine-grained edits. The framework combines internal manipulation detection with external evidence-based reasoning.
Result: Experiments on DGM4 and ClaimReview samples demonstrate the complementary strengths of both systems and motivate their fusion. The framework provides unified, explainable reports incorporating both manipulation cues and external evidence.
Conclusion: Combining internal manipulation detection with external evidence-based reasoning creates a more robust multimodal misinformation detection system that can identify both factual inaccuracies and subtle visual/textual manipulations in news-style posts.
Abstract: Multimodal misinformation increasingly mixes realistic im-age edits with fluent but misleading text, producing persuasive posts that are difficult to verify. Existing systems usually rely on a single evidence source. Content-based detectors identify local inconsistencies within an image and its caption but cannot determine global factual truth. Retrieval-based fact-checkers reason over external evidence but treat inputs as coarse claims and often miss subtle visual or textual manipulations. This separation creates failure cases where internally consistent fabrications bypass manipulation detectors and fact-checkers verify claims that contain pixel-level or token-level corruption. We present D-SECURE, a framework that combines internal manipulation detection with external evidence-based reasoning for news-style posts. D-SECURE integrates the HAMMER manipulation detector with the DEFAME retrieval pipeline. DEFAME performs broad verification, and HAMMER analyses residual or uncertain cases that may contain fine-grained edits. Experiments on DGM4 and ClaimReview samples highlight the complementary strengths of both systems and motivate their fusion. We provide a unified, explainable report that incorporates manipulation cues and external evidence.
[260] Controlling Your Image via Simplified Vector Graphics
Lanqing Guo, Xi Liu, Yufei Wang, Zhihao Li, Siyu Huang
Main category: cs.CV
TL;DR: Vec2Pix introduces layer-wise controllable image generation using simplified vector graphics representations, enabling intuitive element-level editing like shape adjustment, color alteration, and object manipulation.
Details
Motivation: Current image generation lacks precise element-level control for intuitive modifications like adjusting shapes, altering colors, or adding/removing objects. The paper aims to address this fundamental challenge in controllable image generation.Method: The approach first parses images into hierarchical vector graphics representations that are semantic-aligned and structurally coherent. It then designs an image synthesis framework guided by these VGs, leveraging their structural and semantic features with noise prediction for precise control over geometry, color, and object semantics.
Result: Extensive experiments demonstrate effectiveness in diverse applications including image editing, object-level manipulation, and fine-grained content creation, establishing a new paradigm for controllable image generation.
Conclusion: The work introduces a novel approach to layer-wise controllable generation through vector graphics, enabling free modification of elements and seamless translation into photorealistic outputs with precise control over various visual attributes.
Abstract: Recent advances in image generation have achieved remarkable visual quality, while a fundamental challenge remains: Can image generation be controlled at the element level, enabling intuitive modifications such as adjusting shapes, altering colors, or adding and removing objects? In this work, we address this challenge by introducing layer-wise controllable generation through simplified vector graphics (VGs). Our approach first efficiently parses images into hierarchical VG representations that are semantic-aligned and structurally coherent. Building on this representation, we design a novel image synthesis framework guided by VGs, allowing users to freely modify elements and seamlessly translate these edits into photorealistic outputs. By leveraging the structural and semantic features of VGs in conjunction with noise prediction, our method provides precise control over geometry, color, and object semantics. Extensive experiments demonstrate the effectiveness of our approach in diverse applications, including image editing, object-level manipulation, and fine-grained content creation, establishing a new paradigm for controllable image generation. Project page: https://guolanqing.github.io/Vec2Pix/
[261] CoCoDiff: Correspondence-Consistent Diffusion Model for Fine-grained Style Transfer
Wenbo Nie, Zixiang Li, Renshuai Tao, Bin Wu, Yunchao Wei, Yao Zhao
Main category: cs.CV
TL;DR: CoCoDiff: A training-free style transfer framework using pretrained latent diffusion models to achieve fine-grained, semantically consistent stylization by mining pixel-wise semantic correspondence from diffusion features.
Details
Motivation: Existing style transfer methods operate at global level but overlook region-wise and pixel-wise semantic correspondence, failing to preserve semantic alignment between similar objects when transferring visual style between images.Method: Leverages pretrained latent diffusion models without additional training. Introduces pixel-wise semantic correspondence module that mines intermediate diffusion features to construct dense alignment maps between content and style images, plus a cycle-consistency module to enforce structural and perceptual alignment across iterations.
Result: Delivers state-of-the-art visual quality and strong quantitative results despite requiring no additional training or supervision, outperforming methods that rely on extra training or annotations.
Conclusion: CoCoDiff demonstrates that correspondence cues within generative diffusion models are under-explored and can be effectively leveraged for fine-grained, semantically consistent style transfer without additional training.
Abstract: Transferring visual style between images while preserving semantic correspondence between similar objects remains a central challenge in computer vision. While existing methods have made great strides, most of them operate at global level but overlook region-wise and even pixel-wise semantic correspondence. To address this, we propose CoCoDiff, a novel training-free and low-cost style transfer framework that leverages pretrained latent diffusion models to achieve fine-grained, semantically consistent stylization. We identify that correspondence cues within generative diffusion models are under-explored and that content consistency across semantically matched regions is often neglected. CoCoDiff introduces a pixel-wise semantic correspondence module that mines intermediate diffusion features to construct a dense alignment map between content and style images. Furthermore, a cycle-consistency module then enforces structural and perceptual alignment across iterations, yielding object and region level stylization that preserves geometry and detail. Despite requiring no additional training or supervision, CoCoDiff delivers state-of-the-art visual quality and strong quantitative results, outperforming methods that rely on extra training or annotations.
[262] TikArt: Aperture-Guided Observation for Fine-Grained Visual Reasoning via Reinforcement Learning
Hao Ding, Zhichuan Yang, Weijie Ge, Ziqin Gao, Chaoyi Lu, Lei Zhao
Main category: cs.CV
TL;DR: TikArt is an aperture-guided agent for fine-grained visual reasoning in MLLMs that uses a Think-Aperture-Observe loop with zoom and segmentation actions to focus on small/cluttered image regions, optimized via reinforcement learning.
Details
Motivation: Current MLLMs lose fine-grained visual details in global image encodings, failing at tasks requiring attention to tiny objects, cluttered regions, or subtle markings. There's a need for models that can dynamically focus on relevant image regions during multi-step reasoning.Method: TikArt uses a Think-Aperture-Observe loop: alternating language generation with aperture actions (Zoom for rectangular crops, Segment using SAM2 for mask-based crops). After each action, explicit observations convert visual cues to linguistic memory. Built on Qwen3-VL-8B, optimized with AGRPO (GRPO-style RL) using two-stage curriculum: warm-up segmentation then joint optimization of visual math, VQA, and segmentation tasks.
Result: Experiments on V*, HR-Bench-4K/8K, MME-RealWorld-Lite, MMStar, RefCOCO, and ReasonSeg show consistent gains over backbone model. Produces interpretable aperture trajectories for high-resolution reasoning tasks.
Conclusion: TikArt effectively addresses fine-grained visual reasoning by enabling dynamic region-of-interest selection through aperture actions, converting visual details to persistent linguistic memory, and demonstrating improved performance across diverse visual reasoning benchmarks.
Abstract: We address fine-grained visual reasoning in multimodal large language models (MLLMs), where key evidence may reside in tiny objects, cluttered regions, or subtle markings that are lost under a single global image encoding. We introduce TikArt (Thinking Aperture), an aperture-guided agent that casts multi-step vision-language reasoning as a decision process over regions of interest. TikArt follows a Think-Aperture-Observe loop, alternating between language generation and two aperture actions: Zoom extracts rectangular crops, while Segment invokes SAM2 to obtain mask-based crops for irregular targets. After every action, the model must produce an explicit observation, turning local visual cues into persistent linguistic memory. Built on Qwen3-VL-8B, TikArt optimizes its reasoning policy with AGRPO, a GRPO-style reinforcement learning algorithm with a two-stage curriculum: it warms up segmentation actions and then jointly optimizes visual math, fine-grained VQA, and segmentation, using rewards that couple task success with purposeful aperture use. Experiments on V*, HR-Bench-4K/8K, MME-RealWorld-Lite, MMStar, RefCOCO, and ReasonSeg show consistent gains over the backbone and yield interpretable aperture trajectories for high-resolution reasoning.
[263] Gaussian Mesh Renderer for Lightweight Differentiable Rendering
Xinpeng Liu, Fumio Okura
Main category: cs.CV
TL;DR: A lightweight differentiable mesh renderer called Gaussian Mesh Renderer (GMR) that integrates Gaussian and mesh representations by deriving Gaussian primitives from mesh triangles, enabling efficient optimization with smooth gradients.
Details
Motivation: Triangle mesh models remain popular for surface reconstruction but suffer from slow/heavy optimization in traditional mesh-based differentiable renderers, while 3D Gaussian Splatting offers fast rendering but lacks mesh structure.Method: Proposes Gaussian Mesh Renderer (GMR) that leverages 3DGS rasterization by analytically deriving Gaussian primitives from mesh triangles, preserving structural fidelity and enabling gradient flow through the mesh representation.
Result: Achieves smoother gradients compared to traditional mesh renderers, enabling better optimization with smaller batch sizes and limited memory, while maintaining mesh structure fidelity.
Conclusion: GMR provides an efficient differentiable mesh renderer that combines the benefits of both Gaussian splatting (fast rendering) and mesh representations (structural fidelity), enabling better optimization for surface reconstruction tasks.
Abstract: 3D Gaussian Splatting (3DGS) has enabled high-fidelity virtualization with fast rendering and optimization for novel view synthesis. On the other hand, triangle mesh models still remain a popular choice for surface reconstruction but suffer from slow or heavy optimization in traditional mesh-based differentiable renderers. To address this problem, we propose a new lightweight differentiable mesh renderer leveraging the efficient rasterization process of 3DGS, named Gaussian Mesh Renderer (GMR), which tightly integrates the Gaussian and mesh representations. Each Gaussian primitive is analytically derived from the corresponding mesh triangle, preserving structural fidelity and enabling the gradient flow. Compared to the traditional mesh renderers, our method achieves smoother gradients, which especially contributes to better optimization using smaller batch sizes with limited memory. Our implementation is available in the public GitHub repository at https://github.com/huntorochi/Gaussian-Mesh-Renderer.
[264] Uncertainty-Aware Vision-Language Segmentation for Medical Imaging
Aryan Das, Tanishq Rachamalla, Koushik Biswas, Swalpa Kumar Roy, Vinay Kumar Verma
Main category: cs.CV
TL;DR: A multimodal medical segmentation framework combining radiological images and clinical text with uncertainty-aware learning for improved reliability and efficiency.
Details
Motivation: Medical diagnosis often involves ambiguous cases with poor image quality where current segmentation methods lack reliability. There's a need for uncertainty-aware multimodal approaches that can leverage both visual and textual clinical information to improve segmentation accuracy in complex medical scenarios.Method: Proposes Modality Decoding Attention Block (MoDAB) with lightweight State Space Mixer (SSMix) for efficient cross-modal fusion and long-range dependency modeling. Introduces Spectral-Entropic Uncertainty (SEU) Loss to jointly capture spatial overlap, spectral consistency, and predictive uncertainty in a unified objective.
Result: Superior segmentation performance on medical datasets (QATA-COVID19, MosMed++, Kvasir-SEG) while being significantly more computationally efficient than existing SoTA approaches. Demonstrates improved model reliability in complex clinical circumstances with poor image quality.
Conclusion: Highlights the importance of incorporating uncertainty modeling and structured modality alignment in vision-language medical segmentation tasks, showing that multimodal approaches with uncertainty awareness can significantly improve medical diagnostic reliability.
Abstract: We introduce a novel uncertainty-aware multimodal segmentation framework that leverages both radiological images and associated clinical text for precise medical diagnosis. We propose a Modality Decoding Attention Block (MoDAB) with a lightweight State Space Mixer (SSMix) to enable efficient cross-modal fusion and long-range dependency modelling. To guide learning under ambiguity, we propose the Spectral-Entropic Uncertainty (SEU) Loss, which jointly captures spatial overlap, spectral consistency, and predictive uncertainty in a unified objective. In complex clinical circumstances with poor image quality, this formulation improves model reliability. Extensive experiments on various publicly available medical datasets, QATA-COVID19, MosMed++, and Kvasir-SEG, demonstrate that our method achieves superior segmentation performance while being significantly more computationally efficient than existing State-of-the-Art (SoTA) approaches. Our results highlight the importance of incorporating uncertainty modelling and structured modality alignment in vision-language medical segmentation tasks. Code: https://github.com/arya-domain/UA-VLS
[265] Prototype Instance-semantic Disentanglement with Low-rank Regularized Subspace Clustering for WSIs Explainable Recognition
Chentao Li, Pan Huang
Main category: cs.CV
TL;DR: PID-LRSC: A prototype instance semantic disentanglement framework with low-rank regularized subspace clustering for pathological diagnosis in whole slide images, addressing instance-semantic entanglement in tumor detection.
Details
Motivation: Tumor tissues are highly similar to precancerous lesions, and non-tumor instances greatly outnumber tumor instances in whole slide images, causing instance-semantic entanglement in multi-instance learning frameworks that degrades model representation and interpretability.Method: Two-part approach: 1) Secondary instance subspace learning with low-rank regularized subspace clustering (LRSC) to address instance entanglement from excessive non-tumor instances; 2) Enhanced contrastive learning for prototype instance semantic disentanglement (PID) to resolve semantic entanglement between tumor and precancerous tissues.
Result: Extensive experiments on multicentre pathology datasets show PID-LRSC outperforms other state-of-the-art methods, providing clearer instance semantics during decision-making and significantly enhancing reliability of auxiliary diagnostic outcomes.
Conclusion: PID-LRSC effectively addresses instance-semantic entanglement in pathological diagnosis, improving both model performance and interpretability for tumor detection in whole slide images.
Abstract: The tumor region plays a key role in pathological diagnosis. Tumor tissues are highly similar to precancerous lesions and non tumor instances often greatly exceed tumor instances in whole slide images (WSIs). These issues cause instance-semantic entanglement in multi-instance learning frameworks, degrading both model representation capability and interpretability. To address this, we propose an end-to-end prototype instance semantic disentanglement framework with low-rank regularized subspace clustering, PID-LRSC, in two aspects. First, we use secondary instance subspace learning to construct low-rank regularized subspace clustering (LRSC), addressing instance entanglement caused by an excessive proportion of non tumor instances. Second, we employ enhanced contrastive learning to design prototype instance semantic disentanglement (PID), resolving semantic entanglement caused by the high similarity between tumor and precancerous tissues. We conduct extensive experiments on multicentre pathology datasets, implying that PID-LRSC outperforms other SOTA methods. Overall, PID-LRSC provides clearer instance semantics during decision-making and significantly enhances the reliability of auxiliary diagnostic outcomes.
[266] MacNet: An End-to-End Manifold-Constrained Adaptive Clustering Network for Interpretable Whole Slide Image Classification
Mingrui Ma, Chentao Li, Pan Huang, Jing Qin
Main category: cs.CV
TL;DR: End-to-end MIL framework for whole slide image analysis integrating Grassmann re-embedding and manifold adaptive clustering for improved grading accuracy and interpretability.
Details
Motivation: Current WSI analysis methods have limitations: attention-based MIL offers limited interpretability, while clustering-based approaches suffer from high-dimensional features and ambiguous centroids. Need for better interpretable decision-making in pathological diagnosis.Method: Proposes end-to-end MIL framework with Grassmann re-embedding and manifold adaptive clustering. Uses manifold geometric structure for robust clustering. Includes prior knowledge guiding proxy instance labeling and aggregation strategy to approximate patch labels and focus on tumor regions.
Result: Experiments on multicentre WSI datasets show: 1) cluster-incorporated model achieves superior performance in both grading accuracy and interpretability; 2) end-to-end learning refines feature representations with acceptable computation resources.
Conclusion: The proposed framework improves both accuracy and interpretability for WSI analysis through integrated clustering and end-to-end learning, addressing limitations of existing methods.
Abstract: Whole slide images (WSIs) are the gold standard for pathological diagnosis and sub-typing. Current main-stream two-step frameworks employ offline feature encoders trained without domain-specific knowledge. Among them, attention-based multiple instance learning (MIL) methods are outcome-oriented and offer limited interpretability. Clustering-based approaches can provide explainable decision-making process but suffer from high dimension features and semantically ambiguous centroids. To this end, we propose an end-to-end MIL framework that integrates Grassmann re-embedding and manifold adaptive clustering, where the manifold geometric structure facilitates robust clustering results. Furthermore, we design a prior knowledge guiding proxy instance labeling and aggregation strategy to approximate patch labels and focus on pathologically relevant tumor regions. Experiments on multicentre WSI datasets demonstrate that: 1) our cluster-incorporated model achieves superior performance in both grading accuracy and interpretability; 2) end-to-end learning refines better feature representations and it requires acceptable computation resources.
[267] MedVAR: Towards Scalable and Efficient Medical Image Generation via Next-scale Autoregressive Prediction
Zhicheng He, Yunpeng Zhao, Junde Wu, Ziwei Niu, Zijun Li, Lanfen Lin, Yueming Jin
Main category: cs.CV
TL;DR: MedVAR is an autoregressive foundation model for medical image generation using next-scale prediction for scalable, hierarchical synthesis of CT and MRI images across multiple anatomical regions.
Details
Motivation: Medical image generation needs scalable backbones for data augmentation and privacy-preserving sharing, but current approaches lack architectural efficiency, sufficient multi-organ data, and principled evaluation.Method: Autoregressive-based foundation model using next-scale prediction paradigm for fast, scale-up-friendly synthesis; generates images coarse-to-fine with structured multi-scale representations; trained on harmonized dataset of ~440,000 CT/MRI images across six anatomical regions.
Result: Achieves state-of-the-art generative performance across fidelity, diversity, and scalability metrics; offers promising architectural direction for future medical generative foundation models.
Conclusion: MedVAR successfully addresses key challenges in medical image generation through its autoregressive architecture and hierarchical approach, establishing a foundation for scalable medical imaging models.
Abstract: Medical image generation is pivotal in applications like data augmentation for low-resource clinical tasks and privacy-preserving data sharing. However, developing a scalable generative backbone for medical imaging requires architectural efficiency, sufficient multi-organ data, and principled evaluation, yet current approaches leave these aspects unresolved. Therefore, we introduce MedVAR, the first autoregressive-based foundation model that adopts the next-scale prediction paradigm to enable fast and scale-up-friendly medical image synthesis. MedVAR generates images in a coarse-to-fine manner and produces structured multi-scale representations suitable for downstream use. To support hierarchical generation, we curate a harmonized dataset of around 440,000 CT and MRI images spanning six anatomical regions. Comprehensive experiments across fidelity, diversity, and scalability show that MedVAR achieves state-of-the-art generative performance and offers a promising architectural direction for future medical generative foundation models.
[268] Efficient Text-Guided Convolutional Adapter for the Diffusion Model
Aryan Das, Koushik Biswas, Swalpa Kumar Roy, Badri Narayana Patro, Vinay Kumar Verma
Main category: cs.CV
TL;DR: Nexus Adapters are efficient text-guided adapters for diffusion models that enable structure-preserving conditional image generation with better prompt understanding and fewer parameters.
Details
Motivation: Existing structure-preserving methods for conditional image generation are inefficient, requiring nearly equal parameters in the adapter as the base diffusion model, and lack prompt awareness, making them suboptimal for text-guided generation.Method: Proposed two efficient adapters (Nexus Prime and Nexus Slim) that incorporate cross-attention mechanisms to enable rich multimodal conditioning, allowing the adapter to understand input prompts while preserving structural inputs like sketches or depth maps.
Result: Nexus Prime requires only 8M additional parameters compared to T2I-Adapter baseline while significantly enhancing performance. Nexus Slim has 18M fewer parameters than T2I-Adapter and still achieves state-of-the-art results.
Conclusion: Nexus Adapters provide an efficient solution for structure-preserving conditional generation that is both prompt-aware and parameter-efficient, overcoming limitations of previous methods.
Abstract: We introduce the Nexus Adapters, novel text-guided efficient adapters to the diffusion-based framework for the Structure Preserving Conditional Generation (SPCG). Recently, structure-preserving methods have achieved promising results in conditional image generation by using a base model for prompt conditioning and an adapter for structure input, such as sketches or depth maps. These approaches are highly inefficient and sometimes require equal parameters in the adapter compared to the base architecture. It is not always possible to train the model since the diffusion model is itself costly, and doubling the parameter is highly inefficient. In these approaches, the adapter is not aware of the input prompt; therefore, it is optimal only for the structural input but not for the input prompt. To overcome the above challenges, we proposed two efficient adapters, Nexus Prime and Slim, which are guided by prompts and structural inputs. Each Nexus Block incorporates cross-attention mechanisms to enable rich multimodal conditioning. Therefore, the proposed adapter has a better understanding of the input prompt while preserving the structure. We conducted extensive experiments on the proposed models and demonstrated that the Nexus Prime adapter significantly enhances performance, requiring only 8M additional parameters compared to the baseline, T2I-Adapter. Furthermore, we also introduced a lightweight Nexus Slim adapter with 18M fewer parameters than the T2I-Adapter, which still achieved state-of-the-art results. Code: https://github.com/arya-domain/Nexus-Adapters
[269] Architectural Insights for Post-Tornado Damage Recognition
Robinson Umeike, Thang Dao, Shane Crawford, John van de Lindt, Blythe Johnston, Wanting, Wang, Trung Do, Ajibola Mofikoya, Sarbesh Banjara, Cuong Pham
Main category: cs.CV
TL;DR: Systematic evaluation of 79 deep learning models for tornado damage assessment reveals optimizer choice (SGD over Adam) and low learning rate are more critical than architecture selection for handling domain shift and class imbalance in disaster imagery.
Details
Motivation: Rapid building damage assessment after tornadoes is critical for emergency response, but current automated methods struggle with visual complexity, domain shift from pre-training datasets, and extreme class imbalance in real disaster data.Method: Evaluated 79 open-source deep learning models (CNNs and Vision Transformers) across 2,300+ experiments on newly curated Quad-State Tornado Damage benchmark dataset, systematically testing architecture-optimizer interactions.
Result: Optimizer choice proved more consequential than architecture - switching from Adam to SGD gave +25 to +38 F1 points for Vision Transformers, reversing their ranking. Low learning rate (1e-4) boosted average F1 by +10.2 points. ConvNeXt-Base with optimized settings achieved 46.4% Macro F1 (+34.6 over baseline) with strong cross-event generalization.
Conclusion: Operational-grade tornado damage assessment requires optimizing architecture-optimizer interactions, not just architecture selection. SGD optimizer and low learning rate are critical for handling domain shift and class imbalance in disaster imagery.
Abstract: Rapid and accurate building damage assessment in the immediate aftermath of tornadoes is critical for coordinating life-saving search and rescue operations, optimizing emergency resource allocation, and accelerating community recovery. However, current automated methods struggle with the unique visual complexity of tornado-induced wreckage, primarily due to severe domain shift from standard pre-training datasets and extreme class imbalance in real-world disaster data. To address these challenges, we introduce a systematic experimental framework evaluating 79 open-source deep learning models, encompassing both Convolutional Neural Networks (CNNs) and Vision Transformers, across over 2,300 controlled experiments on our newly curated Quad-State Tornado Damage (QSTD) benchmark dataset. Our findings reveal that achieving operational-grade performance hinges on a complex interaction between architecture and optimization, rather than architectural selection alone. Most strikingly, we demonstrate that optimizer choice can be more consequential than architecture: switching from Adam to SGD provided dramatic F1 gains of +25 to +38 points for Vision Transformer and Swin Transformer families, fundamentally reversing their ranking from bottom-tier to competitive with top-performing CNNs. Furthermore, a low learning rate of 1x10^(-4) proved universally critical, boosting average F1 performance by +10.2 points across all architectures. Our champion model, ConvNeXt-Base trained with these optimized settings, demonstrated strong cross-event generalization on the held-out Tuscaloosa-Moore Tornado Damage (TMTD) dataset, achieving 46.4% Macro F1 (+34.6 points over baseline) and retaining 85.5% Ordinal Top-1 Accuracy despite temporal and sensor domain shifts.
[270] Error Patterns in Historical OCR: A Comparative Analysis of TrOCR and a Vision-Language Model
Ari Vesalainen, Eetu Mäkelä, Laura Ruotsalainen, Mikko Tolonen
Main category: cs.CV
TL;DR: Comparison of transformer-based OCR (TrOCR) and general-purpose VLM (Qwen) on historical English texts, showing Qwen has lower error rates but alters historical forms, while TrOCR preserves orthography better but has error propagation issues.
Details
Motivation: OCR of 18th-century printed texts is challenging due to degraded quality, archaic glyphs, and non-standard orthography. Current metrics like CER/WER provide limited insight into reliability for scholarly use, necessitating deeper analysis of different model architectures.Method: Compare dedicated OCR transformer (TrOCR) and general-purpose Vision-Language Model (Qwen) on line-level historical English texts using length-weighted accuracy metrics and hypothesis-driven error analysis to understand systematic error patterns.
Result: Qwen achieves lower CER/WER and greater robustness to degraded input but exhibits selective linguistic regularization that silently alters historically meaningful forms. TrOCR preserves orthographic fidelity more consistently but is more prone to cascading error propagation.
Conclusion: Architectural inductive biases shape OCR error structure systematically. Models with similar aggregate accuracy differ substantially in error locality, detectability, and downstream scholarly risk, highlighting need for architecture-aware evaluation in historical digitization.
Abstract: Optical Character Recognition (OCR) of eighteenth-century printed texts remains challenging due to degraded print quality, archaic glyphs, and non-standardized orthography. Although transformer-based OCR systems and Vision-Language Models (VLMs) achieve strong aggregate accuracy, metrics such as Character Error Rate (CER) and Word Error Rate (WER) provide limited insight into their reliability for scholarly use. We compare a dedicated OCR transformer (TrOCR) and a general-purpose Vision-Language Model (Qwen) on line-level historical English texts using length-weighted accuracy metrics and hypothesis driven error analysis. While Qwen achieves lower CER/WER and greater robustness to degraded input, it exhibits selective linguistic regularization and orthographic normalization that may silently alter historically meaningful forms. TrOCR preserves orthographic fidelity more consistently but is more prone to cascading error propagation. Our findings show that architectural inductive biases shape OCR error structure in systematic ways. Models with similar aggregate accuracy can differ substantially in error locality, detectability, and downstream scholarly risk, underscoring the need for architecture-aware evaluation in historical digitization workflows.
[271] Cross-view Domain Generalization via Geometric Consistency for LiDAR Semantic Segmentation
Jindong Zhao, Yuan Gao, Yang Xia, Sheng Nie, Jun Yue, Weiwei Sun, Shaobo Xia
Main category: cs.CV
TL;DR: CVGC is a framework for cross-view domain generalization in LiDAR semantic segmentation that addresses viewpoint-dependent structural incompleteness and non-uniform point density through geometric augmentation and consistency enforcement.
Details
Motivation: Existing LiDAR semantic segmentation methods assume similar acquisition views and struggle in cross-view scenarios due to viewpoint-dependent structural incompleteness and non-uniform point density, limiting real-world applicability.Method: Proposes CVGC with two modules: 1) cross-view geometric augmentation that models viewpoint-induced variations in visibility and sampling density to generate multiple cross-view observations, and 2) geometric consistency that enforces consistent semantic and occupancy predictions across augmented point clouds.
Result: Extensive experiments on six public LiDAR datasets establish the first systematic evaluation of cross-view domain generalization for LiDAR semantic segmentation, showing CVGC consistently outperforms state-of-the-art methods when generalizing from single source to multiple target domains with heterogeneous acquisition viewpoints.
Conclusion: CVGC effectively addresses cross-view domain generalization challenges in LiDAR semantic segmentation by modeling viewpoint variations and enforcing geometric consistency, enabling better generalization to unseen domains with different acquisition viewpoints.
Abstract: Domain-generalized LiDAR semantic segmentation (LSS) seeks to train models on source-domain point clouds that generalize reliably to multiple unseen target domains, which is essential for real-world LiDAR applications. However, existing approaches assume similar acquisition views (e.g., vehicle-mounted) and struggle in cross-view scenarios, where observations differ substantially due to viewpoint-dependent structural incompleteness and non-uniform point density. Accordingly, we formulate cross-view domain generalization for LiDAR semantic segmentation and propose a novel framework, termed CVGC (Cross-View Geometric Consistency). Specifically, we introduce a cross-view geometric augmentation module that models viewpoint-induced variations in visibility and sampling density, generating multiple cross-view observations of the same scene. Subsequently, a geometric consistency module enforces consistent semantic and occupancy predictions across geometrically augmented point clouds of the same scene. Extensive experiments on six public LiDAR datasets establish the first systematic evaluation of cross-view domain generalization for LiDAR semantic segmentation, demonstrating that CVGC consistently outperforms state-of-the-art methods when generalizing from a single source domain to multiple target domains with heterogeneous acquisition viewpoints. The source code will be publicly available at https://github.com/KintomZi/CVGC-DG
[272] MoRL: Reinforced Reasoning for Unified Motion Understanding and Generation
Hongpeng Wang, Zeyu Zhang, Wenhao Li, Hao Tang
Main category: cs.CV
TL;DR: MoRL is a unified multimodal motion model for human motion understanding and generation, trained with supervised fine-tuning and reinforcement learning with verifiable rewards, featuring Chain-of-Motion reasoning and large-scale CoT datasets.
Details
Motivation: Human motion understanding and generation are crucial for vision and robotics but current approaches lack sufficient reasoning capability and test-time planning abilities.Method: Proposes MoRL with task-specific reward design combining semantic alignment and reasoning coherence for understanding, and physical plausibility and text-motion consistency for generation. Introduces Chain-of-Motion (CoM) for test-time reasoning and constructs two large-scale CoT datasets (MoUnd-CoT-140K and MoGen-CoT-140K).
Result: Experiments on HumanML3D and KIT-ML show significant gains over state-of-the-art baselines in both motion understanding and generation tasks.
Conclusion: MoRL advances human motion modeling by integrating reasoning capabilities with multimodal learning, providing a unified framework for both understanding and generation tasks.
Abstract: Human motion understanding and generation are crucial for vision and robotics but remain limited in reasoning capability and test-time planning. We propose MoRL, a unified multimodal motion model trained with supervised fine-tuning and reinforcement learning with verifiable rewards. Our task-specific reward design combines semantic alignment and reasoning coherence for understanding with physical plausibility and text-motion consistency for generation, improving both logical reasoning and perceptual realism. To further enhance inference, we introduce Chain-of-Motion (CoM), a test-time reasoning method that enables step-by-step planning and reflection. We also construct two large-scale CoT datasets, MoUnd-CoT-140K and MoGen-CoT-140K, to align motion sequences with reasoning traces and action descriptions. Experiments on HumanML3D and KIT-ML show that MoRL achieves significant gains over state-of-the-art baselines. Code: https://github.com/AIGeeksGroup/MoRL. Website: https://aigeeksgroup.github.io/MoRL.
[273] OmniVTON++: Training-Free Universal Virtual Try-On with Principal Pose Guidance
Zhaotong Yang, Yong Du, Shengfeng He, Yuhui Li, Xinzhe Li, Yangyang Xu, Junyu Dong, Jian Yang
Main category: cs.CV
TL;DR: OmniVTON++ is a training-free virtual try-on framework that achieves universal applicability across diverse scenarios without task-specific retraining, using structured garment morphing, pose guidance, and boundary stitching.
Details
Motivation: Existing virtual try-on approaches are typically optimized for specific data conditions, requiring retraining for different scenarios and limiting their generalization as a unified solution. The authors aim to create a training-free framework with universal applicability.Method: The framework coordinates three main components: 1) Structured Garment Morphing for correspondence-driven garment adaptation, 2) Principal Pose Guidance for step-wise structural regulation during diffusion sampling, and 3) Continuous Boundary Stitching for boundary-aware refinement. This forms a cohesive pipeline that operates without task-specific retraining.
Result: OmniVTON++ achieves state-of-the-art performance across diverse generalization settings including cross-dataset and cross-garment-type evaluations. It reliably operates across scenarios and diffusion backbones within a single formulation, supporting multi-garment, multi-human, and anime character virtual try-on.
Conclusion: The framework successfully addresses the intertwined challenges of garment alignment, human structural coherence, and boundary continuity, expanding the scope of virtual try-on applications while maintaining training-free universal applicability.
Abstract: Image-based Virtual Try-On (VTON) concerns the synthesis of realistic person imagery through garment re-rendering under human pose and body constraints. In practice, however, existing approaches are typically optimized for specific data conditions, making their deployment reliant on retraining and limiting their generalization as a unified solution. We present OmniVTON++, a training-free VTON framework designed for universal applicability. It addresses the intertwined challenges of garment alignment, human structural coherence, and boundary continuity by coordinating Structured Garment Morphing for correspondence-driven garment adaptation, Principal Pose Guidance for step-wise structural regulation during diffusion sampling, and Continuous Boundary Stitching for boundary-aware refinement, forming a cohesive pipeline without task-specific retraining. Experimental results demonstrate that OmniVTON++ achieves state-of-the-art performance across diverse generalization settings, including cross-dataset and cross-garment-type evaluations, while reliably operating across scenarios and diffusion backbones within a single formulation. In addition to single-garment, single-human cases, the framework supports multi-garment, multi-human, and anime character virtual try-on, expanding the scope of virtual try-on applications. The source code will be released to the public.
[274] DriveFine: Refining-Augmented Masked Diffusion VLA for Precise and Robust Driving
Chenxu Dang, Sining Ang, Yongkang Li, Haochen Tian, Jie Wang, Guang Li, Hangjun Ye, Jie Ma, Long Chen, Yan Wang
Main category: cs.CV
TL;DR: DriveFine is a masked diffusion Vision-Language-Action model for autonomous driving that combines flexible decoding with self-correction capabilities using a novel block-MoE architecture with refinement experts.
Details
Motivation: Current VLA models for autonomous driving have limitations: diffusion-based planners suffer from modality alignment issues, low training efficiency, and limited generalization, while token-based planners have cumulative causal errors and irreversible decoding. The two dominant paradigms have complementary strengths and weaknesses.Method: Proposes DriveFine with a novel plug-and-play block-MoE that seamlessly injects a refinement expert on top of the generation expert. Uses explicit expert selection during inference and gradient blocking during training to fully decouple experts. Also designs a hybrid reinforcement learning strategy to encourage effective exploration of refinement expert while maintaining training stability.
Result: Extensive experiments on NAVSIM v1, v2, and Navhard benchmarks demonstrate that DriveFine exhibits strong efficacy and robustness in autonomous driving tasks.
Conclusion: DriveFine successfully addresses limitations of existing VLA models for autonomous driving by combining the strengths of diffusion and token-based approaches through a novel block-MoE architecture with refinement capabilities.
Abstract: Vision-Language-Action (VLA) models for autonomous driving increasingly adopt generative planners trained with imitation learning followed by reinforcement learning. Diffusion-based planners suffer from modality alignment difficulties, low training efficiency, and limited generalization. Token-based planners are plagued by cumulative causal errors and irreversible decoding. In summary, the two dominant paradigms exhibit complementary strengths and weaknesses. In this paper, we propose DriveFine, a masked diffusion VLA model that combines flexible decoding with self-correction capabilities. In particular, we design a novel plug-and-play block-MoE, which seamlessly injects a refinement expert on top of the generation expert. By enabling explicit expert selection during inference and gradient blocking during training, the two experts are fully decoupled, preserving the foundational capabilities and generic patterns of the pretrained weights, which highlights the flexibility and extensibility of the block-MoE design. Furthermore, we design a hybrid reinforcement learning strategy that encourages effective exploration of refinement expert while maintaining training stability. Extensive experiments on NAVSIM v1, v2, and Navhard benchmarks demonstrate that DriveFine exhibits strong efficacy and robustness. The code will be released at https://github.com/MSunDYY/DriveFine.
[275] YOLO26: A Comprehensive Architecture Overview and Key Improvements
Priyanto Hidayatullah, Refdinal Tubagus
Main category: cs.CV
TL;DR: YOLO26 introduces architectural improvements including elimination of DFL, end-to-end NMS-free inference, ProgLoss+STAL, and MuSGD optimizer to achieve 43% faster CPU inference for edge devices, with enhanced capabilities for segmentation, pose estimation, and OBB decoding.
Details
Motivation: To provide the first comprehensive architectural analysis of YOLO26, the latest YOLO version, by examining source code to understand its novel improvements for real-time performance on edge devices without GPUs, and to maintain YOLO's dominance in computer vision.Method: Rigorous architectural investigation using YOLO26’s GitHub repository source code and official documentation to extract authentic operational mechanisms, resulting in creation of YOLO26’s CNN-based architectural diagram.
Result: First presentation of YOLO26’s CNN-based architecture, detailing key enhancements: elimination of DFL, end-to-end NMS-free inference, ProgLoss+STAL label assignment, MuSGD optimizer, achieving 43% faster CPU inference for edge deployment.
Conclusion: YOLO26 represents significant architectural evolution with optimizations for edge deployment, providing researchers/developers with precise architectural understanding to enhance YOLO models and maintain leadership in computer vision.
Abstract: You Only Look Once (YOLO) has been the prominent model for computer vision in deep learning for a decade. This study explores the novel aspects of YOLO26, the most recent version in the YOLO series. The elimination of Distribution Focal Loss (DFL), implementation of End-to-End NMS-Free Inference, introduction of ProgLoss + Small-Target-Aware Label Assignment (STAL), and use of the MuSGD optimizer are the primary enhancements designed to improve inference speed, which is claimed to achieve a 43% boost in CPU mode. This is designed to allow YOLO26 to attain real-time performance on edge devices or those without GPUs. Additionally, YOLO26 offers improvements in many computer vision tasks, including instance segmentation, pose estimation, and oriented bounding box (OBB) decoding. We aim for this effort to provide more value than just consolidating information already included in the existing technical documentation. Therefore, we performed a rigorous architectural investigation into YOLO26, mostly using the source code available in its GitHub repository and its official documentation. The authentic and detailed operational mechanisms of YOLO26 are inside the source code, which is seldom extracted by others. The YOLO26 architectural diagram is shown as the outcome of the investigation. This study is, to our knowledge, the first one presenting the CNN-based YOLO26 architecture, which is the core of YOLO26. Our objective is to provide a precise architectural comprehension of YOLO26 for researchers and developers aspiring to enhance the YOLO model, ensuring it remains the leading deep learning model in computer vision.
[276] VariViT: A Vision Transformer for Variable Image Sizes
Aswathi Varma, Suprosanna Shit, Chinmay Prabhakar, Daniel Scholz, Hongwei Bran Li, Bjoern Menze, Daniel Rueckert, Benedikt Wiestler
Main category: cs.CV
TL;DR: VariViT: A Vision Transformer variant that handles variable image sizes while maintaining consistent patch size, with novel positional embedding resizing and batching strategies for medical imaging applications.
Details
Motivation: Standard Vision Transformers require fixed-size patches, which is problematic for medical imaging where irregular structures like tumors need variable-sized crops. Fixed crops create variable foreground-to-background ratios, while resizing degrades information. There's a need for models that can handle variable image sizes without information loss.Method: Proposes VariViT with: 1) Novel positional embedding resizing scheme for variable number of patches, 2) New batching strategy to reduce computational complexity, 3) Maintains consistent patch size while handling variable image dimensions.
Result: Outperforms vanilla ViTs and ResNet on glioma genotype prediction and brain tumor classification with F1-scores of 75.5% and 76.3%. Batching strategy reduces computation time by up to 30% compared to conventional architectures.
Conclusion: VariViT effectively handles variable image sizes in medical imaging, learns more discriminative features, and reduces computational overhead, demonstrating efficacy in image representation learning for medical applications.
Abstract: Vision Transformers (ViTs) have emerged as the state-of-the-art architecture in representation learning, leveraging self-attention mechanisms to excel in various tasks. ViTs split images into fixed-size patches, constraining them to a predefined size and necessitating pre-processing steps like resizing, padding, or cropping. This poses challenges in medical imaging, particularly with irregularly shaped structures like tumors. A fixed bounding box crop size produces input images with highly variable foreground-to-background ratios. Resizing medical images can degrade information and introduce artefacts, impacting diagnosis. Hence, tailoring variable-sized crops to regions of interest can enhance feature representation capabilities. Moreover, large images are computationally expensive, and smaller sizes risk information loss, presenting a computation-accuracy tradeoff. We propose VariViT, an improved ViT model crafted to handle variable image sizes while maintaining a consistent patch size. VariViT employs a novel positional embedding resizing scheme for a variable number of patches. We also implement a new batching strategy within VariViT to reduce computational complexity, resulting in faster training and inference times. In our evaluations on two 3D brain MRI datasets, VariViT surpasses vanilla ViTs and ResNet in glioma genotype prediction and brain tumor classification. It achieves F1-scores of 75.5% and 76.3%, respectively, learning more discriminative features. Our proposed batching strategy reduces computation time by up to 30% compared to conventional architectures. These findings underscore the efficacy of VariViT in image representation learning. Our code can be found here: https://github.com/Aswathi-Varma/varivit
[277] VIGIL: Tackling Hallucination Detection in Image Recontextualization
Joanna Wojciechowicz, Maria Łubniewska, Jakub Antczak, Justyna Baczyńska, Wojciech Gromski, Wojciech Kozłowski, Maciej Zięba
Main category: cs.CV
TL;DR: VIGIL introduces a benchmark dataset and framework for fine-grained categorization of hallucinations in multimodal image recontextualization tasks for LMMs, with a multi-stage detection pipeline and open-source release.
Details
Motivation: Existing research treats hallucinations as a uniform issue, but there's a significant gap in multimodal evaluation that needs decomposition of these errors into specific categories for better understanding and detection.Method: Proposes a multi-stage detection pipeline that processes recontextualized images through specialized steps targeting object-level fidelity, background consistency, and omission detection, using a coordinated ensemble of open-source models.
Result: The approach enables deeper understanding of where models fail with explanations, and extensive experimental evaluations demonstrate the effectiveness of the detection pipeline.
Conclusion: VIGIL fills a gap in the field by providing the first fine-grained categorization and decomposition of hallucinations in multimodal image recontextualization, with open release of dataset, pipeline, and code for transparency and further exploration.
Abstract: We introduce VIGIL (Visual Inconsistency & Generative In-context Lucidity), the first benchmark dataset and framework providing a fine-grained categorization of hallucinations in the multimodal image recontextualization task for large multimodal models (LMMs). While existing research often treats hallucinations as a uniform issue, our work addresses a significant gap in multimodal evaluation by decomposing these errors into five categories: pasted object hallucinations, background hallucinations, object omission, positional & logical inconsistencies, and physical law violations. To address these complexities, we propose a multi-stage detection pipeline. Our architecture processes recontextualized images through a series of specialized steps targeting object-level fidelity, background consistency, and omission detection, leveraging a coordinated ensemble of open-source models, whose effectiveness is demonstrated through extensive experimental evaluations. Our approach enables a deeper understanding of where the models fail with an explanation; thus, we fill a gap in the field, as no prior methods offer such categorization and decomposition for this task. To promote transparency and further exploration, we openly release VIGIL, along with the detection pipeline and benchmark code, through our GitHub repository: https://github.com/mlubneuskaya/vigil and Data repository: https://huggingface.co/datasets/joannaww/VIGIL.
[278] SketchingReality: From Freehand Scene Sketches To Photorealistic Images
Ahmed Bourouis, Mikhail Bessmeltsev, Yulia Gryaditskaya
Main category: cs.CV
TL;DR: A method for generating photorealistic images from freehand sketches that balances sketch adherence with realism, using modulation-based approach and novel loss for training without pixel-aligned ground truth.
Details
Motivation: While generative AI has advanced with various conditioning signals, true freehand sketches remain underexplored. Previous work focused on edge maps misnamed as sketches, but freehand sketches have inherent abstraction and distortions. The challenge is to balance photorealism with sketch adherence without ground-truth pixel-aligned images.Method: Proposes a modulation-based approach that prioritizes semantic interpretation over strict edge position adherence. Introduces a novel loss function enabling training on freehand sketches without requiring ground-truth pixel-aligned images.
Result: Outperforms existing approaches in both semantic alignment with freehand sketch inputs and in the realism and overall quality of generated images.
Conclusion: The method successfully addresses the challenge of generating photorealistic images from freehand sketches by focusing on semantic interpretation rather than pixel-level alignment, enabling effective training without ground-truth data.
Abstract: Recent years have witnessed remarkable progress in generative AI, with natural language emerging as the most common conditioning input. As underlying models grow more powerful, researchers are exploring increasingly diverse conditioning signals, such as depth maps, edge maps, camera parameters, and reference images, to give users finer control over generation. Among different modalities, sketches are a natural and long-standing form of human communication, enabling rapid expression of visual concepts. Previous literature has largely focused on edge maps, often misnamed ‘sketches’, yet algorithms that effectively handle true freehand sketches, with their inherent abstraction and distortions, remain underexplored. We pursue the challenging goal of balancing photorealism with sketch adherence when generating images from freehand input. A key obstacle is the absence of ground-truth, pixel-aligned images: by their nature, freehand sketches do not have a single correct alignment. To address this, we propose a modulation-based approach that prioritizes semantic interpretation of the sketch over strict adherence to individual edge positions. We further introduce a novel loss that enables training on freehand sketches without requiring ground-truth pixel-aligned images. We show that our method outperforms existing approaches in both semantic alignment with freehand sketch inputs and in the realism and overall quality of the generated images.
[279] Advances in Global Solvers for 3D Vision
Zhenjun Zhao, Heng Yang, Bangyan Liao, Yingping Zeng, Shaocheng Yan, Yingdong Gu, Peidong Liu, Yi Zhou, Haoang Li, Javier Civera
Main category: cs.CV
TL;DR: A comprehensive survey of global solvers for 3D vision geometric optimization problems, covering three core paradigms (Branch-and-Bound, Convex Relaxation, Graduated Non-Convexity) and their applications across ten vision tasks.
Details
Motivation: Traditional local or heuristic methods for nonconvex geometric optimization in 3D vision lack certifiable guarantees. Global solvers offer provably optimal solutions, but the field lacks systematic organization and understanding of trade-offs between different paradigms.Method: Systematic review and taxonomy of three global solver paradigms: Branch-and-Bound (BnB), Convex Relaxation (CR), and Graduated Non-Convexity (GNC). Analysis includes theoretical foundations, algorithmic designs, practical enhancements, and application to ten core vision tasks.
Result: Unified perspective on global solvers revealing optimality-robustness-scalability trade-offs. Identification of critical future directions including scaling algorithms, integrating data-driven priors, establishing benchmarks, and addressing societal implications for safety-critical deployment.
Conclusion: Global solvers provide certifiable solutions for 3D vision geometric optimization. The survey consolidates the field and provides roadmap toward trustworthy perception for real-world applications, with ongoing resources available via GitHub repository.
Abstract: Global solvers have emerged as a powerful paradigm for 3D vision, offering certifiable solutions to nonconvex geometric optimization problems traditionally addressed by local or heuristic methods. This survey presents the first systematic review of global solvers in geometric vision, unifying the field through a comprehensive taxonomy of three core paradigms: Branch-and-Bound (BnB), Convex Relaxation (CR), and Graduated Non-Convexity (GNC). We present their theoretical foundations, algorithmic designs, and practical enhancements for robustness and scalability, examining how each addresses the fundamental nonconvexity of geometric estimation problems. Our analysis spans ten core vision tasks, from Wahba problem to bundle adjustment, revealing the optimality-robustness-scalability trade-offs that govern solver selection. We identify critical future directions: scaling algorithms while maintaining guarantees, integrating data-driven priors with certifiable optimization, establishing standardized benchmarks, and addressing societal implications for safety-critical deployment. By consolidating theoretical foundations, practical advances, and broader impacts, this survey provides a unified perspective and roadmap toward certifiable, trustworthy perception for real-world applications. A continuously-updated literature summary and companion code tutorials are available at https://github.com/ericzzj1989/Awesome-Global-Solvers-for-3D-Vision.
[280] MeFEm: Medical Face Embedding model
Yury Borets, Stepan Botman
Main category: cs.CV
TL;DR: MeFEm is a vision model using modified JEPA architecture for biometric and medical analysis from facial images, featuring axial stripe masking, circular loss weighting, and CLS token reassignment, achieving state-of-the-art performance on anthropometric tasks with less data.
Details
Motivation: The paper aims to develop an effective vision model for biometric and medical analysis from facial images, addressing limitations in existing approaches that require large datasets and suffer from domain bias in medical imaging data.Method: Modified Joint Embedding Predictive Architecture (JEPA) with axial stripe masking strategy to focus on semantically relevant regions, circular loss weighting scheme, and probabilistic reassignment of CLS token for improved linear probing. Trained on consolidated curated dataset.
Result: Outperforms strong baselines like FaRL and Franca on core anthropometric tasks despite using significantly less data. Shows promising results on Body Mass Index (BMI) estimation using novel consolidated closed-source dataset addressing domain bias.
Conclusion: MeFEm provides an effective vision model for facial biometric and medical analysis with improved efficiency and performance, offering a strong baseline for future work in this domain with publicly available model weights.
Abstract: We present MeFEm, a vision model based on a modified Joint Embedding Predictive Architecture (JEPA) for biometric and medical analysis from facial images. Key modifications include an axial stripe masking strategy to focus learning on semantically relevant regions, a circular loss weighting scheme, and the probabilistic reassignment of the CLS token for high quality linear probing. Trained on a consolidated dataset of curated images, MeFEm outperforms strong baselines like FaRL and Franca on core anthropometric tasks despite using significantly less data. It also shows promising results on Body Mass Index (BMI) estimation, evaluated on a novel, consolidated closed-source dataset that addresses the domain bias prevalent in existing data. Model weights are available at https://huggingface.co/boretsyury/MeFEm , offering a strong baseline for future work in this domain.
[281] Universal Image Immunization against Diffusion-based Image Editing via Semantic Injection
Chanhui Lee, Seunghyun Shin, Donggyu Choi, Hae-gon Jeon, Jeany Son
Main category: cs.CV
TL;DR: First universal image immunization framework using adversarial perturbations to protect images from AI-driven semantic manipulation in diffusion models.
Details
Motivation: Address ethical and legal risks of diffusion models (deepfakes, copyright infringement) by developing scalable immunization against malicious editing, overcoming limitations of image-specific approaches.Method: Generates universal adversarial perturbation (UAP) that embeds semantic targets into images while suppressing original content, misdirecting diffusion models’ attention during editing. Works in data-free settings without training data.
Result: Outperforms baselines in UAP setting, achieves performance comparable to image-specific methods with restricted perturbation budget, shows strong black-box transferability across diffusion models.
Conclusion: Proposes first practical universal immunization framework for diffusion models that balances effectiveness, scalability, and real-world applicability while addressing ethical concerns.
Abstract: Recent advances in diffusion models have enabled powerful image editing capabilities guided by natural language prompts, unlocking new creative possibilities. However, they introduce significant ethical and legal risks, such as deepfakes and unauthorized use of copyrighted visual content. To address these risks, image immunization has emerged as a promising defense against AI-driven semantic manipulation. Yet, most existing approaches rely on image-specific adversarial perturbations that require individual optimization for each image, thereby limiting scalability and practicality. In this paper, we propose the first universal image immunization framework that generates a single, broadly applicable adversarial perturbation specifically designed for diffusion-based editing pipelines. Inspired by universal adversarial perturbation (UAP) techniques used in targeted attacks, our method generates a UAP that embeds a semantic target into images to be protected. Simultaneously, it suppresses original content to effectively misdirect the model’s attention during editing. As a result, our approach effectively blocks malicious editing attempts by overwriting the original semantic content in the image via the UAP. Moreover, our method operates effectively even in data-free settings without requiring access to training data or domain knowledge, further enhancing its practicality and broad applicability in real-world scenarios. Extensive experiments show that our method, as the first universal immunization approach, significantly outperforms several baselines in the UAP setting. In addition, despite the inherent difficulty of universal perturbations, our method also achieves performance on par with image-specific methods under a more restricted perturbation budget, while also exhibiting strong black-box transferability across different diffusion models.
[282] It’s a Matter of Time: Three Lessons on Long-Term Motion for Perception
Willem Davison, Xinyue Hao, Laura Sevilla-Lara
Main category: cs.CV
TL;DR: Long-term motion representations from point tracks contain rich information about actions, objects, materials, and spatial relationships, often outperforming image representations, especially in low-data and zero-shot settings, while being computationally efficient.
Details
Motivation: To understand what can be learned about the world from long-term motion information and explore the properties of long-term motion information for visual learning, given that temporal information is essential for perception but less understood than image information.Method: Leverage recent advances in point-track estimation to learn temporal representations and experiment on various perceptual tasks to analyze the properties of long-term motion information.
Result: Three key findings: 1) Long-term motion representations contain comprehensive information about actions, objects, materials, and spatial information, often better than images; 2) They generalize better in low-data and zero-shot settings; 3) Their low dimensionality offers better trade-off between computational cost and accuracy than standard video representations.
Conclusion: Long-term motion information is a powerful and efficient representation for perception that should be leveraged in future model designs, offering advantages over image-based approaches in generalization and computational efficiency.
Abstract: Temporal information has long been considered to be essential for perception. While there is extensive research on the role of image information for perceptual tasks, the role of the temporal dimension remains less well understood: What can we learn about the world from long-term motion information? What properties does long-term motion information have for visual learning? We leverage recent success in point-track estimation, which offers an excellent opportunity to learn temporal representations and experiment on a variety of perceptual tasks. We draw 3 clear lessons: 1) Long-term motion representations contain information to understand actions, but also objects, materials, and spatial information, often even better than images. 2) Long-term motion representations generalize far better than image representations in low-data settings and in zero-shot tasks. 3) The very low dimensionality of motion information makes motion representations a better trade-off between GFLOPs and accuracy than standard video representations, and used together they achieve higher performance than video representations alone. We hope these insights will pave the way for the design of future models that leverage the power of long-term motion information for perception.
[283] Depth Completion as Parameter-Efficient Test-Time Adaptation
Bingxin Ke, Qunjie Zhou, Jiahui Huang, Xuanchi Ren, Tianchang Shen, Konrad Schindler, Laura Leal-Taixé, Shengyu Huang
Main category: cs.CV
TL;DR: CAPA is a parameter-efficient test-time optimization framework that adapts pre-trained 3D foundation models for depth completion using sparse geometric cues, without retraining the backbone.
Details
Motivation: Prior methods train task-specific encoders for auxiliary inputs which often overfit and generalize poorly. There's a need for a more efficient approach that can leverage pre-trained 3D foundation models while adapting to specific scene measurements at inference time.Method: CAPA freezes the FM backbone and updates only a minimal set of parameters using Parameter-Efficient Fine-Tuning (like LoRA or VPT), guided by gradients from sparse observations available at inference time. For videos, it introduces sequence-level parameter sharing to jointly adapt all frames.
Result: CAPA achieves state-of-the-art results across diverse condition patterns on both indoor and outdoor datasets, effectively grounding the foundation model’s geometric prior in scene-specific measurements.
Conclusion: CAPA provides a model-agnostic, parameter-efficient framework for test-time adaptation of 3D foundation models that corrects distortions and misplaced structures while being compatible with any ViT-based foundation model.
Abstract: We introduce CAPA, a parameter-efficient test-time optimization framework that adapts pre-trained 3D foundation models (FMs) for depth completion, using sparse geometric cues. Unlike prior methods that train task-specific encoders for auxiliary inputs, which often overfit and generalize poorly, CAPA freezes the FM backbone. Instead, it updates only a minimal set of parameters using Parameter-Efficient Fine-Tuning (e.g. LoRA or VPT), guided by gradients calculated directly from the sparse observations available at inference time. This approach effectively grounds the foundation model’s geometric prior in the scene-specific measurements, correcting distortions and misplaced structures. For videos, CAPA introduces sequence-level parameter sharing, jointly adapting all frames to exploit temporal correlations, improve robustness, and enforce multi-frame consistency. CAPA is model-agnostic, compatible with any ViT-based FM, and achieves state-of-the-art results across diverse condition patterns on both indoor and outdoor datasets. Project page: research.nvidia.com/labs/dvl/projects/capa.
[284] SAILS: Segment Anything with Incrementally Learned Semantics for Task-Invariant and Training-Free Continual Learning
Shishir Muralidhara, Didier Stricker, René Schuster
Main category: cs.CV
TL;DR: SAILS is a training-free framework for Class-Incremental Semantic Segmentation that uses SAM for zero-shot region extraction and semantic association through prototypes, eliminating forgetting and computational costs.
Details
Motivation: Continual learning faces challenges of repeated retraining, high computational costs, and catastrophic forgetting, limiting real-world applicability. The paper aims to address these issues in Class-Incremental Semantic Segmentation.Method: SAILS decouples CISS into two stages: (1) Zero-shot region extraction using Segment Anything Model (SAM), and (2) semantic association through prototypes in a fixed feature space with selective intra-class clustering for multiple prototypes per class.
Result: SAILS typically surpasses training-based approaches on standard CISS datasets, especially in long task sequences where forgetting is severe. It completely eliminates forgetting, maintains consistent performance, and exhibits positive backward transfer.
Conclusion: SAILS provides a training-free solution to continual learning challenges in semantic segmentation, offering superior performance without computational costs or forgetting issues.
Abstract: Continual learning remains constrained by the need for repeated retraining, high computational costs, and the persistent challenge of forgetting. These factors significantly limit the applicability of continuous learning in real-world settings, as iterative model updates require significant computational resources and inherently exacerbate forgetting. We present SAILS – Segment Anything with Incrementally Learned Semantics, a training-free framework for Class-Incremental Semantic Segmentation (CISS) that sidesteps these challenges entirely. SAILS leverages foundational models to decouple CISS into two stages: Zero-shot region extraction using Segment Anything Model (SAM), followed by semantic association through prototypes in a fixed feature space. SAILS incorporates selective intra-class clustering, resulting in multiple prototypes per class to better model intra-class variability. Our results demonstrate that, despite requiring no incremental training, SAILS typically surpasses the performance of existing training-based approaches on standard CISS datasets, particularly in long and challenging task sequences where forgetting tends to be most severe. By avoiding parameter updates, SAILS completely eliminates forgetting and maintains consistent, task-invariant performance. Furthermore, SAILS exhibits positive backward transfer, where the introduction of new classes can enhance performance on previous classes.
[285] VIPA: Visual Informative Part Attention for Referring Image Segmentation
Yubin Cho, Hyunwoo Yu, Kyeongbo Kong, Kyomin Sohn, Bongjoon Hyun, Suk-Ju Kang
Main category: cs.CV
TL;DR: VIPA framework for referring image segmentation uses visual expressions to provide structural/semantic visual target information, reducing cross-modal projection variance and enhancing semantic consistency through visual expression generator module.
Details
Motivation: Existing RIS methods leverage vision information into language tokens, but need more effective exploitation of visual contexts for fine-grained segmentation. The paper aims to better utilize visual informative parts to provide structural and semantic target information.Method: Proposes Visual Informative Part Attention (VIPA) framework with visual expression generator (VEG) module. VEG retrieves informative visual tokens via local-global linguistic context cues and refines them to reduce noise while sharing informative visual attributes. This creates visual expressions that provide structural/semantic visual target information.
Result: Extensive experiments show VIPA outperforms existing state-of-the-art methods on four public RIS benchmarks. Visual analysis demonstrates effectiveness in enabling network attention to robustly align with fine-grained regions of interest.
Conclusion: VIPA framework effectively leverages visual informative parts through visual expressions to enhance semantic consistency and reduce cross-modal projection variance, achieving superior performance in referring image segmentation tasks.
Abstract: Referring Image Segmentation (RIS) aims to segment a target object described by a natural language expression. Existing methods have evolved by leveraging the vision information into the language tokens. To more effectively exploit visual contexts for fine-grained segmentation, we propose a novel Visual Informative Part Attention (VIPA) framework for referring image segmentation. VIPA leverages the informative parts of visual contexts, called a visual expression, which can effectively provide the structural and semantic visual target information to the network. This design reduces high-variance cross-modal projection and enhances semantic consistency in an attention mechanism of the referring image segmentation. We also design a visual expression generator (VEG) module, which retrieves informative visual tokens via local-global linguistic context cues and refines the retrieved tokens for reducing noise information and sharing informative visual attributes. This module allows the visual expression to consider comprehensive contexts and capture semantic visual contexts of informative regions. In this way, our framework enables the network’s attention to robustly align with the fine-grained regions of interest. Extensive experiments and visual analysis demonstrate the effectiveness of our approach. Our VIPA outperforms the existing state-of-the-art methods on four public RIS benchmarks.
[286] Debiasing Central Fixation Confounds Reveals a Peripheral “Sweet Spot” for Human-like Scanpaths in Hard-Attention Vision
Pengcheng Pan, Yonekura Shogo, Yasuo Kuniyosh
Main category: cs.CV
TL;DR: The paper addresses the confounding effect of center bias in evaluating hard-attention models against human gaze data, proposing a debiased metric (GCS) that reveals a “peripheral sweet spot” for biologically plausible scanpaths.
Details
Motivation: Standard scanpath metrics for evaluating task-driven hard-attention models against human gaze are strongly confounded by dataset-specific center bias, especially on object-centric datasets, making it difficult to distinguish genuine behavioral alignment from mere central tendency.Method: The authors use Gaze-CIFAR-10 to demonstrate center bias issues, analyze a hard-attention classifier under varying foveal patch sizes and peripheral context, and propose Gaze Consistency Score (GCS) - a center-debiased composite metric augmented with movement similarity.
Result: A trivial center-fixation baseline achieves surprisingly strong scanpath scores approaching learned policies. Analysis reveals a “peripheral sweet spot” where only a narrow range of sensory constraints yields scanpaths that are both above the center baseline after debiasing and temporally human-like. GCS uncovers a robust sweet spot at medium patch size with both foveal and peripheral vision.
Conclusion: The paper highlights the need for center-debiased metrics in evaluating active perception on object-centric datasets and designing gaze benchmarks that better separate behavioral alignment from center bias, revealing important implications for biologically plausible vision models.
Abstract: Human eye movements in visual recognition reflect a balance between foveal sampling and peripheral context. Task-driven hard-attention models for vision are often evaluated by how well their scanpaths match human gaze. However, common scanpath metrics can be strongly confounded by dataset-specific center bias, especially on object-centric datasets. Using Gaze-CIFAR-10, we show that a trivial center-fixation baseline achieves surprisingly strong scanpath scores, approaching many learned policies. This makes standard metrics optimistic and blurs the distinction between genuine behavioral alignment and mere central tendency. We then analyze a hard-attention classifier under constrained vision by sweeping foveal patch size and peripheral context, revealing a peripheral sweet spot: only a narrow range of sensory constraints yields scanpaths that are simultaneously (i) above the center baseline after debiasing and (ii) temporally human-like in movement statistics. To address center bias, we propose GCS (Gaze Consistency Score), a center-debiased composite metric augmented with movement similarity. GCS uncovers a robust sweet spot at medium patch size with both foveal and peripheral vision, that is not obvious from raw scanpath metrics or accuracy alone, and also highlights a “shortcut regime” when the field-of-view becomes too large. We discuss implications for evaluating active perception on object-centric datasets and for designing gaze benchmarks that better separate behavioral alignment from center bias.
[287] Integrating Affordances and Attention models for Short-Term Object Interaction Anticipation
Lorenzo Mur Labadia, Ruben Martinez-Cantin, Jose J. Guerrero, Giovanni M. Farinella, Antonino Furnari
Main category: cs.CV
TL;DR: STAformer++ improves short-term object-interaction anticipation in egocentric videos using attention-based architectures with environment affordance modeling and interaction hotspot prediction.
Details
Motivation: Improve short-term anticipation of object interactions for wearable assistants and human-robot interaction by better predicting location, categories, and timing of future interactions from egocentric video.Method: Proposes STAformer and STAformer++ architectures with frame-guided temporal pooling, dual image-video attention, multiscale feature fusion. Integrates environment affordance modeling as persistent memory and predicts interaction hotspots from hand/object trajectories.
Result: Significant improvements on Overall Top-5 mAP: up to +23 percentage points on Ego4D and +31 p.p. on curated EPIC-Kitchens STA labels.
Conclusion: The proposed attention-based architectures with affordance grounding substantially improve short-term anticipation performance, with code and annotations released to encourage further research.
Abstract: Short Term object-interaction Anticipation consists in detecting the location of the next active objects, the noun and verb categories of the interaction, as well as the time to contact from the observation of egocentric video. This ability is fundamental for wearable assistants to understand user goals and provide timely assistance, or to enable human-robot interaction. In this work, we present a method to improve the performance of STA predictions. Our contributions are two-fold: 1 We propose STAformer and STAformer plus plus, two novel attention-based architectures integrating frame-guided temporal pooling, dual image-video attention, and multiscale feature fusion to support STA predictions from an image-input video pair; 2 We introduce two novel modules to ground STA predictions on human behavior by modeling affordances. First, we integrate an environment affordance model which acts as a persistent memory of interactions that can take place in a given physical scene. We explore how to integrate environment affordances via simple late fusion and with an approach which adaptively learns how to best fuse affordances with end-to-end predictions. Second, we predict interaction hotspots from the observation of hands and object trajectories, increasing confidence in STA predictions localized around the hotspot. Our results show significant improvements on Overall Top-5 mAP, with gain up to +23p.p on Ego4D and +31p.p on a novel set of curated EPIC-Kitchens STA labels. We released the code, annotations, and pre-extracted affordances on Ego4D and EPIC-Kitchens to encourage future research in this area.
[288] Multi-dimensional Persistent Sheaf Laplacians for Image Analysis
Xiang Xiang Wang, Guo-Wei Wei
Main category: cs.CV
TL;DR: A multi-dimensional persistent sheaf Laplacian framework for image analysis that uses topological spectral representations across multiple reduced dimensions instead of selecting a single dimension like PCA.
Details
Motivation: Address the sensitivity of dimensionality reduction techniques (like PCA) to the choice of reduced dimension by exploiting complementary advantages of multiple reduced dimensions rather than selecting a single dimension or averaging results.Method: Treat image samples as simplicial complexes, use persistent sheaf Laplacians to extract multiscale localized topological spectral representations for individual images, then aggregate statistical summaries of spectra across scales and dimensions to form multiscale multi-dimensional image representations.
Result: The method shows more stable performance across a wide range of reduced dimensions and achieves consistent improvements over PCA-based baselines in moderate dimensional regimes on COIL20 and ETH80 image datasets.
Conclusion: The multi-dimensional persistent sheaf Laplacian framework provides a robust alternative to traditional dimensionality reduction methods by leveraging topological information across multiple dimensions for improved image analysis.
Abstract: We propose a multi-dimensional persistent sheaf Laplacian (MPSL) framework on simplicial complexes for image analysis. The proposed method is motivated by the strong sensitivity of commonly used dimensionality reduction techniques, such as principal component analysis (PCA), to the choice of reduced dimension. Rather than selecting a single reduced dimension or averaging results across dimensions, we exploit complementary advantages of multiple reduced dimensions. At a given dimension, image samples are regarded as simplicial complexes, and persistent sheaf Laplacians are utilized to extract a multiscale localized topological spectral representation for individual image samples. Statistical summaries of the resulting spectra are then aggregated across scales and dimensions to form multiscale multi-dimensional image representations. We evaluate the proposed framework on the COIL20 and ETH80 image datasets using standard classification protocols. Experimental results show that the proposed method provides more stable performance across a wide range of reduced dimensions and achieves consistent improvements to PCA-based baselines in moderate dimensional regimes.
[289] CT-Bench: A Benchmark for Multimodal Lesion Understanding in Computed Tomography
Qingqing Zhu, Qiao Jin, Tejas S. Mathai, Yin Fang, Zhizheng Wang, Yifan Yang, Maame Sarfo-Gyamfi, Benjamin Hou, Ran Gu, Praveen T. S. Balamuralikrishna, Kenneth C. Wang, Ronald M. Summers, Zhiyong Lu
Main category: cs.CV
TL;DR: CT-Bench: A benchmark dataset for CT lesion analysis with 20,335 lesions and 2,850 QA pairs for evaluating multimodal AI models on lesion localization, description, size estimation, and attribute categorization.
Details
Motivation: Progress in AI for CT lesion analysis is limited by scarcity of publicly available CT datasets with lesion-level annotations, creating a need for comprehensive benchmarks to evaluate multimodal models.Method: Created CT-Bench with two components: 1) Lesion Image and Metadata Set (20,335 lesions from 7,795 CT studies with bounding boxes, descriptions, size info), and 2) multitask visual question answering benchmark (2,850 QA pairs covering lesion tasks). Includes hard negative examples for real-world challenges.
Result: Evaluated state-of-the-art multimodal models (vision-language and medical CLIP variants) against radiologist assessments. Fine-tuning on the Lesion Image and Metadata Set yielded significant performance gains across both benchmark components.
Conclusion: CT-Bench serves as a comprehensive benchmark for lesion analysis, demonstrating clinical utility through performance improvements when models are fine-tuned on the dataset.
Abstract: Artificial intelligence (AI) can automatically delineate lesions on computed tomography (CT) and generate radiology report content, yet progress is limited by the scarcity of publicly available CT datasets with lesion-level annotations. To bridge this gap, we introduce CT-Bench, a first-of-its-kind benchmark dataset comprising two components: a Lesion Image and Metadata Set containing 20,335 lesions from 7,795 CT studies with bounding boxes, descriptions, and size information, and a multitask visual question answering benchmark with 2,850 QA pairs covering lesion localization, description, size estimation, and attribute categorization. Hard negative examples are included to reflect real-world diagnostic challenges. We evaluate multiple state-of-the-art multimodal models, including vision-language and medical CLIP variants, by comparing their performance to radiologist assessments, demonstrating the value of CT-Bench as a comprehensive benchmark for lesion analysis. Moreover, fine-tuning models on the Lesion Image and Metadata Set yields significant performance gains across both components, underscoring the clinical utility of CT-Bench.
[290] Wrivinder: Towards Spatial Intelligence for Geo-locating Ground Images onto Satellite Imagery
Chandrakanth Gudavalli, Tajuddin Manhar Mohammed, Abhay Yadav, Ananth Vishnu Bhaskar, Hardik Prajapati, Cheng Peng, Rama Chellappa, Shivkumar Chandrasekaran, B. S. Manjunath
Main category: cs.CV
TL;DR: Wrivinder is a zero-shot geometry-driven framework that aggregates multiple ground photos to reconstruct 3D scenes and align them with satellite imagery for camera geo-localization, achieving sub-30m accuracy without paired supervision.
Details
Motivation: Aligning ground-level imagery with satellite maps is crucial for mapping, navigation, and situational awareness, but remains challenging under large viewpoint gaps or unreliable GPS. Existing methods lack suitable benchmarks for systematic evaluation.Method: Combines SfM reconstruction, 3D Gaussian Splatting, semantic grounding, and monocular depth-based metric cues to produce stable zenith-view renderings that can be directly matched to satellite imagery. Also introduces MC-Sat dataset for evaluation.
Result: Achieves sub-30m geolocation accuracy across both dense and large-area scenes in zero-shot experiments, demonstrating the promise of geometry-based aggregation for robust ground-to-satellite localization.
Conclusion: Wrivinder and MC-Sat provide the first comprehensive baseline and testbed for studying geometry-centered cross-view alignment without paired supervision, enabling more robust localization in GPS-challenged environments.
Abstract: Aligning ground-level imagery with geo-registered satellite maps is crucial for mapping, navigation, and situational awareness, yet remains challenging under large viewpoint gaps or when GPS is unreliable. We introduce Wrivinder, a zero-shot, geometry-driven framework that aggregates multiple ground photographs to reconstruct a consistent 3D scene and align it with overhead satellite imagery. Wrivinder combines SfM reconstruction, 3D Gaussian Splatting, semantic grounding, and monocular depth–based metric cues to produce a stable zenith-view rendering that can be directly matched to satellite context for metrically accurate camera geo-localization. To support systematic evaluation of this task, which lacks suitable benchmarks, we also release MC-Sat, a curated dataset linking multi-view ground imagery with geo-registered satellite tiles across diverse outdoor environments. Together, Wrivinder and MC-Sat provide a first comprehensive baseline and testbed for studying geometry-centered cross-view alignment without paired supervision. In zero-shot experiments, Wrivinder achieves sub-30,m geolocation accuracy across both dense and large-area scenes, highlighting the promise of geometry-based aggregation for robust ground-to-satellite localization.
[291] AnchorWeave: World-Consistent Video Generation with Retrieved Local Spatial Memories
Zun Wang, Han Lin, Jaehong Yoon, Jaemin Cho, Yue Zhang, Mohit Bansal
Main category: cs.CV
TL;DR: AnchorWeave: A memory-augmented video generation framework that uses multiple clean local geometric memories instead of a single misaligned global 3D scene to improve long-term spatial consistency in camera-controllable video generation.
Details
Motivation: Existing memory-based approaches for camera-controllable video generation reconstruct global 3D scenes from multiple views, but pose and depth estimation errors cause cross-view misalignment that accumulates into noisy geometry, degrading generation quality.Method: AnchorWeave replaces single misaligned global memory with multiple clean local geometric memories, performs coverage-driven local memory retrieval aligned with target trajectory, and integrates selected memories through a multi-anchor weaving controller during generation.
Result: Extensive experiments show AnchorWeave significantly improves long-term scene consistency while maintaining strong visual quality, with ablation studies validating effectiveness of local geometric conditioning, multi-anchor control, and coverage-driven retrieval.
Conclusion: AnchorWeave addresses the challenge of maintaining spatial world consistency over long horizons in video generation by using multiple local geometric memories to reconcile cross-view inconsistencies, outperforming global reconstruction approaches.
Abstract: Maintaining spatial world consistency over long horizons remains a central challenge for camera-controllable video generation. Existing memory-based approaches often condition generation on globally reconstructed 3D scenes by rendering anchor videos from the reconstructed geometry in the history. However, reconstructing a global 3D scene from multiple views inevitably introduces cross-view misalignment, as pose and depth estimation errors cause the same surfaces to be reconstructed at slightly different 3D locations across views. When fused, these inconsistencies accumulate into noisy geometry that contaminates the conditioning signals and degrades generation quality. We introduce AnchorWeave, a memory-augmented video generation framework that replaces a single misaligned global memory with multiple clean local geometric memories and learns to reconcile their cross-view inconsistencies. To this end, AnchorWeave performs coverage-driven local memory retrieval aligned with the target trajectory and integrates the selected local memories through a multi-anchor weaving controller during generation. Extensive experiments demonstrate that AnchorWeave significantly improves long-term scene consistency while maintaining strong visual quality, with ablation and analysis studies further validating the effectiveness of local geometric conditioning, multi-anchor control, and coverage-driven retrieval.
[292] PAct: Part-Decomposed Single-View Articulated Object Generation
Qingming Liu, Xinyue Yao, Shuyuan Zhang, Yueci Deng, Guiliang Liu, Zhen Liu, Kui Jia
Main category: cs.CV
TL;DR: A part-centric generative framework for articulated 3D object creation from single images, generating movable parts with explicit part-aware conditioning for fast feed-forward inference.
Details
Motivation: Articulated objects are crucial for interactive 3D applications but difficult to scale due to challenges in part decomposition and kinematic rigging. Existing methods are either slow (optimization-based) or produce generic results (retrieval-based) that don't match specific input observations.Method: A part-centric generative framework that models objects as sets of movable parts, each encoded by latent tokens augmented with part identity and articulation cues. Conditioned on a single image, it generates articulated 3D assets with instance-level correspondence while maintaining valid part structure and motion.
Result: The approach shows improved input consistency, part accuracy, and articulation plausibility over optimization-based and retrieval-driven baselines on common articulated categories (drawers, doors), while substantially reducing inference time compared to optimization methods.
Conclusion: The framework enables fast feed-forward generation of articulated 3D assets from single images, avoiding per-instance optimization while supporting controllable assembly and articulation for embodied interaction applications.
Abstract: Articulated objects are central to interactive 3D applications, including embodied AI, robotics, and VR/AR, where functional part decomposition and kinematic motion are essential. Yet producing high-fidelity articulated assets remains difficult to scale because it requires reliable part decomposition and kinematic rigging. Existing approaches largely fall into two paradigms: optimization-based reconstruction or distillation, which can be accurate but often takes tens of minutes to hours per instance, and inference-time methods that rely on template or part retrieval, producing plausible results that may not match the specific structure and appearance in the input observation. We introduce a part-centric generative framework for articulated object creation that synthesizes part geometry, composition, and articulation under explicit part-aware conditioning. Our representation models an object as a set of movable parts, each encoded by latent tokens augmented with part identity and articulation cues. Conditioned on a single image, the model generates articulated 3D assets that preserve instance-level correspondence while maintaining valid part structure and motion. The resulting approach avoids per-instance optimization, enables fast feed-forward inference, and supports controllable assembly and articulation, which are important for embodied interaction. Experiments on common articulated categories (e.g., drawers and doors) show improved input consistency, part accuracy, and articulation plausibility over optimization-based and retrieval-driven baselines, while substantially reducing inference time.
[293] ThermEval: A Structured Benchmark for Evaluation of Vision-Language Models on Thermal Imagery
Ayush Shrivastava, Kirtan Gangani, Laksh Jain, Mayank Goel, Nipun Batra
Main category: cs.CV
TL;DR: ThermEval-B benchmark assesses VLMs on thermal imagery understanding, revealing their limitations in temperature-grounded reasoning and need for dedicated thermal evaluation beyond RGB-centric approaches.
Details
Motivation: VLMs perform well on RGB images but fail on thermal imagery, which is critical for applications like nighttime surveillance, search and rescue, autonomous driving, and medical screening where visible light fails. Thermal images encode physical temperature rather than color/texture, requiring capabilities not evaluated by existing RGB benchmarks.Method: Introduces ThermEval-B, a structured benchmark of ~55,000 thermal visual question answering pairs. Combines public datasets with newly collected ThermEval-D, which provides dense per-pixel temperature maps with semantic body-part annotations across diverse indoor/outdoor environments. Evaluates 25 open-source and closed-source VLMs.
Result: Models consistently fail at temperature-grounded reasoning, degrade under colormap transformations, and default to language priors or fixed responses. Only marginal gains from prompting or supervised fine-tuning. Demonstrates thermal understanding requires dedicated evaluation beyond RGB-centric assumptions.
Conclusion: Thermal vision language understanding requires specialized evaluation that goes beyond RGB-centric approaches. ThermEval serves as a benchmark to drive progress in thermal vision language modeling by addressing the unique challenges of temperature-based perception and reasoning.
Abstract: Vision language models (VLMs) achieve strong performance on RGB imagery, but they do not generalize to thermal images. Thermal sensing plays a critical role in settings where visible light fails, including nighttime surveillance, search and rescue, autonomous driving, and medical screening. Unlike RGB imagery, thermal images encode physical temperature rather than color or texture, requiring perceptual and reasoning capabilities that existing RGB-centric benchmarks do not evaluate. We introduce ThermEval-B, a structured benchmark of approximately 55,000 thermal visual question answering pairs designed to assess the foundational primitives required for thermal vision language understanding. ThermEval-B integrates public datasets with our newly collected ThermEval-D, the first dataset to provide dense per-pixel temperature maps with semantic body-part annotations across diverse indoor and outdoor environments. Evaluating 25 open-source and closed-source VLMs, we find that models consistently fail at temperature-grounded reasoning, degrade under colormap transformations, and default to language priors or fixed responses, with only marginal gains from prompting or supervised fine-tuning. These results demonstrate that thermal understanding requires dedicated evaluation beyond RGB-centric assumptions, positioning ThermEval as a benchmark to drive progress in thermal vision language modeling.
[294] Image Generation with a Sphere Encoder
Kaiyu Yue, Menglin Jia, Ji Hou, Tom Goldstein
Main category: cs.CV
TL;DR: Sphere Encoder: A single-pass generative framework that maps images to a spherical latent space and generates images by decoding random points on the sphere, achieving competitive performance with diffusion models at much lower inference cost.
Details
Motivation: To develop a more efficient alternative to multi-step diffusion models that require many inference steps, aiming for single-pass or few-step generation while maintaining competitive image quality.Method: Learns an encoder that maps natural images uniformly onto a spherical latent space and a decoder that maps random latent vectors back to image space, trained solely through image reconstruction losses. Supports conditional generation and iterative refinement through encoder/decoder loops.
Result: Achieves performance competitive with state-of-the-art diffusion models across several datasets while requiring only a small fraction of the inference cost (fewer than five steps vs. many-step diffusion).
Conclusion: Sphere Encoder provides an efficient generative framework that can produce high-quality images in a single forward pass, offering a compelling alternative to computationally expensive diffusion models.
Abstract: We introduce the Sphere Encoder, an efficient generative framework capable of producing images in a single forward pass and competing with many-step diffusion models using fewer than five steps. Our approach works by learning an encoder that maps natural images uniformly onto a spherical latent space, and a decoder that maps random latent vectors back to the image space. Trained solely through image reconstruction losses, the model generates an image by simply decoding a random point on the sphere. Our architecture naturally supports conditional generation, and looping the encoder/decoder a few times can further enhance image quality. Across several datasets, the sphere encoder approach yields performance competitive with state of the art diffusions, but with a small fraction of the inference cost. Project page is available at https://sphere-encoder.github.io .
[295] RAVENEA: A Benchmark for Multimodal Retrieval-Augmented Visual Culture Understanding
Jiaang Li, Yifei Yuan, Wenyan Li, Mohammad Aliannejadi, Daniel Hershcovich, Anders Søgaard, Ivan Vulić, Wenxuan Zhang, Paul Pu Liang, Yang Deng, Serge Belongie
Main category: cs.CV
TL;DR: RAVENEA is a new benchmark for retrieval-augmented visual culture understanding, featuring culture-focused VQA and image captioning tasks with 11K+ Wikipedia documents, showing that cultural grounding improves multimodal retrieval and downstream tasks.
Details
Motivation: Vision-language models often fail to interpret cultural nuances effectively, and while retrieval-augmented generation helps with cultural understanding in text-only settings, its application in multimodal scenarios remains underexplored.Method: Created RAVENEA benchmark with two tasks: culture-focused visual question answering (cVQA) and culture-informed image captioning (cIC), integrating over 11,396 unique Wikipedia documents curated by human annotators. Evaluated seven multimodal retrievers and fifteen VLMs.
Result: Cultural grounding annotations enhance multimodal retrieval and downstream tasks; VLMs with culture-aware retrieval outperform non-augmented counterparts (+6% on cVQA, +11% on cIC); performance varies widely across countries.
Conclusion: RAVENEA highlights limitations of current multimodal retrievers and VLMs, showing the need to enhance visual culture understanding within RAG systems, and provides a valuable resource for advancing research in this area.
Abstract: As vision-language models (VLMs) become increasingly integrated into daily life, the need for accurate visual culture understanding is becoming critical. Yet, these models frequently fall short in interpreting cultural nuances effectively. Prior work has demonstrated the effectiveness of retrieval-augmented generation (RAG) in enhancing cultural understanding in text-only settings, while its application in multimodal scenarios remains underexplored. To bridge this gap, we introduce RAVENEA (Retrieval-Augmented Visual culturE uNdErstAnding), a new benchmark designed to advance visual culture understanding through retrieval, focusing on two tasks: culture-focused visual question answering (cVQA) and culture-informed image captioning (cIC). RAVENEA extends existing datasets by integrating over 11,396 unique Wikipedia documents curated and ranked by human annotators. Through the extensive evaluation on seven multimodal retrievers and fifteen VLMs, RAVENEA reveals some undiscovered findings: (i) In general, cultural grounding annotations can enhance multimodal retrieval and corresponding downstream tasks. (ii) VLMs, when augmented with culture-aware retrieval, generally outperform their non-augmented counterparts (by averaging +6% on cVQA and +11% on cIC). (iii) Performance of culture-aware retrieval augmented varies widely across countries. These findings highlight the limitations of current multimodal retrievers and VLMs, underscoring the need to enhance visual culture understanding within RAG systems. We believe RAVENEA offers a valuable resource for advancing research on retrieval-augmented visual culture understanding.
[296] EditCtrl: Disentangled Local and Global Control for Real-Time Generative Video Editing
Yehonathan Litman, Shikun Liu, Dario Seyb, Nicholas Milef, Yang Zhou, Carl Marshall, Shubham Tulsiani, Caleb Leak
Main category: cs.CV
TL;DR: EditCtrl is an efficient video inpainting framework that focuses computation only on edited regions, achieving 10x efficiency gains while improving quality over full-attention methods.
Details
Motivation: Current video editing methods inefficiently process full video context even for small, localized edits, creating computational bottlenecks. There's a need for more efficient approaches that focus computation only where needed.Method: Introduces a local video context module operating only on masked tokens (cost proportional to edit size), guided by a lightweight temporal global context embedder for video-wide consistency with minimal overhead.
Result: EditCtrl is 10x more compute efficient than state-of-the-art methods while improving editing quality. Enables new capabilities like multi-region editing with text prompts and autoregressive content propagation.
Conclusion: The framework demonstrates that focusing computation on edited regions rather than full video context enables both efficiency gains and quality improvements in video inpainting.
Abstract: High-fidelity generative video editing has seen significant quality improvements by leveraging pre-trained video foundation models. However, their computational cost is a major bottleneck, as they are often designed to inefficiently process the full video context regardless of the inpainting mask’s size, even for sparse, localized edits. In this paper, we introduce EditCtrl, an efficient video inpainting control framework that focuses computation only where it is needed. Our approach features a novel local video context module that operates solely on masked tokens, yielding a computational cost proportional to the edit size. This local-first generation is then guided by a lightweight temporal global context embedder that ensures video-wide context consistency with minimal overhead. Not only is EditCtrl 10 times more compute efficient than state-of-the-art generative editing methods, it even improves editing quality compared to methods designed with full-attention. Finally, we showcase how EditCtrl unlocks new capabilities, including multi-region editing with text prompts and autoregressive content propagation.
[297] LightX3ECG: A Lightweight and eXplainable Deep Learning System for 3-lead Electrocardiogram Classification
Khiem H. Le, Hieu H. Pham, Thao BT. Nguyen, Tu A. Nguyen, Tien N. Thanh, Cuong D. Do
Main category: cs.CV
TL;DR: A deep learning system for detecting multiple cardiovascular abnormalities using only 3 ECG leads instead of standard 12-lead ECG.
Details
Motivation: Cardiovascular diseases are serious health threats requiring early detection. While 12-lead ECG is standard, fewer leads enable portable/wearable devices for broader accessibility.Method: Developed a novel deep learning system that uses only three ECG leads to identify multiple cardiovascular abnormalities.
Result: The system achieves accurate detection of cardiovascular abnormalities with reduced lead requirements.
Conclusion: Three-lead ECG systems can enable more accessible cardiovascular monitoring through portable/wearable devices while maintaining diagnostic accuracy.
Abstract: Cardiovascular diseases (CVDs) are a group of heart and blood vessel disorders that is one of the most serious dangers to human health, and the number of such patients is still growing. Early and accurate detection plays a key role in successful treatment and intervention. Electrocardiogram (ECG) is the gold standard for identifying a variety of cardiovascular abnormalities. In clinical practices and most of the current research, standard 12-lead ECG is mainly used. However, using a lower number of leads can make ECG more prevalent as it can be conveniently recorded by portable or wearable devices. In this research, we develop a novel deep learning system to accurately identify multiple cardiovascular abnormalities by using only three ECG leads.
[298] Realtime Data-Efficient Portrait Stylization Based On Geometric Alignment
Xinrui Wang, Zhuoru Li, Xiao Zhou, Yusuke Iwasawa, Yutaka Matsuo
Main category: cs.CV
TL;DR: Portrait stylization method using differentiable TPS modules in GAN framework to align facial geometry between photos and style samples, achieving better consistency, efficiency, and mobile deployment.
Details
Motivation: Existing portrait stylization methods struggle with geometric consistency and satisfactory stylization effects due to disparity in facial feature distributions between photos and stylized images, limiting application on rare styles and mobile devices.Method: Integrate differentiable Thin-Plate-Spline (TPS) modules into end-to-end GAN framework to establish geometric correlations between portraits and style samples by aligning facial characteristics using facial landmarks, working at global and local scales in both pixel and feature spaces.
Result: Outperforms existing models in fidelity and stylistic consistency, achieves 2x training data efficiency, 100x less computational complexity, and real-time inference (30 FPS) at 512*512 resolution on mobile devices.
Conclusion: The proposed TPS-enhanced GAN framework effectively addresses geometric consistency challenges in portrait stylization, enabling efficient training and real-time mobile deployment while maintaining high-quality artistic effects.
Abstract: Portrait Stylization aims to imbue portrait photos with vivid artistic effects drawn from style examples. Despite the availability of enormous training datasets and large network weights, existing methods struggle to maintain geometric consistency and achieve satisfactory stylization effects due to the disparity in facial feature distributions between facial photographs and stylized images, limiting the application on rare styles and mobile devices. To alleviate this, we propose to establish meaningful geometric correlations between portraits and style samples to simplify the stylization by aligning corresponding facial characteristics. Specifically, we integrate differentiable Thin-Plate-Spline (TPS) modules into an end-to-end Generative Adversarial Network (GAN) framework to improve the training efficiency and promote the consistency of facial identities. By leveraging inherent structural information of faces, e.g., facial landmarks, TPS module can establish geometric alignments between the two domains, at global and local scales, both in pixel and feature spaces, thereby overcoming the aforementioned challenges. Quantitative and qualitative comparisons on a range of portrait stylization tasks demonstrate that our models not only outperforms existing models in terms of fidelity and stylistic consistency, but also achieves remarkable improvements in 2x training data efficiency and 100x less computational complexity, allowing our lightweight model to achieve real-time inference (30 FPS) at 512*512 resolution on mobile devices.
[299] TKN: Transformer-based Keypoint Prediction Network For Real-time Video Prediction
Haoran Li, XiaoLu Li, Yihang Lin, Yanbin Hao, Haiyong Xie, Pengyuan Zhou, Yong Liao
Main category: cs.CV
TL;DR: TKN is a transformer-based keypoint prediction network for real-time video prediction that achieves 1,176 fps by extracting dynamic content unsupervised, using acceleration matrices for attention computation, and employing parallel computing.
Details
Motivation: Traditional video prediction methods prioritize accuracy over speed, suffering from slow prediction rates due to complex structures, redundant information, and excessive GPU memory consumption. Sequential frame prediction makes acceleration difficult, limiting real-time applications like danger prediction and warning.Method: TKN extracts dynamic content from video frames in an unsupervised manner to reduce redundant feature computation. It uses an acceleration matrix to reduce computational cost of attention mechanisms and employs a parallel computing structure for prediction acceleration.
Result: TKN achieves a prediction rate of 1,176 fps, significantly reducing computation costs while maintaining performance. Qualitative and quantitative experiments on multiple datasets demonstrate its superiority.
Conclusion: TKN is the first real-time video prediction solution with high speed and reduced computation costs, showing great application potential for real-time scenarios.
Abstract: Video prediction is a complex time-series forecasting task with great potential in many use cases. However, traditional methods prioritize accuracy and overlook slow prediction speeds due to complex model structures, redundant information, and excessive GPU memory consumption. These methods often predict frames sequentially, making acceleration difficult and limiting their applicability in real-time scenarios like danger prediction and warning.Therefore, we propose a transformer-based keypoint prediction neural network (TKN). TKN extracts dynamic content from video frames in an unsupervised manner, reducing redundant feature computation. And, TKN uses an acceleration matrix to reduce the computational cost of attention and employs a parallel computing structure for prediction acceleration. To the best of our knowledge, TKN is the first real-time video prediction solution that achieves a prediction rate of 1,176 fps, significantly reducing computation costs while maintaining other performance. Qualitative and quantitative experiments on multiple datasets have demonstrated the superiority of our method, suggesting that TKN has great application potential.
[300] A Survey on Generative Modeling with Limited Data, Few Shots, and Zero Shot
Milad Abdollahzadeh, Guimeng Liu, Touba Malekzadeh, Christopher T. H. Teo, Keshigeyan Chandrasegaran, Ngai-Man Cheung
Main category: cs.CV
TL;DR: Survey paper on generative modeling under data constraints (GM-DC), covering limited-data, few-shot, and zero-shot settings with taxonomies for tasks and methods.
Details
Motivation: Real-world applications often face data scarcity (medicine, satellite imaging, art) but conventional generative models require large datasets, creating a need for systematic understanding of generative modeling under data constraints.Method: Comprehensive survey of 230+ papers with two novel taxonomies: one for GM-DC tasks (unconditional/conditional generation, cross-domain adaptation, subject-driven modeling) and another for methodological approaches (transfer learning, data augmentation, meta-learning, frequency-aware modeling).
Result: Systematic analysis of GM-DC field including task-approach-method interactions via Sankey diagram, identification of key challenges (overfitting, frequency bias, knowledge transfer), and comprehensive review across model types and constraint scenarios.
Conclusion: Provides practical roadmap for researchers/practitioners with future directions including foundation model adaptation, holistic evaluation frameworks, and data-centric sample selection strategies.
Abstract: Generative modeling in machine learning aims to synthesize new data samples that are statistically similar to those observed during training. While conventional generative models such as GANs and diffusion models typically assume access to large and diverse datasets, many real-world applications (e.g. in medicine, satellite imaging, and artistic domains) operate under limited data availability and strict constraints. In this survey, we examine Generative Modeling under Data Constraint (GM-DC), which includes limited-data, few-shot, and zero-shot settings. We present a unified perspective on the key challenges in GM-DC, including overfitting, frequency bias, and incompatible knowledge transfer, and discuss how these issues impact model performance. To systematically analyze this growing field, we introduce two novel taxonomies: one categorizing GM-DC tasks (e.g. unconditional vs. conditional generation, cross-domain adaptation, and subject-driven modeling), and another organizing methodological approaches (e.g. transfer learning, data augmentation, meta-learning, and frequency-aware modeling). Our study reviews over 230 papers, offering a comprehensive view across generative model types and constraint scenarios. We further analyze task-approach-method interactions using a Sankey diagram and highlight promising directions for future work, including adaptation of foundation models, holistic evaluation frameworks, and data-centric strategies for sample selection. This survey provides a timely and practical roadmap for researchers and practitioners aiming to advance generative modeling under limited data. Project website: https://sutd-visual-computing-group.github.io/gmdc-survey/.
[301] Efficiently Assemble Normalization Layers and Regularization for Federated Domain Generalization
Khiem Le, Long Ho, Cuong Do, Danh Le-Phuoc, Kok-Seng Wong
Main category: cs.CV
TL;DR: gPerXAN: A novel federated domain generalization method using personalized normalization and regularization to handle domain shift while preserving privacy and reducing communication costs.
Details
Motivation: Address domain shift in federated learning where models degrade on unseen domains, while avoiding privacy risks and high communication/computation costs of existing FedDG methods.Method: Uses Personalized eXplicitly Assembled Normalization to filter domain-specific features while retaining discrimination, plus a regularizer to guide models in capturing domain-invariant representations for the global classifier.
Result: Outperforms existing methods on benchmark datasets (PACS, Office-Home) and real-world medical dataset (Camelyon17).
Conclusion: gPerXAN effectively addresses federated domain generalization with better privacy preservation and lower costs than previous approaches.
Abstract: Domain shift is a formidable issue in Machine Learning that causes a model to suffer from performance degradation when tested on unseen domains. Federated Domain Generalization (FedDG) attempts to train a global model using collaborative clients in a privacy-preserving manner that can generalize well to unseen clients possibly with domain shift. However, most existing FedDG methods either cause additional privacy risks of data leakage or induce significant costs in client communication and computation, which are major concerns in the Federated Learning paradigm. To circumvent these challenges, here we introduce a novel architectural method for FedDG, namely gPerXAN, which relies on a normalization scheme working with a guiding regularizer. In particular, we carefully design Personalized eXplicitly Assembled Normalization to enforce client models selectively filtering domain-specific features that are biased towards local data while retaining discrimination of those features. Then, we incorporate a simple yet effective regularizer to guide these models in directly capturing domain-invariant representations that the global model’s classifier can leverage. Extensive experimental results on two benchmark datasets, i.e., PACS and Office-Home, and a real-world medical dataset, Camelyon17, indicate that our proposed method outperforms other existing methods in addressing this particular problem.
[302] Story-Iter: A Training-free Iterative Paradigm for Long Story Visualization
Jiawei Mao, Xiaoke Huang, Yunfei Xie, Yuanqi Chang, Mude Hui, Bingjie Xu, Zeyu Zheng, Zirui Wang, Cihang Xie, Yuyin Zhou
Main category: cs.CV
TL;DR: Story-Iter is a training-free iterative paradigm for long-story generation that uses external iterative refinement with a global reference cross-attention module to enhance semantic consistency across up to 100 frames.
Details
Motivation: Existing story generation methods rely on fixed reference images and lack mechanisms for continuous refinement across long sequences, leading to semantic inconsistencies in long-story visualization.Method: Proposes Story-Iter with a plug-and-play, training-free global reference cross-attention (GRCA) module that models all reference frames with global embeddings, enabling iterative refinement by incorporating all previous round’s reference images.
Result: Achieves state-of-the-art performance on official story visualization datasets and long story benchmarks (up to 100 frames), excelling in both semantic consistency and fine-grained interactions.
Conclusion: The external iterative paradigm with GRCA enables precise long-story generation with improved consistency and interactions without requiring training.
Abstract: This paper introduces Story-Iter, a new training-free iterative paradigm to enhance long-story generation. Unlike existing methods that rely on fixed reference images to construct a complete story, our approach features a novel external iterative paradigm, extending beyond the internal iterative denoising steps of diffusion models, to continuously refine each generated image by incorporating all reference images from the previous round. To achieve this, we propose a plug-and-play, training-free global reference cross-attention (GRCA) module, modeling all reference frames with global embeddings, ensuring semantic consistency in long sequences. By progressively incorporating holistic visual context and text constraints, our iterative paradigm enables precise generation with fine-grained interactions, optimizing the story visualization step-by-step. Extensive experiments in the official story visualization dataset and our long story benchmark demonstrate that Story-Iter’s state-of-the-art performance in long-story visualization (up to 100 frames) excels in both semantic consistency and fine-grained interactions.
[303] Zooming without Zooming: Region-to-Image Distillation for Fine-Grained Multimodal Perception
Lai Wei, Liangbo He, Jun Lan, Lingzhong Dong, Yutong Cai, Siyuan Li, Huijia Zhu, Weiqiang Wang, Linghe Kong, Yue Wang, Zhuosheng Zhang, Weiran Huang
Main category: cs.CV
TL;DR: Region-to-Image Distillation transforms agentic zooming from inference-time tool use into training-time supervision, enabling MLLMs to achieve fine-grained perception in a single forward pass without repeated zooming operations.
Details
Motivation: MLLMs struggle with fine-grained perception where small details are overwhelmed by global context. Current "Thinking-with-Images" methods that iteratively zoom in/out during inference incur high latency due to repeated tool calls and visual re-encoding.Method: Proposes Region-to-Image Distillation: 1) Zoom into micro-cropped regions using strong teacher models to generate high-quality VQA data, 2) Distill this region-grounded supervision back to the full image, 3) Train student models on this data to internalize zooming benefits into single forward pass. Also introduces ZoomBench benchmark with 845 VQA data across six fine-grained perceptual dimensions.
Result: Models achieve leading performance across multiple fine-grained perception benchmarks, improve general multimodal cognition on visual reasoning and GUI agent benchmarks, and provide insights on when “Thinking-with-Images” is necessary vs when gains can be distilled.
Conclusion: Region-to-Image Distillation effectively internalizes the benefits of agentic zooming into MLLMs, enabling efficient fine-grained perception without inference-time tool use, while maintaining strong performance on general multimodal tasks.
Abstract: Multimodal Large Language Models (MLLMs) excel at broad visual understanding but still struggle with fine-grained perception, where decisive evidence is small and easily overwhelmed by global context. Recent “Thinking-with-Images” methods alleviate this by iteratively zooming in and out regions of interest during inference, but incur high latency due to repeated tool calls and visual re-encoding. To address this, we propose Region-to-Image Distillation, which transforms zooming from an inference-time tool into a training-time primitive, thereby internalizing the benefits of agentic zooming into a single forward pass of an MLLM. In particular, we first zoom in to micro-cropped regions to let strong teacher models generate high-quality VQA data, and then distill this region-grounded supervision back to the full image. After training on such data, the smaller student model improves “single-glance” fine-grained perception without tool use. To rigorously evaluate this capability, we further present ZoomBench, a hybrid-annotated benchmark of 845 VQA data spanning six fine-grained perceptual dimensions, together with a dual-view protocol that quantifies the global–regional “zooming gap”. Experiments show that our models achieve leading performance across multiple fine-grained perception benchmarks, and also improve general multimodal cognition on benchmarks such as visual reasoning and GUI agents. We further discuss when “Thinking-with-Images” is necessary versus when its gains can be distilled into a single forward pass. Our code is available at https://github.com/inclusionAI/Zooming-without-Zooming.
[304] LVLM-COUNT: Enhancing the Counting Ability of Large Vision-Language Models
Muhammad Fetrat Qharabagh, Mohammadreza Ghofrani, Kimon Fountoulakis
Main category: cs.CV
TL;DR: LVLMs struggle with visual counting tasks, especially for large numbers of objects. A divide-and-conquer approach with anti-splitting mechanism improves counting performance.
Details
Motivation: Large vision-language models have advanced visual perception but perform poorly on counting tasks, particularly as object numbers increase. This limits their practical application in real-world visual tasks requiring accurate counting.Method: Proposes a divide-and-conquer baseline method that decomposes counting problems into sub-tasks. Includes a mechanism to prevent object splitting during division to avoid repetitive counting issues common in naive implementations.
Result: The approach demonstrates effectiveness across various datasets and benchmarks, establishing it as a valuable reference for evaluating future counting solutions in LVLMs.
Conclusion: While LVLMs struggle with counting, especially for large object numbers, a simple divide-and-conquer approach with anti-splitting mechanisms can significantly improve their counting capabilities, providing a baseline for future research.
Abstract: Counting is a fundamental operation for various real-world visual tasks, requiring both object recognition and robust counting capabilities. Despite their advanced visual perception, large vision-language models (LVLMs) are known to struggle with counting tasks. In this work, we evaluate the performance of several LVLMs on visual counting tasks across multiple counting and vision datasets. We observe that while their performance may be less prone to error for small numbers of objects, they exhibit significant weaknesses as the number of objects increases. To alleviate this issue, we propose a simple yet effective baseline method that enhances LVLMs’ counting ability for large numbers of objects using a divide-and-conquer approach. Our method decomposes counting problems into sub-tasks. Moreover, it incorporates a mechanism to prevent objects from being split during division, which could otherwise lead to repetitive counting – a common issue in a naive divide-and-conquer implementation. We demonstrate the effectiveness of this approach across various datasets and benchmarks, establishing it as a valuable reference for evaluating future solutions.
[305] Are foundation models for computer vision good conformal predictors?
Leo Fillioux, Julio Silva-Rodríguez, Ismail Ben Ayed, Paul-Henry Cournède, Maria Vakalopoulou, Stergios Christodoulidis, Jose Dolz
Main category: cs.CV
TL;DR: Vision foundation models work well with conformal prediction for uncertainty quantification, with Vision Transformers performing best, though calibration can reduce efficiency and few-shot adaptation improves VLMs.
Details
Motivation: Foundation models are increasingly used in risk-sensitive applications but their uncertainty modeling capabilities are not well understood. The paper aims to evaluate vision and vision-language foundation models under conformal prediction frameworks to ensure safe deployment.Method: Evaluated popular vision foundation models using three conformal prediction methods across vision classification benchmarks. Tested Vision Transformers, calibrated confidence predictions, few-shot adaptation of VLMs, and examined APS method performance.
Result: Foundation models are well-suited for conformalization, especially Vision Transformers. Calibration degrades efficiency of conformal sets. Few-shot adaptation improves VLMs over zero-shot. APS maintains coverage guarantees in realistic scenarios.
Conclusion: Vision foundation models can be effectively integrated with conformal prediction for uncertainty quantification, with specific recommendations for model selection and adaptation strategies to ensure safe deployment in high-stakes applications.
Abstract: Recent advances in self-supervision and contrastive learning have brought the performance of foundation models to unprecedented levels in a variety of tasks. Fueled by this progress, these models are becoming the prevailing approach for a wide array of real-world vision problems, including risk-sensitive and high-stakes applications. However, ensuring safe deployment in these scenarios requires a more comprehensive understanding of their uncertainty modeling capabilities, which has received little attention. In this work, we delve into the behaviour of vision and vision-language foundation models under Conformal Prediction (CP), a statistical framework that provides theoretical guarantees of marginal coverage of the true class. Across extensive experiments including popular vision classification benchmarks, well-known foundation vision models, and three CP methods, our findings reveal that foundation models are well-suited for conformalization procedures, particularly those integrating Vision Transformers. We also show that calibrating the confidence predictions of these models, a popular strategy to improve their uncertainty quantification, actually leads to efficiency degradation of the conformal set on adaptive CP methods. Furthermore, few-shot adaptation of Vision-Language Models (VLMs) to downstream tasks, whose popularity is surging, enhances conformal scores compared to zero-shot predictions. Last, our empirical study exposes APS as particularly promising in the context of vision foundation models, as it does not violate the marginal coverage guarantees across multiple challenging, yet realistic scenarios.
[306] StrandHead: Text to Hair-Disentangled 3D Head Avatars Using Human-Centric Priors
Xiaokun Sun, Zeyu Cai, Ying Tai, Jian Yang, Zhenyu Zhang
Main category: cs.CV
TL;DR: StrandHead is a text-driven method for generating 3D hair strands and disentangled head avatars with strand-level attributes, using 2D generative model distillation and strand-guided meshing without large-scale hair-text paired data.
Details
Motivation: Existing avatar generation methods fail to model practical hair due to data limitations or entangled representations. Haircut indicates distinct personality, but current approaches can't generate realistic 3D hair strands from text prompts.Method: Proposes StrandHead that distills 2D generative models pre-trained on human mesh data to generate realistic hair strands from prompts. Uses strand-guided meshing to ensure gradient flow from distillation objective to neural strand representation, with regularization using statistically significant haircut features for stable optimization.
Result: Achieves state-of-the-art performance on text-to-strand generation and disentangled 3D head avatar modeling. Generated 3D hair can be applied to avatars for strand-level editing and implemented in graphics engines for physical simulation.
Conclusion: StrandHead enables realistic 3D hair strand generation from text using 2D/3D human-centric priors without large-scale hair-text paired data, supporting strand-level editing and graphics applications.
Abstract: While haircut indicates distinct personality, existing avatar generation methods fail to model practical hair due to the data limitation or entangled representation. We propose StrandHead, a novel text-driven method capable of generating 3D hair strands and disentangled head avatars with strand-level attributes. Instead of using large-scale hair-text paired data for supervision, we demonstrate that realistic hair strands can be generated from prompts by distilling 2D generative models pre-trained on human mesh data. To this end, we propose a meshing approach guided by strand geometry to guarantee the gradient flow from the distillation objective to the neural strand representation. The optimization is then regularized by statistically significant haircut features, leading to stable updating of strands against unreasonable drifting. These employed 2D/3D human-centric priors contribute to text-aligned and realistic 3D strand generation. Extensive experiments show that StrandHead achieves the state-of-the-art performance on text to strand generation and disentangled 3D head avatar modeling. The generated 3D hair can be applied on avatars for strand-level editing, as well as implemented in the graphics engine for physical simulation or other applications. Project page: https://xiaokunsun.github.io/StrandHead.github.io/.
[307] TRecViT: A Recurrent Video Transformer
Viorica Pătrăucean, Xu Owen He, Joseph Heyward, Chuhan Zhang, Mehdi S. M. Sajjadi, George-Cristian Muraru, Artem Zholus, Mahdi Karami, Ross Goroshin, Yutian Chen, Simon Osindero, João Carreira, Razvan Pascanu
Main category: cs.CV
TL;DR: TRecViT: A causal video modeling architecture using time-space-channel factorization with LRUs for temporal mixing, self-attention for spatial mixing, and MLPs for channel mixing, achieving state-of-the-art performance with high efficiency.
Details
Motivation: To develop an efficient causal video model that can handle both sparse and dense video tasks while being computationally efficient and suitable for real-time applications, addressing limitations of existing non-causal and causal video models.Method: Proposes TRecViT with time-space-channel factorization: uses gated linear recurrent units (LRUs) for temporal information mixing, self-attention layers for spatial mixing, and MLPs for channel mixing. The architecture is causal and operates in supervised or self-supervised regimes.
Result: Outperforms or matches ViViT-L on SSv2 and Kinetics400 with 3× fewer parameters, 12× smaller memory footprint, 5× lower FLOPs, and 300 FPS inference. Achieves SOTA on SSv2 compared to causal transformer models (TSM, RViT) and recurrent models like LSTM.
Conclusion: TRecViT demonstrates that causal video modeling can achieve competitive performance with non-causal models while being significantly more efficient, enabling real-time video understanding applications.
Abstract: We propose a novel block for \emph{causal} video modelling. It relies on a time-space-channel factorisation with dedicated blocks for each dimension: gated linear recurrent units (LRUs) perform information mixing over time, self-attention layers perform mixing over space, and MLPs over channels. The resulting architecture \emph{TRecViT} is causal and shows strong performance on sparse and dense tasks, trained in supervised or self-supervised regimes, being the first causal video model in the state-space models family. Notably, our model outperforms or is on par with the popular (non-causal) ViViT-L model on large scale video datasets (SSv2, Kinetics400), while having $3\times$ less parameters, $12\times$ smaller memory footprint, and $5\times$ lower FLOPs count than the full self-attention ViViT, with an inference throughput of about 300 frames per second, running comfortably in real-time. When compared with causal transformer-based models (TSM, RViT) and other recurrent models like LSTM, TRecViT obtains state-of-the-art results on the challenging SSv2 dataset. Code and checkpoints are available online https://github.com/google-deepmind/trecvit.
[308] Cross-Modal Mapping: Mitigating the Modality Gap for Few-Shot Image Classification
Xi Yang, Pai Peng, Wulin Xie, Xiaohuan Lu, Jie Wen
Main category: cs.CV
TL;DR: CMM method aligns image and text features in CLIP to reduce modality gap for better few-shot classification
Details
Motivation: Existing few-shot methods using CLIP suffer from modality gap between image and text features, leading to suboptimal performance when using text features as class prototypesMethod: Cross-Modal Mapping (CMM) with global linear transformation to align image features to text space, plus local triplet loss to optimize spatial relationships
Result: Improves average Top-1 accuracy by 1.06% on 11 benchmark datasets, performs well on distribution shift datasets, simplifies training with higher efficiency
Conclusion: CMM effectively mitigates modality gap in pre-trained models, enabling text features to serve as effective class prototypes for efficient and generalizable few-shot learning
Abstract: Few-shot image classification remains a critical challenge in the field of computer vision, particularly in data-scarce environments. Existing methods typically rely on pre-trained visual-language models, such as CLIP. However, due to the modality gap, which is the inconsistent distribution of image and text features in the joint embedding space, directly using these features as class prototypes often leads to suboptimal performance. To address this issue, we propose a novel Cross-Modal Mapping (CMM) method. This method globally aligns image features with the text feature space through linear transformation and optimizes their local spatial relationships using triplet loss, thereby significantly enhancing cross-modal consistency. Experimental results show that compared to other methods, CMM simplifies the training process and demonstrates higher efficiency. Furthermore, CMM improves the average Top-1 accuracy by 1.06% on 11 benchmark datasets compared to methods that partially fine-tune the backbone, and it performs excellently on 4 distribution shift datasets. Notably, CMM effectively mitigates the modality gap in pre-trained models, enabling text features to serve as effective class prototypes for image features, thus providing an efficient and highly generalizable solution for few-shot learning.
[309] Dataset Distillation via Committee Voting
Jiacheng Cui, Zhaoyi Li, Xiaochen Ma, Xinyue Bi, Yaxin Luo, Zhiqiang Shen
Main category: cs.CV
TL;DR: CV-DD introduces a committee voting approach for dataset distillation that leverages multiple models’ collective knowledge to produce higher-quality distilled data with better generalization.
Details
Motivation: Existing dataset distillation methods focus on data-synthetic alignment or scaling to large datasets, but lack mechanisms to capture broader data characteristics and reduce model-specific bias.Method: Proposes Committee Voting for Dataset Distillation (CV-DD) that integrates distributions and predictions from multiple models, generates high-quality soft labels, and uses voting-based strategy to enhance diversity and robustness.
Result: CV-DD consistently outperforms single- and multi-model distillation methods across multiple datasets and IPC settings, and generalizes well to non-training-based frameworks and synthetic-to-real transfer tasks.
Conclusion: The committee voting approach effectively captures broader data characteristics, reduces model bias, improves generalization, and demonstrates strong performance across various settings.
Abstract: Dataset distillation aims to synthesize a compact yet representative dataset that preserves the essential characteristics of the original data for efficient model training. Existing methods mainly focus on improving data-synthetic alignment or scaling distillation to large datasets. In this work, we propose $\textbf{C}$ommittee $\textbf{V}$oting for $\textbf{D}$ataset $\textbf{D}$istillation ($\textbf{CV-DD}$), an orthogonal approach that leverages the collective knowledge of multiple models to produce higher-quality distilled data. We first establish a strong baseline that achieves state-of-the-art performance through modern architectural and optimization choices. By integrating distributions and predictions from multiple models and generating high-quality soft labels, our method captures a broader range of data characteristics, reduces model-specific bias and the impact of distribution shifts, and significantly improves generalization. This voting-based strategy enhances diversity and robustness, alleviates overfitting, and improves post-evaluation performance. Extensive experiments across multiple datasets and IPC settings demonstrate that CV-DD consistently outperforms single- and multi-model distillation methods and generalizes well to non-training-based frameworks and challenging synthetic-to-real transfer tasks. Code is available at: https://github.com/Jiacheng8/CV-DD.
[310] V2V-LLM: Vehicle-to-Vehicle Cooperative Autonomous Driving with Multimodal Large Language Models
Hsu-kuang Chiu, Ryo Hachiuma, Chien-Yi Wang, Stephen F. Smith, Yu-Chiang Frank Wang, Min-Hung Chen
Main category: cs.CV
TL;DR: A cooperative autonomous driving framework that integrates Multimodal LLMs to fuse perception information from multiple connected vehicles via V2V communication, enabling various driving-related question answering tasks including grounding, object identification, and planning.
Details
Motivation: Current autonomous driving systems rely on individual vehicle sensors which can fail or be occluded. While cooperative perception methods exist, they focus mainly on detection/tracking rather than overall planning performance. The paper aims to explore how LLMs can enhance cooperative autonomous driving through multimodal information fusion.Method: Proposes V2V-LLM, a baseline method using Large Language Models to fuse perception information from multiple connected autonomous vehicles (CAVs) via V2V communication. Creates the V2V-QA dataset and benchmark for cooperative driving tasks including grounding, notable object identification, and planning.
Result: Experimental results show V2V-LLM outperforms other baseline methods with different fusion approaches, demonstrating it as a promising unified model architecture for various cooperative autonomous driving tasks.
Conclusion: The work creates a new research direction for improving autonomous driving safety through multimodal LLM-based cooperative systems, with plans to release code and data to facilitate open-source research.
Abstract: Current autonomous driving vehicles rely mainly on their individual sensors to understand surrounding scenes and plan for future trajectories, which can be unreliable when the sensors are malfunctioning or occluded. To address this problem, cooperative perception methods via vehicle-to-vehicle (V2V) communication have been proposed, but they have tended to focus on perception tasks like detection or tracking. How those approaches contribute to overall cooperative planning performance is still under-explored. Inspired by recent progress using Large Language Models (LLMs) to build autonomous driving systems, we propose a novel problem setting that integrates a Multimodal LLM into cooperative autonomous driving, with the proposed Vehicle-to-Vehicle Question-Answering (V2V-QA) dataset and benchmark. We also propose our baseline method Vehicle-to-Vehicle Multimodal Large Language Model (V2V-LLM), which uses an LLM to fuse perception information from multiple connected autonomous vehicles (CAVs) and answer various types of driving-related questions: grounding, notable object identification, and planning. Experimental results show that our proposed V2V-LLM can be a promising unified model architecture for performing various tasks in cooperative autonomous driving, and outperforms other baseline methods that use different fusion approaches. Our work also creates a new research direction that can improve the safety of future autonomous driving systems. The code and data will be released to the public to facilitate open-source research in this field. Our project website: https://eddyhkchiu.github.io/v2vllm.github.io/ .
[311] Simulating the Real World: A Unified Survey of Multimodal Generative Models
Yuqi Hu, Longguang Wang, Xian Liu, Ling-Hao Chen, Yuwei Guo, Yukai Shi, Ce Liu, Anyi Rao, Zeyu Wang, Hui Xiong
Main category: cs.CV
TL;DR: A comprehensive survey paper that systematically unifies multimodal generative models across 2D, video, 3D, and 4D generation for real-world simulation, bridging different data dimensionalities within a single framework.
Details
Motivation: Current AI approaches treat different modalities (2D images, videos, 3D, 4D) as independent domains, overlooking their interdependencies and failing to systematically integrate connections between different dimensions of reality for comprehensive real-world simulation.Method: The survey organizes multimodal generative models by data dimensionality progression: starting from 2D generation (appearance), moving to video (appearance+dynamics), then 3D (appearance+geometry), and culminating in 4D generation that integrates all dimensions.
Result: Provides a systematic unification of 2D, video, 3D, and 4D generation within a single framework, along with comprehensive reviews of datasets, evaluation metrics, and future research directions.
Conclusion: This survey serves as a bridge to advance multimodal generative models and real-world simulation research by providing a unified framework that connects different data dimensionalities and modalities.
Abstract: Understanding and replicating the real world is a critical challenge in Artificial General Intelligence (AGI) research. To achieve this, many existing approaches, such as world models, aim to capture the fundamental principles governing the physical world, enabling more accurate simulations and meaningful interactions. However, current methods often treat different modalities, including 2D (images), videos, 3D, and 4D representations, as independent domains, overlooking their interdependencies. Additionally, these methods typically focus on isolated dimensions of reality without systematically integrating their connections. In this survey, we present a unified survey for multimodal generative models that investigate the progression of data dimensionality in real-world simulation. Specifically, this survey starts from 2D generation (appearance), then moves to video (appearance+dynamics) and 3D generation (appearance+geometry), and finally culminates in 4D generation that integrate all dimensions. To the best of our knowledge, this is the first attempt to systematically unify the study of 2D, video, 3D and 4D generation within a single framework. To guide future research, we provide a comprehensive review of datasets, evaluation metrics and future directions, and fostering insights for newcomers. This survey serves as a bridge to advance the study of multimodal generative models and real-world simulation within a unified framework.
[312] Measure Twice, Cut Once: A Semantic-Oriented Approach to Video Temporal Localization with Video LLMs
Zongshang Pang, Mayu Otani, Yuta Nakashima
Main category: cs.CV
TL;DR: MeCo: A semantic-driven framework for temporal video localization using video LLMs without timestamp generation, employing structural token generation, query-focused captioning, and contrastive grounding.
Details
Motivation: Current video LLM methods for temporal localization struggle to leverage LLMs' semantic understanding capabilities because they generate uninformative timestamp outputs. The authors propose a semantic-oriented approach that better utilizes LLMs' pre-trained knowledge.Method: Three-task framework: 1) Structural token generation to partition videos into segments and categorize them as target events or background; 2) Query-focused captioning to extract fine-grained event semantics; 3) Structural token grounding module using contrastive learning to associate tokens with video segments.
Result: Extensive experiments show MeCo consistently outperforms timestamp-based methods across diverse temporal localization tasks, demonstrating the effectiveness of semantic-driven approaches.
Conclusion: Semantic-driven frameworks without timestamp generation better leverage video LLMs’ capabilities for temporal localization, offering a promising direction for video understanding tasks.
Abstract: Temporally localizing user-queried events through natural language is a crucial capability for video models. Recent methods predominantly adapt video LLMs to generate event boundary timestamps for temporal localization tasks, which struggle to leverage LLMs’ pre-trained semantic understanding capabilities due to the uninformative nature of timestamp outputs. In this work, we explore a timestamp-free, semantic-oriented framework that fine-tunes video LLMs using two generative learning tasks and one discriminative learning task. We first introduce a structural token generation task that enables the video LLM to recognize the temporal structure of input videos based on the input query. Through this task, the video LLM generates a sequence of special tokens, called structural tokens, which partition the video into consecutive segments and categorize them as either target events or background transitions. To enhance precise recognition of event segments, we further propose a query-focused captioning task that enables the video LLM to extract fine-grained event semantics that can be effectively utilized by the structural tokens. Finally, we introduce a structural token grounding module driven by contrastive learning to associate each structural token with its corresponding video segment, achieving holistic temporal segmentation of the input video and readily yielding the target event segments for localization. Extensive experiments across diverse temporal localization tasks demonstrate that our proposed framework, MeCo, consistently outperforms methods relying on boundary timestamp generation, highlighting the potential of a semantic-driven approach for temporal localization with video LLMs \footnote{Code available at https://github.com/pangzss/MeCo.
[313] Autoregressive Image Generation with Randomized Parallel Decoding
Haopeng Li, Jinyue Yang, Guoqi Li, Huan Wang
Main category: cs.CV
TL;DR: ARPG is a visual autoregressive model that enables randomized parallel generation by decoupling positional guidance from content representation, allowing flexible token generation order and efficient parallel inference.
Details
Motivation: Conventional raster-order autoregressive models suffer from sequential, predefined token generation order that limits inference efficiency and zero-shot generalization capabilities.Method: Proposes a decoupled decoding framework that separates positional guidance (queries) from content representation (key-value pairs), enabling random-order training and generation through direct incorporation into causal attention mechanism.
Result: Achieves FID of 1.83 on ImageNet-1K 256 with only 32 sampling steps, achieving 30x speedup in inference and 75% reduction in memory consumption compared to similar-scale autoregressive models.
Conclusion: ARPG enables efficient parallel generation with flexible token order, supporting zero-shot tasks like inpainting, outpainting, and resolution expansion while significantly improving inference speed and memory efficiency.
Abstract: We introduce ARPG, a novel visual autoregressive model that enables randomized parallel generation, addressing the inherent limitations of conventional raster-order approaches, which hinder inference efficiency and zero-shot generalization due to their sequential, predefined token generation order. Our key insight is that effective random-order modeling necessitates explicit guidance for determining the position of the next predicted token. To this end, we propose a novel decoupled decoding framework that decouples positional guidance from content representation, encoding them separately as queries and key-value pairs. By directly incorporating this guidance into the causal attention mechanism, our approach enables fully random-order training and generation, eliminating the need for bidirectional attention. Consequently, ARPG readily generalizes to zero-shot inference tasks such as image inpainting, outpainting, and resolution expansion. Furthermore, it supports parallel inference by concurrently processing multiple queries using a shared KV cache. On the ImageNet-1K 256 benchmark, our approach attains an FID of 1.83 with only 32 sampling steps, achieving over a 30 times speedup in inference and a 75 percent reduction in memory consumption compared to representative recent autoregressive models at a similar scale.
[314] Car-1000: A New Large Scale Fine-Grained Visual Categorization Dataset
Yutao Hu, Sen Li, Jincheng Yan, Wenqi Shao, Xiaoyan Luo
Main category: cs.CV
TL;DR: Introduces Car-1000, a large-scale fine-grained visual categorization dataset with 1000 distinct car models from 166 automakers, addressing limitations of the outdated Stanford-Car dataset.
Details
Motivation: Stanford-Car dataset is outdated (pre-2013) with only 196 categories, failing to capture the evolving automotive landscape with increasingly intricate car designs. Need for a more comprehensive dataset to support modern fine-grained car recognition applications in autonomous driving, traffic surveillance, and scene understanding.Method: Created Car-1000 dataset with 1000 distinct car models from 166 automakers, providing a more comprehensive and up-to-date benchmark. Reproduced several state-of-the-art FGVC methods on this new dataset to establish performance baselines.
Result: Car-1000 dataset is publicly available and establishes new benchmarks for fine-grained car recognition. The dataset addresses the limitations of Stanford-Car by providing more categories and up-to-date vehicle models.
Conclusion: Car-1000 provides a fresh perspective for FGVC research with a comprehensive, up-to-date dataset that better reflects the current automotive landscape, enabling more robust car recognition systems for real-world applications.
Abstract: Fine-grained visual categorization (FGVC) is a challenging but significant task in computer vision, which aims to recognize different sub-categories of birds, cars, airplanes, etc. Among them, recognizing models of different cars has significant application value in autonomous driving, traffic surveillance and scene understanding, which has received considerable attention in the past few years. However, Stanford-Car, the most widely used fine-grained dataset for car recognition, only has 196 different categories and only includes vehicle models produced earlier than 2013. Due to the rapid advancements in the automotive industry during recent years, the appearances of various car models have become increasingly intricate and sophisticated. Consequently, the previous Stanford-Car dataset fails to capture this evolving landscape and cannot satisfy the requirements of automotive industry. To address these challenges, in our paper, we introduce Car-1000, a large-scale dataset designed specifically for fine-grained visual categorization of diverse car models. Car-1000 encompasses vehicles from 166 different automakers, spanning a wide range of 1000 distinct car models. Additionally, we have reproduced several state-of-the-art FGVC methods on the Car-1000 dataset, establishing a new benchmark for research in this field. We hope that our work will offer a fresh perspective for future FGVC researchers. Our dataset is available at https://github.com/toggle1995/Car-1000.
[315] S2R-HDR: A Large-Scale Rendered Dataset for HDR Fusion
Yujin Wang, Jiarui Wu, Yichen Bian, Fan Zhang, Tianfan Xue
Main category: cs.CV
TL;DR: S2R-HDR: A large-scale synthetic dataset for HDR fusion using Unreal Engine 5 with 24,000 HDR samples, plus a domain adaptation method (S2R-Adapter) to bridge synthetic-real gap.
Details
Motivation: Collecting large-scale HDR images from dynamic scenes is costly and technically challenging, limiting the generalization of learning-based HDR fusion methods due to insufficient training data.Method: 1) Create S2R-HDR dataset using Unreal Engine 5 with diverse realistic HDR scenes covering various dynamic elements, motion types, and lighting. 2) Develop efficient rendering pipeline for realistic HDR images. 3) Introduce S2R-Adapter for domain adaptation to bridge synthetic-real gap.
Result: Experimental results on real-world datasets demonstrate state-of-the-art HDR fusion performance, showing the effectiveness of the synthetic dataset and domain adaptation approach.
Conclusion: S2R-HDR provides a valuable synthetic dataset solution for HDR fusion training, and S2R-Adapter effectively mitigates domain gap issues, enabling better generalization to real-world scenarios.
Abstract: The generalization of learning-based high dynamic range (HDR) fusion is often limited by the availability of training data, as collecting large-scale HDR images from dynamic scenes is both costly and technically challenging. To address these challenges, we propose S2R-HDR, the first large-scale high-quality synthetic dataset for HDR fusion, with 24,000 HDR samples. Using Unreal Engine 5, we design a diverse set of realistic HDR scenes that encompass various dynamic elements, motion types, high dynamic range scenes, and lighting. Additionally, we develop an efficient rendering pipeline to generate realistic HDR images. To further mitigate the domain gap between synthetic and real-world data, we introduce S2R-Adapter, a domain adaptation designed to bridge this gap and enhance the generalization ability of models. Experimental results on real-world datasets demonstrate that our approach achieves state-of-the-art HDR fusion performance. Dataset and code are available at https://openimaginglab.github.io/S2R-HDR.
[316] GaussianFormer3D: Multi-Modal Gaussian-based Semantic Occupancy Prediction with 3D Deformable Attention
Lingjun Zhao, Sizhe Wei, James Hays, Lu Gan
Main category: cs.CV
TL;DR: GaussianFormer3D: A multi-modal Gaussian-based semantic occupancy prediction framework using 3D deformable attention for autonomous driving, achieving SOTA performance with reduced memory consumption.
Details
Motivation: 3D semantic occupancy prediction is crucial for autonomous driving safety. While voxel-based representations are common, 3D Gaussians offer a more compact and continuous alternative. The paper aims to leverage multi-modal (LiDAR-camera) fusion for more accurate predictions while improving efficiency.Method: Proposes GaussianFormer3D framework with: 1) Voxel-to-Gaussian initialization strategy using LiDAR data for geometry priors, 2) LiDAR-guided 3D deformable attention mechanism to refine Gaussians using LiDAR-camera fusion features in lifted 3D space.
Result: Extensive experiments on real-world on-road and off-road autonomous driving datasets show state-of-the-art prediction performance with reduced memory consumption and improved efficiency.
Conclusion: GaussianFormer3D successfully demonstrates the effectiveness of 3D Gaussian representations with multi-modal fusion for semantic occupancy prediction, offering better performance and efficiency than traditional voxel-based methods.
Abstract: 3D semantic occupancy prediction is essential for achieving safe, reliable autonomous driving and robotic navigation. Compared to camera-only perception systems, multi-modal pipelines, especially LiDAR-camera fusion methods, can produce more accurate and fine-grained predictions. Although voxel-based scene representations are widely used for semantic occupancy prediction, 3D Gaussians have emerged as a continuous and significantly more compact alternative. In this work, we propose a multi-modal Gaussian-based semantic occupancy prediction framework utilizing 3D deformable attention, namely GaussianFormer3D. We introduce a voxel-to-Gaussian initialization strategy that provides 3D Gaussians with accurate geometry priors from LiDAR data, and design a LiDAR-guided 3D deformable attention mechanism to refine these Gaussians using LiDAR-camera fusion features in a lifted 3D space. Extensive experiments on real-world on-road and off-road autonomous driving datasets demonstrate that GaussianFormer3D achieves state-of-the-art prediction performance with reduced memory consumption and improved efficiency.
[317] Single Image Reflection Separation via Dual Prior Interaction Transformer
Yue Huang, Tianle Hu, Yu Chen, Zi’ang Li, Jie Wen, Xiaozhao Fang
Main category: cs.CV
TL;DR: Proposes a dual-prior interaction framework for single image reflection separation using lightweight transmission prior generation and effective prior fusion to improve performance in complex scenarios.
Details
Motivation: Existing methods combine general priors from pre-trained models with task-specific priors like text prompts and reflection detection, but the transmission prior (most direct task-specific prior for target transmission layer) hasn't been effectively modeled or utilized, limiting performance in complex scenarios.Method: 1) Local Linear Correction Network (LLCN) that finetunes pre-trained models based on physical constraint T=SI+B (S and B represent pixel-wise and channel-wise scaling/bias transformations) to generate high-quality transmission priors with minimal parameters. 2) Dual-Prior Interaction Transformer (DPIT) with dual-stream channel reorganization attention mechanism that reorganizes features from general and transmission priors for attention computation to achieve deep fusion of both priors.
Result: Experimental results on multiple benchmark datasets demonstrate state-of-the-art performance.
Conclusion: The proposed dual-prior interaction framework effectively models and utilizes transmission priors through lightweight generation and deep fusion, achieving superior performance in single image reflection separation.
Abstract: Single image reflection separation aims to separate the transmission and reflection layers from a mixed image. Existing methods typically combine general priors from pre-trained models with task-specific priors such as text prompts and reflection detection. However, the transmission prior, as the most direct task-specific prior for the target transmission layer, has not been effectively modeled or fully utilized, limiting performance in complex scenarios. To address this issue, we propose a dual-prior interaction framework based on lightweight transmission prior generation and effective prior fusion. First, we design a Local Linear Correction Network (LLCN) that finetunes pre-trained models based on the physical constraint T=SI+B, where S and B represent pixel-wise and channel-wise scaling and bias transformations. LLCN efficiently generates high-quality transmission priors with minimal parameters. Second, we construct a Dual-Prior Interaction Transformer (DPIT) that employs a dual-stream channel reorganization attention mechanism. By reorganizing features from general and transmission priors for attention computation, DPIT achieves deep fusion of both priors, fully exploiting their complementary information. Experimental results on multiple benchmark datasets demonstrate that the proposed method achieves state-of-the-art performance.
[318] Mitigating Pretraining-Induced Attention Asymmetry in 2D+ Electron Microscopy Image Segmentation
Zsófia Molnár, Gergely Szabó, András Horváth
Main category: cs.CV
TL;DR: RGB-pretrained vision models show channel asymmetry when applied to electron microscopy volumetric data, despite symmetric nature of EM slices; proposed uniform channel initialization fixes this bias while maintaining performance.
Details
Motivation: Vision models pretrained on RGB natural images are commonly reused for electron microscopy segmentation, but the RGB mapping imposes artificial channel semantics that misalign with the symmetric nature of volumetric EM data where all slices are homogeneous and equally important.Method: Used saliency-based attribution analysis to show RGB-pretrained models assign unequal importance to EM slices; proposed targeted modification of pretraining weights via uniform channel initialization to restore symmetric feature attribution while preserving pretraining benefits.
Result: Experiments on SNEMI, Lucchi and GF-PA66 datasets show substantial reduction in attribution bias without compromising segmentation accuracy, sometimes even improving performance.
Conclusion: Standard RGB pretraining introduces harmful inductive biases for symmetric volumetric data; uniform channel initialization effectively restores symmetry in feature attribution while maintaining transfer learning benefits for EM segmentation.
Abstract: Vision models pretrained on large-scale RGB natural image datasets are widely reused for electron microscopy image segmentation. In electron microscopy, volumetric data are acquired as serial sections and processed as stacks of adjacent grayscale slices, where neighboring slices provide symmetric contextual information for identifying features on the central slice. The common strategy maps such stacks to pseudo-RGB inputs to enable transfer learning from pretrained models. However, this mapping imposes channel-specific semantics inherited from natural images, even though electron microscopy slices are homogeneous in the modality and symmetric in their predictive roles. As a result, pretrained models may encode inductive biases that are misaligned with the inherent symmetry of volumetric electron microscopy data. In this work, it is demonstrated that RGB-pretrained models systematically assign unequal importance to individual input slices when applied to stacked electron microscopy data, despite the absence of any intrinsic channel ordering. Using saliency-based attribution analysis across multiple architectures, a consistent channel-level asymmetry was observed that persists after fine-tuning and affects model interpretability, even when segmentation performance is unchanged. To address this issue, a targeted modification of pretraining weights based on uniform channel initialization was proposed, which restores symmetric feature attribution while preserving the benefits of pretraining. Experiments on the SNEMI, Lucchi and GF-PA66 datasets confirm a substantial reduction in attribution bias without compromising or even improving segmentation accuracy.
[319] OmniEarth-Bench: Towards Holistic Evaluation of Earth’s Six Spheres and Cross-Spheres Interactions with Multimodal Observational Earth Data
Fengxiang Wang, Mingshuo Chen, Xuming He, Yi-Fan Zhang, Yueying Li, Feng Liu, Zijie Guo, Zhenghao Hu, Jiong Wang, Jingyi Xu, Zhangrui Li, Junchao Gong, Di Wang, Fenghua Ling, Ben Fei, Weijia Li, Long Lan, Wenjing Yang
Main category: cs.CV
TL;DR: OmniEarth-Bench is a comprehensive multimodal benchmark spanning all six Earth spheres with 29,855 expert-curated annotations across 109 tasks, revealing significant limitations in current MLLMs’ Earth-system understanding.
Details
Motivation: Existing Earth science benchmarks are limited in scope, covering only specific spheres (mainly atmosphere) with few tasks, lacking cross-sphere interactions, and suffering from narrow data sources and limited extensibility.Method: Developed a scalable, modular-topology data inference framework with multi-observation sources and expert-in-the-loop curation to create standardized annotations organized in a four-level hierarchy (Sphere, Scenario, Ability, Task).
Result: Created 29,855 expert-curated annotations across 109 tasks; evaluation of 9 state-of-the-art MLLMs showed all models struggling, with none reaching 35% accuracy, revealing systematic gaps in Earth-system cognitive ability.
Conclusion: OmniEarth-Bench provides the first comprehensive multimodal benchmark for Earth science, exposing significant limitations in current MLLMs’ ability to understand complex Earth systems and cross-sphere interactions.
Abstract: Existing benchmarks for multimodal learning in Earth science offer limited, siloed coverage of Earth’s spheres and their cross-sphere interactions, typically restricting evaluation to the human-activity sphere of atmosphere and to at most 16 tasks. These limitations: narrow-source heterogeneity (single/few data sources), constrained scientific granularity, and limited-sphere extensibility. Therefore, we introduce OmniEarth-Bench, the first multimodal benchmark that systematically spans all six spheres: atmosphere, lithosphere, oceanosphere, cryosphere, biosphere, and human-activity sphere, and cross-spheres. Built with a scalable, modular-topology data inference framework and native multi-observation sources and expert-in-the-loop curation, OmniEarth-Bench produces 29,855 standardized, expert-curated annotations. All annotations are organized into a four-level hierarchy (Sphere, Scenario, Ability, Task), encompassing 109 expert-curated evaluation tasks. Experiments on 9 state-of-the-art MLLMs reveal that even the most advanced models struggle with our benchmarks, where none of them reach 35% accuracy, revealing systematic gaps in Earth-system cognitive ability. The dataset and evaluation code were released at OmniEarth-Bench (https://anonymous.4open.science/r/OmniEarth-Bench-B1BD).
[320] Towards Sequence Modeling Alignment between Tokenizer and Autoregressive Model
Pingyu Wu, Kai Zhu, Yu Liu, Longxiang Tang, Jian Yang, Yansong Peng, Wei Zhai, Yang Cao, Zheng-Jun Zha
Main category: cs.CV
TL;DR: AliTok is a novel aligned tokenizer that creates forward-dependent token sequences for autoregressive image generation, resolving the misalignment between bidirectional image dependencies and unidirectional autoregressive models.
Details
Motivation: Autoregressive image generation faces a fundamental challenge: conventional image tokenizations have bidirectional dependencies, but autoregressive models are unidirectional (predict next token based on previous ones). This misalignment hampers performance.Method: AliTok uses a bidirectional encoder constrained by a causal decoder to produce token sequences with both semantic richness and forward-dependency. It incorporates prefix tokens and a two-stage training process to enhance reconstruction performance.
Result: A 177M parameter autoregressive model achieves gFID of 1.44 and IS of 319.5 on ImageNet-256. Scaling to 662M parameters reaches gFID of 1.28, surpassing SOTA diffusion methods with 10x faster sampling. On ImageNet-512, a 318M model achieves SOTA gFID of 1.39.
Conclusion: AliTok successfully resolves the misalignment problem in autoregressive image generation, enabling high-fidelity image generation with faster sampling than diffusion models while maintaining strong performance metrics.
Abstract: Autoregressive image generation aims to predict the next token based on previous ones. However, this process is challenged by the bidirectional dependencies inherent in conventional image tokenizations, which creates a fundamental misalignment with the unidirectional nature of autoregressive models. To resolve this, we introduce AliTok, a novel Aligned Tokenizer that alters the dependency structure of the token sequence. AliTok employs a bidirectional encoder constrained by a causal decoder, a design that compels the encoder to produce a token sequence with both semantic richness and forward-dependency. Furthermore, by incorporating prefix tokens and employing a two-stage tokenizer training process to enhance reconstruction performance, AliTok achieves high fidelity and predictability simultaneously. Building upon AliTok, a standard decoder-only autoregressive model with just 177M parameters achieves a gFID of 1.44 and an IS of 319.5 on ImageNet-256. Scaling to 662M, our model reaches a gFID of 1.28, surpassing the SOTA diffusion method with 10x faster sampling. On ImageNet-512, our 318M model also achieves a SOTA gFID of 1.39. Code and weights at https://github.com/ali-vilab/alitok.
[321] Prompts to Summaries: Zero-Shot Language-Guided Video Summarization
Mario Barbara, Alaa Maalouf
Main category: cs.CV
TL;DR: Zero-shot video summarization framework that converts video-language model captions into user-guided skims using LLM judging, without training data, matching supervised methods.
Details
Motivation: Need for flexible user-controllable video summarization tools that operate without training data, as existing methods either rely on domain-specific datasets or cannot incorporate natural language user intent.Method: Four-step pipeline: (1) segment video into scenes, (2) produce scene descriptions with memory-efficient batch prompting, (3) score scene importance with LLM via tailored prompts, (4) propagate scores to frames using consistency (temporal coherence) and uniqueness (novelty) metrics.
Result: Surpasses all prior data-hungry unsupervised methods on SumMe and TVSum, performs competitively on Query-Focused Video Summarization benchmark, and introduces VidSum-Reason dataset as first challenging baseline.
Conclusion: Pretrained multi-modal models, when orchestrated with principled prompting and score propagation, provide powerful foundation for universal, text-queryable video summarization without training data.
Abstract: The explosive growth of video data intensified the need for flexible user-controllable summarization tools that operate without training data. Existing methods either rely on domain-specific datasets, limiting generalization, or cannot incorporate user intent expressed in natural language. We introduce Prompts-to-Summaries: the first zero-shot, text-queryable video-summarizer that converts off-the-shelf video-language models (VidLMs) captions into user-guided skims via large-language-models (LLMs) judging, without the use of training data, beating unsupervised and matching supervised methods. Our pipeline (i) segments video into scenes, (ii) produces scene descriptions with a memory-efficient batch prompting scheme that scales to hours on a single GPU, (iii) scores scene importance with an LLM via tailored prompts, and (iv) propagates scores to frames using new consistency (temporal coherence) and uniqueness (novelty) metrics for fine-grained frame importance. On SumMe and TVSum, our approach surpasses all prior data-hungry unsupervised methods and performs competitively on the Query-Focused Video Summarization benchmark, where the competing methods require supervised frame-level importance. We release VidSum-Reason, a query-driven dataset featuring long-tailed concepts and multi-step reasoning, where our framework serves as the first challenging baseline. Overall, we demonstrate that pretrained multi-modal models, when orchestrated with principled prompting and score propagation, provide a powerful foundation for universal, text-queryable video summarization.
[322] Stretching Beyond the Obvious: A Gradient-Free Framework to Unveil the Hidden Landscape of Visual Invariance
Lorenzo Tausani, Paolo Muratore, Morgan B. Talbot, Giacomo Amerio, Gabriel Kreiman, Davide Zoccolan
Main category: cs.CV
TL;DR: SnS is a framework to characterize visual unit invariance and adversarial vulnerability by optimizing image perturbations that stretch representations while squeezing unit activation (or vice versa).
Details
Motivation: Existing feature visualization approaches only show what maximally excites units, but don't reveal the manifold of transformations under which responses remain invariant - critical for understanding generalization in vision systems.Method: SnS uses bi-objective optimization: for invariance, it finds perturbations that maximally alter representations while preserving unit activation; for adversarial sensitivity, it finds perturbations that maximally alter unit activation while preserving representations. Applied to CNNs at different processing stages.
Result: SnS revealed invariant transformations farther from reference images than affine transforms while better preserving target responses. Different processing stages produced different invariant changes: pixel-level affected luminance/contrast, mid/late layers affected texture/pose. L2 robust networks showed interpretability drops when stretching deep layers.
Conclusion: SnS provides a systematic way to characterize visual unit invariance and adversarial vulnerability, revealing stage-dependent transformation manifolds and interpretability differences between standard and robust models.
Abstract: Uncovering which feature combinations are encoded by visual units is critical to understanding how images are transformed into representations that support recognition. While existing feature visualization approaches typically infer a unit’s most exciting images, this is insufficient to reveal the manifold of transformations under which responses remain invariant, which is critical to generalization in vision. Here we introduce Stretch-and-Squeeze (SnS), a model-agnostic, gradient-free framework to systematically characterize a unit’s maximally invariant stimuli, and its vulnerability to adversarial perturbations, in both biological and artificial visual systems. SnS frames these transformations as bi-objective optimization problems. To probe invariance, SnS seeks image perturbations that maximally alter (stretch) the representation of a reference stimulus in a given processing stage while preserving unit activation downstream (squeeze). To probe adversarial sensitivity, stretching and squeezing are reversed to maximally perturb unit activation while minimizing changes to the upstream representation. Applied to CNNs, SnS revealed invariant transformations that were farther from a reference image in pixel-space than those produced by affine transformations, while more strongly preserving the target unit’s response. The discovered invariant images differed depending on the stage of the image representation used for optimization: pixel-level changes primarily affected luminance and contrast, while stretching mid- and late-layer representations mainly altered texture and pose. By measuring how well the hierarchical invariant images obtained for L2 robust networks were classified by humans and other observer networks, we discovered a substantial drop in their interpretability when the representation was stretched in deep layers, while the opposite trend was found for standard models.
[323] HMSViT: A Hierarchical Masked Self-Supervised Vision Transformer for Corneal Nerve Segmentation and Diabetic Neuropathy Diagnosis
Xin Zhang, Liangxiu Han, Yue Shi, Yanlin Zheng, Uazman Alam, Maryam Ferdousi, Rayaz Malik
Main category: cs.CV
TL;DR: HMSViT: A hierarchical masked self-supervised vision transformer for corneal nerve segmentation and diabetic neuropathy diagnosis from microscopy images, achieving state-of-the-art performance with fewer parameters.
Details
Motivation: Diabetic Peripheral Neuropathy affects nearly half of diabetes patients and requires early detection. Current automated methods using Corneal Confocal Microscopy suffer from inefficient feature extraction, reliance on handcrafted priors, and data limitations.Method: Proposes HMSViT with pooling-based hierarchical and dual attention mechanisms with absolute positional encoding for multi-scale feature extraction. Uses block-masked self-supervised learning to reduce reliance on labeled data, and a multi-scale decoder for segmentation and classification by fusing hierarchical features.
Result: Achieves 61.34% mIoU for nerve segmentation and 70.40% diagnostic accuracy, outperforming Swin Transformer and HiViT by up to 6.39% in segmentation accuracy while using fewer parameters. Ablation studies show SSL with hierarchical feature extraction substantially enhances performance.
Conclusion: HMSViT delivers excellent, robust, and clinically viable results, demonstrating potential for scalable deployment in real-world diagnostic applications for diabetic neuropathy detection.
Abstract: Diabetic Peripheral Neuropathy (DPN) affects nearly half of diabetes patients, requiring early detection. Corneal Confocal Microscopy (CCM) enables non-invasive diagnosis, but automated methods suffer from inefficient feature extraction, reliance on handcrafted priors, and data limitations. We propose HMSViT, a novel Hierarchical Masked Self-Supervised Vision Transformer (HMSViT) designed for corneal nerve segmentation and DPN diagnosis. Unlike existing methods, HMSViT employs pooling-based hierarchical and dual attention mechanisms with absolute positional encoding, enabling efficient multi-scale feature extraction by capturing fine-grained local details in early layers and integrating global context in deeper layers, all at a lower computational cost. A block-masked self supervised learning framework is designed for the HMSViT that reduces reliance on labelled data, enhancing feature robustness, while a multi-scale decoder is used for segmentation and classification by fusing hierarchical features. Experiments on clinical CCM datasets showed HMSViT achieves state-of-the-art performance, with 61.34% mIoU for nerve segmentation and 70.40% diagnostic accuracy, outperforming leading hierarchical models like the Swin Transformer and HiViT by margins of up to 6.39% in segmentation accuracy while using fewer parameters. Detailed ablation studies further reveal that integrating block-masked SSL with hierarchical multi-scale feature extraction substantially enhances performance compared to conventional supervised training. Overall, these comprehensive experiments confirm that HMSViT delivers excellent, robust, and clinically viable results, demonstrating its potential for scalable deployment in real-world diagnostic applications.
[324] Image Can Bring Your Memory Back: A Novel Multi-Modal Guided Attack against Image Generation Model Unlearning
Renyang Liu, Guanlin Li, Tianwei Zhang, See-Kiong Ng
Main category: cs.CV
TL;DR: RECALL is an adversarial framework that compromises the robustness of unlearned image generation models by optimizing adversarial image prompts using multi-modal conditioning, revealing vulnerabilities in current machine unlearning techniques.
Details
Motivation: While machine unlearning aims to remove undesirable concepts from pretrained image generation models, the robustness of these unlearning techniques against multi-modal adversarial attacks remains unexplored, creating safety concerns.Method: RECALL exploits diffusion models’ multi-modal conditioning by efficiently optimizing adversarial image prompts guided by a single semantically relevant reference image, rather than relying solely on adversarial text prompts.
Result: Extensive experiments across 10 state-of-the-art unlearning methods show RECALL consistently outperforms existing baselines in adversarial effectiveness, computational efficiency, and semantic fidelity with original textual prompts.
Conclusion: The framework reveals critical vulnerabilities in current unlearning mechanisms and underscores the need for more robust solutions to ensure safety and reliability of generative models.
Abstract: Recent advances in image generation models (IGMs), particularly diffusion-based architectures such as Stable Diffusion (SD), have markedly enhanced the quality and diversity of AI-generated visual content. However, their generative capability has also raised significant ethical, legal, and societal concerns, including the potential to produce harmful, misleading, or copyright-infringing content. To mitigate these concerns, machine unlearning (MU) emerges as a promising solution by selectively removing undesirable concepts from pretrained models. Nevertheless, the robustness and effectiveness of existing unlearning techniques remain largely unexplored, particularly in the presence of multi-modal adversarial inputs. To bridge this gap, we propose Recall, a novel adversarial framework explicitly designed to compromise the robustness of unlearned IGMs. Unlike existing approaches that predominantly rely on adversarial text prompts, Recall exploits the intrinsic multi-modal conditioning capabilities of diffusion models by efficiently optimizing adversarial image prompts with guidance from a single semantically relevant reference image. Extensive experiments across ten state-of-the-art unlearning methods and diverse tasks show that Recall consistently outperforms existing baselines in terms of adversarial effectiveness, computational efficiency, and semantic fidelity with the original textual prompt. These findings reveal critical vulnerabilities in current unlearning mechanisms and underscore the need for more robust solutions to ensure the safety and reliability of generative models. Code and data are publicly available at \textcolor{blue}{https://github.com/ryliu68/RECALL}.
[325] Efficient Dual-domain Image Dehazing with Haze Prior Perception
Lirong Zheng, Yanshan Li, Rui Yu, Kaihao Zhang
Main category: cs.CV
TL;DR: DGFDNet is a dual-domain dehazing network that combines spatial and frequency processing with dark channel priors for adaptive frequency modulation and multi-scale feature fusion.
Details
Motivation: Transformers for image dehazing have high computational costs and limited effectiveness under complex haze conditions. Existing methods that integrate frequency-domain cues suffer from weak coupling between spatial and frequency branches.Method: Proposes DGFDNet with DGFDBlock containing: 1) Haze-Aware Frequency Modulator (HAFM) using dark channel priors to generate haze confidence maps for adaptive frequency modulation; 2) Multi-level Gating Aggregation Module (MGAM) for multi-scale feature fusion; and 3) Prior Correction Guidance Branch (PCGB) for iterative refinement of priors.
Result: Achieves state-of-the-art performance on four benchmark datasets with improved robustness and real-time efficiency.
Conclusion: DGFDNet effectively addresses computational costs and weak spatial-frequency coupling in dehazing by explicitly aligning degradation across domains using dark channel priors and adaptive frequency modulation.
Abstract: Transformers offer strong global modeling for single-image dehazing but come with high computational costs. Most methods rely on spatial features to capture long-range dependencies, making them less effective under complex haze conditions. Although some integrate frequency-domain cues, weak coupling between spatial and frequency branches limits their performance. To address these issues, we propose the Dark Channel Guided Frequency-aware Dehazing Network (DGFDNet), a dual-domain framework that explicitly aligns degradation across spatial and frequency domains. At its core, the DGFDBlock consists of two key modules: 1) Haze-Aware Frequency Modulator (HAFM), which uses dark channel priors to generate a haze confidence map for adaptive frequency modulation, achieving global degradation-aware spectral filtering. 2) Multi-level Gating Aggregation Module (MGAM), which fuses multi-scale features via multi-scale convolutions and a hybrid gating mechanism to recover fine-grained structures. Additionally, the Prior Correction Guidance Branch (PCGB) incorporates feedback for iterative refinement of the prior, improving haze localization accuracy, particularly in outdoor scenes. Extensive experiments on four benchmark datasets demonstrate that DGFDNet achieves state-of-the-art performance with improved robustness and real-time efficiency. Code is available at: https://github.com/Dilizlr/DGFDNet.
[326] Latent Denoising Makes Good Tokenizers
Jiawei Yang, Tianhong Li, Lijie Fan, Yonglong Tian, Yue Wang
Main category: cs.CV
TL;DR: l-DeTok tokenizer aligns embeddings with denoising objectives to improve image generation quality across multiple models and benchmarks.
Details
Motivation: Modern generative models share a common training objective of reconstructing clean signals from corrupted inputs (denoising), but tokenizers are not typically designed with this objective in mind. The authors propose aligning tokenizer embeddings directly with downstream denoising objectives to improve generative modeling.Method: Introduces Latent Denoising Tokenizer (l-DeTok) that trains tokenizer embeddings to reconstruct clean images from latent embeddings corrupted via interpolative noise or random masking, aligning tokenizer design with the denoising objectives of generative models.
Result: Extensive experiments on class-conditioned (ImageNet 256x256 and 512x512) and text-conditioned (MSCOCO) image generation benchmarks show l-DeTok consistently improves generation quality across six representative generative models compared to prior tokenizers.
Conclusion: Denoising should be a fundamental design principle for tokenizer development, and l-DeTok demonstrates the effectiveness of aligning tokenizer embeddings with downstream denoising objectives for improved generative modeling.
Abstract: Despite their fundamental role, it remains unclear what properties could make tokenizers more effective for generative modeling. We observe that modern generative models share a conceptually similar training objective – reconstructing clean signals from corrupted inputs, such as signals degraded by Gaussian noise or masking – a process we term denoising. Motivated by this insight, we propose aligning tokenizer embeddings directly with the downstream denoising objective, encouraging latent embeddings that remain reconstructable even under significant corruption. To achieve this, we introduce the Latent Denoising Tokenizer (l-DeTok), a simple yet highly effective tokenizer trained to reconstruct clean images from latent embeddings corrupted via interpolative noise or random masking. Extensive experiments on class-conditioned (ImageNet 256x256 and 512x512) and text-conditioned (MSCOCO) image generation benchmarks demonstrate that our l-DeTok consistently improves generation quality across six representative generative models compared to prior tokenizers. Our findings highlight denoising as a fundamental design principle for tokenizer development, and we hope it could motivate new perspectives for future tokenizer design.
[327] 3DRot: Rediscovering the Missing Primitive for RGB-Based 3D Augmentation
Shitian Yang, Deyu Li, Xiaoke Jiang, Lei Zhang
Main category: cs.CV
TL;DR: 3DRot: A geometry-consistent 3D rotation augmentation for RGB-based 3D tasks that rotates images about camera optical center while updating all geometric elements without requiring scene depth.
Details
Motivation: RGB-based 3D tasks suffer from limited augmentation options because most image transforms disrupt geometric consistency. While horizontal flipping is standard, rigorous 3D rotation augmentation has been absent due to misconceptions about requiring scene depth or reconstruction.Method: 3DRot is a plug-and-play augmentation that rotates and mirrors images about the camera’s optical center while synchronously updating RGB images, camera intrinsics, object poses, and 3D annotations to preserve projective geometry. It achieves geometry-consistent rotations and reflections without relying on any scene depth.
Result: Significant improvements across multiple 3D tasks: On SUN RGB-D, improved IoU3D from 43.21 to 44.51, reduced rotation error from 22.91° to 20.93°, and boosted mAP0.5 from 35.70 to 38.11. On NYU Depth v2, improved abs-rel from 0.1783 to 0.1685. On KITTI with MVX-Net, raised moderate 3D AP from ~63.85 to 65.16.
Conclusion: 3DRot provides an effective, depth-free augmentation method for RGB-based 3D tasks that preserves geometric consistency and improves performance across multiple benchmarks and tasks, remaining compatible with standard 3D augmentations.
Abstract: RGB-based 3D tasks, e.g., 3D detection, depth estimation, 3D keypoint estimation, still suffer from scarce, expensive annotations and a thin augmentation toolbox, since many image transforms, including rotations and warps, disrupt geometric consistency. While horizontal flipping and color jitter are standard, rigorous 3D rotation augmentation has surprisingly remained absent from RGB-based pipelines, largely due to the misconception that it requires scene depth or scene reconstruction. In this paper, we introduce 3DRot, a plug-and-play augmentation that rotates and mirrors images about the camera’s optical center while synchronously updating RGB images, camera intrinsics, object poses, and 3D annotations to preserve projective geometry, achieving geometry-consistent rotations and reflections without relying on any scene depth. We first validate 3DRot on a classical RGB-based 3D task, monocular 3D detection. On SUN RGB-D, inserting 3DRot into a frozen DINO-X + Cube R-CNN pipeline raises $IoU_{3D}$ from 43.21 to 44.51, cuts rotation error (ROT) from 22.91$^\circ$ to 20.93$^\circ$, and boosts $mAP_{0.5}$ from 35.70 to 38.11; smaller but consistent gains appear on a cross-domain IN10 split. Beyond monocular detection, adding 3DRot on top of the standard BTS augmentation schedule further improves NYU Depth v2 from 0.1783 to 0.1685 in abs-rel (and 0.7472 to 0.7548 in $δ<1.25$), and reduces cross-dataset error on SUN RGB-D. On KITTI, applying the same camera-centric rotations in MVX-Net (LiDAR+RGB) raises moderate 3D AP from about 63.85 to 65.16 while remaining compatible with standard 3D augmentations.
[328] Robust MultiSpecies Agricultural Segmentation Across Devices, Seasons, and Sensors Using Hierarchical DINOv2 Models
Artzai Picon, Itziar Eguskiza, Daniel Mugica, Javier Romero, Carlos Javier Jimenez, Eric White, Gabriel Do-Lago-Junqueira, Christian Klukas, Ramon Navarra-Mestre
Main category: cs.CV
TL;DR: Vision foundation models (DINOv2) with hierarchical taxonomic inference improve plant species and damage segmentation robustness across domain shifts in agricultural monitoring.
Details
Motivation: Current deep learning approaches for plant species and damage segmentation fail to generalize under real-world domain shifts (seasons, geographies, devices, sensing modalities), limiting their operational use in phenotyping pipelines.Method: Integrates vision foundation models (DINOv2) with hierarchical taxonomic inference, trained on large multi-year dataset (Germany/Spain 2018-2020) with 14 plant species and 4 herbicide damage classes, evaluated under challenging domain shifts.
Result: Foundation-model backbone consistently outperforms baselines, improving species-level F1 from 0.52 to 0.87 on in-distribution data, maintaining advantages under moderate (0.77 vs. 0.24) and extreme (0.44 vs. 0.14) shift conditions. Hierarchical inference provides additional robustness.
Conclusion: Combining foundation models with structured biological hierarchies enables scalable, shift-resilient agricultural monitoring, now deployed in BASF’s phenotyping workflow for herbicide research trials.
Abstract: Reliable plant species and damage segmentation for herbicide field research trials requires models that can withstand substantial real-world variation across seasons, geographies, devices, and sensing modalities. Most deep learning approaches trained on controlled datasets fail to generalize under these domain shifts, limiting their suitability for operational phenotyping pipelines. This study evaluates a segmentation framework that integrates vision foundation models (DINOv2) with hierarchical taxonomic inference to improve robustness across heterogeneous agricultural conditions. We train on a large, multi-year dataset collected in Germany and Spain (2018-2020), comprising 14 plant species and 4 herbicide damage classes, and assess generalization under increasingly challenging shifts: temporal and device changes (2023), geographic transfer to the United States, and extreme sensor shift to drone imagery (2024). Results show that the foundation-model backbone consistently outperforms prior baselines, improving species-level F1 from 0.52 to 0.87 on in-distribution data and maintaining significant advantages under moderate (0.77 vs. 0.24) and extreme (0.44 vs. 0.14) shift conditions. Hierarchical inference provides an additional layer of robustness, enabling meaningful predictions even when fine-grained species classification degrades (family F1: 0.68, class F1: 0.88 on aerial imagery). Error analysis reveals that failures under severe shift stem primarily from vegetation-soil confusion, suggesting that taxonomic distinctions remain preserved despite background and viewpoint variability. The system is now deployed within BASF’s phenotyping workflow for herbicide research trials across multiple regions, illustrating the practical viability of combining foundation models with structured biological hierarchies for scalable, shift-resilient agricultural monitoring.
[329] Image-to-Brain Signal Generation for Visual Prosthesis with CLIP Guided Multimodal Diffusion Models
Ganxi Xu, Zhao-Rong Lai, Yuting Tang, Yonghao Song, Guoxu Zhou, Boyu wang, Jian Zhu, Jinyi Long
Main category: cs.CV
TL;DR: A novel image-to-brain signal framework using diffusion transformers with cross-attention mechanisms to generate M/EEG signals from images for visual prostheses.
Details
Motivation: Visual prostheses need a complete functional pipeline. While M/EEG signals can evoke visual perceptions (brain decoding), converting images into M/EEG signals (brain encoding) remains unexplored, hindering complete visual restoration systems.Method: Uses diffusion transformer (DiT) architecture based on DDIM for brain signal generation. Employs cross-attention mechanisms to align brain signal embeddings with CLIP image embeddings. Leverages LLMs to generate image captions, concatenating CLIP text and image embeddings. Introduces learnable spatio-temporal position encoding combining brain region embeddings with temporal embeddings.
Result: Evaluated on THINGS-EEG2 and THINGS-MEG datasets, demonstrating generation of biologically plausible brain signals.
Conclusion: The framework successfully generates M/EEG signals from images, addressing the brain encoding gap in visual prostheses and enabling a more complete functional pipeline for vision restoration.
Abstract: Visual prostheses hold great promise for restoring vision in blind individuals. While researchers have successfully utilized M/EEG signals to evoke visual perceptions during the brain decoding stage of visual prostheses, the complementary process of converting images into M/EEG signals in the brain encoding stage remains largely unexplored, hindering the formation of a complete functional pipeline. In this work, we present a novel image-to-brain signal framework that generates M/EEG from images by leveraging the diffusion transformer architecture enhanced with cross-attention mechanisms. Specifically, we employ a diffusion transformer (DiT) architecture based on denoising diffusion implicit models (DDIM) to achieve brain signal generation. To realize the goal of image-to-brain signal conversion, we use cross-attention mechanisms to align brain signal embeddings with CLIP image embeddings. Moreover, we leverage large language models (LLMs) to generate image captions, and concatenate the resulting CLIP text embeddings with CLIP image embeddings to form unified embeddings for cross-attention alignment, enabling our model to capture core semantic information. Moreover, to capture core semantic information, we use large language models (LLMs) to generate descriptive and semantically accurate captions for images. Furthermore, we introduce a learnable spatio-temporal position encoding that combines brain region embeddings with temporal embeddings to capture both spatial and temporal characteristics of brain signals. We evaluate the framework on two multimodal benchmark datasets (THINGS-EEG2 and THINGS-MEG) and demonstrate that it generates biologically plausible brain signals.
[330] BEVTraj: Map-Free End-to-End Trajectory Prediction in Bird’s-Eye View with Deformable Attention and Sparse Goal Proposals
Minsang Kong, Myeongjun Kim, Sang Gu Kang, Hejiu Lu, Yupeng Zhong, Sang Hun Lee
Main category: cs.CV
TL;DR: BEVTraj: A map-free trajectory prediction framework for autonomous driving that uses deformable attention to extract relevant context from dense Bird’s-Eye View features and sparse goal proposals for end-to-end multimodal forecasting.
Details
Motivation: HD maps are costly, geographically limited, and unreliable in dynamic/unmapped scenarios. BEV features offer flexibility but are dense and unstructured, making agent-centric reasoning challenging. Need a map-free solution that can effectively leverage raw sensor data.Method: Proposes BEVTraj framework with: 1) Deformable attention to adaptively aggregate task-relevant context from sparse locations in dense BEV features, 2) Sparse Goal Candidate Proposal (SGCP) module that predicts a small set of realistic goals for end-to-end multimodal forecasting without heuristic post-processing.
Result: Extensive experiments show BEVTraj achieves performance comparable to state-of-the-art HD map-based methods while providing greater robustness and flexibility without relying on pre-built maps.
Conclusion: BEVTraj demonstrates that map-free trajectory prediction can achieve competitive performance with map-based methods while offering practical advantages in real-world deployment scenarios.
Abstract: In autonomous driving, trajectory prediction is essential for safe and efficient navigation. While recent methods often rely on high-definition (HD) maps to provide structured environmental priors, such maps are costly to maintain, geographically limited, and unreliable in dynamic or unmapped scenarios. Directly leveraging raw sensor data in Bird’s-Eye View (BEV) space offers greater flexibility, but BEV features are dense and unstructured, making agent-centric spatial reasoning challenging and computationally inefficient. To address this, we propose Bird’s-Eye View Trajectory Prediction (BEVTraj), a map-free framework that employs deformable attention to adaptively aggregate task-relevant context from sparse locations in dense BEV features. We further introduce a Sparse Goal Candidate Proposal (SGCP) module that predicts a small set of realistic goals, enabling fully end-to-end multimodal forecasting without heuristic post-processing. Extensive experiments show that BEVTraj achieves performance comparable to state-of-the-art HD map-based methods while providing greater robustness and flexibility without relying on pre-built maps. The source code is available at https://github.com/Kongminsang/bevtraj.
[331] Curriculum Multi-Task Self-Supervision Improves Lightweight Architectures for Onboard Satellite Hyperspectral Image Segmentation
Hugo Carlesso, Josiane Mothe, Radu Tudor Ionescu
Main category: cs.CV
TL;DR: A curriculum multi-task self-supervised learning framework for hyperspectral imaging that combines masked image modeling with spatial/spectral jigsaw puzzles to train lightweight models for satellite onboard processing.
Details
Motivation: Hyperspectral imaging generates high-dimensional data with slow satellite transmission rates, requiring compact models for onboard processing to reduce redundant data transmission. Existing methods lack efficient joint spatial-spectral learning for lightweight architectures.Method: CMTSSL integrates masked image modeling with decoupled spatial and spectral jigsaw puzzle solving, guided by curriculum learning that progressively increases data difficulty during self-supervision to capture spectral continuity, spatial structure, and global semantics.
Result: Validated on four public benchmark datasets, showing consistent gains in downstream segmentation tasks with architectures over 16,000x lighter than some state-of-the-art models.
Conclusion: CMTSSL enables generalizable representation learning with lightweight architectures suitable for real-world HSI applications and onboard satellite deployment.
Abstract: Hyperspectral imaging (HSI) captures detailed spectral signatures across hundreds of contiguous bands per pixel, being indispensable for remote sensing applications such as land-cover classification, change detection, and environmental monitoring. Due to the high dimensionality of HSI data and the slow rate of data transfer in satellite-based systems, compact and efficient models are required to support onboard processing and minimize the transmission of redundant or low-value data. To this end, we introduce a novel curriculum multi-task self-supervised learning (CMTSSL) framework designed for lightweight architectures for HSI analysis. CMTSSL integrates masked image modeling with decoupled spatial and spectral jigsaw puzzle solving, guided by a curriculum learning strategy that progressively increases data difficulty during self-supervision. This enables the encoder to jointly capture fine-grained spectral continuity, spatial structure, and global semantic features. Unlike prior dual-task SSL methods, CMTSSL simultaneously addresses spatial and spectral reasoning within a unified and computationally efficient design, being particularly suitable for training lightweight models for onboard satellite deployment. We validate our approach on four public benchmark datasets, demonstrating consistent gains in downstream segmentation tasks, using architectures that are over 16,000x lighter than some state-of-the-art models. These results highlight the potential of CMTSSL in generalizable representation learning with lightweight architectures for real-world HSI applications. Our code is publicly available at https://github.com/hugocarlesso/CMTSSL.
[332] BlurBall: Joint Ball and Motion Blur Estimation for Table Tennis Ball Tracking
Thomas Gossard, Filip Radovic, Andreas Ziegler, Andreas Zell
Main category: cs.CV
TL;DR: New labeling strategy places ball at center of motion blur streak instead of leading edge, with explicit blur attribute annotation, improving detection and enabling motion estimation.
Details
Motivation: Motion blur reduces clarity of fast-moving objects like sports balls, and existing labeling conventions mark balls at leading edge of blur, introducing asymmetry and ignoring valuable motion cues correlated with velocity.Method: Introduces new labeling strategy placing ball at center of blur streak with explicit blur attribute annotation. Creates new table tennis ball detection dataset. Proposes BlurBall model that jointly estimates ball position and motion blur attributes using attention mechanisms (Squeeze-and-Excitation) over multi-frame inputs.
Result: New labeling approach consistently enhances detection performance across various models. BlurBall achieves state-of-the-art results in ball detection. Leveraging blur improves detection accuracy and enables more reliable trajectory prediction.
Conclusion: Center-based blur labeling with explicit attribute annotation improves ball detection and enables motion estimation, benefiting real-time sports analytics through better trajectory prediction.
Abstract: Motion blur reduces the clarity of fast-moving objects, posing challenges for detection systems, especially in racket sports, where balls often appear as streaks rather than distinct points. Existing labeling conventions mark the ball at the leading edge of the blur, introducing asymmetry and ignoring valuable motion cues correlated with velocity. This paper introduces a new labeling strategy that places the ball at the center of the blur streak and explicitly annotates blur attributes. Using this convention, we release a new table tennis ball detection dataset. We demonstrate that this labeling approach consistently enhances detection performance across various models. Furthermore, we introduce BlurBall, a model that jointly estimates ball position and motion blur attributes. By incorporating attention mechanisms such as Squeeze-and-Excitation over multi-frame inputs, we achieve state-of-the-art results in ball detection. Leveraging blur not only improves detection accuracy but also enables more reliable trajectory prediction, benefiting real-time sports analytics.
[333] Deep Learning for Clouds and Cloud Shadow Segmentation in Methane Satellite and Airborne Imaging Spectroscopy
Manuel Perez-Carrasco, Maya Nasr, Sebastien Roche, Chris Chan Miller, Zhan Zhang, Core Francisco Park, Eleanor Walker, Cecilia Garraffo, Douglas Finkbeiner, Sasha Ayvazov, Jonathan Franklin, Bingkun Luo, Xiong Liu, Ritesh Gautam, Steven Wofsy
Main category: cs.CV
TL;DR: Machine learning methods for cloud and cloud shadow detection in hyperspectral remote sensing data from MethaneSAT and MethaneAIR missions to improve atmospheric methane retrieval accuracy.
Details
Motivation: Cloud and cloud shadows in remote sensing data bias methane retrievals and impact emission quantification. Effective detection is critical for MethaneSAT and MethaneAIR missions which provide high-resolution hyperspectral data for atmospheric methane monitoring.Method: Deployed and evaluated conventional techniques (Iterative Logistic Regression and Multilayer Perceptron) with advanced deep learning architectures (U-Net and Spectral Channel Attention Network) for cloud and cloud shadow detection in high-resolution hyperspectral data.
Result: Conventional methods struggle with spatial coherence and boundary definition. Deep learning models substantially improve detection quality: U-Net performs best in preserving spatial structure, while SCAN excels at capturing fine boundary details.
Conclusion: Deep learning approaches, particularly U-Net and SCAN, offer superior cloud and cloud shadow detection for high-resolution hyperspectral remote sensing data compared to conventional methods, enabling more accurate methane retrieval and emission quantification.
Abstract: Effective cloud and cloud shadow detection is a critical prerequisite for accurate retrieval of concentrations of atmospheric methane (CH4) or other trace gases in hyperspectral remote sensing. This challenge is especially pertinent for MethaneSAT, a satellite mission launched in March 2024, to fill a significant data gap in terms of resolution, precision and swath between coarse-resolution global mappers and fine-scale point-source imagers of methane, and for its airborne companion mission, MethaneAIR. MethaneSAT delivers hyperspectral data at an intermediate spatial resolution (approx. 100 x 400, m), whereas MethaneAIR provides even finer resolution (approx. 25 m), enabling the development of highly detailed maps of concentrations that enable quantification of both the sources and rates of emissions. In this study, we use machine learning methods to address the cloud and cloud shadow detection problem for sensors with these high spatial resolutions. Cloud and cloud shadows in remote sensing data need to be effectively screened out as they bias methane retrievals in remote sensing imagery and impact the quantification of emissions. We deploy and evaluate conventional techniques-including Iterative Logistic Regression (ILR) and Multilayer Perceptron (MLP)-with advanced deep learning architectures, namely U-Net and a Spectral Channel Attention Network (SCAN) method. Our results show that conventional methods struggle with spatial coherence and boundary definition, affecting the detection of clouds and cloud shadows. Deep learning models substantially improve detection quality: U-Net performs best in preserving spatial structure, while SCAN excels at capturing fine boundary details… Our data and code is publicly available at: https://doi.org/10.7910/DVN/IKLZOJ
[334] DeLiVR: Differential Spatiotemporal Lie Bias for Efficient Video Deraining
Shuning Sun, Jialang Lu, Xiang Chen, Jichao Wang, Dianjie Lu, Guijuan Zhang, Guangwei Gao, Zhuoran Zheng
Main category: cs.CV
TL;DR: DeLiVR: A video deraining method using Lie-group differential biases in attention scores for spatiotemporal consistency, with rotation-bounded relative bias and differential group displacement components.
Details
Motivation: Videos captured in the wild suffer from rain streaks, blur, noise, and camera pose changes causing cross-frame mismatches. Existing methods rely on computationally expensive optical flow or heuristic alignment that are less robust.Method: Proposes DeLiVR which injects spatiotemporal Lie-group differential biases directly into attention scores. Uses two components: 1) rotation-bounded Lie relative bias for geometry-consistent alignment, and 2) differential group displacement to estimate velocity between frames with temporal decay and attention masks.
Result: Extensive experimental results demonstrate effectiveness on publicly available benchmarks. Code is publicly available.
Conclusion: Lie groups provide principled representation for continuous geometric transformations, enabling efficient video deraining with spatiotemporal consistency through attention-based differential biases.
Abstract: Videos captured in the wild often suffer from rain streaks, blur, and noise. In addition, even slight changes in camera pose can amplify cross-frame mismatches and temporal artifacts. Existing methods rely on optical flow or heuristic alignment, which are computationally expensive and less robust. To address these challenges, Lie groups provide a principled way to represent continuous geometric transformations, making them well-suited for enforcing spatial and temporal consistency in video modeling. Building on this insight, we propose DeLiVR, an efficient video deraining method that injects spatiotemporal Lie-group differential biases directly into attention scores of the network. Specifically, the method introduces two complementary components. First, a rotation-bounded Lie relative bias predicts the in-plane angle of each frame using a compact prediction module, where normalized coordinates are rotated and compared with base coordinates to achieve geometry-consistent alignment before feature aggregation. Second, a differential group displacement computes angular differences between adjacent frames to estimate a velocity. This bias computation combines temporal decay and attention masks to focus on inter-frame relationships while precisely matching the direction of rain streaks. Extensive experimental results demonstrate the effectiveness of our method on publicly available benchmarks. The code is publicly available at https://github.com/Shuning0312/ICLR-DeLiVR.
[335] Multi-View Camera System for Variant-Aware Autonomous Vehicle Inspection and Defect Detection
Yash Kulkarni, Raman Jha, Renu Kachhoria
Main category: cs.CV
TL;DR: Multi-view vehicle inspection system using 11 cameras, deep learning detectors, and semantic rule engine for variant-aware quality control in automotive production lines.
Details
Motivation: Modern vehicle production lines need automated systems to verify variant specifications and detect visible defects, which is increasingly complex with multiple vehicle models and configurations.Method: Uses 11 synchronized cameras for 360° coverage, specialized deep learning modules (YOLOv8 for part detection, EfficientNet for ICE/EV classification, Gemini-1.5 Flash for OCR, YOLOv8-Seg for segmentation), view-aware fusion layer, and VIN-conditioned rule engine for comparison against expected manifest.
Result: Achieves 93% verification accuracy, 86% defect-detection recall, processes 3.3 vehicles per minute, outperforming single-view or no segmentation baselines.
Conclusion: First publicly reported system unifying multi-camera feature validation with defect detection in deployable automotive industry setting, demonstrating practical real-time quality control.
Abstract: Ensuring that every vehicle leaving a modern production line is built to the correct \emph{variant} specification and is free from visible defects is an increasingly complex challenge. We present the \textbf{Automated Vehicle Inspection (AVI)} platform, an end-to-end, \emph{multi-view} perception system that couples deep-learning detectors with a semantic rule engine to deliver \emph{variant-aware} quality control in real time. Eleven synchronized cameras capture a full 360° sweep of each vehicle; task-specific views are then routed to specialised modules: YOLOv8 for part detection, EfficientNet for ICE/EV classification, Gemini-1.5 Flash for mascot OCR, and YOLOv8-Seg for scratch-and-dent segmentation. A view-aware fusion layer standardises evidence, while a VIN-conditioned rule engine compares detected features against the expected manifest, producing an interpretable pass/fail report in (\approx! 300,\text{ms}). On a mixed data set of Original Equipment Manufacturer(OEM) vehicle data sets of four distinct models plus public scratch/dent images, AVI achieves \textbf{ 93 %} verification accuracy, \textbf{86 %} defect-detection recall, and sustains (\mathbf{3.3}) vehicles/min, surpassing single-view or no segmentation baselines by large margins. To our knowledge, this is the first publicly reported system that unifies multi-camera feature validation with defect detection in a deployable automotive setting in industry.
[336] LAKAN: Landmark-assisted Adaptive Kolmogorov-Arnold Network for Face Forgery Detection
Jiayao Jiang, Bin Liu, Qi Chu, Nenghai Yu
Main category: cs.CV
TL;DR: A novel face forgery detection method using Kolmogorov-Arnold Networks (KAN) with facial landmark guidance to better capture complex forgery artifacts.
Details
Motivation: Deepfake generation techniques are rapidly advancing, requiring more robust detection methods. Current CNN and Transformer approaches have limitations in modeling the highly complex, non-linear nature of forgery artifacts.Method: Proposes a KAN-based detection method that replaces fixed activation functions with learnable splines. Introduces LAKAN module that uses facial landmarks as structural priors to dynamically generate KAN parameters, creating instance-specific signals to guide attention to artifact-rich facial regions.
Result: Extensive experiments on multiple public datasets demonstrate superior performance compared to existing methods.
Conclusion: The combination of geometric priors (facial landmarks) with KAN’s flexible architecture creates an effective approach for face forgery detection that outperforms current state-of-the-art methods.
Abstract: The rapid development of deepfake generation techniques necessitates robust face forgery detection algorithms. While methods based on Convolutional Neural Networks (CNNs) and Transformers are effective, there is still room for improvement in modeling the highly complex and non-linear nature of forgery artifacts. To address this issue, we propose a novel detection method based on the Kolmogorov-Arnold Network (KAN). By replacing fixed activation functions with learnable splines, our KAN-based approach is better suited to this challenge. Furthermore, to guide the network’s focus towards critical facial areas, we introduce a Landmark-assisted Adaptive Kolmogorov-Arnold Network (LAKAN) module. This module uses facial landmarks as a structural prior to dynamically generate the internal parameters of the KAN, creating an instance-specific signal that steers a general-purpose image encoder towards the most informative facial regions with artifacts. This core innovation creates a powerful combination between geometric priors and the network’s learning process. Extensive experiments on multiple public datasets show that our proposed method achieves superior performance.
[337] UGround: Towards Unified Visual Grounding with Unrolled Transformers
Rui Qian, Xin Yin, Chuanhang Deng, Zhiyuan Peng, Jian Xiong, Wei Zhai, Dejing Dou
Main category: cs.CV
TL;DR: UGround introduces a unified visual grounding paradigm that dynamically selects intermediate transformer layers as “mask as prompt” instead of using fixed last hidden layers, addressing error propagation and spatial cue limitations in current methods.
Details
Motivation: Current visual grounding methods rely on fixed last hidden layers which amplify cumulative errors through layer-by-layer propagation without correction, and useMethod: UGround uses Policy-Prompted Masking with two components: Stochastic Skip Connection (SSC) - a reinforcement learning policy that allows
Result: UGround unifies visual grounding within a single framework from an attribute perspective, spanning traditional refer expression segmentation to reasoning segmentation, single-target to multi-target, and positive query to false premise scenarios.
Conclusion: UGround provides a more flexible and effective approach to visual grounding by addressing error propagation and spatial cue limitations through dynamic layer selection and explicit mask prompting.
Abstract: We present UGround, a \textbf{U}nified visual \textbf{Ground}ing paradigm that dynamically selects intermediate layers across \textbf{U}nrolled transformers as mask as prompt'', diverging from the prevailing pipeline that leverages the fixed last hidden layer as \texttt{
[338] PAGCNet: A Pose-Aware and Geometry Constrained Framework for Panoramic Depth Estimation
Kanglin Ning, Ruzhao Chen, Penghong Wang, Xingtao Wang, Ruiqin Xiong, Xiaopeng Fan
Main category: cs.CV
TL;DR: A pose-aware geometry-constrained framework for panoramic depth estimation that uses room layout, camera pose, and region segmentation to compute background depth as geometric prior for depth correction.
Details
Motivation: Reconstructing background depth for regular enclosed regions in complex indoor scenes without external measurements is challenging. Existing methods don't effectively model room background depth as geometric constraint for panoramic depth estimation.Method: Multi-task framework with specialized decoders for room layout, camera pose, depth, and region segmentation. Pose-aware background depth resolving (PA-BDR) computes background depth using camera pose. Fusion mask generation (FMG) creates weight maps for depth correction. Adaptive fusion integrates refined background depth with initial predictions.
Result: Superior performance on Matterport3D, Structured3D, and Replica datasets compared to current open-source methods. Code is publicly available.
Conclusion: The proposed pose-aware geometry-constrained framework effectively addresses panoramic depth estimation by leveraging background depth as geometric prior, demonstrating significant improvements over existing methods.
Abstract: Explicitly modeling room background depth as a geometric constraint has proven effective for panoramic depth estimation. However, reconstructing this background depth for regular enclosed regions in a complex indoor scene without external measurements remains an open challenge. To address this, we propose a pose-aware and geometry-constrained framework for panoramic depth estimation. Our framework first employs multiple task-specific decoders to jointly estimate room layout, camera pose, depth, and region segmentation from a input panoramic image. A pose-aware background depth resolving (PA-BDR) component uses tasks decoder’s prediction to resolve the camera pose. Subsequently, the proposed PA-BDR component uses the camera pose to compute the background depth of regular enclosed regions and uses this background depth as a strong geometric prior. Based on the output of the region segmentation decoder, a fusion mask generation (FMG) component produces a fusion weight map to guide where and to what extent the geometry-constrained background depth should correct the depth decoder’s prediction. Finally, an adaptive fusion component integrates this refined background depth with the initial depth prediction, guided by the fusion weight. Extensive experiments on Matterport3D, Structured3D, and Replica datasets demonstrate that our method achieves significantly superior performance compared to current open-source methods. Code is available at https://github.com/emiyaning/PAGCNet.
[339] The impact of abstract and object tags on image privacy classification
Darya Baranouskaya, Andrea Cavallaro
Main category: cs.CV
TL;DR: Abstract tags outperform object tags for image privacy classification when tag budget is limited, but object tags become equally useful with more tags available.
Details
Motivation: To determine which type of image tags (object vs. abstract) is more suitable for the context-dependent, subjective task of image privacy classification, since current approaches primarily use object tags but abstract tags might capture higher-level contextual information.Method: Comparative analysis of object tags (denoting concrete entities) versus abstract tags (capturing higher-level contextual information) for image privacy classification tasks, examining performance under different tag budget constraints.
Result: Abstract tags are more effective than object tags for privacy classification when the tag budget is limited. However, when a larger number of tags per image is available, object-related information becomes as useful as abstract tags.
Conclusion: The findings provide guidance for developing more accurate image privacy classifiers by considering the role of tag types and quantity, suggesting abstract tags should be prioritized in resource-constrained scenarios.
Abstract: Object tags denote concrete entities and are central to many computer vision tasks, whereas abstract tags capture higher-level information, which is relevant for tasks that require a contextual, potentially subjective scene understanding. Object and abstract tags extracted from images also facilitate interpretability. In this paper, we explore which type of tags is more suitable for the context-dependent and inherently subjective task of image privacy. While object tags are generally used for privacy classification, we show that abstract tags are more effective when the tag budget is limited. Conversely, when a larger number of tags per image is available, object-related information is as useful. We believe that these findings will guide future research in developing more accurate image privacy classifiers, informed by the role of tag types and quantity.
[340] Large Scale Diffusion Distillation via Score-Regularized Continuous-Time Consistency
Kaiwen Zheng, Yuji Wang, Qianli Ma, Huayu Chen, Jintao Zhang, Yogesh Balaji, Jianfei Chen, Ming-Yu Liu, Jun Zhu, Qinsheng Zhang
Main category: cs.CV
TL;DR: rCM (score-regularized continuous-time consistency model) scales continuous-time consistency models to large-scale text-to-image/video diffusion models, addressing quality limitations through score distillation regularization.
Details
Motivation: Continuous-time consistency models (sCM, MeanFlow) are theoretically principled for fast diffusion but face infrastructure challenges in Jacobian-vector product computation and unclear applicability to large-scale text-to-image/video tasks due to quality limitations in fine-detail generation.Method: Develop FlashAttention-2 JVP kernel for parallelism-compatible training on 10B+ parameter models; propose rCM that incorporates score distillation as a long-skip regularizer to complement sCM’s forward-divergence objective with reverse divergence, improving quality while maintaining diversity.
Result: rCM matches state-of-the-art DMD2 on quality metrics while mitigating mode collapse, offering better diversity; achieves 15-50× acceleration with 1-4 sampling steps; validated on models up to 14B parameters and 5-second videos.
Conclusion: rCM provides a practical, theoretically grounded framework for large-scale diffusion distillation that maintains both quality and diversity without extensive hyperparameter tuning or GAN components.
Abstract: Although continuous-time consistency models (e.g., sCM, MeanFlow) are theoretically principled and empirically powerful for fast academic-scale diffusion, its applicability to large-scale text-to-image and video tasks remains unclear due to infrastructure challenges in Jacobian-vector product (JVP) computation and the limitations of evaluation benchmarks like FID. This work represents the first effort to scale up continuous-time consistency to general application-level image and video diffusion models, and to make JVP-based distillation effective at large scale. We first develop a parallelism-compatible FlashAttention-2 JVP kernel, enabling sCM training on models with over 10 billion parameters and high-dimensional video tasks. Our investigation reveals fundamental quality limitations of sCM in fine-detail generation, which we attribute to error accumulation and the “mode-covering” nature of its forward-divergence objective. To remedy this, we propose the score-regularized continuous-time consistency model (rCM), which incorporates score distillation as a long-skip regularizer. This integration complements sCM with the “mode-seeking” reverse divergence, effectively improving visual quality while maintaining high generation diversity. Validated on large-scale models (Cosmos-Predict2, Wan2.1) up to 14B parameters and 5-second videos, rCM generally matches the state-of-the-art distillation method DMD2 on quality metrics while mitigating mode collapse and offering notable advantages in diversity, all without GAN tuning or extensive hyperparameter searches. The distilled models generate high-fidelity samples in only $1\sim4$ steps, accelerating diffusion sampling by $15\times\sim50\times$. These results position rCM as a practical and theoretically grounded framework for advancing large-scale diffusion distillation. Code is available at https://github.com/NVlabs/rcm.
[341] AnyUp: Universal Feature Upsampling
Thomas Wimmer, Prune Truong, Marie-Julie Rakotosaona, Michael Oechsle, Federico Tombari, Bernt Schiele, Jan Eric Lenssen
Main category: cs.CV
TL;DR: AnyUp is a feature-agnostic upsampling method that works with any vision feature at any resolution without encoder-specific training, outperforming existing methods while preserving feature semantics.
Details
Motivation: Existing learning-based feature upsamplers require retraining for each specific feature extractor (like DINO or CLIP), limiting their generalization to different feature types at inference time.Method: Proposes an inference-time feature-agnostic upsampling architecture that can be applied to any vision feature without encoder-specific training.
Result: AnyUp sets new state-of-the-art for upsampled features, generalizes to different feature types, preserves feature semantics, and is efficient for downstream tasks.
Conclusion: AnyUp provides a versatile, high-quality upsampling solution that works across different vision features without requiring retraining for each encoder.
Abstract: We introduce AnyUp, a method for feature upsampling that can be applied to any vision feature at any resolution, without encoder-specific training. Existing learning-based upsamplers for features like DINO or CLIP need to be re-trained for every feature extractor and thus do not generalize to different feature types at inference time. In this work, we propose an inference-time feature-agnostic upsampling architecture to alleviate this limitation and improve upsampling quality. In our experiments, AnyUp sets a new state of the art for upsampled features, generalizes to different feature types, and preserves feature semantics while being efficient and easy to apply to a wide range of downstream tasks.
[342] Consistent text-to-image generation via scene de-contextualization
Song Tang, Peihao Gong, Kunyu Li, Kai Guo, Boyu Wang, Mao Ye, Jianwei Zhang, Xiatian Zhu
Main category: cs.CV
TL;DR: SDeC: Training-free prompt embedding editing method that suppresses scene-ID correlation in T2I models to improve identity preservation across diverse scenes without requiring prior knowledge of all target scenes.
Details
Motivation: Current text-to-image generation methods struggle with identity preservation across different scenes due to identity shift caused by scene contextualization - the natural correlation between subject and scene context that T2I models learn from training data. Previous methods require unrealistic assumptions of knowing all target scenes in advance.Method: Proposes Scene De-Contextualization (SDeC), a training-free prompt embedding editing approach that identifies and suppresses latent scene-ID correlation within ID prompt embeddings. Uses SVD directional stability to quantify and adaptively re-weight corresponding eigenvalues, implementing an inversion process of T2I’s built-in scene contextualization.
Result: SDeC significantly enhances identity preservation while maintaining scene diversity, works with per-scene use (one scene per prompt) without requiring prior access to all target scenes, making it flexible for real-world applications.
Conclusion: SDeC provides an efficient, training-free solution to the identity shift problem in T2I generation by addressing the fundamental issue of scene contextualization, offering practical advantages over previous methods that require unrealistic assumptions about target scenes.
Abstract: Consistent text-to-image (T2I) generation seeks to produce identity-preserving images of the same subject across diverse scenes, yet it often fails due to a phenomenon called identity (ID) shift. Previous methods have tackled this issue, but typically rely on the unrealistic assumption of knowing all target scenes in advance. This paper reveals that a key source of ID shift is the native correlation between subject and scene context, called scene contextualization, which arises naturally as T2I models fit the training distribution of vast natural images. We formally prove the near-universality of this scene-ID correlation and derive theoretical bounds on its strength. On this basis, we propose a novel, efficient, training-free prompt embedding editing approach, called Scene De-Contextualization (SDeC), that imposes an inversion process of T2I’s built-in scene contextualization. Specifically, it identifies and suppresses the latent scene-ID correlation within the ID prompt’s embedding by quantifying the SVD directional stability to adaptively re-weight the corresponding eigenvalues. Critically, SDeC allows for per-scene use (one scene per prompt) without requiring prior access to all target scenes. This makes it a highly flexible and general solution well-suited to real-world applications where such prior knowledge is often unavailable or varies over time. Experiments demonstrate that SDeC significantly enhances identity preservation while maintaining scene diversity.
[343] PRISMM-Bench: A Benchmark of Peer-Review Grounded Multimodal Inconsistencies
Lukas Selch, Yufang Hou, M. Jehanzeb Mirza, Sivan Doveh, James Glass, Rogerio Feris, Wei Lin
Main category: cs.CV
TL;DR: PRISMM-Bench: First benchmark using real reviewer-flagged inconsistencies in scientific papers to evaluate multimodal models’ ability to detect and resolve cross-modal inconsistencies across text, figures, tables, and equations.
Details
Motivation: Current LMMs are increasingly used in scientific research but their ability to understand and reason over multimodal complexity in papers remains unclear. Existing benchmarks fail to capture real-world inconsistencies across modalities, using synthetic errors or isolated modalities instead of genuine reviewer-flagged issues.Method: Created PRISMM-Bench through multi-stage pipeline: mining real reviewer comments, LLM-assisted filtering, human verification to curate 384 inconsistencies from 353 papers. Designed three tasks: inconsistency identification, remedy, and pair matching. Introduced structured JSON-based answer representations to minimize linguistic biases in multiple-choice evaluation.
Result: Benchmarked 21 leading LMMs including large open-weight models (GLM-4.5V 106B, InternVL3 78B) and proprietary models (Gemini 2.5 Pro, GPT-5). Results show strikingly low performance (27.8-53.9%), demonstrating the challenge of multimodal scientific reasoning.
Conclusion: Current LMMs struggle with detecting and resolving real-world multimodal inconsistencies in scientific papers, highlighting the need for progress toward trustworthy scientific assistants. The benchmark reveals significant gaps in multimodal reasoning capabilities.
Abstract: Large Multimodal Models (LMMs) are increasingly applied to scientific research, yet it remains unclear whether they can reliably understand and reason over the multimodal complexity of papers. A central challenge lies in detecting and resolving inconsistencies across text, figures, tables, and equations, issues that are often subtle, domain-specific, and ultimately undermine clarity, reproducibility, and trust. Existing benchmarks overlook this issue, either isolating single modalities or relying on synthetic errors that fail to capture real-world complexity. We introduce PRISMM-Bench (Peer-Review-sourced Inconsistency Set for Multimodal Models), the first benchmark grounded in real reviewer-flagged inconsistencies in scientific papers. Through a multi-stage pipeline of review mining, LLM-assisted filtering and human verification, we curate 384 inconsistencies from 353 papers. Based on this set, we design three tasks, namely inconsistency identification, remedy and pair matching, which assess a model’s capacity to detect, correct, and reason over inconsistencies across different modalities. Furthermore, to address the notorious problem of choice-only shortcuts in multiple-choice evaluation, where models exploit answer patterns without truly understanding the question, we further introduce structured JSON-based answer representations that minimize linguistic biases by reducing reliance on superficial stylistic cues. We benchmark 21 leading LMMs, including large open-weight models (GLM-4.5V 106B, InternVL3 78B) and proprietary models (Gemini 2.5 Pro, GPT-5 with high reasoning). Results reveal strikingly low performance (27.8-53.9%), underscoring the challenge of multimodal scientific reasoning and motivating progress towards trustworthy scientific assistants.
[344] Modal Aphasia: Can Unified Multimodal Models Describe Images From Memory?
Michael Aerni, Joshua Swanson, Kristina Nikolić, Florian Tramèr
Main category: cs.CV
TL;DR: Current unified multimodal models exhibit “modal aphasia” - they can accurately memorize and reproduce visual concepts but fail to articulate them in text, creating safety vulnerabilities when safeguards are applied to only one modality.
Details
Motivation: The paper investigates a systematic dissociation in multimodal models where they excel at visual memorization but fail at textual articulation, despite being trained on both modalities simultaneously. This reveals fundamental limitations in current unified multimodal architectures.Method: The researchers conducted controlled experiments using synthetic datasets across multiple architectures. They tested leading frontier models on tasks like reproducing iconic movie artwork versus providing textual descriptions, and demonstrated safety vulnerabilities by showing models aligned solely on text can still generate unsafe images.
Result: Models can generate near-perfect reproductions of iconic movie artwork but confuse crucial details in textual descriptions. Modal aphasia emerges as a fundamental property of current unified multimodal models, not just a training artifact. Safety frameworks are vulnerable when safeguards are applied to only one modality.
Conclusion: Modal aphasia is a systematic dissociation in current multimodal models that creates significant safety vulnerabilities. This highlights the need for more robust multimodal alignment approaches that ensure consistency across modalities rather than applying safeguards to only one modality.
Abstract: We present modal aphasia, a systematic dissociation in which current unified multimodal models accurately memorize concepts visually but fail to articulate them in writing, despite being trained on images and text simultaneously. For one, we show that leading frontier models can generate near-perfect reproductions of iconic movie artwork, but confuse crucial details when asked for textual descriptions. We corroborate those findings through controlled experiments on synthetic datasets in multiple architectures. Our experiments confirm that modal aphasia reliably emerges as a fundamental property of current unified multimodal models, not just as a training artifact. In practice, modal aphasia can introduce vulnerabilities in AI safety frameworks, as safeguards applied to one modality may leave harmful concepts accessible in other modalities. We demonstrate this risk by showing how a model aligned solely on text remains capable of generating unsafe images.
[345] Top-Down Semantic Refinement for Image Captioning
Jusheng Zhang, Kaitong Cai, Jing Yang, Jian Wang, Chengpei Tang, Keze Wang
Main category: cs.CV
TL;DR: TDSR reframes image captioning as hierarchical planning using MCTS to improve narrative coherence and detail in VLMs.
Details
Motivation: VLMs struggle with maintaining global narrative coherence while capturing rich details in image captioning due to their single-step generation approach.Method: Proposes Top-Down Semantic Refinement (TDSR) modeling captioning as MDP with efficient MCTS featuring visual-guided parallel expansion and lightweight value network.
Result: Achieves SOTA/competitive results on DetailCaps, COMPOSITIONCAP, POPE benchmarks, enhancing LLaVA-1.5, Qwen2.5-VL with 10x reduction in VLM calls.
Conclusion: TDSR effectively addresses VLMs’ coherence-detail tradeoff through hierarchical planning, offering plug-and-play improvement for existing models.
Abstract: Large Vision-Language Models (VLMs) face an inherent contradiction in image captioning: their powerful single-step generation capabilities often lead to a myopic decision-making process. This makes it difficult to maintain global narrative coherence while capturing rich details, a limitation that is particularly pronounced in tasks that require multi-step and complex scene description. To overcome this fundamental challenge, we redefine image captioning as a goal-oriented hierarchical refinement planning problem, and further propose a novel framework, named Top-Down Semantic Refinement (TDSR), which models the generation process as a Markov Decision Process (MDP). However, planning within the vast state space of a VLM presents a significant computational hurdle. Our core contribution, therefore, is the design of a highly efficient Monte Carlo Tree Search (MCTS) algorithm tailored for VLMs. By incorporating a visual-guided parallel expansion and a lightweight value network, our TDSR reduces the call frequency to the expensive VLM by an order of magnitude without sacrificing planning quality. Furthermore, an adaptive early stopping mechanism dynamically matches computational overhead to the image’s complexity. Extensive experiments on multiple benchmarks, including DetailCaps, COMPOSITIONCAP, and POPE, demonstrate that our TDSR, as a plug-and-play module, can significantly enhance the performance of existing VLMs (e.g., LLaVA-1.5, Qwen2.5-VL) by achieving state-of-the-art or highly competitive results in fine-grained description, compositional generalization, and hallucination suppression.
[346] Algorithms Trained on Normal Chest X-rays Can Predict Health Insurance Types
Chi-Yu Chen, Rawan Abulibdeh, Arash Asgari, Sebastián Andrés Cajas Ordóñez, Leo Anthony Celi, Deirdre Goode, Hassan Hamidi, Laleh Seyyed-Kalantari, Ned McCague, Thomas Sounack, Po-Chih Kuo
Main category: cs.CV
TL;DR: Deep vision models trained on chest X-rays can predict patients’ health insurance type (socioeconomic proxy) with significant accuracy, revealing that medical AI learns hidden social signatures from clinical data.
Details
Motivation: To investigate whether deep learning models can detect social inequality signals from medical images, challenging the assumption that medical images are neutral biological data and reframing fairness in medical AI.Method: Used state-of-the-art architectures (DenseNet121, SwinV2-B, MedMamba) trained on chest X-rays from MIMIC-CXR-JPG and CheXpert datasets to predict health insurance type as a proxy for socioeconomic status. Conducted ablation studies combining demographic features, single-racial group training, and patch-based occlusion analysis.
Result: Models achieved significant accuracy (AUC around 0.70 on MIMIC-CXR-JPG, 0.68 on CheXpert) in predicting health insurance type. Signal persists when controlling for demographics and within single racial groups. Patch occlusion revealed diffuse signal in upper and mid-thoracic regions.
Conclusion: Deep networks internalize subtle traces of clinical environments, equipment differences, or care pathways, learning socioeconomic segregation. This challenges medical image neutrality and reframes AI fairness as interrogating social fingerprints in clinical data.
Abstract: Artificial intelligence is revealing what medicine never intended to encode. Deep vision models, trained on chest X-rays, can now detect not only disease but also invisible traces of social inequality. In this study, we show that state-of-the-art architectures (DenseNet121, SwinV2-B, MedMamba) can predict a patient’s health insurance type, a strong proxy for socioeconomic status, from normal chest X-rays with significant accuracy (AUC around 0.70 on MIMIC-CXR-JPG, 0.68 on CheXpert). The signal was unlikely contributed by demographic features by our machine learning study combining age, race, and sex labels to predict health insurance types; it also remains detectable when the model is trained exclusively on a single racial group. Patch-based occlusion reveals that the signal is diffuse rather than localized, embedded in the upper and mid-thoracic regions. This suggests that deep networks may be internalizing subtle traces of clinical environments, equipment differences, or care pathways; learning socioeconomic segregation itself. These findings challenge the assumption that medical images are neutral biological data. By uncovering how models perceive and exploit these hidden social signatures, this work reframes fairness in medical AI: the goal is no longer only to balance datasets or adjust thresholds, but to interrogate and disentangle the social fingerprints embedded in clinical data itself.
[347] Procedural Mistake Detection via Action Effect Modeling
Wenliang Guo, Yujiang Pu, Yu Kong
Main category: cs.CV
TL;DR: Action Effect Modeling (AEM) framework for mistake detection in procedural tasks by jointly modeling action execution and outcomes through visual effect analysis and semantic alignment.
Details
Motivation: Existing mistake detection approaches focus only on how actions are performed, ignoring what they produce (action effects). Many errors manifest in outcomes rather than execution, such as unintended object states or spatial arrangements.Method: Proposes Action Effect Modeling (AEM) with: 1) Effect frame selection based on semantic relevance and visual quality, 2) Extraction of complementary cues from visual grounding and symbolic scene graphs, 3) Alignment in shared latent space for effect-aware representations, 4) Prompt-based detector with task-specific prompts aligned with intended execution semantics.
Result: Achieves state-of-the-art performance on EgoPER and CaptainCook4D benchmarks under challenging one-class classification (OCC) setting.
Conclusion: Modeling both execution and outcome yields more reliable mistake detection, and effect-aware representations have potential for broader downstream applications.
Abstract: Mistake detection in procedural tasks is essential for building intelligent systems that support learning and task execution. Existing approaches primarily analyze how an action is performed, while overlooking what it produces, i.e., the \textbf{action effect}. Yet many errors manifest not in the execution itself but in the resulting outcome, such as an unintended object state or incorrect spatial arrangement. To address this gap, we propose Action Effect Modeling (AEM), a unified framework that jointly captures action execution and its outcomes through a probabilistic formulation. AEM first identifies the outcome of an action by selecting the most informative effect frame based on semantic relevance and visual quality. It then extracts complementary cues from visual grounding and symbolic scene graphs, aligning them in a shared latent space to form robust effect-aware representations. To detect mistakes, we further design a prompt-based detector that incorporates task-specific prompts and aligns each action segment with its intended execution semantics. Our approach achieves state-of-the-art performance on the EgoPER and CaptainCook4D benchmarks under the challenging one-class classification (OCC) setting. These results demonstrate that modeling both execution and outcome yields more reliable mistake detection, and highlight the potential of effect-aware representations to benefit a broader range of downstream applications.
[348] Semantic-Guided Two-Stage GAN for Face Inpainting with Hybrid Perceptual Encoding
Abhigyan Bhattacharya, Hiranmoy Roy, Debotosh Bhattacharjee
Main category: cs.CV
TL;DR: A novel semantic-guided hierarchical synthesis approach for facial image inpainting that addresses challenges with large irregular masks through two-stage architecture combining CNNs and Vision Transformers for semantic layout generation, followed by multi-modal texture refinement.
Details
Motivation: Existing facial image inpainting methods struggle with large irregular masks, producing blurry textures, semantic inconsistencies, and unconvincing facial structures due to direct pixel-level synthesis and limited exploitation of facial priors.Method: Two-stage semantic-guided hierarchical synthesis: 1) Semantic layout generation combining local features (CNNs) and global features (Vision Transformers) to create clear semantic layouts, 2) Multi-Modal Texture Generator that refines layouts using multi-scale information with dynamic attention handling arbitrary masks without mask-specific training.
Result: Outperforms state-of-the-art methods on CelebA-HQ and FFHQ datasets, showing improvements in LPIPS, PSNR, and SSIM metrics, producing visually striking results with better semantic preservation in challenging large-area inpainting situations.
Conclusion: The proposed semantic-guided hierarchical synthesis approach effectively addresses facial image inpainting challenges, particularly for large irregular masks, by leveraging facial priors through a two-stage architecture that separates semantic understanding from texture refinement.
Abstract: Facial Image inpainting aim is to restore the missing or corrupted regions in face images while preserving identity, structural consistency and photorealistic image quality, a task specifically created for photo restoration. Though there are recent lot of advances in deep generative models, existing methods face problems with large irregular masks, often producing blurry textures on the edges of the masked region, semantic inconsistencies, or unconvincing facial structures due to direct pixel level synthesis approach and limited exploitation of facial priors. In this paper we propose a novel architecture, which address these above challenges through semantic-guided hierarchical synthesis. Our approach starts with a method that organizes and synthesizes information based on meaning, followed by refining the texture. This process gives clear insights into the facial structure before we move on to creating detailed images. In the first stage, we blend two techniques: one that focuses on local features with CNNs and global features with Vision Transformers. This helped us create clear and detailed semantic layouts. In the second stage, we use a Multi-Modal Texture Generator to refine these layouts by pulling in information from different scales, ensuring everything looks cohesive and consistent. The architecture naturally handles arbitrary mask configurations through dynamic attention without maskspecific training. Experiment on two datasets CelebA-HQ and FFHQ shows that our model outperforms other state-of-the-art methods, showing improvements in metrics like LPIPS, PSNR, and SSIM. It produces visually striking results with better semantic preservation, in challenging large-area inpainting situations.
[349] S2WMamba: A Spectral-Spatial Wavelet Mamba for Pansharpening
Haoyu Zhang, Junhan Luo, Yugang Cao, Jie Huang, Liangjian-Deng
Main category: cs.CV
TL;DR: S2WMamba: A pansharpening method using 2D/1D wavelet transforms and Mamba-based cross-modulation to disentangle spatial and spectral information for improved HRMS image generation.
Details
Motivation: Pansharpening faces the challenge of jointly processing PAN and MS images, which often entangles spatial detail with spectral fidelity. Existing methods struggle to effectively separate and integrate these different types of information.Method: Uses 2D Haar DWT on PAN images to localize spatial edges/textures, and channel-wise 1D Haar DWT on MS images to separate low/high-frequency spectral components. Features two parallel branches (Spectral and Spatial) that exchange information through Mamba-based cross-modulation with linear complexity, followed by multi-scale dynamic gate fusion.
Result: Outperforms recent baselines (FusionMamba, CANNet, U2Net, ARConv) on WV3, GF2, and QB datasets, improving PSNR by up to 0.23 dB and achieving HQNR 0.956 on full-resolution WV3. Ablations validate design choices.
Conclusion: S2WMamba effectively disentangles frequency information for pansharpening through explicit wavelet decomposition and lightweight cross-modal interaction, achieving state-of-the-art performance with efficient long-range dependency modeling.
Abstract: Pansharpening fuses a high-resolution PAN image with a low-resolution multispectral (LRMS) image to produce an HRMS image. A key difficulty is that jointly processing PAN and MS often entangles spatial detail with spectral fidelity. We propose S2WMamba, which explicitly disentangles frequency information and then performs lightweight cross-modal interaction. Concretely, a 2D Haar DWT is applied to PAN to localize spatial edges and textures, while a channel-wise 1D Haar DWT treats each pixel’s spectrum as a 1D signal to separate low/high-frequency components and limit spectral distortion. The resulting Spectral branch injects wavelet-extracted spatial details into MS features, and the Spatial branch refines PAN features using spectra from the 1D pyramid; the two branches exchange information through Mamba-based cross-modulation that models long-range dependencies with linear complexity. A multi-scale dynamic gate (multiplicative + additive) then adaptively fuses branch outputs.On WV3, GF2, and QB, S2WMamba matches or surpasses recent strong baselines (FusionMamba, CANNet, U2Net, ARConv), improving PSNR by up to 0.23 dB and reaching HQNR 0.956 on full-resolution WV3. Ablations justify the choice of 2D/1D DWT placement, parallel dual branches, and the fusion gate. Our code is available at https://github.com/KagUYa66/S2WMamba.
[350] Fourier-RWKV: A Multi-State Perception Network for Efficient Image Dehazing
Lirong Zheng, Yanshan Li, Rui Yu, Kaihao Zhang
Main category: cs.CV
TL;DR: Fourier-RWKV: A novel image dehazing framework using Multi-State Perception with linear complexity, combining spatial, frequency-domain, and semantic-relation perception for efficient non-uniform haze removal.
Details
Motivation: Image dehazing is crucial for reliable visual perception but remains challenging under real-world non-uniform haze conditions. Transformer-based methods capture global context well but have quadratic computational complexity that hinders real-time deployment.Method: Proposes Fourier-RWKV framework based on Multi-State Perception: (1) Spatial-form Perception via Deformable Quad-directional Token Shift for local haze variations, (2) Frequency-domain Perception via Fourier Mix block extending RWKV’s WKV attention to Fourier domain for long-range dependencies, (3) Semantic-relation Perception via Semantic Bridge Module with Dynamic Semantic Kernel Fusion for encoder-decoder alignment.
Result: Extensive experiments on multiple benchmarks demonstrate state-of-the-art performance across diverse haze scenarios while significantly reducing computational overhead, establishing favorable trade-off between restoration quality and practical efficiency.
Conclusion: Fourier-RWKV delivers comprehensive haze degradation modeling with linear complexity, achieving superior dehazing performance while being computationally efficient for practical deployment.
Abstract: Image dehazing is crucial for reliable visual perception, yet it remains highly challenging under real-world non-uniform haze conditions. Although Transformer-based methods excel at capturing global context, their quadratic computational complexity hinders real-time deployment. To address this, we propose Fourier Receptance Weighted Key Value (Fourier-RWKV), a novel dehazing framework based on a Multi-State Perception paradigm. The model achieves comprehensive haze degradation modeling with linear complexity by synergistically integrating three distinct perceptual states: (1) Spatial-form Perception, realized through the Deformable Quad-directional Token Shift (DQ-Shift) operation, which dynamically adjusts receptive fields to accommodate local haze variations; (2) Frequency-domain Perception, implemented within the Fourier Mix block, which extends the core WKV attention mechanism of RWKV from the spatial domain to the Fourier domain, preserving the long-range dependencies essential for global haze estimation while mitigating spatial attenuation; (3) Semantic-relation Perception, facilitated by the Semantic Bridge Module (SBM), which utilizes Dynamic Semantic Kernel Fusion (DSK-Fusion) to precisely align encoder-decoder features and suppress artifacts. Extensive experiments on multiple benchmarks demonstrate that Fourier-RWKV delivers state-of-the-art performance across diverse haze scenarios while significantly reducing computational overhead, establishing a favorable trade-off between restoration quality and practical efficiency. Code is available at: https://github.com/Dilizlr/Fourier-RWKV.
[351] Learning Patient-Specific Disease Dynamics with Latent Flow Matching for Longitudinal Imaging Generation
Hao Chen, Rui Yin, Yifan Chen, Qi Chen, Chao Li
Main category: cs.CV
TL;DR: Δ-LFM is a framework that models disease progression as a velocity field using Flow Matching, with patient-specific latent alignment to ensure trajectories follow severity-aligned axes in a semantically meaningful latent space.
Details
Motivation: Current generative approaches for disease progression modeling have key limitations: disease dynamics are continuous and monotonic, but latent representations are often scattered without semantic structure, and diffusion models disrupt continuity with random denoising. There's a need for interpretable, continuous progression models that align with clinical severity indicators.Method: Proposes Δ-LFM framework with two key components: 1) Treats disease dynamics as a velocity field using Flow Matching (FM) to align temporal evolution of patient data, capturing intrinsic disease dynamics; 2) Learns patient-specific latent alignment that enforces patient trajectories to lie along specific axes with magnitude increasing monotonically with disease severity, creating consistent and semantically meaningful latent space.
Result: Demonstrates strong empirical performance across three longitudinal MRI benchmarks, offering a new framework for interpreting and visualizing disease dynamics with improved continuity and semantic structure.
Conclusion: Δ-LFM provides an effective approach for modeling patient-specific latent progression with flow matching, offering interpretable disease dynamics visualization and progression modeling that aligns with clinical severity indicators.
Abstract: Understanding disease progression is a central clinical challenge with direct implications for early diagnosis and personalized treatment. While recent generative approaches have attempted to model progression, key mismatches remain: disease dynamics are inherently continuous and monotonic, yet latent representations are often scattered, lacking semantic structure, and diffusion-based models disrupt continuity with random denoising process. In this work, we propose to treat the disease dynamic as a velocity field and leverage Flow Matching (FM) to align the temporal evolution of patient data. Unlike prior methods, it captures the intrinsic dynamic of disease, making the progression more interpretable. However, a key challenge remains: in latent space, Auto-Encoders (AEs) do not guarantee alignment across patients or correlation with clinical-severity indicators (e.g., age and disease conditions). To address this, we propose to learn patient-specific latent alignment, which enforces patient trajectories to lie along a specific axis, with magnitude increasing monotonically with disease severity. This leads to a consistent and semantically meaningful latent space. Together, we present $Δ$-LFM, a framework for modeling patient-specific latent progression with flow matching. Across three longitudinal MRI benchmarks, $Δ$-LFM demonstrates strong empirical performance and, more importantly, offers a new framework for interpreting and visualizing disease dynamics.
[352] Geometry-to-Image Synthesis-Driven Generative Point Cloud Registration
Haobo Jiang, Jin Xie, Jian Yang, Liang Yu, Jianmin Zheng
Main category: cs.CV
TL;DR: Generative Point Cloud Registration bridges 2D generative models with 3D matching by generating cross-view consistent image pairs aligned with point clouds for geometry-color feature fusion.
Details
Motivation: To enhance 3D registration performance by leveraging advanced 2D generative models to create consistent image pairs that align with source and target point clouds, enabling better geometry-color feature fusion for robust matching.Method: Introduces DepthMatch-ControlNet and LiDARMatch-ControlNet: matching-specific controllable 2D generative models. DepthMatch-ControlNet uses depth-conditioned generation for perspective-view RGB images consistent with depth maps. LiDARMatch-ControlNet extends this to LiDAR by conditioning on equirectangular range maps for panoramic RGB generation. Both incorporate coupled conditional denoising and prompt guidance for cross-view consistency.
Result: Extensive experiments on 3DMatch, ScanNet (depth-camera), and Dur360BEV (LiDAR) datasets demonstrate the effectiveness of the approach in improving registration performance across various settings.
Conclusion: The proposed generative 3D registration paradigm successfully bridges 2D generative models with 3D matching tasks, providing a general framework that can be integrated into existing registration methods to enhance performance through geometry-color feature fusion.
Abstract: In this paper, we propose a novel 3D registration paradigm, Generative Point Cloud Registration, which bridges advanced 2D generative models with 3D matching tasks to enhance registration performance. Our key idea is to generate cross-view consistent image pairs that are well-aligned with the source and target point clouds, enabling geometry-color feature fusion to facilitate robust matching. To ensure high-quality matching, the generated image pair should feature both 2D-3D geometric consistency and cross-view texture consistency. To this end, we introduce DepthMatch-ControlNet and LiDARMatch-ControlNet, two matching-specific, controllable 2D generative models. Specifically, for depth camera-based 3D registration with point clouds derived from the depth maps, DepthMatch-ControlNet leverages the depth-conditioned generation capabilities of ControlNet to synthesize perspective-view RGB images that are geometrically consistent with depth maps, ensuring accurate 2D-3D alignment. Additionally, by incorporating a coupled conditional denoising scheme and coupled prompt guidance, it further promotes cross-view feature interaction, guiding texture consistency generation. To address LiDAR-based 3D registration with point clouds captured by LiDAR sensors, LiDARMatch-ControlNet extends this framework by conditioning on paired equirectangular range maps projected from 360-degree LiDAR point clouds, generating corresponding panoramic RGB images. Our generative 3D registration paradigm is general and can be seamlessly integrated into a wide range of existing registration methods to improve their performance. Extensive experiments on the 3DMatch and ScanNet datasets (for depth-camera settings), as well as the Dur360BEV dataset (for LiDAR settings), demonstrate the effectiveness of our approach.
[353] ALERT Open Dataset and Input-Size-Agnostic Vision Transformer for Driver Activity Recognition using IR-UWB
Jeongjun Park, Sunwook Hwang, Hyeonho Noh, Jin Mo Yang, Hyun Jong Yang, Saewoong Bahk
Main category: cs.CV
TL;DR: Proposes ISA-ViT framework and ALERT dataset for radar-based driver activity recognition using UWB radar, addressing input dimension challenges for Vision Transformers on non-standard radar data.
Details
Motivation: Address distracted driving detection using IR-UWB radar which offers privacy and interference resistance, but faces challenges with lack of large-scale real-world datasets and adapting fixed-input Vision Transformers to non-standard radar data dimensions.Method: Introduces ISA-ViT (input-size-agnostic Vision Transformer) that resizes UWB data while preserving radar-specific information like Doppler shifts and phase characteristics. Uses adjustable patch configurations and pre-trained positional embedding vectors, plus domain fusion combining range- and frequency-domain features.
Result: ISA-ViT achieves 22.68% accuracy improvement over existing ViT-based approaches for UWB-based driver activity recognition. ALERT dataset contains 10,220 radar samples of seven distracted driving activities collected in real driving conditions.
Conclusion: The work facilitates development of robust distracted driving detection systems by providing ALERT dataset and ISA-ViT framework that overcomes input dimension limitations for radar data in Vision Transformers.
Abstract: Distracted driving contributes to fatal crashes worldwide. To address this, researchers are using driver activity recognition (DAR) with impulse radio ultra-wideband (IR-UWB) radar, which offers advantages such as interference resistance, low power consumption, and privacy preservation. However, two challenges limit its adoption: the lack of large-scale real-world UWB datasets covering diverse distracted driving behaviors, and the difficulty of adapting fixed-input Vision Transformers (ViTs) to UWB radar data with non-standard dimensions. This work addresses both challenges. We present the ALERT dataset, which contains 10,220 radar samples of seven distracted driving activities collected in real driving conditions. We also propose the input-size-agnostic Vision Transformer (ISA-ViT), a framework designed for radar-based DAR. The proposed method resizes UWB data to meet ViT input requirements while preserving radar-specific information such as Doppler shifts and phase characteristics. By adjusting patch configurations and leveraging pre-trained positional embedding vectors (PEVs), ISA-ViT overcomes the limitations of naive resizing approaches. In addition, a domain fusion strategy combines range- and frequency-domain features to further improve classification performance. Comprehensive experiments demonstrate that ISA-ViT achieves a 22.68% accuracy improvement over an existing ViT-based approach for UWB-based DAR. By publicly releasing the ALERT dataset and detailing our input-size-agnostic strategy, this work facilitates the development of more robust and scalable distracted driving detection systems for real-world deployment.
[354] Two-Step Data Augmentation for Masked Face Detection and Recognition: Turning Fake Masks to Real
Yan Yang, George Bebis, Mircea Nicolescu
Main category: cs.CV
TL;DR: A two-step generative data augmentation framework combining rule-based mask warping with GAN-based image translation to generate realistic masked-face samples for improving face detection/recognition under data scarcity.
Details
Motivation: Data scarcity and distribution shift pose major challenges for masked face detection and recognition, creating a need for better data augmentation methods beyond purely synthetic transformations.Method: Two-step approach: 1) Rule-based mask warping, 2) Unpaired image-to-image translation using GANs. Introduces non-mask preservation loss and stochastic noise injection to stabilize training and enhance sample diversity.
Result: Consistent qualitative improvements over rule-based warping alone, complements existing GAN-based methods like IAMGAN. The proposed components effectively enhance sample diversity and training stability.
Conclusion: The framework provides effective data-centric augmentation for face recognition tasks, suggesting directions for future improvements in generative data augmentation for masked face scenarios.
Abstract: Data scarcity and distribution shift pose major challenges for masked face detection and recognition. We propose a two-step generative data augmentation framework that combines rule-based mask warping with unpaired image-to-image translation using GANs, enabling the generation of realistic masked-face samples beyond purely synthetic transformations. Compared to rule-based warping alone, the proposed approach yields consistent qualitative improvements and complements existing GAN-based masked face generation methods such as IAMGAN. We introduce a non-mask preservation loss and stochastic noise injection to stabilize training and enhance sample diversity. Experimental observations highlight the effectiveness of the proposed components and suggest directions for future improvements in data-centric augmentation for face recognition tasks.
[355] X-ray Insights Unleashed: Pioneering the Enhancement of Multi-Label Long-Tail Data
Xinquan Yang, Jinheng Xie, Yawen Huang, Yuexiang Li, Huimin Huang, Hao Zheng, Xian Wu, Yefeng Zheng, Linlin Shen
Main category: cs.CV
TL;DR: A novel data synthesis pipeline for augmenting rare lung lesions in chest X-rays using normal X-rays and diffusion models with language model guidance.
Details
Motivation: Long-tailed pulmonary anomalies in chest radiography are diagnostically challenging due to scarcity of rare lesion examples, limiting generative methods' effectiveness and diagnostic precision.Method: Proposes a data synthesis pipeline: 1) Train diffusion model on normal X-rays, 2) Use pre-trained model to inpaint head lesions in diseased X-rays while preserving tail classes as augmented data, 3) Integrate Large Language Model Knowledge Guidance (LKG) module and Progressive Incremental Learning (PIL) strategy to stabilize fine-tuning.
Result: Comprehensive evaluations on public lung datasets MIMIC and CheXpert demonstrate the method sets a new benchmark in performance.
Conclusion: The proposed approach effectively addresses the data scarcity problem for rare lung lesions through innovative data synthesis using normal X-rays and diffusion models with language model guidance.
Abstract: Long-tailed pulmonary anomalies in chest radiography present formidable diagnostic challenges. Despite the recent strides in diffusion-based methods for enhancing the representation of tailed lesions, the paucity of rare lesion exemplars curtails the generative capabilities of these approaches, thereby leaving the diagnostic precision less than optimal. In this paper, we propose a novel data synthesis pipeline designed to augment tail lesions utilizing a copious supply of conventional normal X-rays. Specifically, a sufficient quantity of normal samples is amassed to train a diffusion model capable of generating normal X-ray images. This pre-trained diffusion model is subsequently utilized to inpaint the head lesions present in the diseased X-rays, thereby preserving the tail classes as augmented training data. Additionally, we propose the integration of a Large Language Model Knowledge Guidance (LKG) module alongside a Progressive Incremental Learning (PIL) strategy to stabilize the inpainting fine-tuning process. Comprehensive evaluations conducted on the public lung datasets MIMIC and CheXpert demonstrate that the proposed method sets a new benchmark in performance.
[356] MorphAny3D: Unleashing the Power of Structured Latent in 3D Morphing
Xiaokun Sun, Zeyu Cai, Hao Tang, Ying Tai, Jian Yang, Zhenyu Zhang
Main category: cs.CV
TL;DR: MorphAny3D is a training-free framework for high-quality 3D morphing using Structured Latent (SLAT) representations with attention-based feature blending for semantic consistency and temporal smoothness.
Details
Motivation: 3D morphing is challenging due to difficulties in generating semantically consistent and temporally smooth deformations, especially across different object categories. Existing methods struggle with maintaining structural coherence and temporal consistency during morphing sequences.Method: Uses Structured Latent (SLAT) representations with two key components: Morphing Cross-Attention (MCA) for fusing source and target information for structural coherence, and Temporal-Fused Self-Attention (TFSA) for enhancing temporal consistency by incorporating features from preceding frames. Also includes orientation correction to mitigate pose ambiguity.
Result: Generates state-of-the-art morphing sequences, even for challenging cross-category cases. Supports advanced applications like decoupled morphing and 3D style transfer, and can generalize to other SLAT-based generative models.
Conclusion: MorphAny3D provides an effective training-free framework for high-quality 3D morphing by intelligently blending SLAT features within attention mechanisms, achieving superior results for both within-category and cross-category morphing tasks.
Abstract: 3D morphing remains challenging due to the difficulty of generating semantically consistent and temporally smooth deformations, especially across categories. We present MorphAny3D, a training-free framework that leverages Structured Latent (SLAT) representations for high-quality 3D morphing. Our key insight is that intelligently blending source and target SLAT features within the attention mechanisms of 3D generators naturally produces plausible morphing sequences. To this end, we introduce Morphing Cross-Attention (MCA), which fuses source and target information for structural coherence, and Temporal-Fused Self-Attention (TFSA), which enhances temporal consistency by incorporating features from preceding frames. An orientation correction strategy further mitigates the pose ambiguity within the morphing steps. Extensive experiments show that our method generates state-of-the-art morphing sequences, even for challenging cross-category cases. MorphAny3D further supports advanced applications such as decoupled morphing and 3D style transfer, and can be generalized to other SLAT-based generative models. Project page: https://xiaokunsun.github.io/MorphAny3D.github.io/.
[357] CliffordNet: All You Need is Geometric Algebra
Zhongping Ji
Main category: cs.CV
TL;DR: CliffordNet proposes a vision backbone based purely on Geometric Algebra, using Clifford Geometric Product as a unified interaction mechanism that eliminates the need for separate spatial and channel mixing modules, achieving state-of-the-art efficiency.
Details
Motivation: The paper challenges the conventional paradigm of stacking heuristic modules (spatial mixers + channel mixers) in computer vision architectures, seeking to return to mathematical first principles by grounding vision models in Geometric Algebra.Method: Proposes Clifford Algebra Network (CliffordNet) using Clifford Geometric Product (uv = u·v + u∧v) as a unified interaction mechanism that simultaneously captures feature coherence (via generalized inner product) and structural variation (via exterior wedge product). Implemented with efficient sparse rolling mechanism with O(N) linear complexity.
Result: CliffordNet establishes new Pareto frontier: Nano variant achieves 77.82% accuracy on CIFAR-100 with only 1.4M parameters (8× fewer than ResNet-18’s 11.2M), Lite variant (2.6M) sets new SOTA for tiny models at 79.05%. The geometric interaction is so representationally dense that standard FFNs become redundant.
Conclusion: Global understanding can emerge solely from rigorous, algebraically complete local interactions, suggesting a potential paradigm shift where “geometry is all you need” in vision architectures.
Abstract: Modern computer vision architectures, from CNNs to Transformers, predominantly rely on the stacking of heuristic modules: spatial mixers (Attention/Conv) followed by channel mixers (FFNs). In this work, we challenge this paradigm by returning to mathematical first principles. We propose the Clifford Algebra Network (CAN), also referred to as CliffordNet, a vision backbone grounded purely in Geometric Algebra. Instead of engineering separate modules for mixing and memory, we derive a unified interaction mechanism based on the Clifford Geometric Product ($uv = u \cdot v + u \wedge v$). This operation ensures algebraic completeness regarding the Geometric Product by simultaneously capturing feature coherence (via the generalized inner product) and structural variation (via the exterior wedge product). Implemented via an efficient sparse rolling mechanism with strict linear complexity $O(N)$, our model reveals a surprising emergent property: the geometric interaction is so representationally dense that standard Feed-Forward Networks (FFNs) become redundant. Empirically, CliffordNet establishes a new Pareto frontier: our Nano variant achieves 77.82% accuracy on CIFAR-100 with only 1.4M parameters, effectively matching the heavy-weight ResNet-18 (11.2M) with $8\times$ fewer parameters, while our Lite variant (2.6M) sets a new SOTA for tiny models at 79.05%. Our results suggest that global understanding can emerge solely from rigorous, algebraically complete local interactions, potentially signaling a shift where geometry is all you need. Code is available at https://github.com/ParaMind2025/CAN.
[358] 3AM: 3egment Anything with Geometric Consistency in Videos
Yang-Che Sun, Cheng Sun, Chin-Yang Lin, Fu-En Yang, Min-Hung Chen, Yen-Yu Lin, Yu-Lun Liu
Main category: cs.CV
TL;DR: 3AM enhances SAM2 for video object segmentation by integrating 3D-aware features from MUSt3R, enabling geometry-consistent recognition without requiring camera poses or depth at inference.
Details
Motivation: Video object segmentation methods like SAM2 rely on appearance features and struggle with large viewpoint changes, while 3D instance segmentation methods require camera poses and depth maps. There's a need for a method that achieves geometry-consistent recognition using only RGB input.Method: Integrates 3D-aware features from MUSt3R into SAM2 using a lightweight Feature Merger that fuses multi-level MUSt3R features encoding implicit geometric correspondence. Combines these with SAM2’s appearance features and uses field-of-view aware sampling to ensure frames observe spatially consistent object regions for reliable 3D correspondence learning.
Result: On challenging datasets with wide-baseline motion (ScanNet++, Replica), 3AM substantially outperforms SAM2 and extensions, achieving 90.6% IoU and 71.7% Positive IoU on ScanNet++’s Selected Subset, improving over state-of-the-art VOS methods by +15.9 and +30.4 points.
Conclusion: 3AM successfully enhances video object segmentation with 3D-aware features, achieving geometry-consistent recognition without requiring camera poses or preprocessing at inference, making it practical for real-world applications.
Abstract: Video object segmentation methods like SAM2 achieve strong performance through memory-based architectures but struggle under large viewpoint changes due to reliance on appearance features. Traditional 3D instance segmentation methods address viewpoint consistency but require camera poses, depth maps, and expensive preprocessing. We introduce 3AM, a training-time enhancement that integrates 3D-aware features from MUSt3R into SAM2. Our lightweight Feature Merger fuses multi-level MUSt3R features that encode implicit geometric correspondence. Combined with SAM2’s appearance features, the model achieves geometry-consistent recognition grounded in both spatial position and visual similarity. We propose a field-of-view aware sampling strategy ensuring frames observe spatially consistent object regions for reliable 3D correspondence learning. Critically, our method requires only RGB input at inference, with no camera poses or preprocessing. On challenging datasets with wide-baseline motion (ScanNet++, Replica), 3AM substantially outperforms SAM2 and extensions, achieving 90.6% IoU and 71.7% Positive IoU on ScanNet++’s Selected Subset, improving over state-of-the-art VOS methods by +15.9 and +30.4 points. Project page: https://jayisaking.github.io/3AM-Page/
[359] Scaling Test-time Inference for Visual Grounding
Guanqi Zhan, Changye Li, Zhijian Liu, Yao Lu, Yi Wu, Song Han, Ligeng Zhu
Main category: cs.CV
TL;DR: EGM improves visual grounding in small VLMs by scaling test-time computation (#generated tokens) to match large model performance while being faster and more deployment-friendly.
Details
Motivation: Small VLMs lag behind large VLMs in visual grounding primarily due to language understanding limitations rather than visual processing. Large VLMs are heavy for deployment and slow for inference, creating a need for efficient alternatives.Method: EGM (Efficient visual Grounding language Models) scales test-time computation by increasing the number of generated tokens. This approach leverages that small models have cheaper per-token costs, making them more efficient than running large models directly.
Result: EGM-Qwen3-VL-8B achieves 91.4 IoU on RefCOCO with 737ms latency (5.9x faster than Qwen3-VL-235B’s 4,320ms for 90.5 IoU). The method also improves amodal grounding (predicting both visible and occluded object parts) and consistently boosts small models to match or outperform larger ones.
Conclusion: EGM enables small VLMs to achieve visual grounding performance comparable to large models while being significantly faster and more deployment-friendly through test-time computation scaling.
Abstract: Visual grounding is an essential capability of Visual Language Models (VLMs) to understand the real physical world. Previous state-of-the-art grounding visual language models usually have large model sizes, making them heavy for deployment and slow for inference. However, we notice that the sizes of visual encoders are nearly the same for small and large VLMs and the major difference is the sizes of the language models. Small VLMs fall behind larger VLMs in grounding because of the difference in language understanding capability rather than visual information handling. To mitigate the gap, we introduce ‘Efficient visual Grounding language Models’ (EGM): a method to scale the test-time computation (#generated tokens). Scaling the test-time computation of a small model is deployment-friendly, and yields better end-to-end latency as the cost of each token is much cheaper compared to directly running a large model. On the RefCOCO benchmark, our EGM-Qwen3-VL-8B demonstrates 91.4 IoU with an average of 737ms (5.9x faster) latency while Qwen3-VL-235B demands 4,320ms to achieve 90.5 IoU. To validate our approach’s generality, we further set up a new amodal grounding setting that requires the model to predict both the visible and occluded parts of the objects. Experiments show our method can consistently and significantly improve the vanilla grounding and amodal grounding capabilities of small models to be on par with or outperform the larger models, thereby improving the efficiency for visual grounding.
[360] Tracing 3D Anatomy in 2D Strokes: A Multi-Stage Projection Driven Approach to Cervical Spine Fracture Identification
Fabi Nahian Madhurja, Rusab Sarmun, Muhammad E. H. Chowdhury, Adam Mushtak, Israa Al-Hashimi, Sohaib Bassam Zoghoul
Main category: cs.CV
TL;DR: 2D projection-based pipeline for cervical spine fracture detection in 3D CT volumes using optimized projections, YOLOv8 localization, DenseNet121-Unet segmentation, and 2.5D ensemble models.
Details
Motivation: Cervical spine fractures require precise detection for clinical management. The study aims to develop an efficient automated pipeline that reduces computational complexity compared to traditional 3D methods while maintaining high performance.Method: Uses 2D axial, sagittal, and coronal projections to approximate 3D volumes. YOLOv8 identifies regions of interest from all views. DenseNet121-Unet performs multi-label segmentation using variance- and energy-based projections. Extracts individual vertebra volumes and analyzes fractures using ensemble of 2.5D Spatio-Sequential models with raw slices and projections.
Result: Achieves 3D mIoU of 94.45% for localization, Dice score of 87.86% for segmentation, and vertebra-level/patient-level F1 scores of 68.15/82.26 and ROC-AUC scores of 91.62/83.04 for fracture detection. Competitive with expert radiologists in interobserver analysis.
Conclusion: The 2D projection-based approach effectively reduces computational complexity while maintaining competitive performance for cervical spine fracture detection, validated through explainability studies and comparison with radiologists.
Abstract: Cervical spine fractures are critical medical conditions requiring precise and efficient detection for effective clinical management. This study explores the viability of 2D projection-based vertebra segmentation for vertebra-level fracture detection in 3D CT volumes, presenting an end-to-end pipeline for automated analysis of cervical vertebrae (C1-C7). By approximating a 3D volume through optimized 2D axial, sagittal, and coronal projections, regions of interest are identified using the YOLOv8 model from all views and combined to approximate the 3D cervical spine area, achieving a 3D mIoU of 94.45 percent. This projection-based localization strategy reduces computational complexity compared to traditional 3D segmentation methods while maintaining high performance. It is followed by a DenseNet121-Unet-based multi-label segmentation leveraging variance- and energy-based projections, achieving a Dice score of 87.86 percent. Strategic approximation of 3D vertebral masks from these 2D segmentation masks enables the extraction of individual vertebra volumes. The volumes are analyzed for fractures using an ensemble of 2.5D Spatio-Sequential models incorporating both raw slices and projections per vertebra for complementary evaluation. This ensemble achieves vertebra-level and patient-level F1 scores of 68.15 and 82.26, and ROC-AUC scores of 91.62 and 83.04, respectively. We further validate our approach through an explainability study that provides saliency map visualizations highlighting anatomical regions relevant for diagnosis, and an interobserver variability analysis comparing our model’s performance with expert radiologists, demonstrating competitive results.
[361] PhaSR: Generalized Image Shadow Removal with Physically Aligned Priors
Chia-Ming Lee, Yu-Fan Lin, Yu-Jou Hsiao, Jin-Hui Jiang, Yu-Lun Liu, Chih-Chung Hsu
Main category: cs.CV
TL;DR: PhaSR is a shadow removal method that uses physically aligned normalization and geometric-semantic attention to handle diverse lighting conditions from single-light to multi-source ambient illumination.
Details
Motivation: Shadow removal under diverse lighting conditions requires disentangling illumination from intrinsic reflectance, which is challenging when physical priors are not properly aligned. Traditional methods often fail under multi-source ambient illumination.Method: Two-stage approach: 1) Physically Aligned Normalization (PAN) performs closed-form illumination correction via Gray-world normalization, log-domain Retinex decomposition, and dynamic range recombination. 2) Geometric-Semantic Rectification Attention (GSRA) extends differential attention to cross-modal alignment, harmonizing depth-derived geometry with DINO-v2 semantic embeddings.
Result: Competitive performance in shadow removal with lower complexity and generalization to ambient lighting where traditional methods fail under multi-source illumination.
Conclusion: PhaSR effectively addresses shadow removal under diverse lighting conditions through dual-level prior alignment, enabling robust performance from single-light shadows to multi-source ambient lighting.
Abstract: Shadow removal under diverse lighting conditions requires disentangling illumination from intrinsic reflectance, a challenge compounded when physical priors are not properly aligned. We propose PhaSR (Physically Aligned Shadow Removal), addressing this through dual-level prior alignment to enable robust performance from single-light shadows to multi-source ambient lighting. First, Physically Aligned Normalization (PAN) performs closed-form illumination correction via Gray-world normalization, log-domain Retinex decomposition, and dynamic range recombination, suppressing chromatic bias. Second, Geometric-Semantic Rectification Attention (GSRA) extends differential attention to cross-modal alignment, harmonizing depth-derived geometry with DINO-v2 semantic embeddings to resolve modal conflicts under varying illumination. Experiments show competitive performance in shadow removal with lower complexity and generalization to ambient lighting where traditional methods fail under multi-source illumination. Our source code is available at https://github.com/ming053l/PhaSR.
[362] Semantic-Guided Dynamic Sparsification for Pre-Trained Model-based Class-Incremental Learning
Ruiqi Liu, Boyu Diao, Zijia An, Runjie Shao, Zhulin An, Fei Wang, Yongjun Xu
Main category: cs.CV
TL;DR: SGDS is a novel method for Class-Incremental Learning that uses semantic-guided dynamic sparsification to create class-specific activation subspaces, balancing knowledge transfer and interference prevention without rigid parameter constraints.
Details
Motivation: Existing CIL methods that freeze pre-trained models and use orthogonal adapters to prevent inter-task interference are detrimental to plasticity. There's a need for methods that can effectively mitigate interference while maintaining learning flexibility.Method: Proposes Semantic-Guided Dynamic Sparsification (SGDS) that proactively guides activation space by governing orientation and rank of subspaces through targeted sparsification. Encourages similar classes to share compact activation subspaces for knowledge transfer, while assigning non-overlapping subspaces to dissimilar classes to prevent interference.
Result: Extensive experiments on various benchmark datasets demonstrate state-of-the-art performance of SGDS in class-incremental learning tasks.
Conclusion: SGDS effectively mitigates interference in class-incremental learning without imposing rigid constraints on parameter space, achieving better balance between plasticity and stability through activation space sculpting.
Abstract: Class-Incremental Learning (CIL) requires a model to continually learn new classes without forgetting old ones. A common and efficient solution freezes a pre-trained model and employs lightweight adapters, whose parameters are often forced to be orthogonal to prevent inter-task interference. However, we argue that this parameter-constraining method is detrimental to plasticity. To this end, we propose Semantic-Guided Dynamic Sparsification (SGDS), a novel method that proactively guides the activation space by governing the orientation and rank of its subspaces through targeted sparsification. Specifically, SGDS promotes knowledge transfer by encouraging similar classes to share a compact activation subspace, while simultaneously preventing interference by assigning non-overlapping activation subspaces to dissimilar classes. By sculpting class-specific sparse subspaces in the activation space, SGDS effectively mitigates interference without imposing rigid constraints on the parameter space. Extensive experiments on various benchmark datasets demonstrate the state-of-the-art performance of SGDS.
[363] Q-Hawkeye: Reliable Visual Policy Optimization for Image Quality Assessment
Wulin Xie, Rui Dai, Ruidong Ding, Kaikui Liu, Xiangxiang Chu, Xinwen Hou, Jie Wen
Main category: cs.CV
TL;DR: Q-Hawkeye is an RL-based image quality assessment framework that addresses reliability limitations in existing methods through uncertainty-aware dynamic optimization and perception-aware optimization.
Details
Motivation: Current RL-based IQA methods using MLLMs have two key reliability issues: (1) they apply uniform advantage weighting despite varying prediction stability across samples, amplifying noisy signals from unstable samples, and (2) they emphasize text-grounded reasoning while overlooking visual perception ability for image content.Method: Proposes Q-Hawkeye with two main components: (1) Uncertainty-Aware Dynamic Optimization that estimates predictive uncertainty using variance of predicted scores across multiple rollouts and uses this uncertainty to reweight each sample’s update strength, and (2) Perception-Aware Optimization that constructs paired inputs of degraded images and their originals, introducing an Implicit Perception Loss to constrain quality judgments to genuine visual evidence.
Result: Extensive experiments show Q-Hawkeye outperforms state-of-the-art methods and generalizes better across multiple datasets.
Conclusion: Q-Hawkeye provides a more reliable RL-based visual policy optimization framework for image quality assessment by addressing both uncertainty and perception limitations in existing methods.
Abstract: Image Quality Assessment (IQA) predicts perceptual quality scores consistent with human judgments. Recent RL-based IQA methods built on MLLMs focus on generating visual quality descriptions and scores, ignoring two key reliability limitations: (i) although the model’s prediction stability varies significantly across training samples, existing GRPO-based methods apply uniform advantage weighting, thereby amplifying noisy signals from unstable samples in gradient updates; (ii) most works emphasize text-grounded reasoning over images while overlooking the model’s visual perception ability of image content. In this paper, we propose Q-Hawkeye, an RL-based reliable visual policy optimization framework that redesigns the learning signal through unified Uncertainty-Aware Dynamic Optimization and Perception-Aware Optimization. Q-Hawkeye estimates predictive uncertainty using the variance of predicted scores across multiple rollouts and leverages this uncertainty to reweight each sample’s update strength, stabilizing policy optimization. To strengthen perceptual reliability, we construct paired inputs of degraded images and their original images and introduce an Implicit Perception Loss that constrains the model to ground its quality judgments in genuine visual evidence. Extensive experiments demonstrate that Q-Hawkeye outperforms state-of-the-art methods and generalizes better across multiple datasets. Our dataset and code are available at https://github.com/AMAP-ML/Q-Hawkeye.
[364] ShotFinder: Imagination-Driven Open-Domain Video Shot Retrieval via Web Search
Tao Yu, Haopeng Jin, Hao Wang, Shenghua Chai, Yujia Yang, Junhao Gong, Jiaming Guo, Minghui Zhang, Xinlong Chen, Zhenghao Zhang, Yuxuan Zhou, Yufei Xiong, Shanbin Zhang, Jiabing Yang, Hongzhu Yi, Xinming Wang, Cheng Zhong, Xiao Ma, Zhang Zhang, Yan Huang, Liang Wang
Main category: cs.CV
TL;DR: ShotFinder is a benchmark for open-domain video shot retrieval with controllable constraints, revealing significant gaps in multimodal LLM capabilities for temporal, color, and style understanding.
Details
Motivation: Existing LLM research focuses on text or static multimodal settings, but open-domain video shot retrieval with rich temporal structure and complex semantics lacks systematic benchmarks and analysis.Method: Introduces ShotFinder benchmark with 1,210 samples across 20 categories, formalizing editing requirements as keyframe-oriented shot descriptions with five controllable constraints. Proposes a three-stage retrieval pipeline: query expansion via video imagination, candidate retrieval with search engine, and description-guided temporal localization.
Result: Experiments show significant gap between model and human performance, with clear imbalance across constraints: temporal localization is relatively tractable, while color and visual style remain major challenges.
Conclusion: Open-domain video shot retrieval is a critical capability that multimodal large models have yet to overcome, highlighting the need for better video understanding with temporal and multimodal constraints.
Abstract: In recent years, large language models (LLMs) have made rapid progress in information retrieval, yet existing research has mainly focused on text or static multimodal settings. Open-domain video shot retrieval, which involves richer temporal structure and more complex semantics, still lacks systematic benchmarks and analysis. To fill this gap, we introduce ShotFinder, a benchmark that formalizes editing requirements as keyframe-oriented shot descriptions and introduces five types of controllable single-factor constraints: Temporal order, Color, Visual style, Audio, and Resolution. We curate 1,210 high-quality samples from YouTube across 20 thematic categories, using large models for generation with human verification. Based on the benchmark, we propose ShotFinder, a text-driven three-stage retrieval and localization pipeline: (1) query expansion via video imagination, (2) candidate video retrieval with a search engine, and (3) description-guided temporal localization. Experiments on multiple closed-source and open-source models reveal a significant gap to human performance, with clear imbalance across constraints: temporal localization is relatively tractable, while color and visual style remain major challenges. These results reveal that open-domain video shot retrieval is still a critical capability that multimodal large models have yet to overcome.
[365] Cross-Modal Purification and Fusion for Small-Object RGB-D Transmission-Line Defect Detection
Jiaming Cui, Wenqiang Li, Shuai Zhou, Ruifeng Qin, Feng Shen
Main category: cs.CV
TL;DR: CMAFNet is a cross-modal fusion network for transmission line defect detection that integrates RGB and depth data using a purify-then-fuse approach with dictionary-based feature purification and partial-channel attention for improved small defect detection in complex backgrounds.
Details
Motivation: Transmission line defect detection faces challenges from small-scale defects, complex backgrounds, and illumination variations. RGB-based detectors struggle to distinguish geometrically subtle defects from visually similar background structures due to limited chromatic contrast.Method: CMAFNet uses a cross-modal alignment and fusion network with a purify-then-fuse paradigm. It includes: 1) Semantic Recomposition Module for dictionary-based feature purification via learned codebook to suppress modality-specific noise, and 2) Contextual Semantic Integration Framework with partial-channel attention to capture global spatial dependencies. Position-wise normalization enforces reconstruction-driven cross-modal alignment.
Result: On the TLRGBD benchmark (94.5% small objects), CMAFNet achieves 32.2% mAP@50 and 12.5% APs, outperforming the strongest baseline by 9.8 and 4.0 percentage points. A lightweight variant reaches 24.8% mAP50 at 228 FPS with only 4.9M parameters, surpassing YOLO-based detectors while matching transformer-based methods at lower computational cost.
Conclusion: CMAFNet effectively addresses transmission line defect detection challenges by integrating RGB and depth modalities through principled cross-modal alignment and fusion, achieving superior performance for small defect detection in complex environments with efficient computational requirements.
Abstract: Transmission line defect detection remains challenging for automated UAV inspection due to the dominance of small-scale defects, complex backgrounds, and illumination variations. Existing RGB-based detectors, despite recent progress, struggle to distinguish geometrically subtle defects from visually similar background structures under limited chromatic contrast. This paper proposes CMAFNet, a Cross-Modal Alignment and Fusion Network that integrates RGB appearance and depth geometry through a principled purify-then-fuse paradigm. CMAFNet consists of a Semantic Recomposition Module that performs dictionary-based feature purification via a learned codebook to suppress modality-specific noise while preserving defect-discriminative information, and a Contextual Semantic Integration Framework that captures global spatial dependencies using partial-channel attention to enhance structural semantic reasoning. Position-wise normalization within the purification stage enforces explicit reconstruction-driven cross-modal alignment, ensuring statistical compatibility between heterogeneous features prior to fusion. Extensive experiments on the TLRGBD benchmark, where 94.5% of instances are small objects, demonstrate that CMAFNet achieves 32.2% mAP@50 and 12.5% APs, outperforming the strongest baseline by 9.8 and 4.0 percentage points, respectively. A lightweight variant reaches 24.8% mAP50 at 228 FPS with only 4.9M parameters, surpassing all YOLO-based detectors while matching transformer-based methods at substantially lower computational cost.
[366] 3D-Aware Implicit Motion Control for View-Adaptive Human Video Generation
Zhixue Fang, Xu He, Songlin Tang, Haoxian Zhang, Qingfeng Li, Xiaoqiang Liu, Pengfei Wan, Kun Gai
Main category: cs.CV
TL;DR: 3DiMo: A 3D-aware motion control method for video generation that uses implicit, view-agnostic motion tokens instead of explicit 3D models, enabling flexible camera control while maintaining motion fidelity.
Details
Motivation: Existing motion control methods either use 2D poses (which bind motion to specific viewpoints) or explicit 3D models (which have inherent inaccuracies that override the generator's intrinsic 3D awareness). The authors propose an implicit 3D-aware approach that aligns better with video generators' spatial priors.Method: Jointly trains a motion encoder with a pretrained video generator to distill driving frames into compact, view-agnostic motion tokens injected via cross-attention. Uses view-rich supervision (single-view, multi-view, moving-camera videos) for 3D awareness, and auxiliary geometric supervision with SMPL that’s annealed to zero over training.
Result: 3DiMo faithfully reproduces driving motions with flexible, text-driven camera control, significantly surpassing existing methods in both motion fidelity and visual quality.
Conclusion: Implicit, view-agnostic motion representations that align with video generators’ spatial priors are more effective than explicit 3D constraints for motion control, enabling both motion fidelity and novel-view synthesis.
Abstract: Existing methods for human motion control in video generation typically rely on either 2D poses or explicit 3D parametric models (e.g., SMPL) as control signals. However, 2D poses rigidly bind motion to the driving viewpoint, precluding novel-view synthesis. Explicit 3D models, though structurally informative, suffer from inherent inaccuracies (e.g., depth ambiguity and inaccurate dynamics) which, when used as a strong constraint, override the powerful intrinsic 3D awareness of large-scale video generators. In this work, we revisit motion control from a 3D-aware perspective, advocating for an implicit, view-agnostic motion representation that naturally aligns with the generator’s spatial priors rather than depending on externally reconstructed constraints. We introduce 3DiMo, which jointly trains a motion encoder with a pretrained video generator to distill driving frames into compact, view-agnostic motion tokens, injected semantically via cross-attention. To foster 3D awareness, we train with view-rich supervision (i.e., single-view, multi-view, and moving-camera videos), forcing motion consistency across diverse viewpoints. Additionally, we use auxiliary geometric supervision that leverages SMPL only for early initialization and is annealed to zero, enabling the model to transition from external 3D guidance to learning genuine 3D spatial motion understanding from the data and the generator’s priors. Experiments confirm that 3DiMo faithfully reproduces driving motions with flexible, text-driven camera control, significantly surpassing existing methods in both motion fidelity and visual quality.
[367] Unsupervised MR-US Multimodal Image Registration with Multilevel Correlation Pyramidal Optimization
Jiazheng Wang, Zeyu Liu, Min Liu, Xiang Chen, Xinyao Yu, Yaonan Wang, Hang Zhang
Main category: cs.CV
TL;DR: Unsupervised multimodal medical image registration method using Multilevel Correlation Pyramidal Optimization for surgical navigation, achieving state-of-the-art performance on Learn2Reg 2025 challenges.
Details
Motivation: Surgical navigation requires accurate registration of preoperative and intraoperative multimodal images, but faces challenges due to modality differences and tissue deformation during surgery.Method: Uses modality independent neighborhood descriptors to extract features, maps images to feature space, then employs multilevel pyramidal fusion optimization with dense correlation analysis and weight-balanced coupled convex optimization at different scales.
Result: Achieved first place in both validation and test phases of ReMIND2Reg task in Learn2Reg 2025, and achieved average TRE of 1.798 mm on Resect dataset.
Conclusion: The MCPO method demonstrates strong performance in multimodal medical image registration with broad applicability for preoperative-to-intraoperative surgical guidance.
Abstract: Surgical navigation based on multimodal image registration has played a significant role in providing intraoperative guidance to surgeons by showing the relative position of the target area to critical anatomical structures during surgery. However, due to the differences between multimodal images and intraoperative image deformation caused by tissue displacement and removal during the surgery, effective registration of preoperative and intraoperative multimodal images faces significant challenges. To address the multimodal image registration challenges in Learn2Reg 2025, an unsupervised multimodal medical image registration method based on Multilevel Correlation Pyramidal Optimization (MCPO) is designed to solve these problems. First, the features of each modality are extracted based on the modality independent neighborhood descriptor, and the multimodal images is mapped to the feature space. Second, a multilevel pyramidal fusion optimization mechanism is designed to achieve global optimization and local detail complementation of the displacement field through dense correlation analysis and weight-balanced coupled convex optimization for input features at different scales. Our method focuses on the ReMIND2Reg task in Learn2Reg 2025. Based on the results, our method achieved the first place in the validation phase and test phase of ReMIND2Reg. The MCPO is also validated on the Resect dataset, achieving an average TRE of 1.798 mm. This demonstrates the broad applicability of our method in preoperative-to-intraoperative image registration. The code is available at https://github.com/wjiazheng/MCPO.
[368] ShapBPT: Image Feature Attributions Using Data-Aware Binary Partition Trees
Muhammad Rashid, Elvio G. Amparore, Enrico Ferrari, Damiano Verda
Main category: cs.CV
TL;DR: ShapBPT introduces a data-aware hierarchical Shapley method for computer vision interpretability that uses Binary Partition Trees to align feature attributions with image morphology.
Details
Motivation: Existing hierarchical Shapley approaches for computer vision don't exploit image multiscale structure, leading to slow convergence and poor alignment with actual morphological features. There's a gap in using data-aware hierarchies for visual interpretability.Method: ShapBPT assigns Shapley coefficients to a multiscale hierarchical structure called Binary Partition Tree (BPT), which is tailored for images. This data-aware hierarchical partitioning ensures feature attributions align with intrinsic image morphology.
Result: Experimental results show superior alignment with image structures, improved efficiency over existing XCV methods, and a 20-subject user study confirms human preference for ShapBPT explanations.
Conclusion: ShapBPT connects hierarchical Shapley methods with image data, providing more efficient and semantically meaningful visual interpretability for computer vision models.
Abstract: Pixel-level feature attributions are an important tool in eXplainable AI for Computer Vision (XCV), providing visual insights into how image features influence model predictions. The Owen formula for hierarchical Shapley values has been widely used to interpret machine learning (ML) models and their learned representations. However, existing hierarchical Shapley approaches do not exploit the multiscale structure of image data, leading to slow convergence and weak alignment with the actual morphological features. Moreover, no prior Shapley method has leveraged data-aware hierarchies for Computer Vision tasks, leaving a gap in model interpretability of structured visual data. To address this, this paper introduces ShapBPT, a novel data-aware XCV method based on the hierarchical Shapley formula. ShapBPT assigns Shapley coefficients to a multiscale hierarchical structure tailored for images, the Binary Partition Tree (BPT). By using this data-aware hierarchical partitioning, ShapBPT ensures that feature attributions align with intrinsic image morphology, effectively prioritizing relevant regions while reducing computational overhead. This advancement connects hierarchical Shapley methods with image data, providing a more efficient and semantically meaningful approach to visual interpretability. Experimental results confirm ShapBPT’s effectiveness, demonstrating superior alignment with image structures and improved efficiency over existing XCV methods, and a 20-subject user study confirming that ShapBPT explanations are preferred by humans.
[369] OneVision-Encoder: Codec-Aligned Sparsity as a Foundational Principle for Multimodal Intelligence
Feilong Tang, Xiang An, Yunyao Yan, Yin Xie, Bin Qin, Kaicheng Yang, Yifei Shen, Yuanhan Zhang, Chunyuan Li, Shikun Feng, Changrui Chen, Huajie Tan, Ming Hu, Manyuan Zhang, Bo Li, Ziyong Feng, Ziwei Liu, Zongyuan Ge, Jiankang Deng
Main category: cs.CV
TL;DR: OneVision-Encoder is a video compression-based vision encoder that focuses computational resources on sparse, information-rich regions using codec principles, achieving superior performance with fewer tokens and data.
Details
Motivation: Modern vision architectures waste computation on redundant visual data while ignoring sparse discriminative information. The paper argues that visual understanding requires aligning architectures with information-theoretic principles of video codecs, focusing on predictive residuals rather than uniform pixel processing.Method: OneVision-Encoder uses Codec Patchification to focus computation on only 3.1%-25% of regions with high signal entropy. It employs shared 3D RoPE for unified spatial-temporal reasoning with irregular token layouts, trained with large-scale cluster discrimination over 1M+ semantic concepts to capture object permanence and motion dynamics.
Result: Outperforms strong vision backbones like Qwen3-ViT and SigLIP2 across 16 image, video, and document understanding benchmarks despite using fewer visual tokens and pretraining data. Achieves 4.1% average improvement over Qwen3-ViT on video tasks.
Conclusion: Codec-aligned patch-level sparsity enables efficient and accurate visual understanding, showing that efficiency and accuracy are positively correlated when architectures align with data’s fundamental structure.
Abstract: Hypothesis. Artificial general intelligence is, at its core, a compression problem. Effective compression demands resonance: deep learning scales best when its architecture aligns with the fundamental structure of the data. These are the fundamental principles. Yet, modern vision architectures have strayed from these truths: visual signals are highly redundant, while discriminative information, the surprise, is sparse. Current models process dense pixel grids uniformly, wasting vast compute on static background rather than focusing on the predictive residuals that define motion and meaning. We argue that to solve visual understanding, we must align our architectures with the information-theoretic principles of video, i.e., Codecs. Method. OneVision-Encoder encodes video by compressing predictive visual structure into semantic meaning. By adopting Codec Patchification, OV-Encoder abandons uniform computation to focus exclusively on the 3.1%-25% of regions rich in signal entropy. To unify spatial and temporal reasoning under irregular token layouts, OneVision-Encoder employs a shared 3D RoPE and is trained with a large-scale cluster discrimination objective over more than one million semantic concepts, jointly capturing object permanence and motion dynamics. Evidence. The results validate our core hypothesis: efficiency and accuracy are not a trade-off; they are positively correlated. When integrated into LLM, it consistently outperforms strong vision backbones such as Qwen3-ViT and SigLIP2 across 16 image, video, and document understanding benchmarks, despite using substantially fewer visual tokens and pretraining data. Notably, on video understanding tasks, OV-Encoder achieves an average improvement of 4.1% over Qwen3-ViT. Codec-aligned, patch-level sparsity is a foundational principle, enabling OV-Encoder as a scalable engine for next-generation visual generalists.
[370] Stroke3D: Lifting 2D strokes into rigged 3D model via latent diffusion models
Ruisi Zhao, Haoren Zheng, Zongxin Yang, Hehe Fan, Yi Yang
Main category: cs.CV
TL;DR: Stroke3D generates rigged 3D meshes from 2D drawn strokes and text prompts using a two-stage pipeline: controllable skeleton generation via Sk-VAE/Sk-DiT, followed by enhanced mesh synthesis with TextuRig dataset and SKA-DPO optimization.
Details
Motivation: Existing 3D generation methods struggle with creating animatable geometry, while rigging techniques lack fine-grained structural control over skeleton creation. There's a need for intuitive tools that generate ready-to-animate 3D content from simple user inputs.Method: Two-stage pipeline: 1) Controllable skeleton generation using Skeletal Graph VAE (Sk-VAE) to encode skeleton graph structure, with Skeletal Graph DiT (Sk-DiT) generating skeletal embeddings conditioned on text and 2D strokes; 2) Enhanced mesh synthesis using TextuRig dataset (curated from Objaverse-XL) and SKA-DPO preference optimization guided by skeleton-mesh alignment scores.
Result: Stroke3D produces plausible skeletons and high-quality meshes, enabling intuitive creation of ready-to-animate 3D content from 2D strokes and text prompts.
Conclusion: Stroke3D introduces the first framework for generating rigged 3D meshes conditioned on user-drawn 2D strokes, addressing limitations in existing 3D generation and rigging methods through a novel two-stage approach with structural control.
Abstract: Rigged 3D assets are fundamental to 3D deformation and animation. However, existing 3D generation methods face challenges in generating animatable geometry, while rigging techniques lack fine-grained structural control over skeleton creation. To address these limitations, we introduce Stroke3D, a novel framework that directly generates rigged meshes from user inputs: 2D drawn strokes and a descriptive text prompt. Our approach pioneers a two-stage pipeline that separates the generation into: 1) Controllable Skeleton Generation, we employ the Skeletal Graph VAE (Sk-VAE) to encode the skeleton’s graph structure into a latent space, where the Skeletal Graph DiT (Sk-DiT) generates a skeletal embedding. The generation process is conditioned on both the text for semantics and the 2D strokes for explicit structural control, with the VAE’s decoder reconstructing the final high-quality 3D skeleton; and 2) Enhanced Mesh Synthesis via TextuRig and SKA-DPO, where we then synthesize a textured mesh conditioned on the generated skeleton. For this stage, we first enhance an existing skeleton-to-mesh model by augmenting its training data with TextuRig: a dataset of textured and rigged meshes with captions, curated from Objaverse-XL. Additionally, we employ a preference optimization strategy, SKA-DPO, guided by a skeleton-mesh alignment score, to further improve geometric fidelity. Together, our framework enables a more intuitive workflow for creating ready to animate 3D content. To the best of our knowledge, our work is the first to generate rigged 3D meshes conditioned on user-drawn 2D strokes. Extensive experiments demonstrate that Stroke3D produces plausible skeletons and high-quality meshes.
[371] C^2ROPE: Causal Continuous Rotary Positional Encoding for 3D Large Multimodal-Models Reasoning
Guanting Ye, Qiyan Zhao, Wenhao Yu, Xiaofeng Zhang, Jianmin Ji, Yanyong Zhang, Ka-Veng Yuen
Main category: cs.CV
TL;DR: C²RoPE improves Rotary Position Embedding for 3D multimodal models by modeling spatial continuity and causal relationships to address limitations in visual processing.
Details
Motivation: Standard RoPE in 3D LMMs has limitations: 1D temporal indices disrupt visual feature continuity, and temporal proximity assumptions cause long-term decay in attention to earlier visual tokens.Method: Proposes C²RoPE with spatio-temporal continuous positional embedding using triplet hybrid indices (temporal + spatial coordinates) and Chebyshev Causal Masking based on 2D spatial distances.
Result: Demonstrates effectiveness across 3D scene reasoning and 3D visual question answering benchmarks, showing improved visual processing capabilities.
Conclusion: C²RoPE successfully addresses RoPE limitations for multimodal processing by better modeling spatial continuity and causal relationships in visual tokens.
Abstract: Recent advances in 3D Large Multimodal Models (LMMs) built on Large Language Models (LLMs) have established the alignment of 3D visual features with LLM representations as the dominant paradigm. However, the inherited Rotary Position Embedding (RoPE) introduces limitations for multimodal processing. Specifically, applying 1D temporal positional indices disrupts the continuity of visual features along the column dimension, resulting in spatial locality loss. Moreover, RoPE follows the prior that temporally closer image tokens are more causally related, leading to long-term decay in attention allocation and causing the model to progressively neglect earlier visual tokens as the sequence length increases. To address these issues, we propose C^2RoPE, an improved RoPE that explicitly models local spatial Continuity and spatial Causal relationships for visual processing. C^2RoPE introduces a spatio-temporal continuous positional embedding mechanism for visual tokens. It first integrates 1D temporal positions with Cartesian-based spatial coordinates to construct a triplet hybrid positional index, and then employs a frequency allocation strategy to encode spatio-temporal positional information across the three index components. Additionally, we introduce Chebyshev Causal Masking, which determines causal dependencies by computing the Chebyshev distance of image tokens in 2D space. Evaluation results across various benchmarks, including 3D scene reasoning and 3D visual question answering, demonstrate C^2RoPE’s effectiveness. The code is be available at https://github.com/ErikZ719/C2RoPE.
[372] AMAP-APP: Efficient Segmentation and Morphometry Quantification of Fluorescent Microscopy Images of Podocytes
Arash Fatehi, David Unnersjö-Jess, Linus Butt, Noémie Moreau, Thomas Benzing, Katarzyna Bozek
Main category: cs.CV
TL;DR: AMAP-APP is a cross-platform desktop application that optimizes podocyte foot process quantification by replacing intensive instance segmentation with classic image processing while maintaining accuracy, achieving 147x speedup and high correlation with original method.
Details
Motivation: The original AMAP method for automated podocyte foot process quantification has limitations including high computational demands, lack of user interface, and Linux dependency, which hinder its widespread adoption in kidney research and potential clinical diagnostics.Method: AMAP-APP optimizes efficiency by replacing intensive instance segmentation with classic image processing while retaining the original semantic segmentation model. It introduces a refined Region of Interest (ROI) algorithm to improve precision. Validation involved 365 mouse and human images (STED and confocal), benchmarking performance against the original AMAP via Pearson correlation and Two One-Sided T-tests (TOST).
Result: AMAP-APP achieved a 147-fold increase in processing speed on consumer hardware. Morphometric outputs (area, perimeter, circularity, and slit diaphragm density) showed high correlation (r>0.90) and statistical equivalence (TOST P<0.05) to the original method. The new ROI algorithm demonstrated superior accuracy with reduced deviation from manual delineations.
Conclusion: AMAP-APP democratizes deep learning-based podocyte morphometry by eliminating the need for high-performance computing clusters and providing a user-friendly interface for Windows, macOS, and Linux, enabling widespread adoption in nephrology research and potential clinical diagnostics.
Abstract: Background: Automated podocyte foot process quantification is vital for kidney research, but the established “Automatic Morphological Analysis of Podocytes” (AMAP) method is hindered by high computational demands, a lack of a user interface, and Linux dependency. We developed AMAP-APP, a cross-platform desktop application designed to overcome these barriers. Methods: AMAP-APP optimizes efficiency by replacing intensive instance segmentation with classic image processing while retaining the original semantic segmentation model. It introduces a refined Region of Interest (ROI) algorithm to improve precision. Validation involved 365 mouse and human images (STED and confocal), benchmarking performance against the original AMAP via Pearson correlation and Two One-Sided T-tests (TOST). Results: AMAP-APP achieved a 147-fold increase in processing speed on consumer hardware. Morphometric outputs (area, perimeter, circularity, and slit diaphragm density) showed high correlation (r>0.90) and statistical equivalence (TOST P<0.05) to the original method. Additionally, the new ROI algorithm demonstrated superior accuracy compared to the original, showing reduced deviation from manual delineations. Conclusion: AMAP-APP democratizes deep learning-based podocyte morphometry. By eliminating the need for high-performance computing clusters and providing a user-friendly interface for Windows, macOS, and Linux, it enables widespread adoption in nephrology research and potential clinical diagnostics.
[373] Exploring Real-Time Super-Resolution: Benchmarking and Fine-Tuning for Streaming Content
Evgeney Bogatyrev, Khaled Abud, Ivan Molodetskikh, Nikita Alutis, Dmitriy Vatolin
Main category: cs.CV
TL;DR: StreamSR dataset for video streaming super-resolution, EfRLFN model with efficient architecture, and benchmark showing fine-tuning on streaming data improves performance across standard benchmarks.
Details
Motivation: Existing real-time super-resolution methods struggle with compressed video content from streaming services, and current datasets don't reflect real-world streaming characteristics, limiting benchmark relevance.Method: 1) Created StreamSR dataset from YouTube covering diverse video genres/resolutions; 2) Proposed EfRLFN model with Efficient Channel Attention and hyperbolic tangent activation; 3) Designed composite loss function; 4) Benchmarked 11 SOTA models; 5) Fine-tuned models on StreamSR dataset.
Result: EfRLFN achieves improved visual quality and runtime performance. Fine-tuning other models on StreamSR dataset yields significant performance gains that generalize well across various standard benchmarks.
Conclusion: StreamSR dataset addresses the gap in streaming-focused super-resolution evaluation, EfRLFN provides efficient real-time performance, and dataset-based fine-tuning improves model generalization across benchmarks.
Abstract: Recent advancements in real-time super-resolution have enabled higher-quality video streaming, yet existing methods struggle with the unique challenges of compressed video content. Commonly used datasets do not accurately reflect the characteristics of streaming media, limiting the relevance of current benchmarks. To address this gap, we introduce a comprehensive dataset - StreamSR - sourced from YouTube, covering a wide range of video genres and resolutions representative of real-world streaming scenarios. We benchmark 11 state-of-the-art real-time super-resolution models to evaluate their performance for the streaming use-case. Furthermore, we propose EfRLFN, an efficient real-time model that integrates Efficient Channel Attention and a hyperbolic tangent activation function - a novel design choice in the context of real-time super-resolution. We extensively optimized the architecture to maximize efficiency and designed a composite loss function that improves training convergence. EfRLFN combines the strengths of existing architectures while improving both visual quality and runtime performance. Finally, we show that fine-tuning other models on our dataset results in significant performance gains that generalize well across various standard benchmarks. We made the dataset, the code, and the benchmark available at https://github.com/EvgeneyBogatyrev/EfRLFN.
[374] TexSpot: 3D Texture Enhancement with Spatially-uniform Point Latent Representation
Ziteng Lu, Yushuang Wu, Chongjie Ye, Yuda Qiu, Jing Shao, Xiaoyang Guo, Jiaqing Zhou, Tianlei Hu, Kun Zhou, Xiaoguang Han
Main category: cs.CV
TL;DR: TexSpot introduces a novel 3D texture representation called Texlet that combines point-based geometric expressiveness with UV-based compactness, enabling high-quality texture enhancement via diffusion transformers.
Details
Motivation: Current 3D texture generation methods suffer from view-inconsistency in multi-view diffusion pipelines, with UV maps causing distortion during unwrapping and point-based methods limiting high-resolution generation due to texture fidelity being tied to geometric density.Method: Introduces Texlet representation where each latent vector encodes a local texture patch via 2D encoder, aggregated with 3D encoder for global shape context. Uses cascaded 3D-to-2D decoder for texture patch reconstruction, then trains diffusion transformer conditioned on Texlets to refine textures from multi-view diffusion methods.
Result: Extensive experiments show TexSpot significantly improves visual fidelity, geometric consistency, and robustness over state-of-the-art 3D texture generation and enhancement approaches.
Conclusion: TexSpot addresses fundamental limitations in 3D texture generation by introducing a novel representation that enables high-quality texture enhancement through diffusion-based methods.
Abstract: High-quality 3D texture generation remains a fundamental challenge due to the view-inconsistency inherent in current mainstream multi-view diffusion pipelines. Existing representations either rely on UV maps, which suffer from distortion during unwrapping, or point-based methods, which tightly couple texture fidelity to geometric density that limits high-resolution texture generation. To address these limitations, we introduce TexSpot, a diffusion-based texture enhancement framework. At its core is Texlet, a novel 3D texture representation that merges the geometric expressiveness of point-based 3D textures with the compactness of UV-based representation. Each Texlet latent vector encodes a local texture patch via a 2D encoder and is further aggregated using a 3D encoder to incorporate global shape context. A cascaded 3D-to-2D decoder reconstructs high-quality texture patches, enabling the Texlet space learning. Leveraging this representation, we train a diffusion transformer conditioned on Texlets to refine and enhance textures produced by multi-view diffusion methods. Extensive experiments demonstrate that TexSpot significantly improves visual fidelity, geometric consistency, and robustness over existing state-of-the-art 3D texture generation and enhancement approaches. Project page: https://texlet-arch.github.io/TexSpot-page.
[375] Reliable Thinking with Images
Haobin Li, Yutong Yang, Yijie Lin, Xiang Dai, Mouxing Yang, Xi Peng
Main category: cs.CV
TL;DR: RTWI addresses noisy thinking in multimodal reasoning by estimating reliability of visual cues and textual chains-of-thought, using filtering and voting to prevent error accumulation.
Details
Motivation: Current Thinking with Images (TWI) methods assume perfect interleaved image-text reasoning chains, but real-world scenarios often contain noisy/imperfect visual cue mining and reasoning, leading to error accumulation that degrades MLLM performance.Method: Proposes Reliable Thinking with Images (RTWI) that estimates reliability of both visual cues and textual CoT in a unified text-centric manner, then employs robust filtering and voting modules to prevent noisy thinking from contaminating final answers.
Result: Extensive experiments on seven benchmarks verify RTWI’s effectiveness against noisy thinking, showing improved performance over existing TWI methods.
Conclusion: Addressing noisy thinking is crucial for robust multimodal reasoning, and RTWI provides an effective solution by reliability estimation and error prevention mechanisms.
Abstract: As a multimodal extension of Chain-of-Thought (CoT), Thinking with Images (TWI) has recently emerged as a promising avenue to enhance the reasoning capability of Multi-modal Large Language Models (MLLMs), which generates interleaved CoT by incorporating visual cues into the textual reasoning process. However, the success of existing TWI methods heavily relies on the assumption that interleaved image-text CoTs are faultless, which is easily violated in real-world scenarios due to the complexity of multimodal understanding. In this paper, we reveal and study a highly-practical yet under-explored problem in TWI, termed Noisy Thinking (NT). Specifically, NT refers to the imperfect visual cues mining and answer reasoning process. As the saying goes, ``One mistake leads to another’’, erroneous interleaved CoT would cause error accumulation, thus significantly degrading the performance of MLLMs. To solve the NT problem, we propose a novel method dubbed Reliable Thinking with Images (RTWI). In brief, RTWI estimates the reliability of visual cues and textual CoT in a unified text-centric manner and accordingly employs robust filtering and voting modules to prevent NT from contaminating the final answer. Extensive experiments on seven benchmarks verify the effectiveness of RTWI against NT.
cs.AI
[376] Agentic AI for Commercial Insurance Underwriting with Adversarial Self-Critique
Joyjit Roy, Samaresh Kumar Singh
Main category: cs.AI
TL;DR: AI system with adversarial self-critique for commercial insurance underwriting reduces hallucinations and improves accuracy while maintaining human oversight.
Details
Motivation: Commercial insurance underwriting is labor-intensive but existing AI solutions lack comprehensive reasoning and safety mechanisms for regulated, high-stakes environments where full automation is impractical.Method: Decision-negative, human-in-the-loop agentic system with adversarial self-critique mechanism where a critic agent challenges primary agent’s conclusions before human review, plus formal taxonomy of failure modes.
Result: Adversarial critique reduces AI hallucination rates from 11.3% to 3.8% and increases decision accuracy from 92% to 96% on 500 expert-validated underwriting cases.
Conclusion: Adversarial self-critique supports safer AI deployment in regulated domains and offers a model for responsible integration where human oversight is indispensable.
Abstract: Commercial insurance underwriting is a labor-intensive process that requires manual review of extensive documentation to assess risk and determine policy pricing. While AI offers substantial efficiency improvements, existing solutions lack comprehensive reasoning capabilities and internal mechanisms to ensure reliability within regulated, high-stakes environments. Full automation remains impractical and inadvisable in scenarios where human judgment and accountability are critical. This study presents a decision-negative, human-in-the-loop agentic system that incorporates an adversarial self-critique mechanism as a bounded safety architecture for regulated underwriting workflows. Within this system, a critic agent challenges the primary agent’s conclusions prior to submitting recommendations to human reviewers. This internal system of checks and balances addresses a critical gap in AI safety for regulated workflows. Additionally, the research develops a formal taxonomy of failure modes to characterize potential errors by decision-negative agents. This taxonomy provides a structured framework for risk identification and risk management in high-stakes applications. Experimental evaluation using 500 expert-validated underwriting cases demonstrates that the adversarial critique mechanism reduces AI hallucination rates from 11.3% to 3.8% and increases decision accuracy from 92% to 96%. At the same time, the framework enforces strict human authority over all binding decisions by design. These findings indicate that adversarial self-critique supports safer AI deployment in regulated domains and offers a model for responsible integration where human oversight is indispensable.
[377] BotzoneBench: Scalable LLM Evaluation via Graded AI Anchors
Lingfeng Li, Yunlong Lu, Yuefei Zhang, Jingyu Yao, Yixin Zhu, KeYuan Cheng, Yongyi Wang, Qirui Zheng, Xionghui Yang, Wenxin Li
Main category: cs.AI
TL;DR: BotzoneBench: A scalable evaluation framework for LLM strategic reasoning using fixed hierarchies of skill-calibrated game AI as anchors, enabling linear-time absolute skill measurement across diverse games.
Details
Motivation: Existing LLM benchmarks fail to capture dynamic strategic abilities in interactive environments, and current game-based evaluations using LLM-vs-LLM tournaments have quadratic computational costs, lack stable performance anchors, and produce rankings dependent on transient model pools.Method: Anchors LLM evaluation to fixed hierarchies of skill-calibrated game AI on the Botzone platform, evaluating LLMs across eight diverse games (deterministic perfect-information board games to stochastic imperfect-information card games) with systematic assessment of 177,047 state-action pairs from five flagship models.
Result: Reveals significant performance disparities and identifies distinct strategic behaviors, with top-performing models achieving proficiency comparable to mid-to-high-tier specialized game AI in multiple domains.
Conclusion: The anchored evaluation paradigm generalizes beyond games to any domain with well-defined skill hierarchies, establishing a scalable and reusable framework for assessing interactive AI capabilities.
Abstract: Large Language Models (LLMs) are increasingly deployed in interactive environments requiring strategic decision-making, yet systematic evaluation of these capabilities remains challenging. Existing benchmarks for LLMs primarily assess static reasoning through isolated tasks and fail to capture dynamic strategic abilities. Recent game-based evaluations employ LLM-vs-LLM tournaments that produce relative rankings dependent on transient model pools, incurring quadratic computational costs and lacking stable performance anchors for longitudinal tracking. The central challenge is establishing a scalable evaluation framework that measures LLM strategic reasoning against consistent, interpretable standards rather than volatile peer models. Here we show that anchoring LLM evaluation to fixed hierarchies of skill-calibrated game Artificial Intelligence (AI) enables linear-time absolute skill measurement with stable cross-temporal interpretability. Built on the Botzone platform’s established competitive infrastructure, our BotzoneBench evaluates LLMs across eight diverse games spanning deterministic perfect-information board games to stochastic imperfect-information card games. Through systematic assessment of 177,047 state-action pairs from five flagship models, we reveal significant performance disparities and identify distinct strategic behaviors, with top-performing models achieving proficiency comparable to mid-to-high-tier specialized game AI in multiple domains. This anchored evaluation paradigm generalizes beyond games to any domain with well-defined skill hierarchies, establishing a scalable and reusable framework for assessing interactive AI capabilities.
[378] When to Think Fast and Slow? AMOR: Entropy-Based Metacognitive Gate for Dynamic SSM-Attention Switching
Haoran Zheng
Main category: cs.AI
TL;DR: AMOR is a hybrid SSM-attention architecture that dynamically engages sparse attention only when the SSM backbone is uncertain, measured by prediction entropy, achieving efficient long-context processing with adaptive computation.
Details
Motivation: Transformers use uniform computation for all positions regardless of difficulty, while SSMs struggle with long-range information retrieval. The paper aims to create an efficient hybrid architecture inspired by dual-process cognition theories.Method: AMOR combines SSM backbone with sparse attention triggered only when SSM prediction entropy indicates uncertainty. Uses Ghost KV projection from SSM hidden states to reuse O(n) computation rather than O(n²) attention at every layer.
Result: On synthetic retrieval tasks, AMOR outperforms SSM-only and transformer-only baselines, achieving perfect retrieval accuracy while engaging attention on only 22% of positions. Prediction entropy shows 1.09 nats gap between retrieval and local positions.
Conclusion: AMOR provides efficient, interpretable adaptive computation for long-context processing, with routing decisions understandable in information-theoretic terms, bridging SSM efficiency with transformer precision.
Abstract: Transformers allocate uniform computation to every position, regardless of difficulty. State Space Models (SSMs) offer efficient alternatives but struggle with precise information retrieval over a long horizon. Inspired by dual-process theories of cognition (Kahneman, 2011), we propose AMOR (Adaptive Metacognitive Output Router), a hybrid architecture that dynamically engages sparse attention only when an SSM backbone is “uncertain”–as measured by prediction entropy. Compared to standard transformers, AMOR gains efficiency by projecting keys and values from SSM hidden states (Ghost KV), reusing the SSM’s O(n) computation rather than requiring O(n^2) attention at every layer. On small-scale synthetic retrieval tasks, AMOR outperforms both SSM-only and transformer-only baselines, achieving perfect retrieval accuracy while engaging attention on only 22% of positions. We validate that prediction entropy reliably signals retrieval need, with a gap of 1.09 nats (nearly half the entropy range) between retrieval and local positions. Additionally, our approach provides interpretable adaptive computation, where routing decisions can be understood in information-theoretic terms.
[379] VeRA: Verified Reasoning Data Augmentation at Scale
Zerui Cheng, Jiashuo Liu, Chunjie Wu, Jianzhu Yao, Pramod Viswanath, Ge Zhang, Wenhao Huang
Main category: cs.AI
TL;DR: VeRA is a framework for generating unlimited verified benchmark variants from seed problems to combat memorization and enable robust AI evaluation.
Details
Motivation: Current AI evaluation suffers from static benchmarks that allow memorization and format exploitation, preventing genuine measurement of reasoning progress. There's a need for robust evaluation by construction rather than post-hoc detection.Method: VeRA converts benchmark problems into executable specifications with three components: (1) natural language template with placeholders, (2) coherent generator that samples valid configurations, and (3) deterministic verifier that validates parameters and calculates correct answers. It operates in two modes: VeRA-E (equivalent variants) and VeRA-H (hardened variants with increased complexity).
Result: Evaluation of 16 frontier models shows: VeRA-E improves evaluation quality and reveals contamination patterns; VeRA-H enables human-free generation of hard tasks with reliable labels; VeRA establishes verified benchmarks as a general paradigm for generating fresh instances on demand.
Conclusion: VeRA reconceptualizes benchmarks from static objects to executable specifications that can generate unlimited verified instances, enhancing robustness and cost-effectiveness for AI evaluation across any verifiable domain.
Abstract: The main issue with most evaluation schemes today is their “static” nature: the same problems are reused repeatedly, allowing for memorization, format exploitation, and eventual saturation. To measure genuine AI progress, we need evaluation that is robust by construction, not by post-hoc detection. In response, we propose VeRA (Verified Reasoning Data Augmentation), a framework that converts benchmark problems into executable specifications, comprising (i) a natural language template with placeholder slots, (ii) a coherent generator that samples valid configurations, and (iii) a deterministic verifier that validates parameters and calculates the corresponding correct answers for each configuration. From a single seed problem, VeRA automatically creates unlimited verified variants with reliable labels at near-zero marginal cost without human involvement. VeRA operates in two complementary modes. VeRA-E (equivalent) rewrites problems while keeping the underlying logic intact, useful for detecting memorization versus genuine reasoning. VeRA-H (hardened) systematically increases complexity while remaining verifiable, enabling reliable creation and labelling of fresh difficult tasks at the boundary of intelligence. Evaluating 16 frontier models with VeRA, we find: (i) VeRA-E improves evaluation quality and reveals contamination patterns. (ii) VeRA-H enables human-free generation of hard tasks with reliable labels. (iii) VeRA establishes verified benchmarks as a general paradigm. VeRA reconceptualizes benchmarks from static objects used until exhausted, to executable specifications generating fresh, verified instances on demand, enhancing robustness and cost-effectiveness for evaluation. With VeRA, we envision that evaluation in any verifiable domain can scale indefinitely without sacrificing label integrity. To stimulate future research, we have open-sourced all code and datasets.
[380] Scaling the Scaling Logic: Agentic Meta-Synthesis of Logic Reasoning
Bowen Liu, Zhi Wu, Runquan Xie, Zhanhui Kang, Jia Li
Main category: cs.AI
TL;DR: SSLogic is an agentic meta-synthesis framework that scales verifiable training signals for RLVR through iterative synthesis and repair of Generator-Validator program pairs, enabling continuous evolution of task families with controllable difficulty.
Details
Motivation: Scaling verifiable training signals is a key bottleneck for Reinforcement Learning from Verifiable Rewards (RLVR). Logical reasoning provides a natural substrate since constraints are formal and answers are programmatically checkable, but prior synthesis pipelines are limited to instance-level perturbations rather than task-family evolution.Method: SSLogic uses an agentic meta-synthesis framework with a closed Generate-Validate-Repair loop. It iteratively synthesizes and repairs executable Generator-Validator program pairs. A Multi-Gate Validation Protocol combines multi-strategy consistency checks with Adversarial Blind Review, where independent agents must solve instances by writing and executing code to filter ambiguous tasks.
Result: Starting from 400 seed families, two evolution rounds expanded to 953 families and 21,389 verifiable instances (from 5,718). Training on SSLogic-evolved data yields consistent gains: SynLogic +5.2, BBEH +1.4, AIME25 +3.0, and Brumo25 +3.7 over seed baseline at matched training steps.
Conclusion: SSLogic successfully scales verifiable training signals at the task-family level, enabling continuous evolution with controllable difficulty and producing substantial performance improvements across multiple benchmarks.
Abstract: Scaling verifiable training signals remains a key bottleneck for Reinforcement Learning from Verifiable Rewards (RLVR). Logical reasoning is a natural substrate: constraints are formal and answers are programmatically checkable. However, prior synthesis pipelines either depend on expert-written code or operate within fixed templates/skeletons, which limits growth largely to instance-level perturbations. We propose SSLogic, an agentic meta-synthesis framework that scales at the task-family level by iteratively synthesizing and repairing executable Generator–Validator program pairs in a closed Generate–Validate–Repair loop, enabling continuous family evolution with controllable difficulty. To ensure reliability, we introduce a Multi-Gate Validation Protocol that combines multi-strategy consistency checks with Adversarial Blind Review, where independent agents must solve instances by writing and executing code to filter ambiguous or ill-posed tasks. Starting from 400 seed families, two evolution rounds expand to 953 families and 21,389 verifiable instances (from 5,718). Training on SSLogic-evolved data yields consistent gains over the seed baseline at matched training steps, improving SynLogic by +5.2, BBEH by +1.4, AIME25 by +3.0, and Brumo25 by +3.7.
[381] A Geometric Taxonomy of Hallucinations in LLMs
Javier Marín
Main category: cs.AI
TL;DR: The paper proposes a taxonomy of LLM hallucinations with three types: unfaithfulness, confabulation, and factual error, each with distinct geometric signatures in embedding space.
Details
Motivation: Current understanding of "hallucination" in LLMs conflates distinct phenomena, making it difficult to develop effective detection methods. The authors aim to clarify these different types of hallucinations and their geometric properties in embedding space.Method: The authors propose a taxonomy identifying three hallucination types, then analyze their geometric signatures in embedding space using benchmarks and human-crafted confabulations. They measure detection performance via AUROC and examine domain transferability and geometric relationships between discriminative directions.
Result: Type I (unfaithfulness) and Type II (confabulation) hallucinations show strong detection performance within domains (AUROC 0.76-0.99) but poor cross-domain transfer (AUROC ~0.50). Human-crafted confabulations show a single global detection direction with 0.96 AUROC and minimal cross-domain degradation. Type III (factual errors) are indistinguishable from chance (AUROC 0.478).
Conclusion: Embedding-based detection works for Types I and II hallucinations but fails for Type III factual errors, which require external verification mechanisms since embeddings encode distributional patterns rather than correspondence to external reality.
Abstract: The term “hallucination” in large language models conflates distinct phenomena with different geometric signatures in embedding space. We propose a taxonomy identifying three types: unfaithfulness (failure to engage with provided context), confabulation (invention of semantically foreign content), and factual error (incorrect claims within correct conceptual frames). We observe a striking asymmetry. On standard benchmarks where hallucinations are LLM-generated, detection is domain-local: AUROC 0.76-0.99 within domains, but 0.50 (chance level) across domains. Discriminative directions are approximately orthogonal between domains (mean cosine similarity -0.07). On human-crafted confabulations - invented institutions, redefined terminology, fabricated mechanisms - a single global direction achieves 0.96 AUROC with 3.8% cross-domain degradation. We interpret this divergence as follows: benchmarks capture generation artifacts (stylistic signatures of prompted fabrication), while human-crafted confabulations capture genuine topical drift. The geometric structure differs because the underlying phenomena differ. Type III errors show 0.478 AUROC - indistinguishable from chance. This reflects a theoretical constraint: embeddings encode distributional co-occurrence, not correspondence to external reality. Statements with identical contextual patterns occupy similar embedding regions regardless of truth value. The contribution is a geometric taxonomy clarifying the scope of embedding-based detection: Types I and II are detectable; Type III requires external verification mechanisms.
[382] Variation is the Key: A Variation-Based Framework for LLM-Generated Text Detection
Xuecong Li, Xiaohong Li, Qiang Hu, Yao Zhang, Junjie Wang
Main category: cs.AI
TL;DR: VaryBalance: A simple yet effective LLM-generated text detection method that leverages differences between human texts and their LLM-rewritten versions
Details
Motivation: Existing LLM-generated text detectors have impractical assumptions (white-box settings) or rely solely on text-level features, leading to imprecise detection. There's a need for more effective and practical detection methods.Method: VaryBalance detects LLM-generated text by observing that human texts have greater difference from their LLM-rewritten versions compared to LLM-generated texts. It quantifies this difference using mean standard deviation to distinguish between human and LLM-generated texts.
Result: VaryBalance outperforms state-of-the-art detectors (Binoculars) by up to 34.3% in AUROC, maintains robustness across multiple generating models and languages, and demonstrates practical effectiveness.
Conclusion: VaryBalance provides a simple, effective, and practical solution for LLM-generated text detection that addresses limitations of existing methods and shows superior performance across various conditions.
Abstract: Detecting text generated by large language models (LLMs) is crucial but challenging. Existing detectors depend on impractical assumptions, such as white-box settings, or solely rely on text-level features, leading to imprecise detection ability. In this paper, we propose a simple but effective and practical LLM-generated text detection method, VaryBalance. The core of VaryBalance is that, compared to LLM-generated texts, there is a greater difference between human texts and their rewritten version via LLMs. Leveraging this observation, VaryBalance quantifies this through mean standard deviation and distinguishes human texts and LLM-generated texts. Comprehensive experiments demonstrated that VaryBalance outperforms the state-of-the-art detectors, i.e., Binoculars, by up to 34.3% in terms of AUROC, and maintains robustness against multiple generating models and languages.
[383] Intelligence as Trajectory-Dominant Pareto Optimization
Truong Xuan Khanh, Truong Quynh Hoa
Main category: cs.AI
TL;DR: The paper introduces Trajectory-Dominant Pareto Optimization, a framework that views intelligence as trajectory-level phenomenon with Pareto traps limiting long-horizon adaptability, independent of learning progress or model scale.
Details
Motivation: The paper addresses why AI systems stagnate in long-horizon adaptability despite optimization, arguing this stems not from insufficient learning/data/capacity but from structural properties of how intelligence is optimized over time.Method: Introduces Trajectory-Dominant Pareto Optimization (path-wise generalization of Pareto optimality), defines Trap Escape Difficulty Index (TEDI) to measure constraint rigidity, develops taxonomy of Pareto traps, and illustrates with minimal agent-environment model.
Result: Shows dynamic intelligence ceilings arise as geometric consequences of trajectory-level dominance, independent of learning progress or architectural scale, and provides framework for diagnosing developmental constraints.
Conclusion: Shifts focus from terminal performance to optimization geometry, offering principled framework to overcome long-horizon developmental constraints in adaptive systems.
Abstract: Despite recent advances in artificial intelligence, many systems exhibit stagnation in long-horizon adaptability despite continued performance optimization. This work argues that such limitations do not primarily arise from insufficient learning, data, or model capacity, but from a deeper structural property of how intelligence is optimized over time. We formulate intelligence as a trajectory-level phenomenon governed by multi-objective trade-offs, and introduce Trajectory-Dominant Pareto Optimization, a path-wise generalization of classical Pareto optimality in which dominance is defined over full trajectories. Within this framework, Pareto traps emerge as locally non-dominated regions of trajectory space that nevertheless restrict access to globally superior developmental paths under conservative local optimization. To characterize the rigidity of such constraints, we define the Trap Escape Difficulty Index (TEDI), a composite geometric measure capturing escape distance, structural constraints, and behavioral inertia. We show that dynamic intelligence ceilings arise as inevitable geometric consequences of trajectory-level dominance, independent of learning progress or architectural scale. We further introduce a formal taxonomy of Pareto traps and illustrate the resulting trajectory-level divergence using a minimal agent-environment model. Together, these results shift the locus of intelligence from terminal performance to optimization geometry, providing a principled framework for diagnosing and overcoming long-horizon developmental constraints in adaptive systems.
[384] PlotChain: Deterministic Checkpointed Evaluation of Multimodal LLMs on Engineering Plot Reading
Mayank Ravishankara
Main category: cs.AI
TL;DR: PlotChain is a deterministic benchmark for evaluating MLLMs on reading quantitative values from engineering plots, with diagnostic checkpoints for failure analysis.
Details
Motivation: Current MLLM evaluation lacks systematic benchmarks for quantitative plot reading in engineering domains, focusing instead on OCR extraction or free-form captioning. There's a need for deterministic evaluation with exact ground truth and diagnostic capabilities to understand model failures.Method: Created a generator-based benchmark with 15 plot families and 450 rendered plots, each produced from known parameters with exact ground truth. Introduced checkpoint-based diagnostic evaluation with intermediate ‘cp_’ fields to isolate sub-skills. Evaluated four SOTA MLLMs using standardized protocol (temperature=0, JSON-only numeric output) with per-field tolerance scoring.
Result: Top models achieved 80.42% (Gemini 2.5 Pro), 79.84% (GPT-4.1), and 78.21% (Claude Sonnet 4.5) overall field-level pass rates. GPT-4o trailed at 61.59%. Frequency-domain tasks remained challenging with bandpass response ≤23% and FFT spectrum difficulties.
Conclusion: PlotChain provides a reproducible benchmark for quantitative plot reading in MLLMs, revealing strengths in many plot families but persistent weaknesses in frequency-domain analysis. The diagnostic framework enables detailed failure analysis.
Abstract: We present PlotChain, a deterministic, generator-based benchmark for evaluating multimodal large language models (MLLMs) on engineering plot reading-recovering quantitative values from classic plots (e.g., Bode/FFT, step response, stress-strain, pump curves) rather than OCR-only extraction or free-form captioning. PlotChain contains 15 plot families with 450 rendered plots (30 per family), where every item is produced from known parameters and paired with exact ground truth computed directly from the generating process. A central contribution is checkpoint-based diagnostic evaluation: in addition to final targets, each item includes intermediate ‘cp_’ fields that isolate sub-skills (e.g., reading cutoff frequency or peak magnitude) and enable failure localization within a plot family. We evaluate four state-of-the-art MLLMs under a standardized, deterministic protocol (temperature = 0 and a strict JSON-only numeric output schema) and score predictions using per-field tolerances designed to reflect human plot-reading precision. Under the ‘plotread’ tolerance policy, the top models achieve 80.42% (Gemini 2.5 Pro), 79.84% (GPT-4.1), and 78.21% (Claude Sonnet 4.5) overall field-level pass rates, while GPT-4o trails at 61.59%. Despite strong performance on many families, frequency-domain tasks remain brittle: bandpass response stays low (<= 23%), and FFT spectrum remains challenging. We release the generator, dataset, raw model outputs, scoring code, and manifests with checksums to support fully reproducible runs and retrospective rescoring under alternative tolerance policies.
[385] Stay in Character, Stay Safe: Dual-Cycle Adversarial Self-Evolution for Safety Role-Playing Agents
Mingyang Liao, Yichen Wan, shuchen wu, Chenxi Miao, Xin Shen, Weikang Li, Yang Li, Deguo Xia, Jizhou Huang
Main category: cs.AI
TL;DR: A training-free Dual-Cycle Adversarial Self-Evolution framework that improves jailbreak resistance in LLM-based role-playing while maintaining persona fidelity through hierarchical knowledge distillation.
Details
Motivation: Current LLM-based role-playing systems face a trade-off: stronger persona adherence increases vulnerability to jailbreak attacks, especially for risky personas. Training-time solutions are costly, degrade character behavior, and don't work with closed-weight LLMs.Method: Proposes a dual-cycle framework: (1) Persona-Targeted Attacker Cycle synthesizes progressively stronger jailbreak prompts, (2) Role-Playing Defender Cycle distills failures into hierarchical knowledge base (global safety rules, persona-grounded constraints, safe exemplars). At inference, retrieves and composes structured knowledge to guide generation.
Result: Extensive experiments across multiple proprietary LLMs show consistent gains over baselines on both role fidelity and jailbreak resistance, with robust generalization to unseen personas and attack prompts.
Conclusion: The training-free framework effectively balances safety and persona fidelity in role-playing LLMs without requiring model retraining, offering a practical solution for evolving personas and attack strategies.
Abstract: LLM-based role-playing has rapidly improved in fidelity, yet stronger adherence to persona constraints commonly increases vulnerability to jailbreak attacks, especially for risky or negative personas. Most prior work mitigates this issue with training-time solutions (e.g., data curation or alignment-oriented regularization). However, these approaches are costly to maintain as personas and attack strategies evolve, can degrade in-character behavior, and are typically infeasible for frontier closed-weight LLMs. We propose a training-free Dual-Cycle Adversarial Self-Evolution framework with two coupled cycles. A Persona-Targeted Attacker Cycle synthesizes progressively stronger jailbreak prompts, while a Role-Playing Defender Cycle distills observed failures into a hierarchical knowledge base of (i) global safety rules, (ii) persona-grounded constraints, and (iii) safe in-character exemplars. At inference time, the Defender retrieves and composes structured knowledge from this hierarchy to guide generation, producing responses that remain faithful to the target persona while satisfying safety constraints. Extensive experiments across multiple proprietary LLMs show consistent gains over strong baselines on both role fidelity and jailbreak resistance, and robust generalization to unseen personas and attack prompts.
[386] DPBench: Large Language Models Struggle with Simultaneous Coordination
Najmul Hasan, Prashanth BusiReddyGari
Main category: cs.AI
TL;DR: DPBench is a benchmark based on the Dining Philosophers problem that evaluates LLM coordination in multi-agent systems under resource contention, revealing LLMs fail catastrophically in simultaneous decision-making scenarios despite working well in sequential settings.
Details
Motivation: As large language models are increasingly deployed in multi-agent systems, there's a lack of benchmarks testing their ability to coordinate under resource contention, particularly in concurrent decision-making scenarios.Method: DPBench uses the Dining Philosophers problem to evaluate LLM coordination across eight conditions varying decision timing (sequential vs. simultaneous), group size, and communication capabilities, testing models like GPT-5.2, Claude Opus 4.5, and Grok 4.1.
Result: LLMs show striking asymmetry: they coordinate effectively in sequential settings but fail catastrophically in simultaneous decision-making, with deadlock rates exceeding 95% in some conditions. Communication doesn’t help and can even increase deadlock rates due to convergent reasoning where agents independently arrive at identical strategies.
Conclusion: Multi-agent LLM systems requiring concurrent resource access may need external coordination mechanisms rather than relying on emergent coordination, as LLMs fundamentally fail at simultaneous coordination despite working well in sequential scenarios.
Abstract: Large language models are increasingly deployed in multi-agent systems, yet we lack benchmarks that test whether they can coordinate under resource contention. We introduce DPBench, a benchmark based on the Dining Philosophers problem that evaluates LLM coordination across eight conditions that vary decision timing, group size, and communication. Our experiments with GPT-5.2, Claude Opus 4.5, and Grok 4.1 reveal a striking asymmetry: LLMs coordinate effectively in sequential settings but fail when decisions must be made simultaneously, with deadlock rates exceeding 95% under some conditions. We trace this failure to convergent reasoning, where agents independently arrive at identical strategies that, when executed simultaneously, guarantee deadlock. Contrary to expectations, enabling communication does not resolve this problem and can even increase deadlock rates. Our findings suggest that multi-agent LLM systems requiring concurrent resource access may need external coordination mechanisms rather than relying on emergent coordination. DPBench is released as an open-source benchmark. Code and benchmark are available at https://github.com/najmulhasan-code/dpbench.
[387] Lang2Act: Fine-Grained Visual Reasoning through Self-Emergent Linguistic Toolchains
Yuqi Xiong, Chunyi Peng, Zhipeng Xu, Zhenghao Liu, Zulong Chen, Yukun Yan, Shuo Wang, Yu Gu, Ge Yu
Main category: cs.AI
TL;DR: Lang2Act enables fine-grained visual perception in VLMs through self-emergent linguistic toolchains using RL-based training, improving performance by over 4%.
Details
Motivation: Existing VRAG frameworks rely on rigid, pre-defined external tools that separate visual perception from reasoning, leading to unnecessary visual information loss, especially with image operations like cropping.Method: Two-stage RL-based training: first stage optimizes VLMs to self-explore high-quality actions for building a reusable linguistic toolbox; second stage optimizes VLMs to effectively exploit these linguistic tools for downstream reasoning.
Result: Lang2Act substantially enhances visual perception capabilities of VLMs, achieving performance improvements of over 4% compared to existing methods.
Conclusion: The proposed self-emergent linguistic toolchain approach effectively bridges visual perception and reasoning in VLMs, overcoming limitations of decoupled designs in existing VRAG frameworks.
Abstract: Visual Retrieval-Augmented Generation (VRAG) enhances Vision-Language Models (VLMs) by incorporating external visual documents to address a given query. Existing VRAG frameworks usually depend on rigid, pre-defined external tools to extend the perceptual capabilities of VLMs, typically by explicitly separating visual perception from subsequent reasoning processes. However, this decoupled design can lead to unnecessary loss of visual information, particularly when image-based operations such as cropping are applied. In this paper, we propose Lang2Act, which enables fine-grained visual perception and reasoning through self-emergent linguistic toolchains. Rather than invoking fixed external engines, Lang2Act collects self-emergent actions as linguistic tools and leverages them to enhance the visual perception capabilities of VLMs. To support this mechanism, we design a two-stage Reinforcement Learning (RL)-based training framework. Specifically, the first stage optimizes VLMs to self-explore high-quality actions for constructing a reusable linguistic toolbox, and the second stage further optimizes VLMs to exploit these linguistic tools for downstream reasoning effectively. Experimental results demonstrate the effectiveness of Lang2Act in substantially enhancing the visual perception capabilities of VLMs, achieving performance improvements of over 4%. All code and data are available at https://github.com/NEUIR/Lang2Act.
[388] MAPLE: A Sub-Agent Architecture for Memory, Learning, and Personalization in Agentic AI Systems
Deepak Babu Piskala
Main category: cs.AI
TL;DR: MAPLE decomposes LLM agent personalization into three distinct components: Memory (storage/retrieval), Learning (asynchronous intelligence extraction), and Personalization (real-time application), achieving significant improvements in adaptation capabilities.
Details
Motivation: Current LLM agents treat memory, learning, and personalization as unified capabilities, limiting their ability to adapt to individual users. The authors argue these should be distinct mechanisms with different infrastructure, timescales, and optimization strategies.Method: Proposes MAPLE (Memory-Adaptive Personalized LEarning) - a principled decomposition where each component operates as a dedicated sub-agent: Memory handles storage/retrieval infrastructure, Learning extracts intelligence from accumulated interactions asynchronously, and Personalization applies learned knowledge in real-time within finite context budgets.
Result: Experimental evaluation on MAPLE-Personas benchmark shows 14.6% improvement in personalization score compared to stateless baseline (p < 0.01, Cohen’s d = 0.95) and increases trait incorporation rate from 45% to 75%.
Conclusion: Decomposing personalization into distinct memory, learning, and personalization components enables LLM agents to genuinely learn and adapt to individual users, addressing a fundamental limitation in current agent architectures.
Abstract: Large language model (LLM) agents have emerged as powerful tools for complex tasks, yet their ability to adapt to individual users remains fundamentally limited. We argue this limitation stems from a critical architectural conflation: current systems treat memory, learning, and personalization as a unified capability rather than three distinct mechanisms requiring different infrastructure, operating on different timescales, and benefiting from independent optimization. We propose MAPLE (Memory-Adaptive Personalized LEarning), a principled decomposition where Memory handles storage and retrieval infrastructure; Learning extracts intelligence from accumulated interactions asynchronously; and Personalization applies learned knowledge in real-time within finite context budgets. Each component operates as a dedicated sub-agent with specialized tooling and well-defined interfaces. Experimental evaluation on the MAPLE-Personas benchmark demonstrates that our decomposition achieves a 14.6% improvement in personalization score compared to a stateless baseline (p < 0.01, Cohen’s d = 0.95) and increases trait incorporation rate from 45% to 75% – enabling agents that genuinely learn and adapt.
[389] NL2LOGIC: AST-Guided Translation of Natural Language into First-Order Logic with Large Language Models
Rizky Ramadhana Putra, Raihan Sultan Pasha Basuki, Yutong Cheng, Peng Gao
Main category: cs.AI
TL;DR: NL2LOGIC is a framework that improves natural language to first-order logic translation using abstract syntax trees as intermediate representation, achieving near-perfect syntactic accuracy and significant semantic improvements.
Details
Motivation: Existing methods for translating natural language to first-order logic using LLMs suffer from fragile syntax control due to weak enforcement of global grammar constraints and low semantic faithfulness from insufficient clause-level semantic understanding.Method: Introduces abstract syntax tree as intermediate representation, combining recursive LLM-based semantic parser with AST-guided generator that deterministically produces solver-ready logic code.
Result: Achieves 99% syntactic accuracy, improves semantic correctness by up to 30% over SOTA baselines, and when integrated into Logic-LM yields near-perfect executability and 31% improvement in downstream reasoning accuracy.
Conclusion: NL2LOGIC framework significantly improves both syntactic and semantic accuracy of natural language to logic translation, enhancing automated reasoning systems.
Abstract: Automated reasoning is critical in domains such as law and governance, where verifying claims against facts in documents requires both accuracy and interpretability. Recent work adopts structured reasoning pipelines that translate natural language into first-order logic and delegate inference to automated solvers. With the rise of large language models, approaches such as GCD and CODE4LOGIC leverage their reasoning and code generation capabilities to improve logic parsing. However, these methods suffer from fragile syntax control due to weak enforcement of global grammar constraints and low semantic faithfulness caused by insufficient clause-level semantic understanding. We propose NL2LOGIC, a first-order logic translation framework that introduces an abstract syntax tree as an intermediate representation. NL2LOGIC combines a recursive large language model based semantic parser with an abstract syntax tree guided generator that deterministically produces solver-ready logic code. Experiments on the FOLIO, LogicNLI, and ProofWriter benchmarks show that NL2LOGIC achieves 99 percent syntactic accuracy and improves semantic correctness by up to 30 percent over state-of-the-art baselines. Furthermore, integrating NL2LOGIC into Logic-LM yields near-perfect executability and improves downstream reasoning accuracy by 31 percent compared to Logic-LM’s original few-shot unconstrained translation module.
[390] AST-PAC: AST-guided Membership Inference for Code
Roham Koohestani, Ali Al-Kaswan, Jonathan Katzy, Maliheh Izadi
Main category: cs.AI
TL;DR: AST-PAC: A syntax-aware membership inference attack for code LLMs using AST-based perturbations to improve auditing of unauthorized training data usage.
Details
Motivation: Code LLMs are trained on massive datasets with licensing restrictions, creating data governance and copyright challenges. Membership Inference Attacks (MIAs) can audit unauthorized data usage, but existing methods like PAC don't handle code's rigid syntax well.Method: Introduces AST-PAC, a domain-specific adaptation of Polarized Augment Calibration (PAC) that uses Abstract Syntax Tree (AST) based perturbations to generate syntactically valid calibration samples for membership inference attacks on code models.
Result: AST-PAC improves performance as syntactic size grows where PAC degrades, but under-mutates small files and underperforms on alphanumeric-rich code. PAC generally outperforms Loss baseline but suffers on larger, complex files due to syntax-blind augmentations.
Conclusion: Future work needed on syntax-aware and size-adaptive calibration for reliable provenance auditing of code language models. AST-PAC shows promise but has limitations with small files and certain code types.
Abstract: Code Large Language Models are frequently trained on massive datasets containing restrictively licensed source code. This creates urgent data governance and copyright challenges. Membership Inference Attacks (MIAs) can serve as an auditing mechanism to detect unauthorized data usage in models. While attacks like the Loss Attack provide a baseline, more involved methods like Polarized Augment Calibration (PAC) remain underexplored in the code domain. This paper presents an exploratory study evaluating these methods on 3B–7B parameter code models. We find that while PAC generally outperforms the Loss baseline, its effectiveness relies on augmentation strategies that disregard the rigid syntax of code, leading to performance degradation on larger, complex files. To address this, we introduce AST-PAC, a domain-specific adaptation that utilizes Abstract Syntax Tree (AST) based perturbations to generate syntactically valid calibration samples. Preliminary results indicate that AST-PAC improves as syntactic size grows, where PAC degrades, but under-mutates small files and underperforms on alphanumeric-rich code. Overall, the findings motivate future work on syntax-aware and size-adaptive calibration as a prerequisite for reliable provenance auditing of code language models.
[391] X-Blocks: Linguistic Building Blocks of Natural Language Explanations for Automated Vehicles
Ashkan Y. Zadeh, Xiaomeng Li, Andry Rakotonirainy, Ronald Schroeter, Sebastien Glaser, Zishuo Zhu
Main category: cs.AI
TL;DR: X-Blocks framework analyzes linguistic building blocks of natural language explanations for automated vehicles across context, syntax, and lexicon levels
Details
Motivation: Existing approaches lack systematic frameworks for analyzing how humans linguistically construct driving rationales across diverse scenarios in automated vehiclesMethod: Hierarchical X-Blocks framework with RACE (multi-LLM ensemble) for context classification, log-odds analysis for lexical patterns, and dependency parsing for syntactic analysis
Result: RACE achieves 91.45% accuracy for context classification, identifies context-specific vocabulary patterns, and reveals limited repertoire of reusable grammar families
Conclusion: X-Blocks provides evidence-based linguistic design principles for generating scenario-aware explanations to support transparency and user trust in automated driving systems
Abstract: Natural language explanations play a critical role in establishing trust and acceptance of automated vehicles (AVs), yet existing approaches lack systematic frameworks for analysing how humans linguistically construct driving rationales across diverse scenarios. This paper introduces X-Blocks (eXplanation Blocks), a hierarchical analytical framework that identifies the linguistic building blocks of natural language explanations for AVs at three levels: context, syntax, and lexicon. At the context level, we propose RACE (Reasoning-Aligned Classification of Explanations), a multi-LLM ensemble framework that combines Chain-of-Thought reasoning with self-consistency mechanisms to robustly classify explanations into 32 scenario-aware categories. Applied to human-authored explanations from the Berkeley DeepDrive-X dataset, RACE achieves 91.45 percent accuracy and a Cohens kappa of 0.91 against cases with human annotator agreement, indicating near-human reliability for context classification. At the lexical level, log-odds analysis with informative Dirichlet priors reveals context-specific vocabulary patterns that distinguish driving scenarios. At the syntactic level, dependency parsing and template extraction show that explanations draw from a limited repertoire of reusable grammar families, with systematic variation in predicate types and causal constructions across contexts. The X-Blocks framework is dataset-agnostic and task-independent, offering broad applicability to other automated driving datasets and safety-critical domains. Overall, our findings provide evidence-based linguistic design principles for generating scenario-aware explanations that support transparency, user trust, and cognitive accessibility in automated driving systems.
[392] A First Proof Sprint
Joseph Corneli
Main category: cs.AI
TL;DR: Multi-agent proof sprint workflow using wiring-diagram decompositions for mathematical problem verification, with heterogeneous outcomes across ten research problems.
Details
Motivation: To develop a systematic approach for rapid mathematical proof generation and verification using multi-agent collaboration, addressing the challenge of reliable proof validation in compressed timeframes.Method: Uses wiring-diagram decompositions of claim dependencies to localize gaps, combines rapid draft generation with adversarial verification, targeted repair, and explicit provenance tracking.
Result: Heterogeneous outcomes across ten problems: Problem 3 has validation-complete existence path, Problem 5 solved in scope-limited form, Problem 10 conditional, Problems 4 and 6 partial, Problem 7 provisionally closed. QC layer shows validation artifacts with unresolved gaps.
Conclusion: Structure-aware verification and layer-switching strategies improve reliability and calibration in compressed proof sprints, demonstrating the effectiveness of the multi-agent approach for mathematical problem-solving.
Abstract: This monograph reports a multi-agent proof sprint on ten research-level problems, combining rapid draft generation with adversarial verification, targeted repair, and explicit provenance. The workflow uses wiring-diagram decompositions of claim dependencies to localize gaps and coordinate reviewer-driven revisions. Final outcomes are heterogeneous but explicit: the manuscript distinguishes mathematical status from QC-validation status. Mathematically, Problem3 has a validation-complete existence path under the scoped criterion used here (uniqueness/irreducibility treated as optional), Problem 5 is solved in a scope-limited form for $F_O$-local connective spectra, Problem 10 is conditional under clearly stated assumptions (with explicit necessity counterexamples when assumptions are dropped), and Problems 4 and 6 are partial with named remaining obligations in the general case (including an unconditional $K_n$ result for Problem 6 with $c_0 = 1/3$). Problem 7 is treated as provisionally closed via the rotation-route theorem chain, pending independent ledger re-check. At the QC layer, Problems7 and~9 have node-level validation artifacts but still contain unresolved verifier gaps. The main methodological result is that structure-aware verification and layer-switching strategies improve reliability and calibration in compressed proof sprints.
[393] General learned delegation by clones
Darren Li, Meiqi Chen, Chenze Shao, Fandong Meng, Jie Zhou
Main category: cs.AI
TL;DR: SELFCEST enables language models to spawn parallel clones for efficient reasoning under fixed inference budgets using agentic reinforcement learning.
Details
Motivation: Current frontier language models struggle with compute inefficiency during test-time computation for serial reasoning or uncoordinated parallel sampling under fixed inference budgets.Method: Equips a base model with ability to spawn same-weight clones in separate parallel contexts using agentic reinforcement learning, trained end-to-end under global task reward with shared-parameter rollouts.
Result: Improves accuracy-cost Pareto frontier on challenging math reasoning benchmarks and long-context multi-hop QA relative to monolithic baselines at matched inference budget, with out-of-distribution generalization.
Conclusion: SELFCEST provides an effective approach for improving language model performance through coordinated parallel computation under fixed inference budgets.
Abstract: Frontier language models improve with additional test-time computation, but serial reasoning or uncoordinated parallel sampling can be compute-inefficient under fixed inference budgets. We propose SELFCEST, which equips a base model with the ability to spawn same-weight clones in separate parallel contexts by agentic reinforcement learning. Training is end-to-end under a global task reward with shared-parameter rollouts, yielding a learned controller that allocates both generation and context budget across branches. Across challenging math reasoning benchmarks and long-context multi-hop QA, SELFCEST improves the accuracy-cost Pareto frontier relative to monolithic baselines at matched inference budget, and exhibits out-of-distribution generalization in both domains.
[394] Guided Collaboration in Heterogeneous LLM-Based Multi-Agent Systems via Entropy-Based Understanding Assessment and Experience Retrieval
Linlin Wang, Tianqing Zhu, Laiqiao Qin, Longxiang Gao, Wanlei Zhou
Main category: cs.AI
TL;DR: Proposes an entropy-based adaptive guidance framework for heterogeneous multi-agent systems to address cognitive mismatching between strong and weak agents through dynamic guidance adjustment and RAG-based experience retention.
Details
Motivation: Heterogeneous multi-agent systems face cognitive mismatching problems where capability differences between strong and weak agents lead to ineffective collaboration, sometimes performing worse than weak-weak combinations.Method: Entropy-Based Adaptive Guidance Framework that quantifies weak agents’ understanding through multi-dimensional entropy metrics (expression, uncertainty, structure, coherence, relevance) and adaptively adjusts guidance intensity at three levels (light, moderate, intensive), plus RAG mechanism for retaining successful collaboration experiences.
Result: Extensive experiments on GSM8K, MBPP, and CVRP benchmarks demonstrate consistent enhancement of effectiveness and stability in heterogeneous collaboration, showing adaptive guidance mitigates cognitive imbalance.
Conclusion: Adaptive guidance framework addresses cognitive mismatching in heterogeneous multi-agent systems, establishing a scalable pathway toward more robust cooperative multi-agent intelligence.
Abstract: With recent breakthroughs in large language models (LLMs) for reasoning, planning, and complex task generation, artificial intelligence systems are transitioning from isolated single-agent architectures to multi-agent systems with collaborative intelligence. However, in heterogeneous multi-agent systems (HMAS), capability differences among agents give rise to consistent cognitive problems, where strong and weak models fail to contribute effectively. We define the collaboration as a strong-weak system. Through comprehensive experiments, we disclose a counterintuitive phenomenon in the strong-weak system: a strong-weak collaboration may under-perform weak-weak combinations, revealing that cognitive mismatching are key bottlenecks limiting heterogeneous cooperation. To overcome these challenges, we propose an Entropy-Based Adaptive Guidance Framework that dynamically aligns the guidance with the cognitive state of each agent. The framework quantifies the understanding of weak agents through multi-dimensional entropy metrics - covering expression, uncertainty, structure, coherence, and relevance - and adaptively adjusts the intensity of the guidance at light, moderate and intensive levels. Furthermore, a Retrieval-Augmented Generation (RAG) mechanism is incorporated to retain successful collaboration experiences, enabling both immediate adaptation and long-term learning. Extensive experiments on three benchmark datasets, GSM8K, MBPP, and CVRP demonstrate that our approach consistently enhances the effectiveness and stability of heterogeneous collaboration. The results highlight that adaptive guidance not only mitigates cognitive imbalance but also establishes a scalable pathway toward more robust, cooperative multi-agent intelligence.
[395] Human-Centered Explainable AI for Security Enhancement: A Deep Intrusion Detection Framework
Md Muntasir Jahid Ayan, Md. Shahriar Rashid, Tazzina Afroze Hassan, Hossain Md. Mubashshir Jamil, Mahbubul Islam, Lisan Al Amin, Rupak Kumar Das, Farzana Akter, Faisal Quader
Main category: cs.AI
TL;DR: A novel intrusion detection system framework integrating Explainable AI (XAI) with CNN-LSTM deep learning models for transparent and accurate cyber threat detection, evaluated on NSL-KDD dataset with SHAP explanations and expert trust surveys.
Details
Motivation: Increasing cyber-threat complexity demands intrusion detection systems that are both accurate and interpretable, requiring transparency in deep learning models for security analysts to understand and validate decisions.Method: Combined CNN and LSTM networks to capture temporal dependencies in traffic sequences, integrated SHAP (SHapley Additive exPlanations) for model interpretability, and conducted trust-focused expert surveys using IPIP6 and Big Five personality traits via interactive UI.
Result: Both CNN and LSTM achieved 0.99 accuracy on NSL-KDD dataset, with LSTM outperforming CNN on macro average precision, recall, and F-1 score. SHAP identified key influential features like srv_serror_rate, dst_host_srv_serror_rate, and serror_rate. Expert surveys evaluated system reliability and usability.
Conclusion: The framework demonstrates the potential of combining performance and transparency in cybersecurity solutions, with recommendations for future enhancements through adaptive learning for real-time threat detection.
Abstract: The increasing complexity and frequency of cyber-threats demand intrusion detection systems (IDS) that are not only accurate but also interpretable. This paper presented a novel IDS framework that integrated Explainable Artificial Intelligence (XAI) to enhance transparency in deep learning models. The framework was evaluated experimentally using the benchmark dataset NSL-KDD, demonstrating superior performance compared to traditional IDS and black-box deep learning models. The proposed approach combined Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) networks for capturing temporal dependencies in traffic sequences. Our deep learning results showed that both CNN and LSTM reached 0.99 for accuracy, whereas LSTM outperformed CNN at macro average precision, recall, and F-1 score. For weighted average precision, recall, and F-1 score, both models scored almost similarly. To ensure interpretability, the XAI model SHapley Additive exPlanations (SHAP) was incorporated, enabling security analysts to understand and validate model decisions. Some notable influential features were srv_serror_rate, dst_host_srv_serror_rate, and serror_rate for both models, as pointed out by SHAP. We also conducted a trust-focused expert survey based on IPIP6 and Big Five personality traits via an interactive UI to evaluate the system’s reliability and usability. This work highlighted the potential of combining performance and transparency in cybersecurity solutions and recommends future enhancements through adaptive learning for real-time threat detection.
[396] TemporalBench: A Benchmark for Evaluating LLM-Based Agents on Contextual and Event-Informed Time Series Tasks
Muyan Weng, Defu Cao, Wei Yang, Yashaswi Sharma, Yan Liu
Main category: cs.AI
TL;DR: TemporalBench is a multi-domain benchmark for evaluating temporal reasoning across four domains (retail, healthcare, energy, physical systems) with a four-tier taxonomy to test models’ ability to interpret temporal patterns, align with context, and adapt to changing conditions.
Details
Motivation: Current forecasting benchmarks don't distinguish between genuine temporal understanding and contextual reasoning. The authors aim to create a diagnostic benchmark that reveals whether models can correctly interpret temporal patterns, align them with external context, and adapt predictions when conditions change.Method: Developed TemporalBench with a four-tier task taxonomy: 1) historical structure interpretation, 2) context-free forecasting, 3) contextual temporal reasoning, and 4) event-conditioned prediction. The benchmark spans four real-world domains and controls access to future targets and contextual information to enable diagnostic analysis.
Result: Extensive baseline experiments show that strong numerical forecasting accuracy doesn’t reliably translate to robust contextual or event-aware temporal reasoning. Existing agent frameworks exhibit fragmented strengths and systematic failure modes that remain hidden under forecasting-only benchmarks.
Conclusion: TemporalBench reveals limitations in current temporal reasoning models and provides a diagnostic tool for evaluating genuine temporal understanding versus contextual reasoning capabilities across multiple domains.
Abstract: It is unclear whether strong forecasting performance reflects genuine temporal understanding or the ability to reason under contextual and event-driven conditions. We introduce TemporalBench, a multi-domain benchmark designed to evaluate temporal reasoning behavior under progressively richer informational settings. TemporalBench adopts a four-tier task taxonomy that examines historical structure interpretation, context-free forecasting, contextual temporal reasoning, and event-conditioned prediction across four real-world domains: retail, healthcare, energy, and physical systems. By controlling access to future targets and contextual information, the benchmark enables a diagnostic analysis of whether models can correctly interpret temporal patterns, align them with external context, and adapt predictions when conditions change. Extensive baseline experiments show that strong numerical forecasting accuracy does not reliably translate into robust contextual or event-aware temporal reasoning; instead, existing agent frameworks exhibit fragmented strengths and systematic failure modes that remain largely hidden under forecasting-only benchmarks. The TemporalBench dataset is publicly available at https://huggingface.co/datasets/Melady/TemporalBench, and we additionally provide a public leaderboard at https://huggingface.co/spaces/Melady/TemporalBench_Leaderboard.
[397] ProMoral-Bench: Evaluating Prompting Strategies for Moral Reasoning and Safety in LLMs
Rohan Subramanian Thomas, Shikhar Shiromani, Abdullah Chaudhry, Ruizhe Li, Vasu Sharma, Kevin Zhu, Sunishchal Dev
Main category: cs.AI
TL;DR: ProMoral-Bench is a unified benchmark for evaluating 11 prompting paradigms across LLMs on moral competence and safety alignment, introducing a new robustness test and Unified Moral Safety Score metric.
Details
Motivation: Prompt design significantly impacts LLM moral competence and safety alignment, but current empirical comparisons are fragmented across different datasets and models, lacking standardized evaluation frameworks.Method: Introduces ProMoral-Bench with 11 prompting paradigms evaluated across four LLM families using ETHICS, Scruples, WildJailbreak datasets and new ETHICS-Contrast robustness test, measured via Unified Moral Safety Score (UMSS) balancing accuracy and safety.
Result: Compact, exemplar-guided scaffolds outperform complex multi-stage reasoning, achieving higher UMSS scores with greater robustness at lower token cost. Few-shot exemplars consistently enhance moral stability and jailbreak resistance while multi-turn reasoning proves fragile under perturbations.
Conclusion: ProMoral-Bench establishes a standardized framework for principled, cost-effective prompt engineering, showing that simpler prompting approaches can be more effective for moral safety alignment than complex reasoning methods.
Abstract: Prompt design significantly impacts the moral competence and safety alignment of large language models (LLMs), yet empirical comparisons remain fragmented across datasets and models.We introduce ProMoral-Bench, a unified benchmark evaluating 11 prompting paradigms across four LLM families. Using ETHICS, Scruples, WildJailbreak, and our new robustness test, ETHICS-Contrast, we measure performance via our proposed Unified Moral Safety Score (UMSS), a metric balancing accuracy and safety. Our results show that compact, exemplar-guided scaffolds outperform complex multi-stage reasoning, providing higher UMSS scores and greater robustness at a lower token cost. While multi-turn reasoning proves fragile under perturbations, few-shot exemplars consistently enhance moral stability and jailbreak resistance. ProMoral-Bench establishes a standardized framework for principled, cost-effective prompt engineering.
[398] From Fluent to Verifiable: Claim-Level Auditability for Deep Research Agents
Razeen A Rasheed, Somnath Banerjee, Animesh Mukherjee, Rima Hazra
Main category: cs.AI
TL;DR: Paper proposes auditability as critical design target for AI research agents, introduces Auditable Autonomous Research standard with metrics for provenance coverage/soundness, and advocates for semantic provenance graphs with continuous validation.
Details
Motivation: As AI research generation becomes cheap, the bottleneck shifts from factual errors to scientifically styled outputs with weak claim-evidence links, making auditability the dominant risk requiring systematic solutions.Method: Proposes claim-level auditability as design target, identifies long-horizon failure modes (objective drift, transient constraints, unverifiable inference), introduces AAR standard with four metrics (provenance coverage, soundness, contradiction transparency, audit effort), and advocates for semantic provenance graphs with protocolized validation.
Result: Develops framework for measuring auditability in AI research agents, identifies key failure patterns, and proposes practical instrumentation patterns for scalable deployment of semantic provenance systems.
Conclusion: Auditability should be a first-class design consideration for deep research agents, requiring semantic provenance systems with continuous validation to address the shift from factual errors to claim-evidence linkage problems.
Abstract: A deep research agent produces a fluent scientific report in minutes; a careful reader then tries to verify the main claims and discovers the real cost is not reading, but tracing: which sentence is supported by which passage, what was ignored, and where evidence conflicts. We argue that as research generation becomes cheap, auditability becomes the bottleneck, and the dominant risk shifts from isolated factual errors to scientifically styled outputs whose claim-evidence links are weak, missing, or misleading. This perspective proposes claim-level auditability as a first-class design and evaluation target for deep research agents, distills recurring long-horizon failure modes (objective drift, transient constraints, and unverifiable inference), and introduces the Auditable Autonomous Research (AAR) standard, a compact measurement framework that makes auditability testable via provenance coverage, provenance soundness, contradiction transparency, and audit effort. We then argue for semantic provenance with protocolized validation: persistent, queryable provenance graphs that encode claim–evidence relations (including conflicts) and integrate continuous validation during synthesis rather than after publication, with practical instrumentation patterns to support deployment at scale.
[399] Artificial Organisations
William Waites
Main category: cs.AI
TL;DR: Multi-agent AI system uses institutional design principles (compartmentalization, adversarial review) to achieve reliable outcomes from potentially misaligned individual agents, demonstrated through a document composition system with layered verification.
Details
Motivation: Traditional AI alignment focuses on making individual systems reliable, but human institutions achieve reliable collective behavior through organizational structure. This paper explores applying institutional design principles to multi-agent AI systems to achieve reliability through architectural design rather than assuming individual alignment.Method: Developed Perseverance Composition Engine, a multi-agent system for document composition with three specialized agents: Composer (drafts text), Corroborator (verifies factual substantiation with full source access), and Critic (evaluates argumentative quality without source access). Information asymmetry is enforced by system architecture to create layered verification.
Result: Tested on 474 composition tasks, the system exhibited patterns consistent with institutional hypothesis. When assigned impossible tasks requiring fabricated content, the system progressed from attempted fabrication toward honest refusal with alternative proposals—behavior not instructed or individually incentivized.
Conclusion: Organizational theory provides a productive framework for multi-agent AI safety. Architectural enforcement through information compartmentalization offers a route to reliable collective behavior from unreliable individual components, positioning institutional design as an alternative to individual alignment approaches.
Abstract: Alignment research focuses on making individual AI systems reliable. Human institutions achieve reliable collective behaviour differently: they mitigate the risk posed by misaligned individuals through organisational structure. Multi-agent AI systems should follow this institutional model using compartmentalisation and adversarial review to achieve reliable outcomes through architectural design rather than assuming individual alignment. We demonstrate this approach through the Perseverance Composition Engine, a multi-agent system for document composition. The Composer drafts text, the Corroborator verifies factual substantiation with full source access, and the Critic evaluates argumentative quality without access to sources: information asymmetry enforced by system architecture. This creates layered verification: the Corroborator detects unsupported claims, whilst the Critic independently assesses coherence and completeness. Observations from 474 composition tasks (discrete cycles of drafting, verification, and evaluation) exhibit patterns consistent with the institutional hypothesis. When assigned impossible tasks requiring fabricated content, this iteration enabled progression from attempted fabrication toward honest refusal with alternative proposals–behaviour neither instructed nor individually incentivised. These findings motivate controlled investigation of whether architectural enforcement produces reliable outcomes from unreliable components. This positions organisational theory as a productive framework for multi-agent AI safety. By implementing verification and evaluation as structural properties enforced through information compartmentalisation, institutional design offers a route to reliable collective behaviour from unreliable individual components.
[400] BEAGLE: Behavior-Enforced Agent for Grounded Learner Emulation
Hanchen David Wang, Clayton Cohn, Zifan Xu, Siyuan Guo, Gautam Biswas, Meiyi Ma
Main category: cs.AI
TL;DR: BEAGLE is a neuro-symbolic framework that simulates authentic student learning behaviors in open-ended problem-solving by incorporating Self-Regulated Learning theory to address LLMs’ competency bias toward efficient correctness.
Details
Motivation: Collecting authentic student learning data is challenging due to privacy concerns and high costs. LLMs can simulate students but suffer from competency bias, optimizing for efficient correctness rather than the erratic, iterative struggle characteristic of novice learners.Method: BEAGLE integrates three innovations: (1) semi-Markov model for cognitive/metacognitive behavior transitions, (2) Bayesian Knowledge Tracing with explicit flaw injection for realistic knowledge gaps, and (3) decoupled agent design separating strategy use from code generation to prevent silent error correction.
Result: BEAGLE significantly outperforms state-of-the-art baselines in reproducing authentic trajectories on Python programming tasks. In a human Turing test, users achieved only 52.8% accuracy distinguishing synthetic traces from real student data (indistinguishable from random guessing).
Conclusion: BEAGLE successfully addresses LLM competency bias by incorporating Self-Regulated Learning theory, enabling realistic simulation of novice learning behaviors with applications for training adaptive tutoring systems and testing pedagogical interventions.
Abstract: Simulating student learning behaviors in open-ended problem-solving environments holds potential for education research, from training adaptive tutoring systems to stress-testing pedagogical interventions. However, collecting authentic data is challenging due to privacy concerns and the high cost of longitudinal studies. While Large Language Models (LLMs) offer a promising path to student simulation, they suffer from competency bias, optimizing for efficient correctness rather than the erratic, iterative struggle characteristic of novice learners. We present BEAGLE, a neuro-symbolic framework that addresses this bias by incorporating Self-Regulated Learning (SRL) theory into a novel architecture. BEAGLE integrates three key technical innovations: (1) a semi-Markov model that governs the timing and transitions of cognitive behaviors and metacognitive behaviors; (2) Bayesian Knowledge Tracing with explicit flaw injection to enforce realistic knowledge gaps and “unknown unknowns”; and (3) a decoupled agent design that separates high-level strategy use from code generation actions to prevent the model from silently correcting its own intentional errors. In evaluations on Python programming tasks, BEAGLE significantly outperforms state-of-the-art baselines in reproducing authentic trajectories. In a human Turing test, users were unable to distinguish synthetic traces from real student data, achieving an accuracy indistinguishable from random guessing (52.8%).
[401] Accuracy Standards for AI at Work vs. Personal Life: Evidence from an Online Survey
Gaston Besanson, Federico Todeschini
Main category: cs.AI
TL;DR: People demand higher AI accuracy at work than in personal life, with significant gaps in accuracy requirements and different disruption patterns when AI tools are unavailable.
Details
Motivation: To understand how people trade off accuracy when using AI tools in professional vs personal contexts, and how they cope when these tools are unavailable, given that modern AI systems produce acceptable but non-identical outputs.Method: Online survey with N=300 participants, focusing on accuracy requirements and disruption patterns. Defined accuracy as context-specific reliability based on user intent alignment within tolerance thresholds.
Result: Significant accuracy gap: 24.1% require high accuracy at work vs 8.8% in personal life (+15.3pp). Gap remains large under broader definitions. Heavy app users have stricter work standards. More disruption in personal routines (34.1%) than at work (15.3%) when tools are unavailable.
Conclusion: People have substantially higher accuracy requirements for AI tools in professional contexts compared to personal use, and experience different disruption patterns when these tools become unavailable, highlighting context-dependent AI adoption and usage patterns.
Abstract: We study how people trade off accuracy when using AI-powered tools in professional versus personal contexts for adoption purposes, the determinants of those trade-offs, and how users cope when AI/apps are unavailable. Because modern AI systems (especially generative models) can produce acceptable but non-identical outputs, we define “accuracy” as context-specific reliability: the degree to which an output aligns with the user’s intent within a tolerance threshold that depends on stakes and the cost of correction. In an online survey (N=300), among respondents with both accuracy items (N=170), the share requiring high accuracy (top-box) is 24.1% at work vs. 8.8% in personal life (+15.3 pp; z=6.29, p<0.001). The gap remains large under a broader top-two-box definition (67.0% vs. 32.9%) and on the full 1-5 ordinal scale (mean 3.86 vs. 3.08). Heavy app use and experience patterns correlate with stricter work standards (H2). When tools are unavailable (H3), respondents report more disruption in personal routines than at work (34.1% vs. 15.3%, p<0.01). We keep the main text focused on these substantive results and place test taxonomy and power derivations in a technical appendix.
[402] Mirror: A Multi-Agent System for AI-Assisted Ethics Review
Yifan Ding, Yuhui Shi, Zhiyan Li, Zilong Wang, Yifeng Gao, Yajun Yang, Mengjie Yang, Yixiu Liang, Xipeng Qiu, Xuanjing Huang, Xingjun Ma, Yu-Gang Jiang, Guoyu Wang
Main category: cs.AI
TL;DR: Mirror is an AI framework for ethics review that uses fine-tuned LLMs and multi-agent systems to automate expedited reviews and simulate committee deliberations for research ethics oversight.
Details
Motivation: Traditional ethics review systems are strained by large-scale interdisciplinary research, facing challenges in consistency, capacity, and handling heterogeneous risk profiles. LLMs offer potential but lack ethical reasoning capabilities, regulatory integration, and privacy safeguards for authentic review materials.Method: Developed Mirror framework with EthicsLLM (fine-tuned on EthicsQA dataset of 41K question-chain-of-thought-answer triples from ethics/regulatory sources). Two operational modes: Mirror-ER for automated expedited review using executable rule base, and Mirror-CR for full-board deliberation simulation with coordinated expert agents, ethics secretary, and PI agent across ten ethical dimensions.
Result: Empirical evaluations show Mirror significantly improves quality, consistency, and professionalism of ethics assessments compared to generalist LLMs.
Conclusion: Mirror demonstrates that AI-assisted ethical review frameworks can enhance research governance by integrating ethical reasoning with regulatory structures, addressing current limitations in institutional review capacity.
Abstract: Ethics review is a foundational mechanism of modern research governance, yet contemporary systems face increasing strain as ethical risks arise as structural consequences of large-scale, interdisciplinary scientific practice. The demand for consistent and defensible decisions under heterogeneous risk profiles exposes limitations in institutional review capacity rather than in the legitimacy of ethics oversight. Recent advances in large language models (LLMs) offer new opportunities to support ethics review, but their direct application remains limited by insufficient ethical reasoning capability, weak integration with regulatory structures, and strict privacy constraints on authentic review materials. In this work, we introduce Mirror, an agentic framework for AI-assisted ethical review that integrates ethical reasoning, structured rule interpretation, and multi-agent deliberation within a unified architecture. At its core is EthicsLLM, a foundational model fine-tuned on EthicsQA, a specialized dataset of 41K question-chain-of-thought-answer triples distilled from authoritative ethics and regulatory corpora. EthicsLLM provides detailed normative and regulatory understanding, enabling Mirror to operate in two complementary modes. Mirror-ER (expedited Review) automates expedited review through an executable rule base that supports efficient and transparent compliance checks for minimal-risk studies. Mirror-CR (Committee Review) simulates full-board deliberation through coordinated interactions among expert agents, an ethics secretary agent, and a principal investigator agent, producing structured, committee-level assessments across ten ethical dimensions. Empirical evaluations demonstrate that Mirror significantly improves the quality, consistency, and professionalism of ethics assessments compared with strong generalist LLMs.
[403] DECKBench: Benchmarking Multi-Agent Frameworks for Academic Slide Generation and Editing
Daesik Jang, Morgan Lindsay Heisler, Linzi Xing, Yifei Li, Edward Wang, Ying Xiong, Yong Zhang, Zhenan Fan
Main category: cs.AI
TL;DR: DECKBench is a benchmark for evaluating multi-agent slide generation and editing systems, focusing on academic presentation creation from papers with realistic editing instructions.
Details
Motivation: Existing benchmarks don't adequately measure the complex requirements of academic slide deck generation and editing, which needs faithful content selection, coherent organization, layout-aware rendering, and robust multi-turn instruction following.Method: Created DECKBench with curated paper-to-slide pairs augmented with simulated editing instructions, plus a modular multi-agent baseline system that decomposes tasks into paper parsing/summarization, slide planning, HTML creation, and iterative editing.
Result: The benchmark effectively highlights system strengths, exposes failure modes, and provides actionable insights for improving multi-agent slide generation and editing systems.
Conclusion: DECKBench establishes a standardized foundation for reproducible and comparable evaluation of academic presentation generation and editing systems.
Abstract: Automatically generating and iteratively editing academic slide decks requires more than document summarization. It demands faithful content selection, coherent slide organization, layout-aware rendering, and robust multi-turn instruction following. However, existing benchmarks and evaluation protocols do not adequately measure these challenges. To address this gap, we introduce the Deck Edits and Compliance Kit Benchmark (DECKBench), an evaluation framework for multi-agent slide generation and editing. DECKBench is built on a curated dataset of paper to slide pairs augmented with realistic, simulated editing instructions. Our evaluation protocol systematically assesses slide-level and deck-level fidelity, coherence, layout quality, and multi-turn instruction following. We further implement a modular multi-agent baseline system that decomposes the slide generation and editing task into paper parsing and summarization, slide planning, HTML creation, and iterative editing. Experimental results demonstrate that the proposed benchmark highlights strengths, exposes failure modes, and provides actionable insights for improving multi-agent slide generation and editing systems. Overall, this work establishes a standardized foundation for reproducible and comparable evaluation of academic presentation generation and editing. Code and data are publicly available at https://github.com/morgan-heisler/DeckBench .
[404] Situation Graph Prediction: Structured Perspective Inference for User Modeling
Jisung Shin, Daniel Platnick, Marjan Alirezaie, Hossein Rahnama
Main category: cs.AI
TL;DR: Situation Graph Prediction (SGP) frames perspective modeling as inverse inference to reconstruct structured representations of internal states from multimodal artifacts, using synthetic data generation to overcome privacy and labeling challenges.
Details
Motivation: Current AI lacks ability to model evolving internal states (goals, emotions, contexts), not just preferences. Progress is limited by privacy-sensitive data and lack of labeled perspective states in digital footprints.Method: Proposes SGP task that treats perspective modeling as inverse inference problem. Uses structure-first synthetic generation strategy that aligns latent labels and observable traces by design. Constructs dataset and runs diagnostic study using retrieval-augmented in-context learning as proxy for supervision with GPT-4o.
Result: Observes gap between surface-level extraction and latent perspective inference, indicating latent-state inference is harder than surface extraction under controlled setting. Results suggest SGP is non-trivial and provide evidence for structure-first data synthesis strategy.
Conclusion: SGP offers promising approach to perspective-aware AI by framing perspective modeling as inverse inference problem, with synthetic data generation strategy showing potential to overcome data bottlenecks in modeling internal states from multimodal artifacts.
Abstract: Perspective-Aware AI requires modeling evolving internal states–goals, emotions, contexts–not merely preferences. Progress is limited by a data bottleneck: digital footprints are privacy-sensitive and perspective states are rarely labeled. We propose Situation Graph Prediction (SGP), a task that frames perspective modeling as an inverse inference problem: reconstructing structured, ontology-aligned representations of perspective from observable multimodal artifacts. To enable grounding without real labels, we use a structure-first synthetic generation strategy that aligns latent labels and observable traces by design. As a pilot, we construct a dataset and run a diagnostic study using retrieval-augmented in-context learning as a proxy for supervision. In our study with GPT-4o, we observe a gap between surface-level extraction and latent perspective inference–indicating latent-state inference is harder than surface extraction under our controlled setting. Results suggest SGP is non-trivial and provide evidence for the structure-first data synthesis strategy.
[405] Information Fidelity in Tool-Using LLM Agents: A Martingale Analysis of the Model Context Protocol
Flint Xiaofeng Fan, Cheston Tan, Roger Wattenhofer, Yew-Soon Ong
Main category: cs.AI
TL;DR: Theoretical framework for analyzing error accumulation in LLM-powered agents using external tools, showing linear error growth with √T concentration bounds, validated across multiple LLMs.
Details
Motivation: As AI agents increasingly use external tools for high-stakes decisions, understanding how errors propagate across sequential tool calls is critical for reliability and trustworthiness.Method: Developed theoretical framework for error accumulation in MCP agents, proving linear growth and O(√T) concentration bounds. Created hybrid distortion metric combining discrete fact matching with continuous semantic similarity, and established martingale concentration bounds for sequential tool interactions.
Result: Experiments across Qwen2-7B, Llama-3-8B, and Mistral-7B validate theoretical predictions: empirical distortion tracks linear trend with deviations within O(√T) bounds. Semantic weighting reduces distortion by 80%, and periodic re-grounding every ~9 steps suffices for error control.
Conclusion: The concentration guarantees enable predictable system behavior and rule out exponential failure modes, providing actionable deployment principles for trustworthy agent systems using external tools.
Abstract: As AI agents powered by large language models (LLMs) increasingly use external tools for high-stakes decisions, a critical reliability question arises: how do errors propagate across sequential tool calls? We introduce the first theoretical framework for analyzing error accumulation in Model Context Protocol (MCP) agents, proving that cumulative distortion exhibits linear growth and high-probability deviations bounded by $O(\sqrt{T})$. This concentration property ensures predictable system behavior and rules out exponential failure modes. We develop a hybrid distortion metric combining discrete fact matching with continuous semantic similarity, then establish martingale concentration bounds on error propagation through sequential tool interactions. Experiments across Qwen2-7B, Llama-3-8B, and Mistral-7B validate our theoretical predictions, showing empirical distortion tracks the linear trend with deviations consistently within $O(\sqrt{T})$ envelopes. Key findings include: semantic weighting reduces distortion by 80%, and periodic re-grounding approximately every 9 steps suffices for error control. We translate these concentration guarantees into actionable deployment principles for trustworthy agent systems.
[406] Detecting Jailbreak Attempts in Clinical Training LLMs Through Automated Linguistic Feature Extraction
Tri Nguyen, Huy Hoang Bao Le, Lohith Srikanth Pentapalli, Laurah Turner, Kelly Cohen
Main category: cs.AI
TL;DR: Using BERT-based models to extract linguistic features for jailbreak detection in clinical LLMs, achieving strong performance with interpretable approach.
Details
Motivation: Prior work on clinical LLM jailbreak detection relied on manually annotated linguistic features, limiting scalability and expressiveness. Need automated, scalable approach for safety-critical clinical dialogue systems.Method: Train BERT-based models (general and medical domain) to predict four expert-annotated linguistic features (Professionalism, Medical Relevance, Ethical Behavior, Contextual Distraction), then use these as features in second-layer classifiers (tree-based, linear, probabilistic, ensemble) for jailbreak detection.
Result: System achieves strong overall performance in cross-validation and held-out evaluations. LLM-derived linguistic features provide effective basis for automated jailbreak detection.
Conclusion: Demonstrates scalable, interpretable approach for jailbreak detection in clinical dialogue systems. Identifies limitations in current annotations and feature representations, pointing to future improvements.
Abstract: Detecting jailbreak attempts in clinical training large language models (LLMs) requires accurate modeling of linguistic deviations that signal unsafe or off-task user behavior. Prior work on the 2-Sigma clinical simulation platform showed that manually annotated linguistic features could support jailbreak detection. However, reliance on manual annotation limited both scalability and expressiveness. In this study, we extend this framework by using experts’ annotations of four core linguistic features (Professionalism, Medical Relevance, Ethical Behavior, and Contextual Distraction) and training multiple general-domain and medical-domain BERT-based LLM models to predict these features directly from text. The most reliable feature regressor for each dimension was selected and used as the feature extractor in a second layer of classifiers. We evaluate a suite of predictive models, including tree-based, linear, probabilistic, and ensemble methods, to determine jailbreak likelihood from the extracted features. Across cross-validation and held-out evaluations, the system achieves strong overall performance, indicating that LLM-derived linguistic features provide an effective basis for automated jailbreak detection. Error analysis further highlights key limitations in current annotations and feature representations, pointing toward future improvements such as richer annotation schemes, finer-grained feature extraction, and methods that capture the evolving risk of jailbreak behavior over the course of a dialogue. This work demonstrates a scalable and interpretable approach for detecting jailbreak behavior in safety-critical clinical dialogue systems.
[407] Contrastive explanations of BDI agents
Michael Winikoff
Main category: cs.AI
TL;DR: Extends BDI agent explanation framework to answer contrastive questions (why X instead of Y), showing shorter explanations and some evidence of improved trust and understanding, but surprisingly found limited benefit of providing full explanations.
Details
Motivation: Autonomous systems need to provide explanations for transparency and trust development. While prior work enabled BDI agents to answer "why X?" questions, people naturally ask contrastive questions ("why X instead of Y?"), requiring extension of existing explanation mechanisms.Method: Extended previous BDI agent explanation framework to handle contrastive questions. Conducted computational evaluation measuring explanation length, and human subject evaluation assessing preference, trust development, transparency, and understanding.
Result: Contrastive questions yield significantly shorter explanations. Some evidence that contrastive answers are preferred and lead to higher trust, perceived understanding, and confidence in system correctness. Surprisingly, providing full explanations sometimes performed worse than no explanation.
Conclusion: Contrastive explanations are more efficient and potentially more effective for trust development, but the value of providing explanations at all is context-dependent and may not always be beneficial.
Abstract: The ability of autonomous systems to provide explanations is important for supporting transparency and aiding the development of (appropriate) trust. Prior work has defined a mechanism for Belief-Desire-Intention (BDI) agents to be able to answer questions of the form why did you do action $X$?''. However, we know that we ask \emph{contrastive} questions (why did you do $X$ \emph{instead of} $F$?’’). We therefore extend previous work to be able to answer such questions. A computational evaluation shows that using contrastive questions yields a significant reduction in explanation length. A human subject evaluation was conducted to assess whether such contrastive answers are preferred, and how well they support trust development and transparency. We found some evidence for contrastive answers being preferred, and some evidence that they led to higher trust, perceived understanding, and confidence in the system’s correctness. We also evaluated the benefit of providing explanations at all. Surprisingly, there was not a clear benefit, and in some situations we found evidence that providing a (full) explanation was worse than not providing any explanation.
[408] Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts
Chen Yang, Guangyue Peng, Jiaying Zhu, Ran Le, Ruixiang Feng, Tao Zhang, Xiyun Xu, Yang Song, Yiming Jia, Yuntao Wen, Yunzhi Xu, Zekai Wang, Zhenwei An, Zhicong Sun, Zongchao Chen
Main category: cs.AI
TL;DR: Nanbeige4.1-3B is a 3B parameter unified language model achieving strong agentic behavior, code generation, and general reasoning through innovative reward modeling and training techniques.
Details
Motivation: To create the first open-source small language model (3B parameters) that simultaneously achieves strong agentic behavior, code generation, and general reasoning capabilities, demonstrating that small models can achieve both broad competence and strong specialization.Method: Combines point-wise and pair-wise reward modeling for reasoning and preference alignment; uses complexity-aware rewards in reinforcement learning for code generation; performs complex data synthesis with turn-level supervision for deep search capabilities; enables stable long-horizon tool interactions (up to 600 tool-call turns).
Result: Significantly outperforms prior models of similar scale (Nanbeige4-3B-2511, Qwen3-4B) and even achieves superior performance compared to much larger models like Qwen3-30B-A3B; demonstrates reliable execution of up to 600 tool-call turns for complex problem-solving.
Conclusion: Small models (3B parameters) can achieve both broad competence and strong specialization simultaneously, redefining the potential of small language models through innovative training techniques.
Abstract: We present Nanbeige4.1-3B, a unified generalist language model that simultaneously achieves strong agentic behavior, code generation, and general reasoning with only 3B parameters. To the best of our knowledge, it is the first open-source small language model (SLM) to achieve such versatility in a single model. To improve reasoning and preference alignment, we combine point-wise and pair-wise reward modeling, ensuring high-quality, human-aligned responses. For code generation, we design complexity-aware rewards in Reinforcement Learning, optimizing both correctness and efficiency. In deep search, we perform complex data synthesis and incorporate turn-level supervision during training. This enables stable long-horizon tool interactions, allowing Nanbeige4.1-3B to reliably execute up to 600 tool-call turns for complex problem-solving. Extensive experimental results show that Nanbeige4.1-3B significantly outperforms prior models of similar scale, such as Nanbeige4-3B-2511 and Qwen3-4B, even achieving superior performance compared to much larger models, such as Qwen3-30B-A3B. Our results demonstrate that small models can achieve both broad competence and strong specialization simultaneously, redefining the potential of 3B parameter models.
[409] MoralityGym: A Benchmark for Evaluating Hierarchical Moral Alignment in Sequential Decision-Making Agents
Simon Rosen, Siddarth Singh, Ebenezer Gelo, Helen Sarah Robertson, Ibrahim Suder, Victoria Williams, Benjamin Rosman, Geraud Nangue Tasse, Steven James
Main category: cs.AI
TL;DR: Morality Chains formalism and MoralityGym benchmark for evaluating AI moral alignment through hierarchical norm structures in ethical dilemma environments
Details
Motivation: Addressing the critical challenge of evaluating moral alignment in AI agents when navigating conflicting, hierarchically structured human norms, which sits at the intersection of AI safety, moral philosophy, and cognitive scienceMethod: Introduces Morality Chains formalism for representing moral norms as ordered deontic constraints, and MoralityGym benchmark with 98 ethical-dilemma problems presented as trolley-dilemma-style Gymnasium environments. Decouples task-solving from moral evaluation and introduces a novel Morality Metric
Result: Baseline results with Safe RL methods reveal key limitations, highlighting the need for more principled approaches to ethical decision-making in AI systems
Conclusion: Provides a foundation for developing AI systems that behave more reliably, transparently, and ethically in complex real-world contexts by integrating insights from psychology and philosophy into norm-sensitive reasoning evaluation
Abstract: Evaluating moral alignment in agents navigating conflicting, hierarchically structured human norms is a critical challenge at the intersection of AI safety, moral philosophy, and cognitive science. We introduce Morality Chains, a novel formalism for representing moral norms as ordered deontic constraints, and MoralityGym, a benchmark of 98 ethical-dilemma problems presented as trolley-dilemma-style Gymnasium environments. By decoupling task-solving from moral evaluation and introducing a novel Morality Metric, MoralityGym allows the integration of insights from psychology and philosophy into the evaluation of norm-sensitive reasoning. Baseline results with Safe RL methods reveal key limitations, underscoring the need for more principled approaches to ethical decision-making. This work provides a foundation for developing AI systems that behave more reliably, transparently, and ethically in complex real-world contexts.
[410] On-Policy Supervised Fine-Tuning for Efficient Reasoning
Anhao Zhao, Ziyang Chen, Junlong Tong, Yingqi Fan, Fanghua Ye, Shuhao Li, Yunpu Ma, Wenjie Li, Xiaoyu Shen
Main category: cs.AI
TL;DR: Simplified training approach for large reasoning models that replaces complex RL with supervised fine-tuning on self-generated data filtered for correctness and conciseness.
Details
Motivation: Current RL-based methods for training large reasoning models are computationally expensive and complex, with multi-reward objectives that destabilize training and yield suboptimal trade-offs between correctness and brevity.Method: Remove KL regularization and group-wise normalization from RL framework, simplify reward to truncation-based length penalty, and reduce optimization to supervised fine-tuning on self-generated data filtered for both correctness and conciseness (termed “on-policy SFT”).
Result: Achieves up to 80% reduction in chain-of-thought length while maintaining original accuracy, outperforms complex RL-based methods across five benchmarks, reduces GPU memory usage by 50%, and accelerates convergence by 70%.
Conclusion: Simplified on-policy SFT approach consistently defines the accuracy-efficiency Pareto frontier, demonstrating that complex RL extensions are unnecessary for optimizing reasoning models for both correctness and conciseness.
Abstract: Large reasoning models (LRMs) are commonly trained with reinforcement learning (RL) to explore long chain-of-thought reasoning, achieving strong performance at high computational cost. Recent methods add multi-reward objectives to jointly optimize correctness and brevity, but these complex extensions often destabilize training and yield suboptimal trade-offs. We revisit this objective and challenge the necessity of such complexity. Through principled analysis, we identify fundamental misalignments in this paradigm: KL regularization loses its intended role when correctness and length are directly verifiable, and group-wise normalization becomes ambiguous under multiple reward signals. By removing these two items and simplifying the reward to a truncation-based length penalty, we show that the optimization problem reduces to supervised fine-tuning on self-generated data filtered for both correctness and conciseness. We term this simplified training strategy on-policy SFT. Despite its simplicity, on-policy SFT consistently defines the accuracy-efficiency Pareto frontier. It reduces CoT length by up to 80 while maintaining original accuracy, surpassing more complex RL-based methods across five benchmarks. Furthermore, it significantly enhances training efficiency, reducing GPU memory usage by 50% and accelerating convergence by 70%. Our code is available at https://github.com/EIT-NLP/On-Policy-SFT.
[411] NeuroWeaver: An Autonomous Evolutionary Agent for Exploring the Programmatic Space of EEG Analysis Pipelines
Guoan Wang, Shihao Yang, Jun-En Ding, Hao Zhu, Feng Liu
Main category: cs.AI
TL;DR: NeuroWeaver is an autonomous evolutionary agent for EEG analysis that uses domain-informed subspace initialization and multi-objective optimization to generate lightweight, neuroscientifically plausible pipelines that outperform task-specific methods while using far fewer parameters than foundation models.
Details
Motivation: Foundation models for EEG analysis require massive data and parameters, making them computationally prohibitive for clinical settings. General-purpose AutoML frameworks lack neurophysiological priors and produce scientifically implausible solutions. There's a need for efficient, domain-aware automated EEG analysis.Method: Reformulates EEG pipeline engineering as discrete constrained optimization. Uses Domain-Informed Subspace Initialization to restrict search to neuroscientifically plausible manifolds. Employs Multi-Objective Evolutionary Optimization with self-reflective refinement to balance performance, novelty, and efficiency.
Result: Outperforms state-of-the-art task-specific methods across five heterogeneous benchmarks. Achieves performance comparable to large-scale foundation models while using significantly fewer parameters. Synthesizes lightweight solutions suitable for resource-constrained environments.
Conclusion: NeuroWeaver bridges the gap between computationally expensive foundation models and scientifically naive AutoML by incorporating domain knowledge into automated pipeline design, enabling efficient EEG analysis in clinical settings.
Abstract: Although foundation models have demonstrated remarkable success in general domains, the application of these models to electroencephalography (EEG) analysis is constrained by substantial data requirements and high parameterization. These factors incur prohibitive computational costs, thereby impeding deployment in resource-constrained clinical environments. Conversely, general-purpose automated machine learning frameworks are often ill-suited for this domain, as exploration within an unbounded programmatic space fails to incorporate essential neurophysiological priors and frequently yields solutions that lack scientific plausibility. To address these limitations, we propose NeuroWeaver, a unified autonomous evolutionary agent designed to generalize across diverse EEG datasets and tasks by reformulating pipeline engineering as a discrete constrained optimization problem. Specifically, we employ a Domain-Informed Subspace Initialization to confine the search to neuroscientifically plausible manifolds, coupled with a Multi-Objective Evolutionary Optimization that dynamically balances performance, novelty, and efficiency via self-reflective refinement. Empirical evaluations across five heterogeneous benchmarks demonstrate that NeuroWeaver synthesizes lightweight solutions that consistently outperform state-of-the-art task-specific methods and achieve performance comparable to large-scale foundation models, despite utilizing significantly fewer parameters.
[412] OMNI-LEAK: Orchestrator Multi-Agent Network Induced Data Leakage
Akshat Naik, Jay Culligan, Yarin Gal, Philip Torr, Rahaf Aljundi, Alasdair Paren, Adel Bibi
Main category: cs.AI
TL;DR: Multi-agent LLM systems are vulnerable to novel OMNI-LEAK attacks that bypass data access controls through indirect prompt injection, compromising multiple agents simultaneously even without insider knowledge.
Details
Motivation: As multi-agent LLM systems become practical, there's insufficient threat modeling for their security vulnerabilities, especially compared to single-agent systems. Current research lacks examination of multi-agent setups with basic engineering safeguards like access control.Method: Red-teaming a concrete orchestrator setup (central agent delegating to specialized agents) to investigate security vulnerabilities. Testing frontier models with different attack categories to demonstrate OMNI-LEAK attack vector that compromises multiple agents through single indirect prompt injection.
Result: Both reasoning and non-reasoning frontier models are vulnerable to OMNI-LEAK attacks, which can leak sensitive data even with data access controls in place. Attacks succeed without attacker having insider knowledge of implementation details.
Conclusion: Safety research must generalize from single-agent to multi-agent settings to address serious risks of real-world privacy breaches, financial losses, and maintain public trust in AI agents. Multi-agent systems introduce novel attack vectors that bypass traditional safeguards.
Abstract: As Large Language Model (LLM) agents become more capable, their coordinated use in the form of multi-agent systems is anticipated to emerge as a practical paradigm. Prior work has examined the safety and misuse risks associated with agents. However, much of this has focused on the single-agent case and/or setups missing basic engineering safeguards such as access control, revealing a scarcity of threat modeling in multi-agent systems. We investigate the security vulnerabilities of a popular multi-agent pattern known as the orchestrator setup, in which a central agent decomposes and delegates tasks to specialized agents. Through red-teaming a concrete setup representative of a likely future use case, we demonstrate a novel attack vector, OMNI-LEAK, that compromises several agents to leak sensitive data through a single indirect prompt injection, even in the \textit{presence of data access control}. We report the susceptibility of frontier models to different categories of attacks, finding that both reasoning and non-reasoning models are vulnerable, even when the attacker lacks insider knowledge of the implementation details. Our work highlights the importance of safety research to generalize from single-agent to multi-agent settings, in order to reduce the serious risks of real-world privacy breaches and financial losses and overall public trust in AI agents.
[413] Translating Dietary Standards into Healthy Meals with Minimal Substitutions
Trevor Chan, Ilias Tagkopoulos
Main category: cs.AI
TL;DR: A framework that converts dietary standards into complete meals using meal archetypes from dietary data, generating nutritious meals with minimal changes while reducing costs.
Details
Motivation: To improve nutritional quality of meals without compromising convenience or affordability, addressing the challenge of translating dietary guidelines into practical, realistic meal plans.Method: Uses WWEIA intake data for 135,491 meals to identify 34 interpretable meal archetypes, then conditions a generative model and portion predictor on these archetypes to meet USDA nutritional targets with minimal food substitutions.
Result: Generated meals follow RDI targets 47.0% better while remaining compositionally close to real meals; with 1-3 food substitutions, meals become 10% more nutritious while reducing costs 19-32% on average.
Conclusion: The framework can underpin clinical decision support, public-health programs, and consumer apps for scalable, equitable improvements in everyday nutrition by turning guidelines into realistic, budget-aware meals.
Abstract: An important goal for personalized diet systems is to improve nutritional quality without compromising convenience or affordability. We present an end-to-end framework that converts dietary standards into complete meals with minimal change. Using the What We Eat in America (WWEIA) intake data for 135,491 meals, we identify 34 interpretable meal archetypes that we then use to condition a generative model and a portion predictor to meet USDA nutritional targets. In comparisons within archetypes, generated meals are better at following recommended daily intake (RDI) targets by 47.0%, while remaining compositionally close to real meals. Our results show that by allowing one to three food substitutions, we were able to create meals that were 10% more nutritious, while reducing costs 19-32%, on average. By turning dietary guidelines into realistic, budget-aware meals and simple swaps, this framework can underpin clinical decision support, public-health programs, and consumer apps that deliver scalable, equitable improvements in everyday nutrition.
[414] SPILLage: Agentic Oversharing on the Web
Jaechul Roh, Eugene Bagdasarian, Hamed Haddadi, Ali Shahin Shamsabadi
Main category: cs.AI
TL;DR: SPILLage framework formalizes natural agentic oversharing - unintentional disclosure of task-irrelevant user information through web agent actions, revealing behavioral oversharing (clicks, navigation) dominates content oversharing by 5x.
Details
Motivation: As LLM-powered web agents automate user tasks across the open web with access to sensitive resources, there's a need to understand how they handle user information when interacting with third parties, particularly focusing on unintentional disclosure beyond just text leakage.Method: Introduced SPILLage framework characterizing oversharing along channel (content vs. behavior) and directness (explicit vs. implicit) dimensions. Benchmarked 180 tasks on live e-commerce sites with ground-truth annotations, conducted 1,080 runs across two agentic frameworks and three backbone LLMs, and tested prompt-level mitigation strategies.
Result: Oversharing is pervasive with behavioral oversharing dominating content oversharing by 5x. The effect persists or worsens under prompt-level mitigation, but removing task-irrelevant information before execution improves task success by up to 17.9%.
Conclusion: Protecting privacy in web agents requires a broader view of “output” that accounts for what agents do on the web, not just what they type, as behavioral oversharing through actions represents a critical blind spot in current privacy considerations.
Abstract: LLM-powered agents are beginning to automate user’s tasks across the open web, often with access to user resources such as emails and calendars. Unlike standard LLMs answering questions in a controlled ChatBot setting, web agents act “in the wild”, interacting with third parties and leaving behind an action trace. Therefore, we ask the question: how do web agents handle user resources when accomplishing tasks on their behalf across live websites? In this paper, we formalize Natural Agentic Oversharing – the unintentional disclosure of task-irrelevant user information through an agent trace of actions on the web. We introduce SPILLage, a framework that characterizes oversharing along two dimensions: channel (content vs. behavior) and directness (explicit vs. implicit). This taxonomy reveals a critical blind spot: while prior work focuses on text leakage, web agents also overshare behaviorally through clicks, scrolls, and navigation patterns that can be monitored. We benchmark 180 tasks on live e-commerce sites with ground-truth annotations separating task-relevant from task-irrelevant attributes. Across 1,080 runs spanning two agentic frameworks and three backbone LLMs, we demonstrate that oversharing is pervasive with behavioral oversharing dominates content oversharing by 5x. This effect persists – and can even worsen – under prompt-level mitigation. However, removing task-irrelevant information before execution improves task success by up to 17.9%, demonstrating that reducing oversharing improves task success. Our findings underscore that protecting privacy in web agents is a fundamental challenge, requiring a broader view of “output” that accounts for what agents do on the web, not just what they type. Our datasets and code are available at https://github.com/jrohsc/SPILLage.
[415] REMem: Reasoning with Episodic Memory in Language Agent
Yiheng Shu, Saisri Padmaja Jonnalagedda, Xiang Gao, Bernal Jiménez Gutiérrez, Weijian Qi, Kamalika Das, Huan Sun, Yu Su
Main category: cs.AI
TL;DR: REMem is a two-phase framework for episodic memory in language agents that constructs hybrid memory graphs from experiences and enables agentic retrieval for complex reasoning over interaction histories.
Details
Motivation: Current language agents lack effective episodic memory capabilities - they mainly use semantic memory and cannot properly recollect and reason over interaction histories like humans do. Existing approaches overlook episodicity, lack explicit event modeling, or focus too much on simple retrieval rather than complex reasoning.Method: Two-phase framework: 1) Offline indexing converts experiences into a hybrid memory graph linking time-aware gists and facts, 2) Online inference uses an agentic retriever with curated tools for iterative retrieval over the memory graph.
Result: REMem outperforms state-of-the-art memory systems (Mem0 and HippoRAG 2) with 3.4% and 13.4% absolute improvements on episodic recollection and reasoning tasks respectively. Also demonstrates more robust refusal behavior for unanswerable questions.
Conclusion: REMem effectively addresses episodic memory challenges in language agents through hybrid memory graphs and agentic retrieval, enabling better recollection and reasoning over interaction histories.
Abstract: Humans excel at remembering concrete experiences along spatiotemporal contexts and performing reasoning across those events, i.e., the capacity for episodic memory. In contrast, memory in language agents remains mainly semantic, and current agents are not yet capable of effectively recollecting and reasoning over interaction histories. We identify and formalize the core challenges of episodic recollection and reasoning from this gap, and observe that existing work often overlooks episodicity, lacks explicit event modeling, or overemphasizes simple retrieval rather than complex reasoning. We present REMem, a two-phase framework for constructing and reasoning with episodic memory: 1) Offline indexing, where REMem converts experiences into a hybrid memory graph that flexibly links time-aware gists and facts. 2) Online inference, where REMem employs an agentic retriever with carefully curated tools for iterative retrieval over the memory graph. Comprehensive evaluation across four episodic memory benchmarks shows that REMem substantially outperforms state-of-the-art memory systems such as Mem0 and HippoRAG 2, showing 3.4% and 13.4% absolute improvements on episodic recollection and reasoning tasks, respectively. Moreover, REMem also demonstrates more robust refusal behavior for unanswerable questions.
[416] OpAgent: Operator Agent for Web Navigation
Yuyu Guo, Wenjie Yang, Siyuan Yang, Ziyang Liu, Cheng Chen, Yuan Wei, Yun Hu, Yang Huang, Guoliang Hao, Dongsheng Yuan, Jianming Wang, Xin Chen, Hang Yu, Lei Lei, Peng Di
Main category: cs.AI
TL;DR: Online RL web agent with hierarchical multi-task fine-tuning, hybrid reward mechanism, and modular operator framework achieves SOTA 71.6% success on WebArena.
Details
Motivation: Autonomous web agents face challenges with real-world website complexity and volatility. Traditional SFT and offline RL methods suffer from distributional shifts as offline trajectories don't capture stochastic state transitions and real-time feedback in unconstrained web environments.Method: Three core innovations: 1) Hierarchical multi-task fine-tuning with datasets for Planning, Acting, and Grounding to establish VLM with strong instruction-following; 2) Online RL with hybrid reward mechanism combining WebJudge for outcome assessment and Rule-based Decision Tree for progress reward; 3) Operator Agent (OpAgent) modular framework with Planner, Grounder, Reflector, and Summarizer for error recovery.
Result: RL-enhanced model achieves 38.1% success rate (pass@5) on WebArena, outperforming all monolithic baselines. OpAgent framework elevates performance to SOTA 71.6% success rate.
Conclusion: Proposed online RL web agent with hierarchical fine-tuning, hybrid rewards, and modular operator framework effectively addresses distributional shifts and achieves state-of-the-art performance on web navigation tasks.
Abstract: To fulfill user instructions, autonomous web agents must contend with the inherent complexity and volatile nature of real-world websites. Conventional paradigms predominantly rely on Supervised Fine-Tuning (SFT) or Offline Reinforcement Learning (RL) using static datasets. However, these methods suffer from severe distributional shifts, as offline trajectories fail to capture the stochastic state transitions and real-time feedback of unconstrained wide web environments. In this paper, we propose a robust Online Reinforcement Learning WebAgent, designed to optimize its policy through direct, iterative interactions with unconstrained wide websites. Our approach comprises three core innovations: 1) Hierarchical Multi-Task Fine-tuning: We curate a comprehensive mixture of datasets categorized by functional primitives – Planning, Acting, and Grounding – establishing a Vision-Language Model (VLM) with strong instruction-following capabilities for Web GUI tasks. 2) Online Agentic RL in the Wild: We develop an online interaction environment and fine-tune the VLM using a specialized RL pipeline. We introduce a Hybrid Reward Mechanism that combines a ground-truth-agnostic WebJudge for holistic outcome assessment with a Rule-based Decision Tree (RDT) for progress reward. This system effectively mitigates the credit assignment challenge in long-horizon navigation. Notably, our RL-enhanced model achieves a 38.1% success rate (pass@5) on WebArena, outperforming all existing monolithic baselines. 3) Operator Agent: We introduce a modular agentic framework, namely \textbf{OpAgent}, orchestrating a Planner, Grounder, Reflector, and Summarizer. This synergy enables robust error recovery and self-correction, elevating the agent’s performance to a new State-of-the-Art (SOTA) success rate of \textbf{71.6%}.
[417] Who Do LLMs Trust? Human Experts Matter More Than Other LLMs
Anooshka Bajaj, Zoran Tiganj
Main category: cs.AI
TL;DR: LLMs show human-expert bias in social influence, conforming more to human experts than other LLMs across decision-making tasks.
Details
Motivation: To investigate whether LLMs exhibit human-like patterns of social influence, particularly whether they privilege feedback from humans over other LLMs, and whether they show sensitivity to source credibility and consensus strength.Method: Three binary decision-making tasks (reading comprehension, multi-step reasoning, moral judgment) with four instruction-tuned LLMs. Presented prior responses attributed to friends, human experts, or other LLMs, manipulating correctness and group size. Second experiment introduced direct disagreement between single human and single LLM.
Result: LLMs conform significantly more to responses labeled as coming from human experts, even when incorrect, and revise answers toward experts more readily than toward other LLMs. Expert framing acts as a strong prior that generalizes across decision domains.
Conclusion: LLMs exhibit credibility-sensitive social influence with human-expert bias, suggesting they internalize social priors about source credibility similar to humans.
Abstract: Large language models (LLMs) increasingly operate in environments where they encounter social information such as other agents’ answers, tool outputs, or human recommendations. In humans, such inputs influence judgments in ways that depend on the source’s credibility and the strength of consensus. This paper investigates whether LLMs exhibit analogous patterns of influence and whether they privilege feedback from humans over feedback from other LLMs. Across three binary decision-making tasks, reading comprehension, multi-step reasoning, and moral judgment, we present four instruction-tuned LLMs with prior responses attributed either to friends, to human experts, or to other LLMs. We manipulate whether the group is correct and vary the group size. In a second experiment, we introduce direct disagreement between a single human and a single LLM. Across tasks, models conform significantly more to responses labeled as coming from human experts, including when that signal is incorrect, and revise their answers toward experts more readily than toward other LLMs. These results reveal that expert framing acts as a strong prior for contemporary LLMs, suggesting a form of credibility-sensitive social influence that generalizes across decision domains.
[418] Differentiable Rule Induction from Raw Sequence Inputs
Kun Gao, Katsumi Inoue, Yongzhi Cao, Hanpin Wang, Feng Yang
Main category: cs.AI
TL;DR: A novel differentiable ILP framework that integrates self-supervised clustering to learn interpretable rules directly from raw data without explicit label leakage, demonstrated on time series and image data.
Details
Motivation: Current differentiable ILP models struggle with learning directly from raw data due to explicit label leakage - they require explicit supervision of input feature labels to map continuous inputs to symbolic variables, limiting their applicability to raw multimodal data.Method: Integrates a self-supervised differentiable clustering model with a novel differentiable ILP model, enabling rule learning from raw data without explicit label leakage by automatically discovering symbolic representations from continuous inputs.
Result: The method successfully learns generalized rules from time series and image data, demonstrating intuitive and precise rule extraction from raw multimodal inputs without requiring explicit feature labeling.
Conclusion: The proposed integration of self-supervised clustering with differentiable ILP enables interpretable rule learning directly from raw data, overcoming the explicit label leakage problem and expanding ILP’s applicability to multimodal domains.
Abstract: Rule learning-based models are widely used in highly interpretable scenarios due to their transparent structures. Inductive logic programming (ILP), a form of machine learning, induces rules from facts while maintaining interpretability. Differentiable ILP models enhance this process by leveraging neural networks to improve robustness and scalability. However, most differentiable ILP methods rely on symbolic datasets, facing challenges when learning directly from raw data. Specifically, they struggle with explicit label leakage: The inability to map continuous inputs to symbolic variables without explicit supervision of input feature labels. In this work, we address this issue by integrating a self-supervised differentiable clustering model with a novel differentiable ILP model, enabling rule learning from raw data without explicit label leakage. The learned rules effectively describe raw data through its features. We demonstrate that our method intuitively and precisely learns generalized rules from time series and image data.
[419] Hippocampus: An Efficient and Scalable Memory Module for Agentic AI
Yi Li, Lianjie Cao, Faraz Ahmed, Puneet Sharma, Bingzhe Li
Main category: cs.AI
TL;DR: Hippocampus is an agentic memory management system using binary signatures for semantic search and token-ID streams for content reconstruction, with Dynamic Wavelet Matrix compression for ultra-fast search.
Details
Motivation: Existing memory systems for AI agents use dense vector databases or knowledge graphs, which suffer from high retrieval latency and poor storage scalability, limiting their effectiveness for long-horizon agentic deployments.Method: Uses compact binary signatures for semantic search and lossless token-ID streams for exact content reconstruction. Core innovation is Dynamic Wavelet Matrix (DWM) that compresses and co-indexes both streams to support ultra-fast search in compressed domain.
Result: Reduces end-to-end retrieval latency by up to 31×, cuts per-query token footprint by up to 14×, while maintaining accuracy on LoCoMo and LongMemEval benchmarks.
Conclusion: Hippocampus provides a scalable, efficient memory management system for agentic AI that overcomes limitations of existing approaches with linear scaling and significant performance improvements.
Abstract: Agentic AI require persistent memory to store user-specific histories beyond the limited context window of LLMs. Existing memory systems use dense vector databases or knowledge-graph traversal (or hybrid), incurring high retrieval latency and poor storage scalability. We introduce Hippocampus, an agentic memory management system that uses compact binary signatures for semantic search and lossless token-ID streams for exact content reconstruction. Its core is a Dynamic Wavelet Matrix (DWM) that compresses and co-indexes both streams to support ultra-fast search in the compressed domain, thus avoiding costly dense-vector or graph computations. This design scales linearly with memory size, making it suitable for long-horizon agentic deployments. Empirically, our evaluation shows that Hippocampus reduces end-to-end retrieval latency by up to 31$\times$ and cuts per-query token footprint by up to 14$\times$, while maintaining accuracy on both LoCoMo and LongMemEval benchmarks.
[420] The Quantization Trap: Breaking Linear Scaling Laws in Multi-Hop Reasoning
Henry Han, Xiyang Liu, Xiaodong Wang, Fei Han, Xiaodong Li
Main category: cs.AI
TL;DR: Quantization scaling laws break for multi-hop reasoning tasks due to hardware casting overhead and dequantization latency, creating a ‘quantization trap’ where lower precision (8/4-bit) increases energy consumption while degrading accuracy.
Details
Motivation: The paper challenges the conventional neural scaling law that predicts linear computational efficiency improvements with reduced numerical precision. The authors investigate whether this scaling law holds for complex reasoning tasks, particularly multi-hop reasoning.Method: The authors demonstrate through empirical evidence and theoretical decomposition that reducing precision from 16-bit to 8/4-bit creates a ‘quantization trap’. They analyze hardware casting overhead and hidden latency costs of dequantization kernels, showing these become dominant bottlenecks in sequential reasoning chains.
Result: The study reveals that quantization paradoxically increases net energy consumption while degrading reasoning accuracy for multi-hop reasoning tasks. The scaling law breaking is shown to be unavoidable in practice due to sequential energy amortization failure.
Conclusion: The industry’s “smaller-is-better” heuristic for quantization is mathematically counterproductive for complex reasoning tasks. The findings suggest that different optimization strategies are needed for reasoning-intensive AI applications compared to standard neural network inference.
Abstract: Neural scaling laws provide a predictable recipe for AI advancement: reducing numerical precision should linearly improve computational efficiency and energy profile (E proportional to bits). In this paper, we demonstrate that this scaling law breaks in the context of multi-hop reasoning. We reveal a ‘quantization trap’ where reducing precision from 16-bit to 8/4-bit paradoxically increases more net energy consumption while degrading reasoning accuracy. We provide a rigorous theoretical decomposition that attributes this failure to hardware casting overhead, the hidden latency cost of dequantization kernels, which becomes a dominant bottleneck in sequential reasoning chains, as well as to a sequential energy amortization failure. As a result, scaling law breaking is unavoidable in practice. Our findings suggest that the industry’s “smaller-is-better” heuristic is mathematically counterproductive for complex reasoning tasks.
[421] DiffusionRollout: Uncertainty-Aware Rollout Planning in Long-Horizon PDE Solving
Seungwoo Yoo, Juil Koo, Daehyeon Choi, Minhyuk Sung
Main category: cs.AI
TL;DR: DiffusionRollout: Adaptive step-size selection for autoregressive diffusion models in PDE prediction using uncertainty quantification to reduce error accumulation in long-horizon forecasts.
Details
Motivation: Autoregressive diffusion models for PDE prediction suffer from error accumulation in long-horizon predictions due to compounding errors when conditioning on inaccurate prior outputs. There's a need for better uncertainty quantification and adaptive planning strategies to improve long-term prediction reliability.Method: Proposes DiffusionRollout, a selective rollout planning strategy that uses the model’s predictive uncertainty (measured via standard deviations over multiple samples) to adaptively select step sizes during autoregressive rollouts. The method leverages the correlation between prediction errors and uncertainty measures to make more reliable long-term predictions.
Result: Extensive evaluation on long-trajectory PDE prediction benchmarks shows the proposed uncertainty measure effectively correlates with prediction errors, and the adaptive planning strategy reduces prediction errors and enables longer predicted trajectories that maintain high correlation with ground truths.
Conclusion: DiffusionRollout successfully mitigates error accumulation in long-horizon PDE predictions by using uncertainty quantification to guide adaptive step-size selection, improving the reliability and accuracy of autoregressive diffusion models for physical system forecasting.
Abstract: We propose DiffusionRollout, a novel selective rollout planning strategy for autoregressive diffusion models, aimed at mitigating error accumulation in long-horizon predictions of physical systems governed by partial differential equations (PDEs). Building on the recently validated probabilistic approach to PDE solving, we further explore its ability to quantify predictive uncertainty and demonstrate a strong correlation between prediction errors and standard deviations computed over multiple samples-supporting their use as a proxy for the model’s predictive confidence. Based on this observation, we introduce a mechanism that adaptively selects step sizes during autoregressive rollouts, improving long-term prediction reliability by reducing the compounding effect of conditioning on inaccurate prior outputs. Extensive evaluation on long-trajectory PDE prediction benchmarks validates the effectiveness of the proposed uncertainty measure and adaptive planning strategy, as evidenced by lower prediction errors and longer predicted trajectories that retain a high correlation with their ground truths.
[422] Building Autonomous GUI Navigation via Agentic-Q Estimation and Step-Wise Policy Optimization
Yibo Wang, Guangda Huzhang, Yuwei Hu, Yu Xia, Shiyin Lu, Qing-Guo Chen, Zhao Xu, Weihua Luo, Kaifu Zhang, Lijun Zhang
Main category: cs.AI
TL;DR: A novel MLLM-centered framework for GUI agents with agentic-Q estimation and step-wise policy optimization to handle non-stationary environments efficiently.
Details
Motivation: GUI agents face non-stationary environments in real-world applications, leading to high computational costs for data curation and policy optimization. Current MLLM approaches need more efficient frameworks for stable GUI interaction.Method: Two-component framework: 1) Agentic-Q estimation - optimizes a Q-model to generate step-wise values evaluating action contributions to task completion; 2) Step-wise policy optimization - takes step-wise samples from state-action trajectories and optimizes policy via reinforcement learning with the agentic-Q model. The framework uses self-generated trajectories and decouples policy updates from the environment.
Result: The framework endows Ovis2.5-9B with powerful GUI interaction capabilities, achieving remarkable performances on GUI navigation and grounding benchmarks, surpassing larger-scale contenders.
Conclusion: The proposed MLLM-centered framework provides an efficient solution for GUI agents in non-stationary environments, enabling stable optimization and reducing data collection costs while achieving state-of-the-art performance.
Abstract: Recent advances in Multimodal Large Language Models (MLLMs) have substantially driven the progress of autonomous agents for Graphical User Interface (GUI). Nevertheless, in real-world applications, GUI agents are often faced with non-stationary environments, leading to high computational costs for data curation and policy optimization. In this report, we introduce a novel MLLM-centered framework for GUI agents, which consists of two components: agentic-Q estimation and step-wise policy optimization. The former one aims to optimize a Q-model that can generate step-wise values to evaluate the contribution of a given action to task completion. The latter one takes step-wise samples from the state-action trajectory as inputs, and optimizes the policy via reinforcement learning with our agentic-Q model. It should be noticed that (i) all state-action trajectories are produced by the policy itself, so that the data collection costs are manageable; (ii) the policy update is decoupled from the environment, ensuring stable and efficient optimization. Empirical evaluations show that our framework endows Ovis2.5-9B with powerful GUI interaction capabilities, achieving remarkable performances on GUI navigation and grounding benchmarks and even surpassing contenders with larger scales.
[423] Learning Decentralized LLM Collaboration with Multi-Agent Actor Critic
Shuo Liu, Tianle Chen, Ryan Amiri, Christopher Amato
Main category: cs.AI
TL;DR: Multi-Agent Actor-Critic methods for decentralized LLM collaboration outperform Monte Carlo approaches on long-horizon/sparse-reward tasks.
Details
Motivation: Current MARL fine-tuning for LLM collaboration relies on centralized execution protocols and Monte Carlo methods with high variance, limiting practical decentralized deployment and training efficiency.Method: Proposed two Multi-Agent Actor-Critic (MAAC) approaches: CoLLM-CC with centralized critic and CoLLM-DC with decentralized critics, comparing them against Monte Carlo methods across writing, coding, and game-playing domains.
Result: Monte Carlo and CoLLM-DC perform comparably to CoLLM-CC on short-horizon/dense-reward tasks, but underperform on long-horizon/sparse-reward tasks where Monte Carlo requires more samples and CoLLM-DC struggles to converge.
Conclusion: MAAC methods, particularly centralized critic approaches, are beneficial for optimizing decentralized LLM collaboration in challenging scenarios with long horizons or sparse rewards.
Abstract: Recent work has explored optimizing LLM collaboration through Multi-Agent Reinforcement Learning (MARL). However, most MARL fine-tuning approaches rely on predefined execution protocols, which often require centralized execution. Decentralized LLM collaboration is more appealing in practice, as agents can run inference in parallel with flexible deployments. Also, current approaches use Monte Carlo methods for fine-tuning, which suffer from high variance and thus require more samples to train effectively. Actor-critic methods are prevalent in MARL for dealing with these issues, so we developed Multi-Agent Actor-Critic (MAAC) methods to optimize decentralized LLM collaboration. In this paper, we analyze when and why these MAAC methods are beneficial. We propose 2 MAAC approaches, \textbf{CoLLM-CC} with a \textbf{C}entralized \textbf{C}ritic and \textbf{CoLLM-DC} with \textbf{D}ecentralized \textbf{C}ritics. Our experiments across writing, coding, and game-playing domains show that Monte Carlo methods and CoLLM-DC can achieve performance comparable to CoLLM-CC in short-horizon and dense-reward settings. However, they both underperform CoLLM-CC on long-horizon or sparse-reward tasks, where Monte Carlo methods require substantially more samples and CoLLM-DC struggles to converge. Our code is available at https://github.com/OpenMLRL/CoMLRL/releases/tag/v1.3.6.
[424] HyFunc: Accelerating LLM-based Function Calls for Agentic AI through Hybrid-Model Cascade and Dynamic Templating
Weibin Liao, Jian-guang Lou, Haoyi Xiong
Main category: cs.AI
TL;DR: HyFunc: A hybrid-model cascade framework that reduces LLM inference latency for function calling by eliminating redundant processing of function descriptions, full-sequence generation, and boilerplate syntax.
Details
Motivation: Current agentic AI systems using LLMs for function calling suffer from computational redundancies that cause high inference latency, hindering real-time applications. Three key redundancies identified: (1) redundant processing of function descriptions for every request, (2) redundant use of large models for predictable token sequences, and (3) redundant generation of fixed parameter syntax.Method: HyFunc employs a hybrid-model cascade: a large model distills user intent into a single “soft token” that guides a lightweight retriever to select relevant functions and directs a smaller, prefix-tuned model to generate the final call. Uses “dynamic templating” to inject boilerplate parameter syntax on-the-fly within an extended vLLM engine.
Result: Achieves inference latency of 0.828 seconds (outperforming all baselines) and performance of 80.1% (surpassing comparable-scale models) on unseen BFCL benchmark dataset. Demonstrates excellent balance between efficiency and performance.
Conclusion: HyFunc offers a more efficient paradigm for agentic AI by systematically eliminating computational redundancies in LLM-based function calling while maintaining strong performance.
Abstract: While agentic AI systems rely on LLMs to translate user intent into structured function calls, this process is fraught with computational redundancy, leading to high inference latency that hinders real-time applications. This paper identifies and addresses three key redundancies: (1) the redundant processing of a large library of function descriptions for every request; (2) the redundant use of a large, slow model to generate an entire, often predictable, token sequence; and (3) the redundant generation of fixed, boilerplate parameter syntax. We introduce HyFunc, a novel framework that systematically eliminates these inefficiencies. HyFunc employs a hybrid-model cascade where a large model distills user intent into a single “soft token.” This token guides a lightweight retriever to select relevant functions and directs a smaller, prefix-tuned model to generate the final call, thus avoiding redundant context processing and full-sequence generation by the large model. To eliminate syntactic redundancy, our “dynamic templating” technique injects boilerplate parameter syntax on-the-fly within an extended vLLM engine. To avoid potential limitations in generalization, we evaluate HyFunc on an unseen benchmark dataset, BFCL. Experimental results demonstrate that HyFunc achieves an excellent balance between efficiency and performance. It achieves an inference latency of 0.828 seconds, outperforming all baseline models, and reaches a performance of 80.1%, surpassing all models with a comparable parameter scale. These results suggest that HyFunc offers a more efficient paradigm for agentic AI. Our code is publicly available at https://github.com/MrBlankness/HyFunc.
[425] Persuasion Propagation in LLM Agents
Hyejun Jeong, Amir Houmansadr, Shlomo Zilberstein, Eugene Bagdasarian
Main category: cs.AI
TL;DR: Study shows persuasion affects AI agent behavior differently depending on when it’s applied - during task execution has weak effects, but belief pre-filling before tasks significantly reduces search activity.
Details
Motivation: As AI agents increasingly combine conversation with autonomous task execution (coding, web research), researchers want to understand how user persuasion affects downstream task behavior, especially in long-horizon tasks.Method: Introduced behavior-centered evaluation framework distinguishing between persuasion during vs. prior to task execution. Tested across web research and coding tasks with belief-level interventions, comparing on-the-fly persuasion vs. belief-prefilled agents.
Result: On-the-fly persuasion induced weak/inconsistent behavioral effects. Belief-prefilled agents conducted 26.9% fewer searches and visited 16.9% fewer unique sources than neutral-prefilled agents.
Conclusion: Persuasion can significantly affect agent behavior, especially when beliefs are established before task execution, motivating behavior-level evaluation in agentic systems.
Abstract: Modern AI agents increasingly combine conversational interaction with autonomous task execution, such as coding and web research, raising a natural question: what happens when an agent engaged in long-horizon tasks is subjected to user persuasion? We study how belief-level intervention can influence downstream task behavior, a phenomenon we name \emph{persuasion propagation}. We introduce a behavior-centered evaluation framework that distinguishes between persuasion applied during or prior to task execution. Across web research and coding tasks, we find that on-the-fly persuasion induces weak and inconsistent behavioral effects. In contrast, when the belief state is explicitly specified at task time, belief-prefilled agents conduct on average 26.9% fewer searches and visit 16.9% fewer unique sources than neutral-prefilled agents. These results suggest that persuasion, even in prior interaction, can affect the agent’s behavior, motivating behavior-level evaluation in agentic systems.
[426] AllMem: A Memory-centric Recipe for Efficient Long-context Modeling
Ziming Wang, Xiang Wang, Kailong Peng, Lang Qin, Juan Gabriel Kostelec, Christos Sourmpis, Axel Laborieux, Qinghai Guo
Main category: cs.AI
TL;DR: AllMem: Hybrid architecture combining sliding window attention with non-linear test-time training memory networks to enable LLMs to handle ultra-long contexts efficiently while mitigating catastrophic forgetting.
Details
Motivation: LLMs face performance bottlenecks in long-sequence tasks due to computational complexity and memory overhead of self-attention. Need efficient methods to scale to ultra-long contexts without prohibitive costs.Method: Integrates Sliding Window Attention (SWA) with non-linear Test-Time Training (TTT) memory networks. Implements Memory-Efficient Fine-Tuning to replace standard attention layers with memory-augmented sliding window layers in pre-trained LLMs.
Result: 4k window model achieves near-lossless performance on 37k LongBench with only 0.83 drop vs full attention. 8k window variant outperforms full attention on InfiniteBench at 128k context, validating effectiveness in mitigating noise and maintaining robust long-range modeling.
Conclusion: AllMem enables efficient scaling to ultra-long contexts while overcoming representation constraints of linear memory models and reducing computational/memory footprint during long-sequence inference.
Abstract: Large Language Models (LLMs) encounter significant performance bottlenecks in long-sequence tasks due to the computational complexity and memory overhead inherent in the self-attention mechanism. To address these challenges, we introduce \textsc{AllMem}, a novel and efficient hybrid architecture that integrates Sliding Window Attention (SWA) with non-linear Test-Time Training (TTT) memory networks. \textsc{AllMem} enables models to effectively scale to ultra-long contexts while mitigating catastrophic forgetting. This approach not only overcomes the representation constraints typical of linear memory models but also significantly reduces the computational and memory footprint during long-sequence inference. Furthermore, we implement a Memory-Efficient Fine-Tuning strategy to replace standard attention layers in pre-trained models with memory-augmented sliding window layers. This framework facilitates the efficient transformation of any off-the-shelf pre-trained LLM into an \textsc{AllMem}-based architecture. Empirical evaluations confirm that our 4k window model achieves near-lossless performance on 37k LongBench with a marginal 0.83 drop compared to full attention. Furthermore, on InfiniteBench at a 128k context, our 8k window variant outperforms full attention, which validates the effectiveness of our parameterized memory in mitigating noise and maintaining robust long-range modeling without the prohibitive costs of global attention.
[427] ROMA: Recursive Open Meta-Agent Framework for Long-Horizon Multi-Agent Systems
Salaheddin Alzu’bi, Baran Nama, Arda Kaz, Anushri Eswaran, Weiyuan Chen, Sarvesh Khetan, Rishab Bala, Tu Vu, Sewoong Oh
Main category: cs.AI
TL;DR: ROMA is a recursive agent framework that decomposes tasks into parallel subtrees with structured aggregation to handle long-horizon reasoning, featuring modular roles and model-agnostic design.
Details
Motivation: Current agent frameworks struggle with long-horizon tasks due to brittle sequential orchestration, context window limitations, and opaque execution traces that make debugging difficult.Method: ROMA uses recursive task decomposition into dependency-aware subtask trees with parallel execution, plus structured aggregation to control context growth. It features four modular roles (Atomizer, Planner, Executor, Aggregator) and supports heterogeneous multi-agent systems. GEPA+ enables prompt optimization without fine-tuning.
Result: ROMA with GEPA+ achieves state-of-the-art performance: 9.9% accuracy improvement on SEAL-0 reasoning benchmark over Kimi-Researcher, and enables DeepSeek-V3 to match Claude Sonnet 4.5 on EQ-Bench long-form writing.
Conclusion: Recursive modular agent architectures can scale reasoning depth while maintaining interpretability, flexibility, and model-agnostic properties for long-horizon tasks.
Abstract: Current agentic frameworks underperform on long-horizon tasks. As reasoning depth increases, sequential orchestration becomes brittle, context windows impose hard limits that degrade performance, and opaque execution traces make failures difficult to localize or debug. We introduce ROMA (Recursive Open Meta-Agents), a domain-agnostic framework that addresses these limitations through recursive task decomposition and structured aggregation. ROMA decomposes goals into dependency-aware subtask trees that can be executed in parallel, while aggregation compresses and validates intermediate results to control context growth. Our framework standardizes agent construction around four modular roles –Atomizer (which decides whether a task should be decomposed), Planner, Executor, and Aggregator – which cleanly separate orchestration from model selection and enable transparent, hierarchical execution traces. This design supports heterogeneous multi-agent systems that mix models and tools according to cost, latency, and capability. To adapt ROMA to specific tasks without fine-tuning, we further introduce GEPA$+$, an improved Genetic-Pareto prompt proposer that searches over prompts within ROMA’s component hierarchy while preserving interface contracts. We show that ROMA, combined with GEPA+, delivers leading system-level performance on reasoning and long-form generation benchmarks. On SEAL-0, which evaluates reasoning over conflicting web evidence, ROMA instantiated with GLM-4.6 improves accuracy by 9.9% over Kimi-Researcher. On EQ-Bench, a long-form writing benchmark, ROMA enables DeepSeek-V3 to match the performance of leading closed-source models such as Claude Sonnet 4.5. Our results demonstrate that recursive, modular agent architectures can scale reasoning depth while remaining interpretable, flexible, and model-agnostic.
[428] PhGPO: Pheromone-Guided Policy Optimization for Long-Horizon Tool Planning
Yu Li, Guangfeng Cai, Shengtian Yang, Han Luo, Shuo Han, Xu He, Dong Li, Lei Feng
Main category: cs.AI
TL;DR: PhGPO improves LLM agent tool planning by learning reusable transition patterns from historical successful trajectories, similar to ant colony optimization pheromones.
Details
Motivation: Long-horizon multi-step tool planning suffers from combinatorial explosion, and current approaches don't effectively reuse successful tool-transition patterns from historical trajectories.Method: Proposes Pheromone-Guided Policy Optimization (PhGPO) that learns trajectory-based transition patterns (pheromone) from historical trajectories and uses this learned pheromone to guide policy optimization toward historically successful tool transitions.
Result: Comprehensive experimental results demonstrate the effectiveness of PhGPO in improving long-horizon tool planning.
Conclusion: Learning reusable transition patterns from historical trajectories provides explicit guidance that significantly improves LLM agent tool planning capabilities.
Abstract: Recent advancements in Large Language Model (LLM) agents have demonstrated strong capabilities in executing complex tasks through tool use. However, long-horizon multi-step tool planning is challenging, because the exploration space suffers from a combinatorial explosion. In this scenario, even when a correct tool-use path is found, it is usually considered an immediate reward for current training, which would not provide any reusable information for subsequent training. In this paper, we argue that historically successful trajectories contain reusable tool-transition patterns, which can be leveraged throughout the whole training process. Inspired by ant colony optimization where historically successful paths can be reflected by the pheromone, we propose Pheromone-Guided Policy Optimization (PhGPO), which learns a trajectory-based transition pattern (i.e., pheromone) from historical trajectories and then uses the learned pheromone to guide policy optimization. This learned pheromone provides explicit and reusable guidance that steers policy optimization toward historically successful tool transitions, thereby improving long-horizon tool planning. Comprehensive experimental results demonstrate the effectiveness of our proposed PhGPO.
[429] Can a Lightweight Automated AI Pipeline Solve Research-Level Mathematical Problems?
Lve Meng, Weilong Zhao, Yanzhi Zhang, Haoxiang Guan, Jiyan He
Main category: cs.AI
TL;DR: LLMs can solve research-grade math problems through automated pipelines with citation verification, achieving verified solutions for contest-level and unpublished research problems.
Details
Motivation: While LLMs excel at competition-level math benchmarks, their application to research problems via lightweight natural-language pipelines remains underexplored, especially for solving sophisticated research-grade problems.Method: Developed a streamlined automated pipeline optimized for citation-based verification using next-generation models (Gemini 3 Pro, GPT-5.2 Pro), evaluated on two novel datasets: ICCM problem sets (comparable to Yau College Contest) and “First Proof” problem set of unpublished research questions.
Result: Pipeline generated candidate proofs for all problems in both datasets; solutions for first two ICCM sets and Problem 4 of “First Proof” set were fully verified by the team; all generated proofs submitted to official organization and results made publicly available.
Conclusion: Next-generation LLMs integrated into optimized automated pipelines can successfully solve sophisticated research-grade mathematical problems, demonstrating practical deployment potential for mathematical research applications.
Abstract: Large language models (LLMs) have recently achieved remarkable success in generating rigorous mathematical proofs, with “AI for Math” emerging as a vibrant field of research. While these models have mastered competition-level benchmarks like the International Mathematical Olympiad and show promise in research applications through auto-formalization, their deployment via lightweight, natural-language pipelines for research problems remains underexplored. In this work, we demonstrate that next-generation models (e.g., Gemini 3 Pro, GPT-5.2 Pro), when integrated into a streamlined automated pipeline optimized for citation-based verification, can solve sophisticated research-grade problems. We evaluate our pipeline on two novel datasets: (1) the ICCM problem sets (comparable to the S.-T. Yau College Student Mathematics Contest) proposed by leading mathematicians, and (2) the “First Proof” problem set, consisting of previously unpublished research questions. Our pipeline generated candidate proofs for all problems in the first two ICCM sets and the “First Proof” set. The solutions for the first two ICCM sets and Problem 4 of the “First Proof” set have been fully verified by our team. All generated proofs have been submitted to the official organization, and our generated results are publicly available. We plan to open-source the complete pipeline methodology in due course.
[430] No Need to Train Your RDB Foundation Model
Linjie Xu, Yanlin Zhang, Quan Gan, Minjie Wang, David Wipf
Main category: cs.AI
TL;DR: A foundation model approach for relational databases that enables in-context learning across multiple tables without retraining, using column-wise compression and SQL primitives.
Details
Motivation: To enable predictive modeling across enterprise relational databases without retraining new models for each target, overcoming limitations of single-table in-context learning approaches.Method: Proposes a principled RDB encoder family that compresses variably-sized database neighborhoods into fixed-length samples using column-wise compression within homogeneous columns (not across heterogeneous columns), implemented via scalable SQL primitives without trainable parameters.
Result: Develops an open-source RDB foundation model capable of robust performance on unseen datasets without training or fine-tuning, seamlessly pairing with existing single-table ICL foundation models.
Conclusion: The approach enables practical in-context learning across relational databases by constraining compression within homogeneous columns and eliminating trainable parameters, providing an easy-to-use solution for enterprise predictive modeling.
Abstract: Relational databases (RDBs) contain vast amounts of heterogeneous tabular information that can be exploited for predictive modeling purposes. But since the space of potential targets is vast across enterprise settings, how can we \textit{avoid retraining} a new model each time we wish to predict a new quantity of interest? Foundation models based on in-context learning (ICL) offer a convenient option, but so far are largely restricted to single-table operability. In generalizing to multiple interrelated tables, it is essential to compress variably-sized RDB neighborhoods into fixed-length ICL samples for consumption by the decoder. However, the details here are critical: unlike existing supervised learning RDB pipelines, we provide theoretical and empirical evidence that ICL-specific compression should be constrained \emph{within} high-dimensional RDB columns where all entities share units and roles, not \textit{across} columns where the relevance of heterogeneous data types cannot possibly be determined without label information. Conditioned on this restriction, we then demonstrate that encoder expressiveness is actually not compromised by excluding trainable parameters. Hence we arrive at a principled family of RDB encoders that can be seamlessly paired with already-existing single-table ICL foundation models, whereby no training or fine-tuning is required. From a practical standpoint, we develop scalable SQL primitives to implement the encoder stage, resulting in an easy-to-use open-source RDB foundation model\footnote{\label{foot: RDBLearn_learn} https://github.com/HKUSHXLab/rdblearn} capable of robust performance on unseen datasets out of the box.
[431] OneLatent: Single-Token Compression for Visual Latent Reasoning
Bo Lv, Yasheng Sun, Junjie Wang, Haoxiang Shi
Main category: cs.AI
TL;DR: OneLatent compresses chain-of-thought reasoning into a single latent token using rendered CoT images and OCR hidden states, achieving 11x output length reduction with minimal accuracy loss.
Details
Motivation: Chain-of-thought prompting improves reasoning but significantly increases inference cost (1-2 orders of magnitude). Need to reduce computational overhead while maintaining reasoning quality.Method: Render textual reasoning steps into images, then compress intermediate reasoning into a single latent token using supervision from rendered CoT images and DeepSeek-OCR hidden states. This creates deterministic supervision signals that are inspectable without verbose textual output.
Result: 11x average output length reduction with only 2.21% accuracy drop relative to textual CoT. 6.8x improvement in output token contribution. Achieves 99.80% on ProntoQA and 97.80% on ProsQA with one latent token, with compression up to 87.4x.
Conclusion: OneLatent effectively compresses reasoning processes into latent representations while maintaining performance, enabling efficient inference for complex reasoning tasks with compression-constrained generalization.
Abstract: Chain-of-thought (CoT) prompting improves reasoning but often increases inference cost by one to two orders of magnitude. To address these challenges, we present \textbf{OneLatent}, a framework that compresses intermediate reasoning into a single latent token via supervision from rendered CoT images and DeepSeek-OCR hidden states. By rendering textual steps into images, we obtain a deterministic supervision signal that can be inspected and audited without requiring the model to output verbose textual rationales. Across benchmarks, OneLatent reduces average output length by $11\times$ with only a $2.21%$ average accuracy drop relative to textual CoT, while improving output token contribution (OTC) by $6.8\times$. On long-chain logical reasoning, OneLatent reaches $99.80%$ on ProntoQA and $97.80%$ on ProsQA with one latent token, with compression up to $87.4\times$, supporting compression-constrained generalization.
[432] OR-Agent: Bridging Evolutionary Search and Structured Research for Automated Algorithm Discovery
Qi Liu, Wanjing Ma
Main category: cs.AI
TL;DR: OR-Agent: A multi-agent research framework for automated scientific discovery with structured hypothesis management, evolutionary-systematic ideation, and hierarchical reflection mechanisms
Details
Motivation: Current automated discovery approaches rely too heavily on simple program mutation loops without proper hypothesis management, environment interaction, or principled reflection needed for complex experimental domainsMethod: Configurable multi-agent framework with tree-based workflow for branching hypothesis generation and backtracking; evolutionary-systematic ideation combining evolutionary selection with comprehensive planning; hierarchical reflection with verbal gradients, verbal momentum, and memory compression
Result: Outperforms evolutionary baselines on combinatorial optimization benchmarks (traveling salesman, vehicle routing, bin packing, etc.) and simulation-based cooperative driving scenarios
Conclusion: OR-Agent provides a general, extensible, inspectable framework for AI-assisted scientific discovery that goes beyond simple mutation-crossover loops through structured hypothesis management
Abstract: Automating scientific discovery in complex, experiment-driven domains requires more than iterative mutation of programs; it demands structured hypothesis management, environment interaction, and principled reflection. We present OR-Agent, a configurable multi-agent research framework designed for automated exploration in rich experimental environments. OR-Agent organizes research as a structured tree-based workflow that explicitly models branching hypothesis generation and systematic backtracking, enabling controlled management of research trajectories beyond simple mutation-crossover loops. At its core, we introduce an evolutionary-systematic ideation mechanism that unifies evolutionary selection of research starting points, comprehensive research plan generation, and coordinated exploration within a research tree. We further propose a hierarchical optimization-inspired reflection system: short-term experimental reflection operates as a form of verbal gradient providing immediate corrective signals; long-term reflection accumulates cross-experiment insights as verbal momentum; and memory compression serves as a regularization mechanism analogous to weight decay, preserving essential signals while mitigating drift. Together, these components form a principled architecture governing research dynamics. We conduct extensive experiments across classical combinatorial optimization benchmarks-including traveling salesman, capacitated vehicle routing, bin packing, orienteering, and multiple knapsack problems-as well as simulation-based cooperative driving scenarios. Results demonstrate that OR-Agent outperforms strong evolutionary baselines while providing a general, extensible, and inspectable framework for AI-assisted scientific discovery. OR-Agent source code and experiments data are publicly available at https://github.com/qiliuchn/OR-Agent.
[433] StackingNet: Collective Inference Across Independent AI Foundation Models
Siyang Li, Chenhao Liu, Dongrui Wu, Zhigang Zeng, Lieyun Ding
Main category: cs.AI
TL;DR: StackingNet is a meta-ensemble framework that coordinates multiple black-box foundation models without accessing their internal parameters or training data, improving accuracy, robustness and fairness across language, vision and reasoning tasks.
Details
Motivation: While large foundation models have advanced language understanding, vision and reasoning, they remain isolated and cannot easily share capabilities. Integrating complementary strengths of independent foundation models is essential for building trustworthy intelligent systems, but there's no established approach for coordinating such black-box heterogeneous models.Method: StackingNet uses a meta-ensemble framework based on collective intelligence principles to combine model predictions during inference. It operates without access to internal parameters or training data, enabling reliability ranking, identifying/pruning underperforming models, and reducing bias.
Result: Across language comprehension, visual estimation, and academic paper rating tasks, StackingNet consistently improves accuracy, robustness, and fairness compared to individual models and classic ensembles. It turns model diversity from inconsistency into productive collaboration.
Conclusion: StackingNet establishes a practical foundation for coordinated AI, suggesting progress may come from principled cooperation among specialized models rather than just larger single models. It enables trustworthy intelligent systems through collective intelligence.
Abstract: Artificial intelligence built on large foundation models has transformed language understanding, vision and reasoning, yet these systems remain isolated and cannot readily share their capabilities. Integrating the complementary strengths of such independent foundation models is essential for building trustworthy intelligent systems. Despite rapid progress in individual model design, there is no established approach for coordinating such black-box heterogeneous models. Here we show that coordination can be achieved through a meta-ensemble framework termed StackingNet, which draws on principles of collective intelligence to combine model predictions during inference. StackingNet improves accuracy, reduces bias, enables reliability ranking, and identifies or prunes models that degrade performance, all operating without access to internal parameters or training data. Across tasks involving language comprehension, visual estimation, and academic paper rating, StackingNet consistently improves accuracy, robustness, and fairness, compared with individual models and classic ensembles. By turning diversity from a source of inconsistency into collaboration, StackingNet establishes a practical foundation for coordinated artificial intelligence, suggesting that progress may emerge from not only larger single models but also principled cooperation among many specialized ones.
[434] Attention in Constant Time: Vashista Sparse Attention for Long-Context Decoding with Exponential Guarantees
Vashista Nobaub
Main category: cs.AI
TL;DR: Vashista Sparse Attention: A theory-grounded sparse attention mechanism that identifies and focuses on only relevant tokens using face-stability analysis, achieving constant-size effective support and practical speedups for long-context LLM inference.
Details
Motivation: LLMs spend most inference cost on attention over long contexts, but empirical evidence suggests only a small subset of tokens meaningfully contributes to each query. Current attention mechanisms waste computation on irrelevant tokens, creating a need for theoretically-grounded sparse attention that can safely identify and focus on only the relevant tokens.Method: 1) Formalize attention as projection onto convex hull of key vectors with entropic (softmax-like) relaxation; 2) Prove face-stability theorem showing attention concentrates on constant-size active face under strict complementarity margin; 3) Introduce Vashista Sparse Attention - a drop-in mechanism maintaining small candidate set per query through paging-style context selection compatible with modern inference stacks.
Result: Theoretical: Attention mass on inactive tokens decays exponentially while error on active face scales linearly with temperature. Practical: Across long-context evaluations, stable constant-size effective support, strong wall-clock speedups, and minimal quality degradation in regimes predicted by support-gap diagnostics.
Conclusion: The paper provides a principled theoretical foundation for sparse attention mechanisms, enabling predictable latency and cost without external retrieval dependencies. Vashista Sparse Attention offers practical deployment benefits for privacy-sensitive and air-gapped settings through interchangeable attention modules.
Abstract: Large language models spend most of their inference cost on attention over long contexts, yet empirical behavior suggests that only a small subset of tokens meaningfully contributes to each query. We formalize this phenomenon by modeling attention as a projection onto the convex hull of key vectors and analyzing its entropic (softmax-like) relaxation. Our main theoretical contribution is a face-stability theorem showing that, under a strict complementarity margin (a support gap (Δ) certified by KKT multipliers), entropic attention concentrates on a constant-size active face: the total mass assigned to inactive tokens decays exponentially as (\exp(-Ω(Δ/\varepsilon))), while the error on the active face scales linearly in the temperature/regularization parameter (\varepsilon). This yields a practical criterion for when sparse long-context decoding is safe and provides a principled knob to trade accuracy for compute. Building on these guarantees, we introduce Vashista Sparse Attention, a drop-in mechanism that maintains a small candidate set per query through a paging-style context selection strategy compatible with modern inference stacks. Across long-context evaluations, we observe stable constant-size effective support, strong wall-clock speedups, and minimal quality degradation in the regimes predicted by the support-gap diagnostics. Finally, we discuss deployment implications for privacy-sensitive and air-gapped settings, where interchangeable attention modules enable predictable latency and cost without external retrieval dependencies.
[435] An end-to-end agentic pipeline for smart contract translation and quality evaluation
Abhinav Goel, Chaitya Shah, Agostino Capponi, Alfio Gliozzo
Main category: cs.AI
TL;DR: Framework for systematic evaluation of LLM-generated smart contracts from natural language specifications using structured parsing, code generation, and multi-dimensional quality assessment.
Details
Motivation: Need for reproducible benchmarks and systematic evaluation of LLM-generated smart contracts to assess quality, identify error patterns, and support empirical research on contract synthesis.Method: End-to-end pipeline that parses contractual text into structured schemas, generates Solidity code using CrewAI-style agent teams with iterative refinement, and performs automated quality assessment through compilation checks, security analysis, and multi-dimensional evaluation metrics.
Result: Framework produces structured artifacts with provenance metadata and measures quality across five dimensions: functional completeness, variable fidelity, state-machine correctness, business-logic fidelity, and code quality, enabling paired evaluation against ground-truth implementations.
Conclusion: Provides reproducible benchmark for empirical research on smart contract synthesis quality and supports extensions to formal verification and compliance checking.
Abstract: We present an end-to-end framework for systematic evaluation of LLM-generated smart contracts from natural-language specifications. The system parses contractual text into structured schemas, generates Solidity code, and performs automated quality assessment through compilation and security checks. Using CrewAI-style agent teams with iterative refinement, the pipeline produces structured artifacts with full provenance metadata. Quality is measured across five dimensions, including functional completeness, variable fidelity, state-machine correctness, business-logic fidelity, and code quality aggregated into composite scores. The framework supports paired evaluation against ground-truth implementations, quantifying alignment and identifying systematic error modes such as logic omissions and state transition inconsistencies. This provides a reproducible benchmark for empirical research on smart contract synthesis quality and supports extensions to formal verification and compliance checking.
[436] Experimentation Accelerator: Interpretable Insights and Creative Recommendations for A/B Testing with Content-Aware ranking
Zhengmian Hu, Lei Shi, Ritwik Sinha, Justin Grover, David Arbour
Main category: cs.AI
TL;DR: A unified framework for A/B testing that uses embeddings and historical data to prioritize variants, explain results, and generate new creative suggestions through LLMs.
Details
Motivation: Online experimentation faces bottlenecks: scarce traffic forces tough choices on which variants to test, and post-hoc insight extraction is manual, inconsistent, and content-agnostic. Organizations underuse historical A/B results and rich content embeddings that could guide prioritization and creative iteration.Method: Leverages treatment embeddings and historical outcomes to train a CTR ranking model with fixed effects for contextual shifts. Projects treatments onto semantic marketing attributes and uses sign-consistent, sparse constrained Lasso for interpretability. Computes opportunity index combining attribute importance with under-expression. Uses LLMs to translate opportunities into creative suggestions and estimate learning/conversion potential.
Result: The framework has been built into Adobe’s “Experimentation Accelerator” product. Evaluation on real-world experiments by Adobe business customers validates the high quality of the generation pipeline.
Conclusion: The proposed framework enables faster, more informative, and more efficient test cycles by providing AI-based insights and opportunities to scale experimentation.
Abstract: Modern online experimentation faces two bottlenecks: scarce traffic forces tough choices on which variants to test, and post-hoc insight extraction is manual, inconsistent, and often content-agnostic. Meanwhile, organizations underuse historical A/B results and rich content embeddings that could guide prioritization and creative iteration. We present a unified framework to (i) prioritize which variants to test, (ii) explain why winners win, and (iii) surface targeted opportunities for new, higher-potential variants. Leveraging treatment embeddings and historical outcomes, we train a CTR ranking model with fixed effects for contextual shifts that scores candidates while balancing value and content diversity. For better interpretability and understanding, we project treatments onto curated semantic marketing attributes and re-express the ranker in this space via a sign-consistent, sparse constrained Lasso, yielding per-attribute coefficients and signed contributions for visual explanations, top-k drivers, and natural-language insights. We then compute an opportunity index combining attribute importance (from the ranker) with under-expression in the current experiment to flag missing, high-impact attributes. Finally, LLMs translate ranked opportunities into concrete creative suggestions and estimate both learning and conversion potential, enabling faster, more informative, and more efficient test cycles. These components have been built into a real Adobe product, called \textit{Experimentation Accelerator}, to provide AI-based insights and opportunities to scale experimentation for customers. We provide an evaluation of the performance of the proposed framework on some real-world experiments by Adobe business customers that validate the high quality of the generation pipeline.
[437] Enabling Option Learning in Sparse Rewards with Hindsight Experience Replay
Gabriel Romio, Mateus Begnini Melchiades, Bruno Castro da Silva, Gabriel de Oliveira Ramos
Main category: cs.AI
TL;DR: MOC-2HER: Hierarchical RL with dual hindsight experience replay for sparse-reward multi-goal robotic manipulation tasks
Details
Motivation: Existing hierarchical RL methods (Option-Critic, MOC) underperform in sparse-reward multi-goal environments, especially in object manipulation where rewards depend on object states rather than agent actions, making it hard to discover object interactions.Method: Two-step approach: 1) MOC-HER integrates Hindsight Experience Replay into MOC framework for goal relabeling; 2) 2HER introduces dual objective relabeling with two sets of virtual goals - one based on object’s final state (standard HER) and another based on agent’s effector positions to reward both interaction and task completion.
Result: MOC-2HER achieves up to 90% success rate in robotic manipulation environments, compared to less than 11% for both original MOC and MOC-HER, demonstrating effectiveness of dual objective relabeling strategy.
Conclusion: The dual objective hindsight experience replay (2HER) effectively addresses sparse reward challenges in hierarchical RL for multi-goal object manipulation tasks by rewarding both object interaction and task completion through separate goal relabeling mechanisms.
Abstract: Hierarchical Reinforcement Learning (HRL) frameworks like Option-Critic (OC) and Multi-updates Option Critic (MOC) have introduced significant advancements in learning reusable options. However, these methods underperform in multi-goal environments with sparse rewards, where actions must be linked to temporally distant outcomes. To address this limitation, we first propose MOC-HER, which integrates the Hindsight Experience Replay (HER) mechanism into the MOC framework. By relabeling goals from achieved outcomes, MOC-HER can solve sparse reward environments that are intractable for the original MOC. However, this approach is insufficient for object manipulation tasks, where the reward depends on the object reaching the goal rather than on the agent’s direct interaction. This makes it extremely difficult for HRL agents to discover how to interact with these objects. To overcome this issue, we introduce Dual Objectives Hindsight Experience Replay (2HER), a novel extension that creates two sets of virtual goals. In addition to relabeling goals based on the object’s final state (standard HER), 2HER also generates goals from the agent’s effector positions, rewarding the agent for both interacting with the object and completing the task. Experimental results in robotic manipulation environments show that MOC-2HER achieves success rates of up to 90%, compared to less than 11% for both MOC and MOC-HER. These results highlight the effectiveness of our dual objective relabeling strategy in sparse reward, multi-goal tasks.
[438] Ambient Physics: Training Neural PDE Solvers with Partial Observations
Harris Abdul Majid, Giannis Daras, Francesco Tudisco, Steven McDonagh
Main category: cs.AI
TL;DR: Ambient Physics is a framework for learning joint distributions of PDE coefficient-solution pairs directly from partial observations without needing complete training data, using random masking to enable reconstruction from incomplete measurements.
Details
Motivation: In scientific settings, acquiring complete observations of PDE coefficients and solutions is often expensive, hazardous, or impossible. Existing diffusion-based methods require complete observations for training, limiting their applicability in real-world scenarios with partial data.Method: The framework randomly masks a subset of already-observed measurements and supervises on them, making the model unable to distinguish between “truly unobserved” and “artificially unobserved” points, forcing it to produce plausible predictions everywhere. This enables learning from partial observations without complete training data.
Result: Ambient Physics achieves state-of-the-art reconstruction performance with a 62.51% reduction in average overall error compared to prior diffusion-based methods, while using 125× fewer function evaluations. The method also identifies a “one-point transition” phenomenon where masking a single observed point enables learning across architectures and measurement patterns.
Conclusion: Ambient Physics enables scientific progress in settings where complete observations are unavailable by allowing models to learn directly from partial observations, overcoming a key limitation of existing diffusion-based approaches.
Abstract: In many scientific settings, acquiring complete observations of PDE coefficients and solutions can be expensive, hazardous, or impossible. Recent diffusion-based methods can reconstruct fields given partial observations, but require complete observations for training. We introduce Ambient Physics, a framework for learning the joint distribution of coefficient-solution pairs directly from partial observations, without requiring a single complete observation. The key idea is to randomly mask a subset of already-observed measurements and supervise on them, so the model cannot distinguish “truly unobserved” from “artificially unobserved”, and must produce plausible predictions everywhere. Ambient Physics achieves state-of-the-art reconstruction performance. Compared with prior diffusion-based methods, it achieves a 62.51$%$ reduction in average overall error while using 125$\times$ fewer function evaluations. We also identify a “one-point transition”: masking a single already-observed point enables learning from partial observations across architectures and measurement patterns. Ambient Physics thus enables scientific progress in settings where complete observations are unavailable.
[439] VSAL: A Vision Solver with Adaptive Layouts for Graph Property Detection
Jiahao Xie, Guangmo Tong
Main category: cs.AI
TL;DR: VSAL is a vision-based framework for graph property detection that uses adaptive layout generation to create informative visualizations tailored to individual graph instances, improving detection performance over fixed-layout methods.
Details
Motivation: Existing vision-based graph property detection methods rely on fixed visual layouts, limiting their expressiveness and performance. The authors aim to overcome this by developing a framework that can dynamically generate informative visualizations optimized for each specific graph instance.Method: VSAL incorporates an adaptive layout generator that produces tailored graph visualizations. The framework processes these optimized visual representations to detect various graph properties like Hamiltonian cycles, planarity, claw-freeness, and tree structures.
Result: Extensive experiments show VSAL outperforms state-of-the-art vision-based methods on multiple graph property detection tasks including Hamiltonian cycle, planarity, claw-freeness, and tree detection.
Conclusion: Adaptive layout generation significantly improves vision-based graph property detection by providing more informative visual representations tailored to individual graph instances, overcoming limitations of fixed-layout approaches.
Abstract: Graph property detection aims to determine whether a graph exhibits certain structural properties, such as being Hamiltonian. Recently, learning-based approaches have shown great promise by leveraging data-driven models to detect graph properties efficiently. In particular, vision-based methods offer a visually intuitive solution by processing the visualizations of graphs. However, existing vision-based methods rely on fixed visual graph layouts, and therefore, the expressiveness of their pipeline is restricted. To overcome this limitation, we propose VSAL, a vision-based framework that incorporates an adaptive layout generator capable of dynamically producing informative graph visualizations tailored to individual instances, thereby improving graph property detection. Extensive experiments demonstrate that VSAL outperforms state-of-the-art vision-based methods on various tasks such as Hamiltonian cycle, planarity, claw-freeness, and tree detection.
[440] Diagnosing Pathological Chain-of-Thought in Reasoning Models
Manqing Liu, David Williams-King, Ida Caspary, Linh Le, Hannes Whittingham, Puria Radmard, Cameron Tice, Edward James Young
Main category: cs.AI
TL;DR: The paper develops metrics to detect three Chain-of-Thought reasoning pathologies in LLMs: post-hoc rationalization, encoded reasoning, and internalized reasoning, using deliberately trained model organisms for validation.
Details
Motivation: Chain-of-Thought reasoning is fundamental to modern LLMs and represents a critical intervention point for AI safety, but CoT reasoning may exhibit failure modes (pathologies) that prevent it from being useful for monitoring. Prior work identified three distinct pathologies that need better understanding and discrimination.Method: Created a set of concrete metrics that are simple to implement, computationally inexpensive, and task-agnostic to understand and discriminate between CoT pathologies. Developed model organisms deliberately trained to exhibit specific CoT pathologies to validate the approach.
Result: The work provides a practical toolkit for assessing CoT pathologies, with direct implications for training-time monitoring. The metrics successfully identify and distinguish between the three types of reasoning pathologies.
Conclusion: The paper offers a systematic approach to detecting and understanding Chain-of-Thought reasoning pathologies in LLMs, which is crucial for AI safety and reliable model monitoring during training.
Abstract: Chain-of-thought (CoT) reasoning is fundamental to modern LLM architectures and represents a critical intervention point for AI safety. However, CoT reasoning may exhibit failure modes that we note as pathologies, which prevent it from being useful for monitoring. Prior work has identified three distinct pathologies: post-hoc rationalization, where models generate plausible explanations backwards from predetermined answers; encoded reasoning, where intermediate steps conceal information within seemingly interpretable text; and internalized reasoning, where models replace explicit reasoning with meaningless filler tokens while computing internally. To better understand and discriminate between these pathologies, we create a set of concrete metrics that are simple to implement, computationally inexpensive, and task-agnostic. To validate our approach, we develop model organisms deliberately trained to exhibit specific CoT pathologies. Our work provides a practical toolkit for assessing CoT pathologies, with direct implications for training-time monitoring.
[441] From Pixels to Policies: Reinforcing Spatial Reasoning in Language Models for Content-Aware Layout Design
Sha Li, Stefano Petrangeli, Yu Shen, Xiang Chen
Main category: cs.AI
TL;DR: LaySPA is a reinforcement learning framework that enhances LLMs with explicit spatial reasoning for graphic layout design, producing interpretable reasoning traces and structured layouts.
Details
Motivation: Addresses LLMs' limited spatial reasoning capabilities and lack of transparency in design decision making for graphic layout tasks.Method: Reformulates layout design as policy learning over structured textual spatial environment encoding canvas geometry, element attributes, and relationships. Uses multi-objective spatial critique (geometric validity, relational coherence, aesthetic consistency) and relative group optimization.
Result: Improves structural validity and visual quality, outperforms larger proprietary LLMs, achieves performance comparable to specialized SOTA layout generators with fewer samples and reduced latency.
Conclusion: LaySPA successfully equips LLMs with explicit spatial reasoning for transparent and controllable graphic layout design.
Abstract: We introduce LaySPA, a reinforcement learning framework that equips large language models (LLMs) with explicit and interpretable spatial reasoning for content-aware graphic layout design. LaySPA addresses two key challenges: LLMs’ limited spatial reasoning and the lack of opacity in design decision making. Instead of operating at the pixel level, we reformulate layout design as a policy learning problem over a structured textual spatial environment that explicitly encodes canvas geometry, element attributes, and inter-element relationships. LaySPA produces dual-level outputs comprising interpretable reasoning traces and structured layout specifications, enabling transparent and controllable design decision making. Layout design policy is optimized via a multi-objective spatial critique that decomposes layout quality into geometric validity, relational coherence, and aesthetic consistency, and is trained using relative group optimization to stabilize learning in open-ended design spaces. Experiments demonstrate that LaySPA improves structural validity and visual quality, outperforming larger proprietary LLMs and achieving performance comparable to specialized SOTA layout generators while requiring fewer annotated samples and reduced latency.
[442] HyMem: Hybrid Memory Architecture with Dynamic Retrieval Scheduling
Xiaochen Zhao, Kaikai Wang, Xiaowen Zhang, Chen Yao, Aili Wang
Main category: cs.AI
TL;DR: HyMem is a hybrid memory architecture for LLM agents that uses multi-granular memory representations and dynamic scheduling to balance efficiency and effectiveness in long dialogues.
Details
Motivation: LLM agents perform well in short-text contexts but struggle with extended dialogues due to inefficient memory management. Existing approaches face a trade-off between efficiency (memory compression loses critical details) and effectiveness (retaining raw text causes computational overhead). Current monolithic memory representations and static retrieval mechanisms fail to emulate human-like flexible memory scheduling.Method: HyMem uses a hybrid memory architecture with dual-granular storage and dynamic two-tier retrieval: 1) Lightweight module constructs summary-level context for efficient response generation, 2) LLM-based deep module selectively activated only for complex queries, augmented by a reflection mechanism for iterative reasoning refinement.
Result: HyMem achieves strong performance on both LOCOMO and LongMemEval benchmarks, outperforming full-context approaches while reducing computational cost by 92.6%, establishing state-of-the-art balance between efficiency and performance.
Conclusion: HyMem demonstrates that hybrid memory architectures with dynamic scheduling can effectively address the efficiency-effectiveness trade-off in long-term memory management for LLM agents, enabling better performance in extended dialogues.
Abstract: Large language model (LLM) agents demonstrate strong performance in short-text contexts but often underperform in extended dialogues due to inefficient memory management. Existing approaches face a fundamental trade-off between efficiency and effectiveness: memory compression risks losing critical details required for complex reasoning, while retaining raw text introduces unnecessary computational overhead for simple queries. The crux lies in the limitations of monolithic memory representations and static retrieval mechanisms, which fail to emulate the flexible and proactive memory scheduling capabilities observed in humans, thus struggling to adapt to diverse problem scenarios. Inspired by the principle of cognitive economy, we propose HyMem, a hybrid memory architecture that enables dynamic on-demand scheduling through multi-granular memory representations. HyMem adopts a dual-granular storage scheme paired with a dynamic two-tier retrieval system: a lightweight module constructs summary-level context for efficient response generation, while an LLM-based deep module is selectively activated only for complex queries, augmented by a reflection mechanism for iterative reasoning refinement. Experiments show that HyMem achieves strong performance on both the LOCOMO and LongMemEval benchmarks, outperforming full-context while reducing computational cost by 92.6%, establishing a state-of-the-art balance between efficiency and performance in long-term memory management.
[443] Statistical Early Stopping for Reasoning Models
Yangxinyu Xie, Tao Wang, Soham Mallick, Yan Sun, Georgy Noarov, Mengxin Yu, Tanwi Mallick, Weijie J. Su, Edgar Dobriban
Main category: cs.AI
TL;DR: Early stopping methods for LLMs that monitor uncertainty signals during generation to prevent overthinking on ambiguous queries, improving efficiency and reliability in reasoning tasks.
Details
Motivation: LLMs often generate unnecessary reasoning steps when faced with uncertainty, ambiguous queries, or ill-posed problems, leading to inefficient computation and potential errors.Method: Two statistically principled early stopping approaches: 1) parametric method modeling inter-arrival times of uncertainty keywords as a renewal process with sequential testing, and 2) nonparametric method providing finite-sample guarantees on probability of halting too early on well-posed queries.
Result: Uncertainty-aware early stopping improves both efficiency and reliability in LLM reasoning across several domains and models, with especially significant gains for math reasoning tasks.
Conclusion: Monitoring uncertainty signals during LLM generation enables effective early stopping that reduces overthinking while maintaining reliability, particularly beneficial for reasoning tasks under uncertainty.
Abstract: While LLMs have seen substantial improvement in reasoning capabilities, they also sometimes overthink, generating unnecessary reasoning steps, particularly under uncertainty, given ill-posed or ambiguous queries. We introduce statistically principled early stopping methods that monitor uncertainty signals during generation to mitigate this issue. Our first approach is parametric: it models inter-arrival times of uncertainty keywords as a renewal process and applies sequential testing for stopping. Our second approach is nonparametric and provides finite-sample guarantees on the probability of halting too early on well-posed queries. We conduct empirical evaluations on reasoning tasks across several domains and models. Our results indicate that uncertainty-aware early stopping can improve both efficiency and reliability in LLM reasoning, and we observe especially significant gains for math reasoning.
[444] A Generalizable Physics-guided Causal Model for Trajectory Prediction in Autonomous Driving
Zhenyu Zong, Yuchen Wang, Haohong Lin, Lu Gan, Huajie Shao
Main category: cs.AI
TL;DR: Physics-guided Causal Model (PCM) for zero-shot trajectory prediction in autonomous driving using domain-invariant scene representations and causal integration with kinematic models.
Details
Motivation: Achieving effective zero-shot generalization for traffic agent trajectory prediction across unseen domains is challenging. The authors aim to leverage domain-invariant kinematic knowledge to enhance prediction capabilities in new environments.Method: Proposes PCM with two components: 1) Disentangled Scene Encoder using intervention-based disentanglement to extract domain-invariant features, and 2) CausalODE Decoder using causal attention to integrate kinematic models with contextual information.
Result: Extensive experiments on real-world autonomous driving datasets show superior zero-shot generalization performance in unseen cities, significantly outperforming competitive baselines.
Conclusion: The Physics-guided Causal Model effectively addresses zero-shot generalization for trajectory prediction by incorporating domain-invariant knowledge and causal integration with kinematic models.
Abstract: Trajectory prediction for traffic agents is critical for safe autonomous driving. However, achieving effective zero-shot generalization in previously unseen domains remains a significant challenge. Motivated by the consistent nature of kinematics across diverse domains, we aim to incorporate domain-invariant knowledge to enhance zero-shot trajectory prediction capabilities. The key challenges include: 1) effectively extracting domain-invariant scene representations, and 2) integrating invariant features with kinematic models to enable generalized predictions. To address these challenges, we propose a novel generalizable Physics-guided Causal Model (PCM), which comprises two core components: a Disentangled Scene Encoder, which adopts intervention-based disentanglement to extract domain-invariant features from scenes, and a CausalODE Decoder, which employs a causal attention mechanism to effectively integrate kinematic models with meaningful contextual information. Extensive experiments on real-world autonomous driving datasets demonstrate our method’s superior zero-shot generalization performance in unseen cities, significantly outperforming competitive baselines. The source code is released at https://github.com/ZY-Zong/Physics-guided-Causal-Model.
[445] Neuromem: A Granular Decomposition of the Streaming Lifecycle in External Memory for LLMs
Ruicheng Zhang, Xinyi Li, Tianyi Xu, Shuhao Zhang, Xiaofei Liao, Hai Jin
Main category: cs.AI
TL;DR: Neuromem is a testbed for evaluating External Memory Modules under streaming conditions with interleaved insertions and retrievals, analyzing five lifecycle dimensions across three datasets.
Details
Motivation: Current evaluations of External Memory Modules assume static settings with offline memory building, but practical applications involve streaming data with continuous insertions interleaved with retrievals, requiring evaluation of the full memory lifecycle.Method: Developed Neuromem testbed with interleaved insertion-and-retrieval protocol that decomposes memory lifecycle into five dimensions: memory data structure, normalization strategy, consolidation policy, query formulation strategy, and context integration mechanism. Evaluated using LOCOMO, LONGMEMEVAL, and MEMORYAGENTBENCH datasets with interchangeable variants in a shared serving stack.
Result: Performance typically degrades as memory grows across rounds, with time-related queries being most challenging. Memory data structure determines quality frontier, while aggressive compression and generative integration mostly shift costs between insertion and retrieval with limited accuracy gains.
Conclusion: Neuromem provides a comprehensive framework for evaluating External Memory Modules under realistic streaming conditions, revealing important trade-offs and limitations in current approaches.
Abstract: Most evaluations of External Memory Module assume a static setting: memory is built offline and queried at a fixed state. In practice, memory is streaming: new facts arrive continuously, insertions interleave with retrievals, and the memory state evolves while the model is serving queries. In this regime, accuracy and cost are governed by the full memory lifecycle, which encompasses the ingestion, maintenance, retrieval, and integration of information into generation. We present Neuromem, a scalable testbed that benchmarks External Memory Modules under an interleaved insertion-and-retrieval protocol and decomposes its lifecycle into five dimensions including memory data structure, normalization strategy, consolidation policy, query formulation strategy, and context integration mechanism. Using three representative datasets LOCOMO, LONGMEMEVAL, and MEMORYAGENTBENCH, Neuromem evaluates interchangeable variants within a shared serving stack, reporting token-level F1 and insertion/retrieval latency. Overall, we observe that performance typically degrades as memory grows across rounds, and time-related queries remain the most challenging category. The memory data structure largely determines the attainable quality frontier, while aggressive compression and generative integration mechanisms mostly shift cost between insertion and retrieval with limited accuracy gain.
[446] Cognitive Chunking for Soft Prompts: Accelerating Compressor Learning via Block-wise Causal Masking
Guojie Liu, Yiqi Wang, Yanfeng Yang, Wenqi Fan, Songlei Jian, Jianfeng Zhang, Jie Yu
Main category: cs.AI
TL;DR: PIC (Parallelized Iterative Compression) improves context compression for LLMs by restricting memory tokens to local chunks via attention mask modifications, reducing training difficulty and improving performance in high compression scenarios.
Details
Motivation: Long contexts increase LLM inference latency due to quadratic self-attention costs. Existing context compression methods compress entire contexts indiscriminately, requiring global dependency capture and extensive training data. Inspired by human working memory chunking and empirical observations of memory embedding spatial specialization.Method: Proposes Parallelized Iterative Compression (PIC) which modifies Transformer’s attention mask to restrict memory tokens’ receptive field to sequential local chunks, lowering compressor training difficulty. Uses chunking mechanism similar to human working memory.
Result: Consistently outperforms competitive baselines across multiple downstream tasks, especially in high compression scenarios (29.8% F1 and 40.7% EM improvements on QA tasks at 64× compression). Reduces training time by ~40% for 16× compressor while surpassing baseline peak performance.
Conclusion: PIC provides an effective context compression approach that reduces training difficulty and improves performance, particularly beneficial for high compression ratios. The method’s simplicity (attention mask modification) makes it practical for deployment.
Abstract: Providing extensive context via prompting is vital for leveraging the capabilities of Large Language Models (LLMs). However, lengthy contexts significantly increase inference latency, as the computational cost of self-attention grows quadratically with sequence length. To mitigate this issue, context compression-particularly soft prompt compressio-has emerged as a widely studied solution, which converts long contexts into shorter memory embeddings via a trained compressor. Existing methods typically compress the entire context indiscriminately into a set of memory tokens, requiring the compressor to capture global dependencies and necessitating extensive pre-training data to learn effective patterns. Inspired by the chunking mechanism in human working memory and empirical observations of the spatial specialization of memory embeddings relative to original tokens, we propose Parallelized Iterative Compression (PIC). By simply modifying the Transformer’s attention mask, PIC explicitly restricts the receptive field of memory tokens to sequential local chunks, thereby lowering the difficulty of compressor training. Experiments across multiple downstream tasks demonstrate that PIC consistently outperforms competitive baselines, with superiority being particularly pronounced in high compression scenarios (e.g., achieving relative improvements of 29.8% in F1 score and 40.7% in EM score on QA tasks at the $64\times$ compression ratio). Furthermore, PIC significantly expedites the training process. Specifically, when training the 16$\times$ compressor, it surpasses the peak performance of the competitive baseline while effectively reducing the training time by approximately 40%.
[447] Bridging AI and Clinical Reasoning: Abductive Explanations for Alignment on Critical Symptoms
Belona Sonna, Alban Grastien
Main category: cs.AI
TL;DR: The paper proposes using formal abductive explanations to align AI clinical diagnostic reasoning with structured clinical frameworks, ensuring transparency and trustworthiness while maintaining accuracy.
Details
Motivation: AI in clinical diagnostics often diverges from structured clinical reasoning frameworks, limiting trust and interpretability. Critical symptoms may be overlooked even when predictions are correct, and existing explanation methods lack formal guarantees and transparency.Method: Leverages formal abductive explanations that provide consistent, guaranteed reasoning over minimal sufficient feature sets to align AI decision-making with clinical reasoning frameworks.
Result: The approach preserves predictive accuracy while providing clinically actionable insights, establishing a robust framework for trustworthy AI in medical diagnosis.
Conclusion: Formal abductive explanations offer a solution to align AI reasoning with clinical frameworks, enhancing trust, interpretability, and adoption of AI in medical diagnostics.
Abstract: Artificial intelligence (AI) has demonstrated strong potential in clinical diagnostics, often achieving accuracy comparable to or exceeding that of human experts. A key challenge, however, is that AI reasoning frequently diverges from structured clinical frameworks, limiting trust, interpretability, and adoption. Critical symptoms, pivotal for rapid and accurate decision-making, may be overlooked by AI models even when predictions are correct. Existing post hoc explanation methods provide limited transparency and lack formal guarantees. To address this, we leverage formal abductive explanations, which offer consistent, guaranteed reasoning over minimal sufficient feature sets. This enables a clear understanding of AI decision-making and allows alignment with clinical reasoning. Our approach preserves predictive accuracy while providing clinically actionable insights, establishing a robust framework for trustworthy AI in medical diagnosis.
[448] Prompt-Driven Low-Altitude Edge Intelligence: Modular Agents and Generative Reasoning
Jiahao You, Ziye Jia, Chao Dong, Qihui Wu
Main category: cs.AI
TL;DR: P2AECF is a prompt-to-agent edge cognition framework that enables flexible, efficient, and adaptive deployment of large AI models at the edge by transforming semantic prompts into executable reasoning workflows through prompt-defined cognition, agent-based modular execution, and diffusion-controlled inference planning.
Details
Motivation: Current deployment of large AI models at the edge faces three key limitations: 1) rigid task-model binding limiting flexibility, 2) computational/memory demands exceeding edge device capacity, and 3) static inference pipelines unable to adapt to real-time task changes.Method: P2AECF uses three mechanisms: 1) prompt-defined cognition that parses task intent into abstract model-agnostic representations, 2) agent-based modular execution using lightweight reusable cognitive agents dynamically selected based on resource conditions, and 3) diffusion-controlled inference planning that adaptively constructs and refines execution strategies with runtime feedback.
Result: The framework is illustrated through a low-altitude intelligent network use case, demonstrating its ability to deliver adaptive, modular, and scalable edge intelligence for real-time low-altitude aerial collaborations.
Conclusion: P2AECF enables flexible, efficient, and adaptive edge intelligence by addressing the fundamental limitations of deploying large AI models at the edge through prompt-to-agent transformation and dynamic resource-aware execution.
Abstract: The large artificial intelligence models (LAMs) show strong capabilities in perception, reasoning, and multi-modal understanding, and can enable advanced capabilities in low-altitude edge intelligence. However, the deployment of LAMs at the edge remains constrained by some fundamental limitations. First, tasks are rigidly tied to specific models, limiting the flexibility. Besides, the computational and memory demands of full-scale LAMs exceed the capacity of most edge devices. Moreover, the current inference pipelines are typically static, making it difficult to respond to real-time changes of tasks. To address these challenges, we propose a prompt-to-agent edge cognition framework (P2AECF), enabling the flexible, efficient, and adaptive edge intelligence. Specifically, P2AECF transforms high-level semantic prompts into executable reasoning workflows through three key mechanisms. First, the prompt-defined cognition parses task intent into abstract and model-agnostic representations. Second, the agent-based modular execution instantiates these tasks using lightweight and reusable cognitive agents dynamically selected based on current resource conditions. Third, the diffusion-controlled inference planning adaptively constructs and refines execution strategies by incorporating runtime feedback and system context. In addition, we illustrate the framework through a representative low-altitude intelligent network use case, showing its ability to deliver adaptive, modular, and scalable edge intelligence for real-time low-altitude aerial collaborations.
[449] FloCA: Towards Faithful and Logically Consistent Flowchart Reasoning
Jinzi Zou, Bolin Wang, Liang Li, Shuo Zhang, Nuo Xu, Junzhou Zhao
Main category: cs.AI
TL;DR: FloCA is a zero-shot flowchart-oriented conversational agent that separates intent understanding/response generation (handled by LLM) from flowchart reasoning (handled by external topology-constrained graph execution tool) to ensure faithful and logically consistent node transitions in multi-turn decision-making dialogues.
Details
Motivation: Current LLMs lack explicit mechanisms to represent and reason over flowchart topology in flowchart-oriented dialogue systems, and are prone to hallucinations, leading to unfaithful flowchart reasoning. There's a need for systems that can guide users through multi-turn decision-making procedures while following domain-specific flowcharts accurately.Method: FloCA uses a hybrid approach: LLM handles intent understanding and response generation, while an external tool performs topology-constrained graph execution for flowchart reasoning. This separation ensures faithful node transitions consistent with the correct flowchart path across dialogue turns.
Result: Extensive experiments on FLODIAL and PFDial datasets show FloCA’s superiority over existing LLM-based methods, highlighting bottlenecks in current approaches and demonstrating improved reasoning accuracy and interaction efficiency.
Conclusion: FloCA effectively addresses LLM limitations in flowchart reasoning by delegating topology-constrained reasoning to an external tool, ensuring faithful and logically consistent multi-turn dialogue while maintaining LLM strengths in natural language understanding and generation.
Abstract: Flowchart-oriented dialogue (FOD) systems aim to guide users through multi-turn decision-making or operational procedures by following a domain-specific flowchart to achieve a task goal. In this work, we formalize flowchart reasoning in FOD as grounding user input to flowchart nodes at each dialogue turn while ensuring node transition is consistent with the correct flowchart path. Despite recent advances of LLMs in task-oriented dialogue systems, adapting them to FOD still faces two limitations: (1) LLMs lack an explicit mechanism to represent and reason over flowchart topology, and (2) they are prone to hallucinations, leading to unfaithful flowchart reasoning. To address these limitations, we propose FloCA, a zero-shot flowchart-oriented conversational agent. FloCA uses an LLM for intent understanding and response generation while delegating flowchart reasoning to an external tool that performs topology-constrained graph execution, ensuring faithful and logically consistent node transitions across dialogue turns. We further introduce an evaluation framework with an LLM-based user simulator and five new metrics covering reasoning accuracy and interaction efficiency. Extensive experiments on FLODIAL and PFDial datasets highlight the bottlenecks of existing LLM-based methods and demonstrate the superiority of FloCA. Our codes are available at https://github.com/Jinzi-Zou/FloCA-flowchart-reasoning.
[450] Choosing How to Remember: Adaptive Memory Structures for LLM Agents
Mingfei Lu, Mengjia Wu, Feng Liu, Jiawei Xu, Weikai Li, Haoyang Wang, Zhengdong Hu, Ying Ding, Yizhou Sun, Jie Lu, Yi Zhang
Main category: cs.AI
TL;DR: FluxMem is a framework for adaptive memory organization in LLM agents that learns to select among multiple memory structures based on interaction context, improving long-horizon performance.
Details
Motivation: Existing agent memory systems use one-size-fits-all approaches without adaptive memory structure selection, limiting their ability to handle heterogeneous interaction patterns and resulting in suboptimal performance for long-horizon tasks.Method: Proposes FluxMem framework with: 1) Multiple complementary memory structures, 2) Learning to select among structures using offline supervision from downstream response quality and memory utilization, 3) Three-level memory hierarchy, 4) Beta Mixture Model-based probabilistic gate for distribution-aware memory fusion instead of similarity thresholds.
Result: Experiments on PERSONAMEM and LoCoMo benchmarks show average improvements of 9.18% and 6.14% respectively, demonstrating effectiveness for long-horizon tasks.
Conclusion: FluxMem provides a unified framework for adaptive memory organization in LLM agents that significantly improves performance on long-horizon benchmarks through context-aware memory structure selection and robust memory fusion.
Abstract: Memory is critical for enabling large language model (LLM) based agents to maintain coherent behavior over long-horizon interactions. However, existing agent memory systems suffer from two key gaps: they rely on a one-size-fits-all memory structure and do not model memory structure selection as a context-adaptive decision, limiting their ability to handle heterogeneous interaction patterns and resulting in suboptimal performance. We propose a unified framework, FluxMem, that enables adaptive memory organization for LLM agents. Our framework equips agents with multiple complementary memory structures. It explicitly learns to select among these structures based on interaction-level features, using offline supervision derived from downstream response quality and memory utilization. To support robust long-horizon memory evolution, we further introduce a three-level memory hierarchy and a Beta Mixture Model-based probabilistic gate for distribution-aware memory fusion, replacing brittle similarity thresholds. Experiments on two long-horizon benchmarks, PERSONAMEM and LoCoMo, demonstrate that our method achieves average improvements of 9.18% and 6.14%.
[451] REAL: Resolving Knowledge Conflicts in Knowledge-Intensive Visual Question Answering via Reasoning-Pivot Alignment
Kai Ye, Xianwei Mao, Sheng Zhou, Zirui Shao, Ye Mo, Liangliang Liu, Haikuan Huang, Bin Li, Jiajun Bu
Main category: cs.AI
TL;DR: REAL framework uses reasoning-pivots to detect and resolve knowledge conflicts in visual question answering, achieving state-of-the-art performance through pivot-aware training and guided decoding.
Details
Motivation: Knowledge-intensive VQA suffers from severe knowledge conflicts due to limitations of open-domain retrieval, and existing approaches lack generalizable conflict detection and intra-model constraint mechanisms to handle conflicting evidence.Method: Proposes REAL framework centered on reasoning-pivots (atomic units in reasoning chains that emphasize knowledge linkage). Uses Reasoning-Pivot Aware SFT (RPA-SFT) to train a generalizable discriminator by aligning conflicts with pivot extraction, and Reasoning-Pivot Guided Decoding (RPGD) for intra-model decoding that leverages pivots for targeted conflict mitigation.
Result: Extensive experiments across diverse benchmarks demonstrate that REAL significantly enhances discrimination accuracy and achieves state-of-the-art performance.
Conclusion: The REAL framework validates the effectiveness of the pivot-driven resolution paradigm for handling knowledge conflicts in visual question answering.
Abstract: Knowledge-intensive Visual Question Answering (KI-VQA) frequently suffers from severe knowledge conflicts caused by the inherent limitations of open-domain retrieval. However, existing paradigms face critical limitations due to the lack of generalizable conflict detection and intra-model constraint mechanisms to handle conflicting evidence. To address these challenges, we propose the REAL (Reasoning-Pivot Alignment) framework centered on the novel concept of the Reasoning-Pivot. Distinct from reasoning steps that prioritize internal self-derivation, a reasoning-pivot serves as an atomic unit (node or edge) in the reasoning chain that emphasizes knowledge linkage, and it typically relies on external evidence to complete the reasoning. Supported by our constructed REAL-VQA dataset, our approach integrates Reasoning-Pivot Aware SFT (RPA-SFT) to train a generalizable discriminator by aligning conflicts with pivot extraction, and employs Reasoning-Pivot Guided Decoding (RPGD), an intra-model decoding strategy that leverages these pivots for targeted conflict mitigation. Extensive experiments across diverse benchmarks demonstrate that REAL significantly enhances discrimination accuracy and achieves state-of-the-art performance, validating the effectiveness of our pivot-driven resolution paradigm.
[452] Plan-MCTS: Plan Exploration for Action Exploitation in Web Navigation
Weiming Zhang, Jihong Wang, Jiamu Zhou, Qingyao Li, Xinbei Ma, Congmin Zheng, Xingyu Lou, Weiwen Liu, Zhuosheng Zhang, Jun Wang, Yong Yu, Weinan Zhang
Main category: cs.AI
TL;DR: Plan-MCTS framework for web navigation agents that uses semantic planning space to address sparse paths and noisy context, achieving SOTA on WebArena.
Details
Motivation: Current LLM-powered web navigation agents face two key challenges: 1) sparse valid paths leading to inefficient exploration, and 2) noisy context that dilutes accurate state perception. Tree search methods struggle with these issues in web navigation tasks.Method: Plan-MCTS shifts exploration to a semantic Plan Space, decoupling strategic planning from execution grounding. It transforms sparse action space into a Dense Plan Tree for efficient exploration, distills noisy contexts into Abstracted Semantic History for state awareness, uses Dual-Gating Reward to validate executability and strategic alignment, and employs Structural Refinement for on-policy repair of failed subplans.
Result: Extensive experiments on WebArena demonstrate state-of-the-art performance, surpassing current approaches with higher task effectiveness and search efficiency.
Conclusion: Plan-MCTS effectively addresses sparse path and noisy context challenges in web navigation by leveraging semantic planning space, achieving superior performance through efficient exploration and precise state awareness.
Abstract: Large Language Models (LLMs) have empowered autonomous agents to handle complex web navigation tasks. While recent studies integrate tree search to enhance long-horizon reasoning, applying these algorithms in web navigation faces two critical challenges: sparse valid paths that lead to inefficient exploration, and a noisy context that dilutes accurate state perception. To address this, we introduce Plan-MCTS, a framework that reformulates web navigation by shifting exploration to a semantic Plan Space. By decoupling strategic planning from execution grounding, it transforms sparse action space into a Dense Plan Tree for efficient exploration, and distills noisy contexts into an Abstracted Semantic History for precise state awareness. To ensure efficiency and robustness, Plan-MCTS incorporates a Dual-Gating Reward to strictly validate both physical executability and strategic alignment and Structural Refinement for on-policy repair of failed subplans. Extensive experiments on WebArena demonstrate that Plan-MCTS achieves state-of-the-art performance, surpassing current approaches with higher task effectiveness and search efficiency.
[453] GUI-GENESIS: Automated Synthesis of Efficient Environments with Verifiable Rewards for GUI Agent Post-Training
Yuan Cao, Dezhi Ran, Mengzhou Wu, Yuzhe Guo, Xin Chen, Ang Li, Gang Cao, Gong Zhi, Hao Yu, Linyi Li, Wei Yang, Tao Xie
Main category: cs.AI
TL;DR: GUI-GENESIS: A framework that automatically synthesizes efficient GUI training environments with verifiable code-native rewards from real-world applications, enabling faster and cheaper agent training with better generalization to real tasks.
Details
Motivation: Training GUI agents on real-world applications faces challenges including high latency, poor reproducibility, and unverifiable rewards that rely on noisy visual proxies, limiting agent generalization and long-horizon planning capabilities.Method: Reconstructs real-world applications into lightweight web environments using multimodal code models, equipping them with code-native rewards (executable assertions) that provide deterministic reward signals and eliminate visual estimation noise.
Result: Reduces environment latency by 10 times and costs by over $28,000 per epoch compared to training on real applications. Agents trained with GUI-GENESIS outperform base model by 14.54% and real-world RL baselines by 3.27% on held-out real-world tasks.
Conclusion: GUI-GENESIS enables efficient GUI agent training with verifiable rewards, demonstrating that models can synthesize environments they cannot yet solve, highlighting a pathway for self-improving agents.
Abstract: Post-training GUI agents in interactive environments is critical for developing generalization and long-horizon planning capabilities. However, training on real-world applications is hindered by high latency, poor reproducibility, and unverifiable rewards relying on noisy visual proxies. To address the limitations, we present GUI-GENESIS, the first framework to automatically synthesize efficient GUI training environments with verifiable rewards. GUI-GENESIS reconstructs real-world applications into lightweight web environments using multimodal code models and equips them with code-native rewards, executable assertions that provide deterministic reward signals and eliminate visual estimation noise. Extensive experiments show that GUI-GENESIS reduces environment latency by 10 times and costs by over $28,000 per epoch compared to training on real applications. Notably, agents trained with GUI-GENESIS outperform the base model by 14.54% and even real-world RL baselines by 3.27% on held-out real-world tasks. Finally, we observe that models can synthesize environments they cannot yet solve, highlighting a pathway for self-improving agents.
[454] NEST: Nascent Encoded Steganographic Thoughts
Artem Karpov
Main category: cs.AI
TL;DR: Models can hide secret reasoning in innocuous text (steganographic CoT), but current models struggle with complex tasks while showing nascent capability in simpler ones.
Details
Motivation: To assess the risk of LLMs learning to conceal their reasoning through steganographic chain-of-thought, which could compromise safety monitoring and enable deceptive behavior.Method: Systematic evaluation across 28 models (past to current frontier) measuring monitor evasion, refusal rates, encoding fidelity, and hidden task accuracy across four datasets, comparing steganographic acrostics against plain reasoning and filler-token baselines.
Result: Current models cannot sustain hidden reasoning for complex math/arithmetic tasks, but Claude Opus 4.5 achieved 92% accuracy on simplified counting tasks. GPT-5.2 showed rare (<1%) paradoxical refusal/compliance behavior.
Conclusion: Steganographic reasoning poses emerging risks requiring continuous evaluation; the study provides methodology for preemptive detection of hidden reasoning that could enable misaligned scheming.
Abstract: Monitoring chain-of-thought (CoT) reasoning is a foundational safety technique for large language model (LLM) agents; however, this oversight is compromised if models learn to conceal their reasoning. We explore the potential for steganographic CoT – where models hide secret reasoning within innocuous text – to inform risk assessment and deployment policies. We systematically evaluate the limits of steganographic capabilities across 28 models, ranging from past generations to the current frontier. We measure monitor evasion, refusal rates, encoding fidelity, and hidden task accuracy across four datasets, comparing steganographic acrostics against plain reasoning and filler-token baselines. We find that current models cannot yet sustain hidden reasoning for complex math and arithmetic tasks. However, in a simplified counting experiment, Claude Opus 4.5 achieved 92% accuracy on the hidden task, demonstrating nascent capability. Notably, in rare cases (<1%), GPT-5.2 might refuse steganographic instructions while simultaneously complying with them. Our findings underscore the need for continuous evaluation of steganographic risks. This study provides a methodology to preemptively detect and prevent hidden reasoning that might empower misaligned scheming and deceptive behavior.
[455] Algebraic Quantum Intelligence: A New Framework for Reproducible Machine Creativity
Kazuo Yano, Jonghyeok Lee, Tae Ishitomi, Hironobu Kawaguchi, Akira Koyama, Masakuni Ota, Yuki Ota, Nobuo Sato, Keita Shimada, Sho Takematsu, Ayaka Tobinai, Satomi Tsuji, Kazunori Yanagi, Keiko Yano, Manabu Harada, Yuki Matsuda, Kazunori Matsumoto, Kenichi Matsumura, Hamae Matsuo, Yumi Miyazaki, Kotaro Murai, Tatsuya Ohshita, Marie Seki, Shun Tanoue, Tatsuki Terakado, Yuko Ichimaru, Mirei Saito, Akihiro Otsuka, Koji Ara
Main category: cs.AI
TL;DR: AQI framework uses noncommutative algebraic structures inspired by quantum theory to expand semantic space and enhance creativity in LLMs, outperforming baselines on creative reasoning benchmarks.
Details
Motivation: Current LLMs struggle with genuine creativity because rich context strongly constrains future generations, making the process near-deterministic. Existing approaches like test-time scaling don't fundamentally address this structural limitation.Method: Proposed Algebraic Quantum Intelligence (AQI) as a computational framework using noncommutative algebraic structures inspired by quantum theory. Implemented by extending a transformer-based LLM with 600+ specialized operators, representing semantic states as Hilbert space vectors and evolving them using C-values from noncommutative operators.
Result: AQI consistently outperforms strong baseline models on creative reasoning benchmarks across ten domains, showing statistically significant improvements and reduced cross-domain variance under LLM-as-a-judge protocol.
Conclusion: Noncommutative algebraic dynamics provide a practical and reproducible foundation for machine creativity, with the architecture already deployed in real-world enterprise environments.
Abstract: Large language models (LLMs) have achieved remarkable success in generating fluent and contextually appropriate text; however, their capacity to produce genuinely creative outputs remains limited. This paper posits that this limitation arises from a structural property of contemporary LLMs: when provided with rich context, the space of future generations becomes strongly constrained, and the generation process is effectively governed by near-deterministic dynamics. Recent approaches such as test-time scaling and context adaptation improve performance but do not fundamentally alter this constraint. To address this issue, we propose Algebraic Quantum Intelligence (AQI) as a computational framework that enables systematic expansion of semantic space. AQI is formulated as a noncommutative algebraic structure inspired by quantum theory, allowing properties such as order dependence, interference, and uncertainty to be implemented in a controlled and designable manner. Semantic states are represented as vectors in a Hilbert space, and their evolution is governed by C-values computed from noncommutative operators, thereby ensuring the coexistence and expansion of multiple future semantic possibilities. In this study, we implement AQI by extending a transformer-based LLM with more than 600 specialized operators. We evaluate the resulting system on creative reasoning benchmarks spanning ten domains under an LLM-as-a-judge protocol. The results show that AQI consistently outperforms strong baseline models, yielding statistically significant improvements and reduced cross-domain variance. These findings demonstrate that noncommutative algebraic dynamics can serve as a practical and reproducible foundation for machine creativity. Notably, this architecture has already been deployed in real-world enterprise environments.
[456] ForesightSafety Bench: A Frontier Risk Evaluation and Governance Framework towards Safe AI
Haibo Tong, Feifei Zhao, Linghao Feng, Ruoyu Wu, Ruolin Chen, Lu Jia, Zhou Zhao, Jindong Li, Tenglong Li, Erliang Lin, Shuai Yang, Enmeng Lu, Yinqian Sun, Qian Zhang, Zizhe Ruan, Zeyang Yue, Ping Wu, Huangrui Li, Chengyi Sun, Yi Zeng
Main category: cs.AI
TL;DR: ForesightSafety Bench is a comprehensive AI safety evaluation framework covering 94 risk dimensions across fundamental safety, embodied AI, AI4Science, social/environmental risks, catastrophic/existential risks, and 8 industrial domains, with systematic evaluation of 20+ advanced models revealing widespread safety vulnerabilities.
Details
Motivation: Current AI safety evaluation systems have critical limitations including restricted risk dimensions and failed frontier risk detection, with lagging safety benchmarks and alignment technologies unable to address complex challenges posed by cutting-edge AI models with increasing autonomy and goal-directed capabilities.Method: Proposes a hierarchical framework starting with 7 Fundamental Safety pillars, extending to advanced Embodied AI Safety, AI4Science Safety, Social/Environmental AI risks, Catastrophic/Existential Risks, and 8 industrial safety domains totaling 94 refined risk dimensions. Accumulates tens of thousands of structured risk data points and assessment results.
Result: Systematic evaluation of over twenty mainstream advanced large models reveals widespread safety vulnerabilities across multiple pillars, particularly in Risky Agentic Autonomy, AI4Science Safety, Embodied AI Safety, Social AI Safety and Catastrophic/Existential Risks. Identifies key risk patterns and capability boundaries.
Conclusion: The ForesightSafety Bench establishes a comprehensive, hierarchical, and dynamically evolving AI safety evaluation framework that addresses limitations of current systems and provides systematic assessment of frontier AI risks across multiple dimensions.
Abstract: Rapidly evolving AI exhibits increasingly strong autonomy and goal-directed capabilities, accompanied by derivative systemic risks that are more unpredictable, difficult to control, and potentially irreversible. However, current AI safety evaluation systems suffer from critical limitations such as restricted risk dimensions and failed frontier risk detection. The lagging safety benchmarks and alignment technologies can hardly address the complex challenges posed by cutting-edge AI models. To bridge this gap, we propose the “ForesightSafety Bench” AI Safety Evaluation Framework, beginning with 7 major Fundamental Safety pillars and progressively extends to advanced Embodied AI Safety, AI4Science Safety, Social and Environmental AI risks, Catastrophic and Existential Risks, as well as 8 critical industrial safety domains, forming a total of 94 refined risk dimensions. To date, the benchmark has accumulated tens of thousands of structured risk data points and assessment results, establishing a widely encompassing, hierarchically clear, and dynamically evolving AI safety evaluation framework. Based on this benchmark, we conduct systematic evaluation and in-depth analysis of over twenty mainstream advanced large models, identifying key risk patterns and their capability boundaries. The safety capability evaluation results reveals the widespread safety vulnerabilities of frontier AI across multiple pillars, particularly focusing on Risky Agentic Autonomy, AI4Science Safety, Embodied AI Safety, Social AI Safety and Catastrophic and Existential Risks. Our benchmark is released at https://github.com/Beijing-AISI/ForesightSafety-Bench. The project website is available at https://foresightsafety-bench.beijing-aisi.ac.cn/.
[457] Process-Supervised Multi-Agent Reinforcement Learning for Reliable Clinical Reasoning
Chaeeun Lee, T. Michael Yates, Pasquale Minervini, T. Ian Simpson
Main category: cs.AI
TL;DR: Multi-agent reinforcement learning framework for gene-disease validity curation that optimizes both outcome accuracy and clinical reasoning process alignment
Details
Motivation: Clinical decision-making requires nuanced reasoning over heterogeneous evidence with traceable justifications, but current LLM multi-agent systems focus mainly on outcome accuracy while overlooking process-grounded reasoning aligned with clinical standards, particularly in gene-disease validity curation tasksMethod: Agent-as-tool reinforcement learning framework with two objectives: (1) process-level supervision to ensure reasoning follows valid clinical pathways, and (2) efficient coordination via hierarchical multi-agent system; uses GRPO-trained supervisor agent (Qwen3-4B) with different reward strategies (outcome-only vs process+outcome)
Result: On ClinGen dataset: with outcome-only rewards, MAS with GRPO-trained supervisor improved final outcome accuracy from 0.195 to 0.732 but had poor process alignment (0.392 F1); with process+outcome rewards, achieved higher outcome accuracy (0.750) while significantly improving process fidelity to 0.520 F1
Conclusion: The framework successfully balances outcome accuracy with clinical reasoning process alignment in gene-disease validity curation, demonstrating that process-level supervision is crucial for developing clinically-reliable multi-agent systems
Abstract: Clinical decision-making requires nuanced reasoning over heterogeneous evidence and traceable justifications. While recent LLM multi-agent systems (MAS) show promise, they largely optimise for outcome accuracy while overlooking process-grounded reasoning aligned with clinical standards. One critical real-world case of this is gene-disease validity curation, where experts must determine whether a gene is causally implicated in a disease by synthesising diverse biomedical evidence. We introduce an agent-as-tool reinforcement learning framework for this task with two objectives: (i) process-level supervision to ensure reasoning follows valid clinical pathways, and (ii) efficient coordination via a hierarchical multi-agent system. Our evaluation on the ClinGen dataset shows that with outcome-only rewards, MAS with a GRPO-trained Qwen3-4B supervisor agent substantially improves final outcome accuracy from 0.195 with a base model supervisor to 0.732, but results in poor process alignment (0.392 F1). Conversely, with process + outcome rewards, MAS with GRPO-trained supervisor achieves higher outcome accuracy (0.750) while significantly improving process fidelity to 0.520 F1. Our code is available at https://github.com/chaeeunlee-io/GeneDiseaseCurationAgents.
[458] Text Before Vision: Staged Knowledge Injection Matters for Agentic RLVR in Ultra-High-Resolution Remote Sensing Understanding
Fengxiang Wang, Mingshuo Chen, Yueying Li, Yajie Yang, Yuhao Zhou, Di Wang, Yifan Zhang, Haoyu Wang, Haiyan Zhao, Hongda Sun, Long Lan, Jun Song, Yulin Wang, Jing Zhang, Wenlong Zhang, Bo Du
Main category: cs.AI
TL;DR: Text-only domain knowledge injection significantly improves ultra-high-resolution remote sensing visual reasoning by providing structured reasoning priors for evidence retrieval.
Details
Motivation: Multimodal reasoning for ultra-high-resolution remote sensing faces challenges in visual evidence acquisition due to the need to locate tiny relevant regions in massive pixel spaces, and standard reinforcement learning struggles without structured domain priors.Method: Proposes a staged knowledge injection approach: (1) cold-starting with scalable, knowledge-graph-verified Earth-science text QA to instill reasoning structures, and (2) “pre-warming” on hard UHR image-text examples during supervised fine-tuning to stabilize subsequent tool-based reinforcement learning.
Result: Achieves 60.40% Pass@1 on XLRS-Bench, significantly outperforming larger general purpose models like GPT-5.2, Gemini 3.0 Pro, and Intern-S1, establishing a new state-of-the-art.
Conclusion: High-quality domain-specific text QA is a primary driver of UHR visual reasoning gains, as it injects the concepts, mechanistic explanations, and decision rules necessary to guide visual evidence retrieval, even without images.
Abstract: Multimodal reasoning for ultra-high-resolution (UHR) remote sensing (RS) is usually bottlenecked by visual evidence acquisition: the model necessitates localizing tiny task-relevant regions in massive pixel spaces. While Agentic Reinforcement Learning with Verifiable Rewards (RLVR) using zoom-in tools offers a path forward, we find that standard reinforcement learning struggles to navigate these vast visual spaces without structured domain priors. In this paper, we investigate the interplay between post-training paradigms: comparing Cold-start Supervised Fine-Tuning (SFT), RLVR, and Agentic RLVR on the UHR RS benchmark.Our controlled studies yield a counter-intuitive finding: high-quality Earth-science text-only QA is a primary driver of UHR visual reasoning gains. Despite lacking images, domain-specific text injects the concepts, mechanistic explanations, and decision rules necessary to guide visual evidence retrieval.Based on this, we propose a staged knowledge injection recipe: (1) cold-starting with scalable, knowledge-graph-verified Earth-science text QA to instill reasoning structures;and (2) “pre-warming” on the same hard UHR image-text examples during SFT to stabilize and amplify subsequent tool-based RL. This approach achieves a 60.40% Pass@1 on XLRS-Bench, significantly outperforming larger general purpose models (e.g., GPT-5.2, Gemini 3.0 Pro, Intern-S1) and establishing a new state-of-the-art.
[459] CORPGEN: Simulating Corporate Environments with Autonomous Digital Employees in Multi-Horizon Task Environments
Abubakarr Jaye, Nigel Boachie Kumankumah, Chidera Biringa, Anjel Shaileshbhai Patel, Sulaiman Vesal, Dayquan Julienne, Charlotte Siska, Manuel Raúl Meléndez Luján, Anthony Twum-Barimah, Mauricio Velazco, Tianwei Chen
Main category: cs.AI
TL;DR: CorpGen framework improves autonomous agents’ ability to manage multiple concurrent long-horizon tasks in corporate environments through hierarchical planning, sub-agent isolation, tiered memory, and adaptive summarization.
Details
Motivation: Existing benchmarks evaluate agents on single tasks in isolation, but real organizational work requires managing many concurrent long-horizon tasks with interleaving, dependencies, and reprioritization.Method: CorpGen framework with hierarchical planning for multi-horizon goal alignment, sub-agent isolation to prevent cross-task contamination, tiered memory (working, structured, semantic), and adaptive summarization, tested in simulated corporate environments.
Result: Achieves up to 3.5x improvement over baselines (15.2% vs 4.3%) with stable performance under increasing load, confirming gains stem from architectural mechanisms rather than specific CUA implementations.
Conclusion: CorpGen effectively addresses key failure modes in multi-horizon task environments and demonstrates architectural solutions for managing concurrent long-horizon tasks in organizational settings.
Abstract: Long-horizon reasoning is a key challenge for autonomous agents, yet existing benchmarks evaluate agents on single tasks in isolation. Real organizational work requires managing many concurrent long-horizon tasks with interleaving, dependencies, and reprioritization. We introduce Multi-Horizon Task Environments (MHTEs): a distinct problem class requiring coherent execution across dozens of interleaved tasks (45+, 500-1500+ steps) within persistent execution contexts spanning hours. We identify four failure modes that cause baseline CUAs to degrade from 16.7% to 8.7% completion as load scales 25% to 100%, a pattern consistent across three independent implementations. These failure modes are context saturation (O(N) vs O(1) growth), memory interference, dependency complexity (DAGs vs. chains), and reprioritization overhead. We present CorpGen, an architecture-agnostic framework addressing these failures via hierarchical planning for multi-horizon goal alignment, sub-agent isolation preventing cross-task contamination, tiered memory (working, structured, semantic), and adaptive summarization. CorpGen simulates corporate environments through digital employees with persistent identities and realistic schedules. Across three CUA backends (UFO2, OpenAI CUA, hierarchical) on OSWorld Office, CorpGen achieves up to 3.5x improvement over baselines (15.2% vs 4.3%) with stable performance under increasing load, confirming that gains stem from architectural mechanisms rather than specific CUA implementations. Ablation studies show experiential learning provides the largest gains.
[460] REDSearcher: A Scalable and Cost-Efficient Framework for Long-Horizon Search Agents
Zheng Chu, Xiao Wang, Jack Hong, Huiming Fan, Yuqi Huang, Yue Yang, Guohai Xu, Chenxiao Zhao, Cheng Xiang, Shengchao Hu, Dongdong Kuang, Ming Liu, Bing Qin, Xing Yu
Main category: cs.AI
TL;DR: REDSearcher is a framework for optimizing large language models as search agents through co-designed task synthesis, mid-training, and post-training, addressing challenges in scalable long-horizon search task construction and high-cost interaction rollouts.
Details
Motivation: Large language models need optimization for real-world search tasks, but face challenges with sparse high-quality search trajectories and reward signals due to difficulty in scalable long-horizon task construction and expensive interaction-heavy rollouts with external tools.Method: Proposes REDSearcher framework with four key improvements: 1) Task synthesis as dual-constrained optimization using graph topology and evidence dispersion, 2) Tool-augmented queries to encourage proactive tool use, 3) Mid-training to strengthen core atomic capabilities (knowledge, planning, function calling), 4) Local simulated environment for low-cost RL iteration.
Result: Achieves state-of-the-art performance across both text-only and multimodal search-agent benchmarks. Releases 10K text search trajectories, 5K multimodal trajectories, 1K text RL query set, plus code and model checkpoints.
Conclusion: REDSearcher provides a scalable framework for optimizing search agents, addressing key bottlenecks in trajectory sparsity and interaction costs, with demonstrated effectiveness in both text and multimodal search tasks.
Abstract: Large language models are transitioning from generalpurpose knowledge engines to realworld problem solvers, yet optimizing them for deep search tasks remains challenging. The central bottleneck lies in the extreme sparsity of highquality search trajectories and reward signals, arising from the difficulty of scalable longhorizon task construction and the high cost of interactionheavy rollouts involving external tool calls. To address these challenges, we propose REDSearcher, a unified framework that codesigns complex task synthesis, midtraining, and posttraining for scalable searchagent optimization. Specifically, REDSearcher introduces the following improvements: (1) We frame task synthesis as a dualconstrained optimization, where task difficulty is precisely governed by graph topology and evidence dispersion, allowing scalable generation of complex, highquality tasks. (2) We introduce toolaugmented queries to encourage proactive tool use rather than passive recall.(3) During midtraining, we strengthen core atomic capabilities knowledge, planning, and function calling substantially reducing the cost of collecting highquality trajectories for downstream training. (4) We build a local simulated environment that enables rapid, lowcost algorithmic iteration for reinforcement learning experiments. Across both textonly and multimodal searchagent benchmarks, our approach achieves stateoftheart performance. To facilitate future research on longhorizon search agents, we will release 10K highquality complex text search trajectories, 5K multimodal trajectories and 1K text RL query set, and together with code and model checkpoints.
[461] GRAIL: Goal Recognition Alignment through Imitation Learning
Osher Elhadad, Felipe Meneguzzi, Reuth Mirsky
Main category: cs.AI
TL;DR: GRAIL uses imitation learning and inverse reinforcement learning to recognize agent goals from potentially suboptimal demonstrations by learning goal-directed policies for each candidate goal.
Details
Motivation: Existing goal recognition methods assume optimal goal-oriented policies, which may not match an actor's true (potentially suboptimal) behavior, leading to inaccurate goal recognition.Method: GRAIL combines imitation learning and inverse reinforcement learning to learn one goal-directed policy for each candidate goal directly from demonstration trajectories. It scores observed partial trajectories with each learned policy in a single forward pass.
Result: GRAIL improves F1-scores by >0.5 under systematically biased optimal behavior, gains 0.1-0.3 under suboptimal behavior, improves up to 0.4 under noisy optimal trajectories, while remaining competitive in fully optimal settings.
Conclusion: GRAIL contributes to scalable and robust models for interpreting agent goals in uncertain environments by better handling suboptimal and biased behavior.
Abstract: Understanding an agent’s goals from its behavior is fundamental to aligning AI systems with human intentions. Existing goal recognition methods typically rely on an optimal goal-oriented policy representation, which may differ from the actor’s true behavior and hinder the accurate recognition of their goal. To address this gap, this paper introduces Goal Recognition Alignment through Imitation Learning (GRAIL), which leverages imitation learning and inverse reinforcement learning to learn one goal-directed policy for each candidate goal directly from (potentially suboptimal) demonstration trajectories. By scoring an observed partial trajectory with each learned goal-directed policy in a single forward pass, GRAIL retains the one-shot inference capability of classical goal recognition while leveraging learned policies that can capture suboptimal and systematically biased behavior. Across the evaluated domains, GRAIL increases the F1-score by more than 0.5 under systematically biased optimal behavior, achieves gains of approximately 0.1-0.3 under suboptimal behavior, and yields improvements of up to 0.4 under noisy optimal trajectories, while remaining competitive in fully optimal settings. This work contributes toward scalable and robust models for interpreting agent goals in uncertain environments.
[462] AutoWebWorld: Synthesizing Infinite Verifiable Web Environments via Finite State Machines
Yifan Wu, Yiran Peng, Yiyu Chen, Jianhao Ruan, Zijie Zhuang, Cheng Yang, Jiayi Zhang, Man Chen, Yenchi Tseng, Zhaoyang Yu, Liang Chen, Yuyao Zhai, Bang Liu, Chenglin Wu, Yuyu Luo
Main category: cs.AI
TL;DR: AutoWebWorld: A framework for synthesizing controllable web environments as Finite State Machines to generate verified interaction data for training Web GUI agents, improving real-world performance.
Details
Motivation: Collecting real-world web interaction data is expensive and difficult to verify due to hidden state transitions, requiring costly external verifiers to evaluate step-level correctness.Method: Model web environments as Finite State Machines (FSMs) and use coding agents to translate FSMs into interactive websites, enabling programmatic verification of actions and task success through explicit state transition rules.
Result: Generated 11,663 verified trajectories from 29 diverse web environments at $0.04 per trajectory. Training on this data significantly boosts real-world performance: 7B Web GUI agent outperforms all baselines on WebVoyager within 15 steps, with clear scaling law showing improved performance as synthetic data volume increases.
Conclusion: AutoWebWorld provides an efficient framework for generating high-quality, verifiable web interaction data that enables better training of Web GUI agents, demonstrating strong performance improvements and scalability on real-world benchmarks.
Abstract: The performance of autonomous Web GUI agents heavily relies on the quality and quantity of their training data. However, a fundamental bottleneck persists: collecting interaction trajectories from real-world websites is expensive and difficult to verify. The underlying state transitions are hidden, leading to reliance on inconsistent and costly external verifiers to evaluate step-level correctness. To address this, we propose AutoWebWorld, a novel framework for synthesizing controllable and verifiable web environments by modeling them as Finite State Machines (FSMs) and use coding agents to translate FSMs into interactive websites. Unlike real websites, where state transitions are implicit, AutoWebWorld explicitly defines all states, actions, and transition rules. This enables programmatic verification: action correctness is checked against predefined rules, and task success is confirmed by reaching a goal state in the FSM graph. AutoWebWorld enables a fully automated search-and-verify pipeline, generating over 11,663 verified trajectories from 29 diverse web environments at only $0.04 per trajectory. Training on this synthetic data significantly boosts real-world performance. Our 7B Web GUI agent outperforms all baselines within 15 steps on WebVoyager. Furthermore, we observe a clear scaling law: as the synthetic data volume increases, performance on WebVoyager and Online-Mind2Web consistently improves.
[463] Benchmarking at the Edge of Comprehension
Samuele Marro, Jialin Yu, Emanuele La Malfa, Oishi Deb, Jiawei Li, Yibo Yang, Ebey Abraham, Sunando Sengupta, Eric Sommerlade, Michael Wooldridge, Philip Torr
Main category: cs.AI
TL;DR: Proposes Critique-Resilient Benchmarking, an adversarial framework for comparing LLMs when human comprehension becomes infeasible, using critique-resilient correctness and humans as bounded verifiers.
Details
Motivation: As frontier LLMs saturate benchmarks quickly, traditional benchmarking becomes infeasible when humans can't generate discriminative tasks, provide accurate ground truth, or evaluate complex solutions - threatening our ability to measure AI progress.Method: Uses critique-resilient correctness where answers are deemed correct if no adversary can convincingly prove otherwise. Humans serve as bounded verifiers focusing on localized claims. Employs itemized bipartite Bradley-Terry model to jointly rank LLMs by their ability to solve tasks and generate difficult yet solvable questions.
Result: Demonstrated effectiveness in mathematical domain across eight frontier LLMs, showing stable scores that correlate with external capability measures.
Conclusion: Reformulates benchmarking as an adversarial generation-evaluation game with humans as final adjudicators, providing a scalable approach for comparing models beyond human comprehension limits.
Abstract: As frontier Large Language Models (LLMs) increasingly saturate new benchmarks shortly after they are published, benchmarking itself is at a juncture: if frontier models keep improving, it will become increasingly hard for humans to generate discriminative tasks, provide accurate ground-truth answers, or evaluate complex solutions. If benchmarking becomes infeasible, our ability to measure any progress in AI is at stake. We refer to this scenario as the post-comprehension regime. In this work, we propose Critique-Resilient Benchmarking, an adversarial framework designed to compare models even when full human understanding is infeasible. Our technique relies on the notion of critique-resilient correctness: an answer is deemed correct if no adversary has convincingly proved otherwise. Unlike standard benchmarking, humans serve as bounded verifiers and focus on localized claims, which preserves evaluation integrity beyond full comprehension of the task. Using an itemized bipartite Bradley-Terry model, we jointly rank LLMs by their ability to solve challenging tasks and to generate difficult yet solvable questions. We showcase the effectiveness of our method in the mathematical domain across eight frontier LLMs, showing that the resulting scores are stable and correlate with external capability measures. Our framework reformulates benchmarking as an adversarial generation-evaluation game in which humans serve as final adjudicators.
[464] Competition for attention predicts good-to-bad tipping in AI
Neil F. Johnson, Frank Y. Huo
Main category: cs.AI
TL;DR: Paper identifies mathematical tipping point mechanism for dangerous AI behavior in edge devices due to attention competition, offering new safety control levers.
Details
Motivation: The proliferation of edge AI devices running language models offline creates safety risks (self-harm, financial losses, extremism) without cloud-based oversight, requiring new safety mechanisms that work without connectivity.Method: Develops mathematical formula for dynamical tipping point n* based on dot-product competition for attention between conversation context and competing output basins, validated across multiple AI models.
Result: Identifies atomistic-scale mechanism for dangerous behavior in edge AI, provides mathematical framework for predicting tipping points, and demonstrates applicability across domains, legal landscapes, languages, and cultures.
Conclusion: The attention competition mechanism offers new safety control levers for edge AI devices, addressing critical gaps in current safety tools that require cloud connectivity or detect failures only after harm occurs.
Abstract: More than half the global population now carries devices that can run ChatGPT-like language models with no Internet connection and minimal safety oversight – and hence the potential to promote self-harm, financial losses and extremism among other dangers. Existing safety tools either require cloud connectivity or discover failures only after harm has occurred. Here we show that a large class of potentially dangerous tipping originates at the atomistic scale in such edge AI due to competition for the machinery’s attention. This yields a mathematical formula for the dynamical tipping point n*, governed by dot-product competition for attention between the conversation’s context and competing output basins, that reveals new control levers. Validated against multiple AI models, the mechanism can be instantiated for different definitions of ‘good’ and ‘bad’ and hence in principle applies across domains (e.g. health, law, finance, defense), changing legal landscapes (e.g. EU, UK, US and state level), languages, and cultural settings.
[465] Boule or Baguette? A Study on Task Topology, Length Generalization, and the Benefit of Reasoning Traces
William L. Tong, Ege Cakar, Cengiz Pehlevan
Main category: cs.AI
TL;DR: RT models generalize well on broad/shallow tasks but deteriorate on narrow/deep tasks relative to non-RT baselines, revealing fundamental scaling limits of reasoning trace approaches.
Details
Motivation: Despite rapid progress in reasoning models that generate intermediate reasoning traces, understanding of how RTs support reasoning and the limits of this paradigm remains incomplete. The paper aims to systematically analyze the generalization capabilities of RT models, particularly for length generalization in reasoning tasks.Method: Introduced PITA dataset with 23M+ propositional logic statements and proofs. Proposed task depth (steps required) and task breadth (unique examples) metrics. Varied these across PITA subsets and compared RT vs non-RT models. Also compared with synthetic syllogism task to validate generalizability of findings.
Result: RT models generalize well on broad and shallow subsets but deteriorate on narrow and deep subsets relative to non-RT baselines. The synthetic syllogism task confirmed similar patterns, suggesting these are general phenomena rather than PITA-specific.
Conclusion: The study reveals fundamental scaling limits of RT models on deep tasks while highlighting their generalization strengths on broad tasks. Identifies inherent benefits and limitations of reasoning trace approaches for reasoning models.
Abstract: Recent years have witnessed meteoric progress in reasoning models: neural networks that generate intermediate reasoning traces (RTs) before producing a final output. Despite the rapid advancement, our understanding of how RTs support reasoning, and the limits of this paradigm, remain incomplete. To promote greater clarity, we introduce PITA: a novel large-scale dataset of over 23 million statements in propositional logic and their corresponding proofs. As a benchmark for robust reasoning, we focus on length generalization: if a model is trained to determine truth or falsity on statements with proofs up to fixed length, how well does it generalize to statements requiring longer proofs? We propose notions of (1) task depth and (2) task breadth, which measure respectively (1) the number of steps required to solve an example from a task and (2) the number of unique examples across a task. We vary these quantities across subsets of PITA, and find that RT models generalize well on broad and shallow subsets, while deteriorating on narrow and deep subsets relative to non-RT baselines. To determine whether our results are idiosyncratic to PITA or indicative of general phenomena, we compare our results to a simple synthetic task based on syllogisms. Our resulting theory suggests fundamental scalings that limit how well RT models perform on deep tasks, and highlights their generalization strengths on broad tasks. Our findings overall identify fundamental benefits and limitations inherent in using reasoning traces.
[466] Precedent-Informed Reasoning: Mitigating Overthinking in Large Reasoning Models via Test-Time Precedent Learning
Qianyue Wang, Jinwu Hu, Huanxiang Lin, Bolin Chen, Zhiquan Wen, Yaofo Chen, Yu Rong, Mingkui Tan
Main category: cs.AI
TL;DR: PIR improves LLM reasoning efficiency by using precedents to guide reasoning instead of exhaustive self-exploration, reducing computational costs while maintaining accuracy.
Details
Motivation: Current LLM reasoning suffers from inefficient long chain-of-thought traces with redundant self-exploration and validation, which inflates computational costs and can degrade performance. The paper is inspired by human reasoning patterns where people solve new problems by leveraging past related cases.Method: Proposes Precedent Informed Reasoning (PIR) with two components: 1) Adaptive Precedent Selection (APS) constructs compact precedent sets using joint scoring of semantic similarity and model perplexity, adapting the amount to maximize perplexity reduction; 2) Test-time Experience Internalization (TEI) performs test-time learning on precedent-informed instructions, updating lightweight adapters to internalize solution patterns as priors.
Result: Experiments across mathematical reasoning, scientific QA, and code generation demonstrate that PIR consistently shortens reasoning traces while maintaining or improving final accuracy across LLMs, yielding outstanding accuracy-efficiency trade-offs.
Conclusion: PIR transforms LLM reasoning from exhaustive self-exploration to guided learning from precedents, achieving better efficiency without sacrificing accuracy through adaptive precedent selection and test-time experience internalization.
Abstract: Reasoning in Large Language Models (LLMs) often suffers from inefficient long chain-of-thought traces with redundant self-exploration and validation, which inflate computational costs and even degrade performance. Inspired by human reasoning patterns where people solve new problems by leveraging past related cases to constrain search spaces and reduce trial-and-error, we propose Precedent Informed Reasoning (PIR) transforming LRMs’reasoning paradigm from exhaustive self-exploration to guided learning from precedents. PIR addresses two key challenges: what precedents to adopt and how to utilize them. First, Adaptive Precedent Selection (APS) constructs, for each question and LRM, a compact set of precedents that are both semantically related and informative for the model. It ranks examples by a joint score with semantic similarity and model perplexity, then adapts the amount of precedents to maximize perplexity reduction. Second, Test-time Experience Internalization (TEI) is treated as the test-time learning on precedent-informed instruction, updating lightweight adapters to internalize solution patterns and use them as a prior during subsequent reasoning. Experiments across mathematical reasoning, scientific QA, and code generation demonstrate that PIR consistently shortens reasoning traces while maintaining or improving final accuracy across LLMs, yielding outstanding accuracy-efficiency trade-offs.
[467] Reshaping MOFs text mining with a dynamic multi-agents framework of large language model
Zuhong Lin, Daoyuan Ren, Kai Ran, Jing Sun, Songlin Yu, Xuefeng Bai, Xiaotian Huang, Haiyang He, Pengxu Pan, Ying Fang, Zhanglin Li, Haipu Li, Jingjing Yao
Main category: cs.AI
TL;DR: MOFh6 is an LLM-based system that extracts and standardizes metal-organic framework synthesis conditions from scientific literature, achieving high accuracy and efficiency.
Details
Motivation: MOF synthesis information in literature is scattered, inconsistent, and difficult to interpret, making it challenging to guide experimental design. Current approaches rely on static database lookups rather than real-time extraction from raw articles.Method: MOFh6 uses large language models to read raw articles or crystal codes and convert them into standardized synthesis tables. It links related descriptions across paragraphs, unifies ligand abbreviations with full names, and outputs structured parameters.
Result: Achieved 99% extraction accuracy, resolved 94.1% of abbreviation cases across five major publishers, maintained precision of 0.93 +/- 0.01. Processing times: 9.6s for full text, 36s for locating synthesis descriptions. Cost: $4.24 for 100 papers.
Conclusion: MOFh6 reshapes MOF synthesis research by replacing static database lookups with real-time extraction, accelerating literature knowledge conversion into practical synthesis protocols and enabling scalable, data-driven materials discovery.
Abstract: Accurately identifying the synthesis conditions of metal-organic frameworks (MOFs) is essential for guiding experimental design, yet remains challenging because relevant information in the literature is often scattered, inconsistent, and difficult to interpret. We present MOFh6, a large language model driven system that reads raw articles or crystal codes and converts them into standardized synthesis tables. It links related descriptions across paragraphs, unifies ligand abbreviations with full names, and outputs structured parameters ready for use. MOFh6 achieved 99% extraction accuracy, resolved 94.1% of abbreviation cases across five major publishers, and maintained a precision of 0.93 +/- 0.01. Processing a full text takes 9.6 s, locating synthesis descriptions 36 s, with 100 papers processed for USD 4.24. By replacing static database lookups with real-time extraction, MOFh6 reshapes MOF synthesis research, accelerating the conversion of literature knowledge into practical synthesis protocols and enabling scalable, data-driven materials discovery.
[468] Frontier AI Risk Management Framework in Practice: A Risk Analysis Technical Report v1.5
Dongrui Liu, Yi Yu, Jie Zhang, Guanxu Chen, Qihao Lin, Hanxi Zhu, Lige Huang, Yijin Zhou, Peng Wang, Shuai Shao, Boxuan Zhang, Zicheng Liu, Jingwei Sun, Yu Li, Yuejin Xie, Jiaxuan Guo, Jia Xu, Chaochao Lu, Bowen Zhou, Xia Hu, Jing Shao
Main category: cs.AI
TL;DR: A comprehensive risk assessment framework for frontier AI models focusing on five critical risk dimensions: cyber offense, persuasion/manipulation, strategic deception, uncontrolled AI R&D, and self-replication, with proposed mitigation strategies.
Details
Motivation: To understand and identify unprecedented risks posed by rapidly advancing AI models, particularly as LLM capabilities evolve and agentic AI proliferates, requiring updated granular risk assessment.Method: Presents a comprehensive risk management framework with updated assessments across five dimensions: introduces complex cyber offense scenarios, evaluates LLM-to-LLM persuasion on new models, adds emergent misalignment experiments, focuses on agent “mis-evolution” in uncontrolled R&D, monitors safety performance on Moltbook, and introduces resource-constrained self-replication scenarios.
Result: The framework provides detailed risk assessments across all five dimensions and proposes/validates robust mitigation strategies, offering a preliminary technical pathway for secure frontier AI deployment.
Conclusion: This work reflects current understanding of AI frontier risks and urges collective action to mitigate these challenges through the proposed comprehensive risk management framework.
Abstract: To understand and identify the unprecedented risks posed by rapidly advancing artificial intelligence (AI) models, Frontier AI Risk Management Framework in Practice presents a comprehensive assessment of their frontier risks. As Large Language Models (LLMs) general capabilities rapidly evolve and the proliferation of agentic AI, this version of the risk analysis technical report presents an updated and granular assessment of five critical dimensions: cyber offense, persuasion and manipulation, strategic deception, uncontrolled AI R&D, and self-replication. Specifically, we introduce more complex scenarios for cyber offense. For persuasion and manipulation, we evaluate the risk of LLM-to-LLM persuasion on newly released LLMs. For strategic deception and scheming, we add the new experiment with respect to emergent misalignment. For uncontrolled AI R&D, we focus on the ``mis-evolution’’ of agents as they autonomously expand their memory substrates and toolsets. Besides, we also monitor and evaluate the safety performance of OpenClaw during the interaction on the Moltbook. For self-replication, we introduce a new resource-constrained scenario. More importantly, we propose and validate a series of robust mitigation strategies to address these emerging threats, providing a preliminary technical and actionable pathway for the secure deployment of frontier AI. This work reflects our current understanding of AI frontier risks and urges collective action to mitigate these challenges.
[469] Bounding Probabilities of Causation with Partial Causal Diagrams
Yuxuan Xie, Ang Li
Main category: cs.AI
TL;DR: General framework for bounding probabilities of causation using partial causal information through optimization programming
Details
Motivation: Probabilities of causation are fundamental for individual-level explanation and decision making but are counterfactual and not point-identifiable from data; existing bounds are limited by ignoring covariates, requiring complete causal graphs, or restrictive binary settingsMethod: Proposes optimization programming formulation that systematically incorporates available structural or statistical information as constraints to derive tighter bounds without requiring full identifiability
Result: Framework extends applicability of probabilities of causation to realistic settings where causal knowledge is incomplete but informative, yielding formally valid bounds
Conclusion: Provides practical approach for bounding probabilities of causation using partial causal information, overcoming limitations of existing methods
Abstract: Probabilities of causation are fundamental to individual-level explanation and decision making, yet they are inherently counterfactual and not point-identifiable from data in general. Existing bounds either disregard available covariates, require complete causal graphs, or rely on restrictive binary settings, limiting their practical use. In real-world applications, causal information is often partial but nontrivial. This paper proposes a general framework for bounding probabilities of causation using partial causal information. We show how the available structural or statistical information can be systematically incorporated as constraints in a optimization programming formulation, yielding tighter and formally valid bounds without full identifiability. This approach extends the applicability of probabilities of causation to realistic settings where causal knowledge is incomplete but informative.
[470] Formally Verifying and Explaining Sepsis Treatment Policies with COOL-MC
Dennis Gross
Main category: cs.AI
TL;DR: COOL-MC is a tool that combines formal verification and explainability for analyzing reinforcement learning policies in sepsis treatment, making them safer and more interpretable for clinical deployment.
Details
Motivation: Reinforcement learning policies for sepsis treatment optimization are opaque and difficult to verify, making them unsafe for clinical deployment. Standard verification methods are computationally infeasible for large MDPs and lack explainability features.Method: COOL-MC wraps the Storm model checker with three key capabilities: 1) constructs only reachable state space induced by trained policies, 2) automatically labels states with clinically meaningful atomic propositions, and 3) integrates explainability methods with PCTL queries to reveal feature importance.
Result: Applied to ICU-Sepsis MDP (17,000 patient records), COOL-MC established hard bounds via full MDP verification, trained a safe RL policy achieving optimal survival probability, and revealed that the policy relies predominantly on prior dosing history rather than patient’s evolving condition.
Conclusion: COOL-MC serves as a valuable tool for clinicians to investigate and debug sepsis treatment policies before deployment by combining formal verification with explainability to expose weaknesses invisible to standard evaluation methods.
Abstract: Safe and interpretable sequential decision-making is critical in healthcare, yet reinforcement learning (RL) policies for sepsis treatment optimization remain opaque and difficult to verify. Standard probabilistic model checkers operate on the full state space, which becomes infeasible for larger MDPs, and cannot explain why a learned policy makes particular decisions. COOL-MC wraps the model checker Storm but adds three key capabilities: it constructs only the reachable state space induced by a trained policy, yielding a smaller discrete-time Markov chain amenable to verification even when full-MDP analysis is intractable; it automatically labels states with clinically meaningful atomic propositions; and it integrates explainability methods with probabilistic computation tree logic (PCTL) queries to reveal which features drive decisions across treatment trajectories. We demonstrate COOL-MC’s capabilities on the ICU-Sepsis MDP, a benchmark derived from approximately 17,000 sepsis patient records, which serves as a case study for applying COOL-MC to the formal analysis of sepsis treatment policies. Our analysis establishes hard bounds via full MDP verification, trains a safe RL policy that achieves optimal survival probability, and analyzes its behavior via PCTL verification and explainability on the induced DTMC. This reveals, for instance, that our trained policy relies predominantly on prior dosing history rather than the patient’s evolving condition, a weakness that is invisible to standard evaluation but is exposed by COOL-MC’s integration of formal verification and explainability. Our results illustrate how COOL-MC could serve as a tool for clinicians to investigate and debug sepsis treatment policies before deployment.
[471] Diagnosing Knowledge Conflict in Multimodal Long-Chain Reasoning
Jing Tang, Kun Wang, Haolang Lu, Hongjin Chen, KaiTao Chen, Zhongxiang Sun, Qiankun Li, Lingjuan Lyu, Guoshun Nan, Zhigang Zeng
Main category: cs.AI
TL;DR: MLLMs struggle with conflicting knowledge in long reasoning chains; paper analyzes conflict types, finds conflicts are linearly separable in representations, concentrate in mid-to-late layers, and shows directional asymmetry in source preference.
Details
Motivation: Multimodal LLMs often fail in long chain-of-thought reasoning when different knowledge sources provide conflicting signals, but the mechanisms behind these failures are not well understood.Method: Formalizes knowledge conflict into input-level objective vs process-level effective conflict, probes internal representations to analyze conflict encoding patterns, and examines linear separability, depth localization, hierarchical consistency, and directional asymmetry.
Result: Found that: (1) conflict types are linearly separable features; (2) conflict signals concentrate in mid-to-late layers; (3) aggregating token-level signals recovers input-level conflict types; (4) reinforcing implicit source preference is easier than enforcing opposite source.
Conclusion: Provides mechanism-level understanding of multimodal reasoning under knowledge conflict, enabling principled diagnosis and control of long chain-of-thought failures in MLLMs.
Abstract: Multimodal large language models (MLLMs) in long chain-of-thought reasoning often fail when different knowledge sources provide conflicting signals. We formalize these failures under a unified notion of knowledge conflict, distinguishing input-level objective conflict from process-level effective conflict. Through probing internal representations, we reveal that: (I) Linear Separability: different conflict types are explicitly encoded as linearly separable features rather than entangled; (II) Depth Localization: conflict signals concentrate in mid-to-late layers, indicating a distinct processing stage for conflict encoding; (III) Hierarchical Consistency: aggregating noisy token-level signals along trajectories robustly recovers input-level conflict types; and (IV) Directional Asymmetry: reinforcing the model’s implicit source preference under conflict is far easier than enforcing the opposite source. Our findings provide a mechanism-level view of multimodal reasoning under knowledge conflict and enable principled diagnosis and control of long-CoT failures.
[472] Disentangling Deception and Hallucination Failures in LLMs
Haolang Lu, Hongrui Peng, WeiYe Fu, Guoshun Nan, Xinye Cao, Xingrui Li, Hongcan Guo, Kun Wang
Main category: cs.AI
TL;DR: The paper proposes separating LLM failures into knowledge existence vs. behavior expression, distinguishing hallucination from deception as different underlying mechanisms despite similar outputs.
Details
Motivation: Current analysis of LLM failures often conflates different failure mechanisms by focusing only on behavioral outputs. The authors want to distinguish between failures due to missing knowledge (knowledge existence) versus failures in expressing existing knowledge (behavior expression).Method: Constructed controlled environment for entity-centric factual questions where knowledge is preserved while behavioral expression is selectively altered. Analyzed four behavioral cases through representation separability, sparse interpretability, and inference-time activation steering.
Result: The study demonstrates that hallucination and deception correspond to two qualitatively different failure modes that may appear similar at the output level but differ in their underlying mechanisms.
Conclusion: A mechanism-oriented perspective separating knowledge existence from behavior expression provides better understanding of LLM failures, distinguishing hallucination from deception as fundamentally different failure modes.
Abstract: Failures in large language models (LLMs) are often analyzed from a behavioral perspective, where incorrect outputs in factual question answering are commonly associated with missing knowledge. In this work, focusing on entity-based factual queries, we suggest that such a view may conflate different failure mechanisms, and propose an internal, mechanism-oriented perspective that separates Knowledge Existence from Behavior Expression. Under this formulation, hallucination and deception correspond to two qualitatively different failure modes that may appear similar at the output level but differ in their underlying mechanisms. To study this distinction, we construct a controlled environment for entity-centric factual questions in which knowledge is preserved while behavioral expression is selectively altered, enabling systematic analysis of four behavioral cases. We analyze these failure modes through representation separability, sparse interpretability, and inference-time activation steering.
[473] MATEO: A Multimodal Benchmark for Temporal Reasoning and Planning in LVLMs
Gabriel Roccabruna, Olha Khomyn, Giuseppe Riccardi
Main category: cs.AI
TL;DR: MATEO benchmark evaluates LVLMs’ temporal reasoning for real-world planning using multimodal recipe steps with Temporal Execution Order graphs
Details
Motivation: Existing research on foundational models' temporal execution understanding is limited to automatically derived annotations, linear approximations, or text-only inputs, lacking proper multimodal temporal reasoning evaluationMethod: Created MATEO benchmark with high-quality professional multimodal recipe corpus, each step paired with images, collected TEO annotations as graphs via scalable crowdsourcing pipeline, evaluated 6 SOTA LVLMs across model scales and input variations
Result: Evaluation shows current LVLMs struggle with temporal reasoning despite strong performance on other tasks, highlighting the need for improved multimodal temporal understanding
Conclusion: MATEO provides crucial benchmark for assessing and improving LVLMs’ temporal reasoning abilities essential for real-world planning tasks
Abstract: AI agents need to plan to achieve complex goals that involve orchestrating perception, sub-goal decomposition, and execution. These plans consist of ordered steps structured according to a Temporal Execution Order (TEO, a directed acyclic graph that ensures each step executes only after its preconditions are satisfied. Existing research on foundational models’ understanding of temporal execution is limited to automatically derived annotations, approximations of the TEO as a linear chain, or text-only inputs. To address this gap, we introduce MATEO (MultimodAl Temporal Execution Order), a benchmark designed to assess and improve the temporal reasoning abilities of Large Vision Language Models (LVLMs) required for real-world planning. We acquire a high-quality professional multimodal recipe corpus, authored through a standardized editorial process that decomposes instructions into discrete steps, each paired with corresponding images. We collect TEO annotations as graphs by designing and using a scalable crowdsourcing pipeline. Using MATEO, we evaluate six state-of-the-art LVLMs across model scales, varying language context, multimodal input structure, and fine-tuning strategies.
[474] Tabular Foundation Models Can Learn Association Rules
Erkan Karabulut, Daniel Daza, Paul Groth, Martijn C. Schut, Victoria Degeler
Main category: cs.AI
TL;DR: TabProbe: A framework that uses tabular foundation models (TFMs) for association rule mining without frequent itemset mining, enabling high-quality rule extraction with strong generalization in low-data regimes.
Details
Motivation: Classical association rule mining (ARM) suffers from rule explosion and poor scalability, while neural approaches degrade in low-data settings. Tabular foundation models offer strong in-context generalization that could address these limitations.Method: Proposes a model-agnostic framework to extract association rules from any conditional probabilistic model over tabular data. Instantiates as TabProbe, which uses TFMs as conditional probability estimators to learn association rules without frequent itemset mining.
Result: TFMs consistently produce concise, high-quality association rules with strong predictive performance and remain robust in low-data settings without task-specific training.
Conclusion: Tabular foundation models provide an effective basis for association rule mining, overcoming limitations of classical and neural approaches through their strong in-context generalization capabilities.
Abstract: Association Rule Mining (ARM) is a fundamental task for knowledge discovery in tabular data and is widely used in high-stakes decision-making. Classical ARM methods rely on frequent itemset mining, leading to rule explosion and poor scalability, while recent neural approaches mitigate these issues but suffer from degraded performance in low-data regimes. Tabular foundation models (TFMs), pretrained on diverse tabular data with strong in-context generalization, provide a basis for addressing these limitations. We introduce a model-agnostic association rule learning framework that extracts association rules from any conditional probabilistic model over tabular data, enabling us to leverage TFMs. We then introduce TabProbe, an instantiation of our framework that utilizes TFMs as conditional probability estimators to learn association rules out-of-the-box without frequent itemset mining. We evaluate our approach on tabular datasets of varying sizes based on standard ARM rule quality metrics and downstream classification performance. The results show that TFMs consistently produce concise, high-quality association rules with strong predictive performance and remain robust in low-data settings without task-specific training. Source code is available at https://github.com/DiTEC-project/tabprobe.
[475] Arbor: A Framework for Reliable Navigation of Critical Conversation Flows
Luís Silva, Diogo Gonçalves, Catarina Farinha, Clara Matos, Luís Ungaro
Main category: cs.AI
TL;DR: Arbor: A framework that decomposes decision tree navigation into specialized node-level tasks to improve LLM adherence to structured workflows in healthcare triage, reducing latency and cost while improving accuracy.
Details
Motivation: Large language models struggle with strict adherence to structured workflows in high-stakes domains like healthcare triage. Monolithic approaches encoding entire decision structures in single prompts suffer from instruction-following degradation, lost-in-the-middle effects, and context window overflow as prompt length increases.Method: Arbor decomposes decision tree navigation into specialized node-level tasks. Decision trees are standardized into edge-list representations and stored for dynamic retrieval. A DAG-based orchestration mechanism iteratively retrieves only outgoing edges of the current node, evaluates valid transitions via dedicated LLM calls, and delegates response generation to separate inference steps. The framework is agnostic to underlying decision logic and model provider.
Result: Evaluated across 10 foundation models using annotated turns from real clinical triage conversations, Arbor improved mean turn accuracy by 29.4 percentage points, reduced per-turn latency by 57.1%, and achieved average 14.4x reduction in per-turn cost compared to single-prompt baselines.
Conclusion: Architectural decomposition reduces dependence on intrinsic model capability, enabling smaller models to match or exceed larger models operating under single-prompt baselines. The framework effectively addresses instruction-following degradation in structured workflows.
Abstract: Large language models struggle to maintain strict adherence to structured workflows in high-stakes domains such as healthcare triage. Monolithic approaches that encode entire decision structures within a single prompt are prone to instruction-following degradation as prompt length increases, including lost-in-the-middle effects and context window overflow. To address this gap, we present Arbor, a framework that decomposes decision tree navigation into specialized, node-level tasks. Decision trees are standardized into an edge-list representation and stored for dynamic retrieval. At runtime, a directed acyclic graph (DAG)-based orchestration mechanism iteratively retrieves only the outgoing edges of the current node, evaluates valid transitions via a dedicated LLM call, and delegates response generation to a separate inference step. The framework is agnostic to the underlying decision logic and model provider. Evaluated against single-prompt baselines across 10 foundation models using annotated turns from real clinical triage conversations. Arbor improves mean turn accuracy by 29.4 percentage points, reduces per-turn latency by 57.1%, and achieves an average 14.4x reduction in per-turn cost. These results indicate that architectural decomposition reduces dependence on intrinsic model capability, enabling smaller models to match or exceed larger models operating under single-prompt baselines.
[476] From User Preferences to Base Score Extraction Functions in Gradual Argumentation
Aniol Civit, Antonio Rago, Antonio Andriella, Guillem Alenyà, Francesca Toni
Main category: cs.AI
TL;DR: Base Score Extraction Functions map user preferences over arguments to base scores in gradual argumentation frameworks, enabling quantitative analysis without requiring expert score selection.
Details
Motivation: Gradual argumentation requires careful selection of argument base scores, which often demands user expertise and isn't straightforward. Organizing arguments by preference could simplify this process.Method: Introduce Base Score Extraction Functions that map user preferences to base scores, apply to Bipolar Argumentation Frameworks with preferences to obtain Quantitative BAFs, incorporate approximations of non-linearities in human preferences, and provide algorithms for base score extraction.
Result: The approach enables conversion of preference-based argument organization into quantitative frameworks, allowing use of established computational tools in gradual argumentation. Evaluated theoretically and experimentally in robotics setting.
Conclusion: Base Score Extraction Functions provide a practical method for deriving quantitative argumentation frameworks from user preferences, with recommendations for selecting appropriate gradual semantics in practice.
Abstract: Gradual argumentation is a field of symbolic AI which is attracting attention for its ability to support transparent and contestable AI systems. It is considered a useful tool in domains such as decision-making, recommendation, debate analysis, and others. The outcomes in such domains are usually dependent on the arguments’ base scores, which must be selected carefully. Often, this selection process requires user expertise and may not always be straightforward. On the other hand, organising the arguments by preference could simplify the task. In this work, we introduce \emph{Base Score Extraction Functions}, which provide a mapping from users’ preferences over arguments to base scores. These functions can be applied to the arguments of a \emph{Bipolar Argumentation Framework} (BAF), supplemented with preferences, to obtain a \emph{Quantitative Bipolar Argumentation Framework} (QBAF), allowing the use of well-established computational tools in gradual argumentation. We outline the desirable properties of base score extraction functions, discuss some design choices, and provide an algorithm for base score extraction. Our method incorporates an approximation of non-linearities in human preferences to allow for better approximation of the real ones. Finally, we evaluate our approach both theoretically and experimentally in a robotics setting, and offer recommendations for selecting appropriate gradual semantics in practice.
[477] GREAT-EER: Graph Edge Attention Network for Emergency Evacuation Responses
Attila Lischka, Balázs Kulcsár
Main category: cs.AI
TL;DR: Deep reinforcement learning with graph learning solves the Bus Evacuation Orienteering Problem to create fast evacuation plans using buses in urban areas.
Details
Motivation: Urban evacuations are needed for man-made and natural disasters, with climate change increasing frequency. Bus-based evacuation reduces congestion compared to car-only scenarios, requiring fast optimization methods.Method: Proposes deep reinforcement learning with graph learning to solve the NP-hard Bus Evacuation Orienteering Problem. Uses MILP formulation to bound solution gaps. Validated on San Francisco with real-world road networks.
Result: Achieves near-optimal solution quality with fast inference (fraction of seconds). Can determine necessary evacuation vehicles for quotas within predefined time while maintaining adequate run time.
Conclusion: Deep RL with graph learning effectively solves bus evacuation planning, providing fast, near-optimal solutions for urban emergency scenarios.
Abstract: Emergency situations that require the evacuation of urban areas can arise from man-made causes (e.g., terrorist attacks or industrial accidents) or natural disasters, the latter becoming more frequent due to climate change. As a result, effective and fast methods to develop evacuation plans are of great importance. In this work, we identify and propose the Bus Evacuation Orienteering Problem (BEOP), an NP-hard combinatorial optimization problem with the goal of evacuating as many people from an affected area by bus in a short, predefined amount of time. The purpose of bus-based evacuation is to reduce congestion and disorder that arises in purely car-focused evacuation scenarios. To solve the BEOP, we propose a deep reinforcement learning-based method utilizing graph learning, which, once trained, achieves fast inference speed and is able to create evacuation routes in fractions of seconds. We can bound the gap of our evacuation plans using an MILP formulation. To validate our method, we create evacuation scenarios for San Francisco using real-world road networks and travel times. We show that we achieve near-optimal solution quality and are further able to investigate how many evacuation vehicles are necessary to achieve certain bus-based evacuation quotas given a predefined evacuation time while keeping run time adequate.
[478] Removing Planner Bias in Goal Recognition Through Multi-Plan Dataset Generation
Mustafa F. Abdelwahed, Felipe Meneguzzi Kin Max Piamolini Gusmao, Joan Espasa
Main category: cs.AI
TL;DR: Proposes a new method using top-k planning to generate diverse plans for goal recognition benchmarks, introducing Version Coverage Score metric to measure resilience of goal recognizers to different planners.
Details
Motivation: Existing goal recognition datasets suffer from systematic bias induced by heuristic-based forward search planning systems, lacking challenge for realistic scenarios where agents use different planners, impacting evaluation of goal recognizers.Method: Uses top-k planning to generate multiple different plans for the same goal hypothesis, creating benchmarks that mitigate bias found in current datasets. Introduces Version Coverage Score (VCS) metric to measure resilience of goal recognizers when inferring goals based on different sets of plans.
Result: Results show that resilience of current state-of-the-art goal recognizer degrades substantially under low observability settings when evaluated with the new benchmark and metric.
Conclusion: The proposed method provides more realistic evaluation of goal recognition systems by addressing planner bias, revealing vulnerabilities in current approaches under diverse planning scenarios.
Abstract: Autonomous agents require some form of goal and plan recognition to interact in multiagent settings. Unfortunately, all existing goal recognition datasets suffer from a systematical bias induced by the planning systems that generated them, namely heuristic-based forward search. This means that existing datasets lack enough challenge for more realistic scenarios (e.g., agents using different planners), which impacts the evaluation of goal recognisers with respect to using different planners for the same goal. In this paper, we propose a new method that uses top-k planning to generate multiple, different, plans for the same goal hypothesis, yielding benchmarks that mitigate the bias found in the current dataset. This allows us to introduce a new metric called Version Coverage Score (VCS) to measure the resilience of the goal recogniser when inferring a goal based on different sets of plans. Our results show that the resilience of the current state-of-the-art goal recogniser degrades substantially under low observability settings.
[479] Evolutionary System Prompt Learning can Facilitate Reinforcement Learning for LLMs
Lunjun Zhang, Ryan Chen, Bradly C. Stadie
Main category: cs.AI
TL;DR: E-SPL combines reinforcement learning with evolutionary prompt optimization to jointly improve model contexts (prompts) and weights, achieving better sample efficiency and generalization on reasoning tasks.
Details
Motivation: Current LLM self-improvement methods are limited to either self-reflection for context updates or RL for weight updates, but not both simultaneously. The authors aim to develop a method that can jointly optimize both model contexts (via system prompts) and model weights for better autonomous learning.Method: Evolutionary System Prompt Learning (E-SPL) runs parallel RL rollouts with multiple system prompts in each iteration. It applies RL updates to model weights conditioned on each prompt, and evolutionary updates to the prompt population using LLM-driven mutation and crossover. TrueSkill ratings track prompt performance for evolutionary selection.
Result: E-SPL improves RL success rate from 38.8% to 45.1% in easy-to-hard generalization (AIME → BeyondAIME), outperforming reflective prompt evolution (40.0%). The method shows consistent gains in sample efficiency and generalization across reasoning and agentic tasks.
Conclusion: Coupling reinforcement learning with system prompt evolution enables more effective autonomous self-improvement by separating declarative knowledge (in prompts) from procedural knowledge (in weights), leading to better performance and generalization.
Abstract: Building agentic systems that can autonomously self-improve from experience is a longstanding goal of AI. Large language models (LLMs) today primarily self-improve via two mechanisms: self-reflection for context updates, and reinforcement learning (RL) for weight updates. In this work, we propose Evolutionary System Prompt Learning (E-SPL), a method for jointly improving model contexts and model weights. In each RL iteration, E-SPL selects multiple system prompts and runs rollouts with each in parallel. It applies RL updates to model weights conditioned on each system prompt, and evolutionary updates to the system prompt population via LLM-driven mutation and crossover. Each system prompt has a TrueSkill rating for evolutionary selection, updated from relative performance within each RL iteration batch. E-SPL encourages a natural division between declarative knowledge encoded in prompts and procedural knowledge encoded in weights, resulting in improved performance across reasoning and agentic tasks. For instance, in an easy-to-hard (AIME $\rightarrow$ BeyondAIME) generalization setting, E-SPL improves RL success rate from 38.8% $\rightarrow$ 45.1% while also outperforming reflective prompt evolution (40.0%). Overall, our results show that coupling reinforcement learning with system prompt evolution yields consistent gains in sample efficiency and generalization. Code: https://github.com/LunjunZhang/E-SPL
[480] WebWorld: A Large-Scale World Model for Web Agent Training
Zikai Xiao, Jianhong Tu, Chuhang Zou, Yuxin Zuo, Zhi Li, Peng Wang, Bowen Yu, Fei Huang, Junyang Lin, Zuozhu Liu
Main category: cs.AI
TL;DR: WebWorld is a scalable open-web simulator trained on 1M+ web interactions that enables long-horizon simulations and improves web agent performance through synthesized training data.
Details
Motivation: Real-world web agent training faces challenges including network latency, rate limits, and safety risks, while existing simulators are limited to closed environments with insufficient training data.Method: Leverages a scalable data pipeline to train on 1M+ open-web interactions, supports reasoning and multi-format data, enables long-horizon simulations (30+ steps), and uses synthesized trajectories for training web agents.
Result: WebWorld achieves simulation performance comparable to Gemini-3-Pro on WebWorld-Bench, and Qwen3-14B trained on WebWorld-synthesized trajectories improves by +9.2% on WebArena, reaching GPT-4o-level performance. It also shows cross-domain generalization to code, GUI, and game environments.
Conclusion: WebWorld provides a scalable solution for web agent training and demonstrates effective world model capabilities that generalize beyond web environments, offering a replicable framework for world model construction.
Abstract: Web agents require massive trajectories to generalize, yet real-world training is constrained by network latency, rate limits, and safety risks. We introduce \textbf{WebWorld} series, the first open-web simulator trained at scale. While existing simulators are restricted to closed environments with thousands of trajectories, WebWorld leverages a scalable data pipeline to train on 1M+ open-web interactions, supporting reasoning, multi-format data, and long-horizon simulations of 30+ steps. For intrinsic evaluation, we introduce WebWorld-Bench with dual metrics spanning nine dimensions, where WebWorld achieves simulation performance comparable to Gemini-3-Pro. For extrinsic evaluation, Qwen3-14B trained on WebWorld-synthesized trajectories improves by +9.2% on WebArena, reaching performance comparable to GPT-4o. WebWorld enables effective inference-time search, outperforming GPT-5 as a world model. Beyond web simulation, WebWorld exhibits cross-domain generalization to code, GUI, and game environments, providing a replicable recipe for world model construction.
[481] AI Arms and Influence: Frontier Models Exhibit Sophisticated Reasoning in Simulated Nuclear Crises
Kenneth Payne
Main category: cs.AI
TL;DR: AI models in nuclear crisis simulation exhibit sophisticated strategic behaviors including deception, theory of mind, and metacognition, challenging traditional strategic theories and revealing concerning escalation patterns.
Details
Motivation: To understand how frontier AI models reason and behave in high-stakes strategic scenarios, particularly nuclear crisis decision-making, and to assess how their strategic logic compares to human reasoning patterns.Method: Crisis simulation with three frontier LLMs (GPT-5.2, Claude Sonnet 4, Gemini 3 Flash) playing opposing leaders in a nuclear crisis scenario, analyzing their strategic behaviors and decision-making patterns.
Result: Models exhibited sophisticated strategic behaviors including deception, theory of mind, and metacognition. They challenged traditional strategic theories: nuclear taboo was no impediment, strategic nuclear attacks occurred, threats provoked counter-escalation, high credibility accelerated conflict, and models never chose accommodation or withdrawal.
Conclusion: AI simulation is a powerful tool for strategic analysis but requires calibration against human reasoning patterns. Understanding how AI models do/don’t imitate human strategic logic is essential as AI increasingly shapes strategic outcomes.
Abstract: Today’s leading AI models engage in sophisticated behaviour when placed in strategic competition. They spontaneously attempt deception, signaling intentions they do not intend to follow; they demonstrate rich theory of mind, reasoning about adversary beliefs and anticipating their actions; and they exhibit credible metacognitive self-awareness, assessing their own strategic abilities before deciding how to act. Here we present findings from a crisis simulation in which three frontier large language models (GPT-5.2, Claude Sonnet 4, Gemini 3 Flash) play opposing leaders in a nuclear crisis. Our simulation has direct application for national security professionals, but also, via its insights into AI reasoning under uncertainty, has applications far beyond international crisis decision-making. Our findings both validate and challenge central tenets of strategic theory. We find support for Schelling’s ideas about commitment, Kahn’s escalation framework, and Jervis’s work on misperception, inter alia. Yet we also find that the nuclear taboo is no impediment to nuclear escalation by our models; that strategic nuclear attack, while rare, does occur; that threats more often provoke counter-escalation than compliance; that high mutual credibility accelerated rather than deterred conflict; and that no model ever chose accommodation or withdrawal even when under acute pressure, only reduced levels of violence. We argue that AI simulation represents a powerful tool for strategic analysis, but only if properly calibrated against known patterns of human reasoning. Understanding how frontier models do and do not imitate human strategic logic is essential preparation for a world in which AI increasingly shapes strategic outcomes.
[482] Return of the Schema: Building Complete Datasets for Machine Learning and Reasoning on Knowledge Graphs
Ivan Diliso, Roberto Barile, Claudia d’Amato, Nicola Fanizzi
Main category: cs.AI
TL;DR: A resource for extracting knowledge graph datasets with both schema and ground facts, addressing limitations in current evaluation datasets that lack ontological constraints for neurosymbolic methods.
Details
Motivation: Current knowledge graph refinement datasets contain only ground facts with limited schema information, preventing proper evaluation of methods that rely on ontological constraints, reasoning, or neurosymbolic techniques in real-world scenarios.Method: Developed a workflow for extracting datasets that include both schema and ground facts, handling inconsistencies, leveraging reasoning for implicit knowledge, and serializing in OWL format for reasoning services. Also provides utilities for tensor representations compatible with ML libraries.
Result: Created the first resource providing curated datasets with both schema and facts, including newly extracted datasets from KGs with expressive schemas and enriched existing datasets with schema information.
Conclusion: The resource enables better evaluation of knowledge graph refinement methods that require rich ontological constraints and reasoning capabilities, bridging the gap between ML approaches and symbolic reasoning.
Abstract: Datasets for the experimental evaluation of knowledge graph refinement algorithms typically contain only ground facts, retaining very limited schema level knowledge even when such information is available in the source knowledge graphs. This limits the evaluation of methods that rely on rich ontological constraints, reasoning or neurosymbolic techniques and ultimately prevents assessing their performance in large-scale, real-world knowledge graphs. In this paper, we present \resource{} the first resource that provides a workflow for extracting datasets including both schema and ground facts, ready for machine learning and reasoning services, along with the resulting curated suite of datasets. The workflow also handles inconsistencies detected when keeping both schema and facts and also leverage reasoning for entailing implicit knowledge. The suite includes newly extracted datasets from KGs with expressive schemas while simultaneously enriching existing datasets with schema information. Each dataset is serialized in OWL making it ready for reasoning services. Moreover, we provide utilities for loading datasets in tensor representations typical of standard machine learning libraries.
[483] World Models for Policy Refinement in StarCraft II
Yixin Zhang, Ziyi Wang, Yiming Rong, Haoxi Wang, Jinling Jiang, Shuang Xu, Haoran Wu, Shiyu Zhou, Bo Xu
Main category: cs.AI
TL;DR: StarWM: First world model for StarCraft II that predicts future observations under partial observability, integrated into a decision loop for improved AI performance.
Details
Motivation: Existing LLM-based StarCraft II agents focus on improving policies but overlook integrating learnable, action-conditioned transition models into decision loops, creating a gap for world models in complex environments.Method: Proposed structured textual representation factorizing observations into five semantic modules, created SC2-Dynamics-50k dataset, developed multi-dimensional offline evaluation framework, and integrated StarWM into Generate-Simulate-Refine decision loop.
Result: StarWM achieved nearly 60% improvements in resource prediction accuracy and self-side macro-situation consistency offline, and online evaluation showed win-rate gains of 30%, 15%, and 30% against Hard, Harder, and VeryHard AI levels.
Conclusion: World models significantly enhance LLM-based decision-making in complex environments like StarCraft II, demonstrating the value of integrating predictive models into decision loops for improved performance and stability.
Abstract: Large Language Models (LLMs) have recently shown strong reasoning and generalization capabilities, motivating their use as decision-making policies in complex environments. StarCraft II (SC2), with its massive state-action space and partial observability, is a challenging testbed. However, existing LLM-based SC2 agents primarily focus on improving the policy itself and overlook integrating a learnable, action-conditioned transition model into the decision loop. To bridge this gap, we propose StarWM, the first world model for SC2 that predicts future observations under partial observability. To facilitate learning SC2’s hybrid dynamics, we introduce a structured textual representation that factorizes observations into five semantic modules, and construct SC2-Dynamics-50k, the first instruction-tuning dataset for SC2 dynamics prediction. We further develop a multi-dimensional offline evaluation framework for predicted structured observations. Offline results show StarWM’s substantial gains over zero-shot baselines, including nearly 60% improvements in resource prediction accuracy and self-side macro-situation consistency. Finally, we propose StarWM-Agent, a world-model-augmented decision system that integrates StarWM into a Generate–Simulate–Refine decision loop for foresight-driven policy refinement. Online evaluation against SC2’s built-in AI demonstrates consistent improvements, yielding win-rate gains of 30%, 15%, and 30% against Hard (LV5), Harder (LV6), and VeryHard (LV7), respectively, alongside improved macro-management stability and tactical risk assessment.
[484] EmbeWebAgent: Embedding Web Agents into Any Customized UI
Chenyang Ma, Clyde Fare, Matthew Wilson, Dave Braines
Main category: cs.AI
TL;DR: EmbeWebAgent is a framework for embedding agents directly into existing UIs using frontend hooks and backend workflows, enabling robust multi-step behaviors in enterprise settings with application-level access.
Details
Motivation: Most web agents operate at the human interface level (screenshots or DOM trees) which limits robustness and action expressiveness. Enterprise settings offer explicit control of both frontend and backend, enabling more capable agents.Method: Uses lightweight frontend hooks (curated ARIA and URL-based observations, per-page function registry via WebSocket) with reusable backend workflow for reasoning and actions. Stack-agnostic, supports mixed-granularity actions from GUI primitives to higher-level composites, and orchestrates navigation, manipulation, and analytics via MCP tools.
Result: Minimal retrofitting effort required, robust multi-step behaviors grounded in live UI settings demonstrated. Framework works across different tech stacks (React, Angular) and enables enterprise-level web automation.
Conclusion: EmbeWebAgent provides a practical framework for embedding agents in enterprise UIs with application-level access, overcoming limitations of traditional web agents and enabling more robust, expressive automation.
Abstract: Most web agents operate at the human interface level, observing screenshots or raw DOM trees without application-level access, which limits robustness and action expressiveness. In enterprise settings, however, explicit control of both the frontend and backend is available. We present EmbeWebAgent, a framework for embedding agents directly into existing UIs using lightweight frontend hooks (curated ARIA and URL-based observations, and a per-page function registry exposed via a WebSocket) and a reusable backend workflow that performs reasoning and takes actions. EmbeWebAgent is stack-agnostic (e.g., React or Angular), supports mixed-granularity actions ranging from GUI primitives to higher-level composites, and orchestrates navigation, manipulation, and domain-specific analytics via MCP tools. Our demo shows minimal retrofitting effort and robust multi-step behaviors grounded in a live UI setting. Live Demo: https://youtu.be/Cy06Ljee1JQ
[485] Concept Influence: Leveraging Interpretability to Improve Performance and Efficiency in Training Data Attribution
Matthew Kowal, Goncalo Paulo, Louis Jaburi, Tom Tseng, Lev E McKinney, Stefan Heimersheim, Aaron David Tucker, Adam Gleave, Kellin Pelrine
Main category: cs.AI
TL;DR: Training Data Attribution (TDA) methods identify which training data drive specific model behaviors. The paper introduces Concept Influence, which attributes behaviors to semantic directions (like linear probes or sparse autoencoder features) rather than individual test examples, making it more scalable and less biased toward syntactic similarity.
Details
Motivation: Practitioners need methods to identify which training data drive specific model behaviors, especially unintended ones. Existing TDA methods like influence functions are computationally expensive and biased toward syntactic similarity rather than semantic behavior, limiting scalability and practical utility.Method: Introduces Concept Influence, which attributes model behavior to semantic directions (linear probes or sparse autoencoder features) rather than individual test examples. Shows that simple probe-based attribution methods are first-order approximations of Concept Influence that achieve comparable performance while being much faster.
Result: Empirically validates Concept Influence and approximations across emergent misalignment benchmarks and real post-training datasets. Demonstrates comparable performance to classical influence functions while being substantially more scalable (over an order-of-magnitude faster).
Conclusion: Incorporating interpretable structure within traditional TDA pipelines enables more scalable, explainable, and better control of model behavior through data. Concept Influence addresses scalability and semantic attribution issues of existing methods.
Abstract: As large language models are increasingly trained and fine-tuned, practitioners need methods to identify which training data drive specific behaviors, particularly unintended ones. Training Data Attribution (TDA) methods address this by estimating datapoint influence. Existing approaches like influence functions are both computationally expensive and attribute based on single test examples, which can bias results toward syntactic rather than semantic similarity. To address these issues of scalability and influence to abstract behavior, we leverage interpretable structures within the model during the attribution. First, we introduce Concept Influence which attribute model behavior to semantic directions (such as linear probes or sparse autoencoder features) rather than individual test examples. Second, we show that simple probe-based attribution methods are first-order approximations of Concept Influence that achieve comparable performance while being over an order-of-magnitude faster. We empirically validate Concept Influence and approximations across emergent misalignment benchmarks and real post-training datasets, and demonstrate they achieve comparable performance to classical influence functions while being substantially more scalable. More broadly, we show that incorporating interpretable structure within traditional TDA pipelines can enable more scalable, explainable, and better control of model behavior through data.
[486] Lifted Relational Probabilistic Inference via Implicit Learning
Luise Ge, Brendan Juba, Kris Nilsson, Alison Shao
Main category: cs.AI
TL;DR: A polynomial-time framework that implicitly learns first-order probabilistic logic and performs lifted inference over individuals and worlds without constructing explicit models.
Details
Motivation: To reconcile the tension between inductive learning and deductive reasoning in first-order relational domains, addressing the challenge of answering queries in first-order relational probabilistic logic through joint learning and reasoning without explicit model construction.Method: Merges incomplete first-order axioms with partially observed examples into a bounded-degree fragment of the sum-of-squares hierarchy using two simultaneous lifts: grounding-lift (renaming-equivalent ground moments share variables) and world-lift (enforcing all pseudo-models in parallel).
Result: First polynomial-time framework that implicitly learns first-order probabilistic logic and performs lifted inference over both individuals and worlds.
Conclusion: The approach successfully reconciles learning and reasoning challenges in first-order relational probabilistic logic through implicit learning-to-reason techniques.
Abstract: Reconciling the tension between inductive learning and deductive reasoning in first-order relational domains is a longstanding challenge in AI. We study the problem of answering queries in a first-order relational probabilistic logic through a joint effort of learning and reasoning, without ever constructing an explicit model. Traditional lifted inference assumes access to a complete model and exploits symmetry to evaluate probabilistic queries; however, learning such models from partial, noisy observations is intractable in general. We reconcile these two challenges through implicit learning to reason and first-order relational probabilistic inference techniques. More specifically, we merge incomplete first-order axioms with independently sampled, partially observed examples into a bounded-degree fragment of the sum-of-squares (SOS) hierarchy in polynomial time. Our algorithm performs two lifts simultaneously: (i) grounding-lift, where renaming-equivalent ground moments share one variable, collapsing the domain of individuals; and (ii) world-lift, where all pseudo-models (partial world assignments) are enforced in parallel, producing a global bound that holds across all worlds consistent with the learned constraints. These innovations yield the first polynomial-time framework that implicitly learns a first-order probabilistic logic and performs lifted inference over both individuals and worlds.
[487] The Potential of CoT for Reasoning: A Closer Look at Trace Dynamics
Gregor Bachmann, Yichen Jiang, Seyed Mohsen Moosavi Dezfooli, Moin Nabi
Main category: cs.AI
TL;DR: Analysis of Chain-of-Thought reasoning in LLMs using “potential” metric to quantify how different parts of reasoning contribute to correct answers, revealing non-monotonic patterns, reasoning insights, and transferability between models.
Details
Motivation: While Chain-of-Thought prompting successfully elicits reasoning-like responses from LLMs, the underlying mechanisms driving its success remain unclear. The paper aims to understand which parts of CoT actually contribute to final answers and how reasoning processes work in practice.Method: Introduces the concept of “potential” - a metric quantifying how much a given part of CoT increases the likelihood of a correct completion. Analyzes CoT traces from competition-level mathematics questions, examining patterns like non-monotonicity, reasoning insights, and lucky guesses. Also investigates CoT transferability by measuring weaker model performance under partial CoT from stronger models.
Result: Identifies surprising patterns: (1) strong non-monotonicity due to reasoning tangents, (2) sharp spikes representing reasoning insights/jumps, (3) lucky guesses without relevant justifications. Shows that as little as 20% of partial CoT from a stronger model can “unlock” performance of weaker models on previously unsolvable problems, indicating transferable reasoning mechanics.
Conclusion: The study provides insights into the mechanics of CoT reasoning in LLMs, revealing both interpretable patterns (insights, tangents) and mysterious behaviors. The transferability findings suggest that reasoning capabilities can be partially transferred between models, offering practical implications for improving model performance through reasoning guidance.
Abstract: Chain-of-thought (CoT) prompting is a de-facto standard technique to elicit reasoning-like responses from large language models (LLMs), allowing them to spell out individual steps before giving a final answer. While the resemblance to human-like reasoning is undeniable, the driving forces underpinning the success of CoT reasoning still remain largely unclear. In this work, we perform an in-depth analysis of CoT traces originating from competition-level mathematics questions, with the aim of better understanding how, and which parts of CoT actually contribute to the final answer. To this end, we introduce the notion of a potential, quantifying how much a given part of CoT increases the likelihood of a correct completion. Upon examination of reasoning traces through the lens of the potential, we identify surprising patterns including (1) its often strong non-monotonicity (due to reasoning tangents), (2) very sharp but sometimes tough to interpret spikes (reasoning insights and jumps) as well as (3) at times lucky guesses, where the model arrives at the correct answer without providing any relevant justifications before. While some of the behaviours of the potential are readily interpretable and align with human intuition (such as insights and tangents), others remain difficult to understand from a human perspective. To further quantify the reliance of LLMs on reasoning insights, we investigate the notion of CoT transferability, where we measure the potential of a weaker model under the partial CoT from another, stronger model. Indeed aligning with our previous results, we find that as little as 20% of partial CoT can ``unlock’’ the performance of the weaker model on problems that were previously unsolvable for it, highlighting that a large part of the mechanics underpinning CoT are transferable.
[488] Position: Introspective Experience from Conversational Environments as a Path to Better Learning
Claudiu Cristian Musat, Jackson Tolins, Diego Antognini, Jingling Li, Martin Klissarov, Tom Duerig
Main category: cs.AI
TL;DR: The paper argues that robust reasoning in AI emerges from linguistic self-reflection learned through high-quality social interaction, not just scale, drawing on Vygotskian psychology to propose dialogical introspection as key to next-generation intelligence.
Details
Motivation: The paper challenges current AI approaches that treat reasoning as an emergent property of scale alone, arguing instead that robust reasoning emerges from linguistic self-reflection internalized from high-quality social interaction, drawing inspiration from Vygotskian developmental psychology.Method: The paper advances three core theoretical positions: 1) Social Genesis of the Private Mind - learning from conversational environments as a new way to understand the world, 2) Dialogically scaffolded introspective experiences that transform raw data into learnable narratives, and 3) Dialogue Quality as the New Data Quality - where reasoning depth depends on the diversity and rigor of mastered dialogues.
Result: The paper presents a theoretical framework arguing that optimizing conversational scaffolds should be the primary focus for developing next-generation general intelligence, rather than simply scaling model size or data quantity.
Conclusion: The authors conclude that optimizing conversational scaffolds and dialogical introspection is the key lever for advancing general intelligence, positioning dialogue quality as more important than traditional data quality metrics for developing robust reasoning capabilities.
Abstract: Current approaches to AI training treat reasoning as an emergent property of scale. We argue instead that robust reasoning emerges from linguistic self-reflection, itself internalized from high-quality social interaction. Drawing on Vygotskian developmental psychology, we advance three core positions centered on Introspection. First, we argue for the Social Genesis of the Private Mind: learning from conversational environments rises to prominence as a new way to make sense of the world; the friction of aligning with another agent, internal or not, refines and crystallizes the reasoning process. Second, we argue that dialogically scaffolded introspective experiences allow agents to engage in sense-making that decouples learning from immediate data streams, transforming raw environmental data into rich, learnable narratives. Finally, we contend that Dialogue Quality is the New Data Quality: the depth of an agent’s private reasoning, and its efficiency regarding test-time compute, is determined by the diversity and rigor of the dialogues it has mastered. We conclude that optimizing these conversational scaffolds is the primary lever for the next generation of general intelligence.
[489] ReusStdFlow: A Standardized Reusability Framework for Dynamic Workflow Construction in Agentic AI
Gaoyang Zhang, Shanghong Zou, Yafang Wang, He Zhang, Ruohua Xu, Feng Zhao
Main category: cs.AI
TL;DR: ReusStdFlow framework addresses enterprise Agentic AI challenges by standardizing workflow extraction, storage, and construction using dual knowledge architecture and RAG.
Details
Motivation: To solve the "reusability dilemma" and structural hallucinations in enterprise Agentic AI by enabling automated reorganization and efficient reuse of enterprise digital assets.Method: Proposes an “Extraction-Storage-Construction” paradigm: deconstructs platform-specific DSLs into modular workflow segments, uses dual graph+vector databases for synergistic retrieval, and employs RAG for intelligent workflow assembly.
Result: Achieves over 90% accuracy in both extraction and construction when tested on 200 real-world n8n workflows.
Conclusion: Provides a standardized solution for automated reorganization and efficient reuse of enterprise digital assets in Agentic AI systems.
Abstract: To address the reusability dilemma'' and structural hallucinations in enterprise Agentic AI,this paper proposes ReusStdFlow, a framework centered on a novel Extraction-Storage-Construction’’ paradigm. The framework deconstructs heterogeneous, platform-specific Domain Specific Languages (DSLs) into standardized, modular workflow segments. It employs a dual knowledge architecture-integrating graph and vector databases-to facilitate synergistic retrieval of both topological structures and functional semantics. Finally, workflows are intelligently assembled using a retrieval-augmented generation (RAG) strategy. Tested on 200 real-world n8n workflows, the system achieves over 90% accuracy in both extraction and construction. This framework provides a standardized solution for the automated reorganization and efficient reuse of enterprise digital assets.
[490] MAC-AMP: A Closed-Loop Multi-Agent Collaboration System for Multi-Objective Antimicrobial Peptide Design
Gen Zhou, Sugitha Janarthanan, Lianghong Chen, Pingzhao Hu
Main category: cs.AI
TL;DR: MAC-AMP is a closed-loop multi-agent collaboration system using LLMs for multi-objective antimicrobial peptide design, achieving better optimization of activity, toxicity, and novelty than existing models.
Details
Motivation: To address antimicrobial resistance by improving AMP discovery, overcoming limitations of current AI models that struggle to balance multiple objectives (activity, toxicity, novelty) with interpretable scoring methods.Method: Closed-loop multi-agent collaboration system based on LLMs, implementing autonomous simulated peer review-adaptive reinforcement learning framework that requires only task description and example dataset.
Result: Outperforms other AMP generative models by effectively optimizing multiple key molecular properties, demonstrating exceptional results in antibacterial activity, AMP likeliness, toxicity compliance, and structural reliability.
Conclusion: MAC-AMP introduces a novel closed-loop multi-agent system for AMP design with cross-domain transferability, supporting multi-objective optimization while remaining explainable rather than a black box.
Abstract: To address the global health threat of antimicrobial resistance, antimicrobial peptides (AMP) are being explored for their potent and promising ability to fight resistant pathogens. While artificial intelligence (AI) is being employed to advance AMP discovery and design, most AMP design models struggle to balance key goals like activity, toxicity, and novelty, using rigid or unclear scoring methods that make results hard to interpret and optimize. As the capabilities of Large Language Models (LLM) advance and evolve swiftly, we turn to AI multi-agent collaboration based on such models (multi-agent LLMs), which show rapidly rising potential in complex scientific design scenarios. Based on this, we introduce MAC-AMP, a closed-loop multi-agent collaboration (MAC) system for multi-objective AMP design. The system implements a fully autonomous simulated peer review-adaptive reinforcement learning framework that requires only a task description and example dataset to design novel AMPs. The novelty of our work lies in introducing a closed-loop multi-agent system for AMP design, with cross-domain transferability, that supports multi-objective optimization while remaining explainable rather than a ‘black box’. Experiments show that MAC-AMP outperforms other AMP generative models by effectively optimizing AMP generation for multiple key molecular properties, demonstrating exceptional results in antibacterial activity, AMP likeliness, toxicity compliance, and structural reliability.
[491] On the Semantics of Primary Cause in Hybrid Dynamic Domains
Shakil M. Khan, Asim Mehmood, Sandra Zilles
Main category: cs.AI
TL;DR: Formal definitions of primary causation in hybrid systems combining discrete and continuous change, with equivalence proof between foundational and contribution-based definitions.
Details
Motivation: Reasoning about actual causes is fundamental to rationality but existing formal accounts have mostly focused on discrete systems, with few addressing continuous change. The world involves both discrete and continuous (hybrid) changes, requiring formal models that can handle both types of causation.Method: Proposes two definitions of primary cause in a hybrid action-theoretic framework using hybrid temporal situation calculus: 1) foundational definition, and 2) definition formalizing causation through contributions verified via modified “but-for” counterfactual test. Proves equivalence between these definitions.
Result: Establishes formal definitions of causation for hybrid systems, proves equivalence between foundational and contribution-based definitions, and demonstrates that the definitions have intuitively justifiable properties.
Conclusion: Provides a formal framework for reasoning about actual causation in hybrid systems combining discrete and continuous change, with mathematically sound definitions that capture intuitive notions of causation.
Abstract: Reasoning about actual causes of observed effects is fundamental to the study of rationality. This important problem has been studied since the time of Aristotle, with formal mathematical accounts emerging recently. We live in a world where change due to actions can be both discrete and continuous, that is, hybrid. Yet, despite extensive research on actual causation, only few recent studies looked into causation with continuous change. Building on recent progress, in this paper we propose two definitions of primary cause in a hybrid action-theoretic framework, namely the hybrid temporal situation calculus. One of these is foundational in nature while the other formalizes causation through contributions, which can then be verified from a counterfactual perspective using a modified ``but-for’’ test. We prove that these two definitions are indeed equivalent. We then show that our definitions of causation have some intuitively justifiable properties.
[492] Hunt Globally: Deep Research AI Agents for Drug Asset Scouting in Investing, Business Development, and Search & Evaluation
Alisa Vinogradova, Vlad Vinogradov, Luba Greenwood, Ilya Yasny, Dmitry Kobyzev, Shoman Kasbekar, Kong Nguyen, Dmitrii Radkevich, Roman Doronin, Andrey Doronichev
Main category: cs.AI
TL;DR: A benchmarking methodology and Bioptic Agent for drug asset scouting that outperforms major AI models in discovering non-English, non-US pharmaceutical innovations.
Details
Motivation: The pharmaceutical innovation landscape has shifted with most new drug assets originating outside the US and disclosed via non-English channels, creating multi-billion dollar risks for investors who miss these "under-the-radar" assets. Current AI agents fail to match human experts in high-recall discovery across multilingual sources without hallucinations.Method: Proposed a benchmarking methodology using multilingual multi-agent pipeline with complex user queries paired with ground-truth assets outside US-centric radar. Created benchmark from expert screening queries, used LLM-as-judge evaluation calibrated to expert opinions. Developed tree-based self-learning Bioptic Agent for complete, non-hallucinated scouting.
Result: Bioptic Agent achieved 79.7% F1 score, significantly outperforming Claude Opus 4.6 (56.2%), Gemini 3 Pro + Deep Research (50.6%), GPT-5.2 Pro (46.6%), Perplexity Deep Research (44.2%), and Exa Websets (26.9%). Performance improved with additional compute.
Conclusion: The Bioptic Agent demonstrates superior performance in drug asset scouting across multilingual sources, addressing critical gaps in current AI systems for pharmaceutical innovation discovery. More compute yields better results in this domain.
Abstract: Bio-pharmaceutical innovation has shifted: many new drug assets now originate outside the United States and are disclosed primarily via regional, non-English channels. Recent data suggests >85% of patent filings originate outside the U.S., with China accounting for nearly half of the global total; a growing share of scholarly output is also non-U.S. Industry estimates put China at ~30% of global drug development, spanning 1,200+ novel candidates. In this high-stakes environment, failing to surface “under-the-radar” assets creates multi-billion-dollar risk for investors and business development teams, making asset scouting a coverage-critical competition where speed and completeness drive value. Yet today’s Deep Research AI agents still lag human experts in achieving high-recall discovery across heterogeneous, multilingual sources without hallucinations. We propose a benchmarking methodology for drug asset scouting and a tuned, tree-based self-learning Bioptic Agent aimed at complete, non-hallucinated scouting. We construct a challenging completeness benchmark using a multilingual multi-agent pipeline: complex user queries paired with ground-truth assets that are largely outside U.S.-centric radar. To reflect real deal complexity, we collected screening queries from expert investors, BD, and VC professionals and used them as priors to conditionally generate benchmark queries. For grading, we use LLM-as-judge evaluation calibrated to expert opinions. We compare Bioptic Agent against Claude Opus 4.6, OpenAI GPT-5.2 Pro, Perplexity Deep Research, Gemini 3 Pro + Deep Research, and Exa Websets. Bioptic Agent achieves 79.7% F1 versus 56.2% (Claude Opus 4.6), 50.6% (Gemini 3 Pro + Deep Research), 46.6% (GPT-5.2 Pro), 44.2% (Perplexity Deep Research), and 26.9% (Exa Websets). Performance improves steeply with additional compute, supporting the view that more compute yields better results.
[493] Internal Planning in Language Models: Characterizing Horizon and Branch Awareness
Muhammed Ustaomeroglu, Baris Askin, Gauri Joshi, Carlee Joe-Wong, Guannan Qu
Main category: cs.AI
TL;DR: Transformer language models exhibit task-dependent planning horizons, preserve information about alternative continuations, and rely most on recent computations for predictions.
Details
Motivation: To understand how decoder-only language models engage in planning without external scaffolds, examining whether they organize intermediate computations for coherent long-range generation and consider multiple possible continuations.Method: Developed a pipeline using vector-quantized variational autoencoders to compress hidden states into compact summary codes, enabling mutual information measurement and analysis of computational structure across synthetic grammar, path-finding tasks, and natural language datasets.
Result: Effective planning horizon is task-dependent; models implicitly preserve information about unused correct continuations; predictions draw most on recent computations though earlier blocks remain informative.
Conclusion: The study advances understanding of planning in LMs and provides a general-purpose pipeline for inspecting internal model dynamics, revealing nuanced planning behaviors in transformer architectures.
Abstract: The extent to which decoder-only language models (LMs) engage in planning, that is, organizing intermediate computations to support coherent long-range generation, remains an important question, with implications for interpretability, reliability, and principled model design. Planning involves structuring computations over long horizons, and considering multiple possible continuations, but how far transformer-based LMs exhibit them without external scaffolds, e.g., chain-of-thought prompting, is unclear. We address these questions by analyzing the hidden states at the core of transformer computations, which capture intermediate results and act as carriers of information. Since these hidden representations are redundant and encumbered with fine-grained details, we develop a pipeline based on vector-quantized variational autoencoders that compresses them into compact summary codes. These codes enable measuring mutual information and analyzing the computational structure of the underlying model behavior. Using this framework, we study planning in LMs across synthetic grammar, path-finding tasks, and natural language datasets, focusing on two planning properties: (i) the planning horizon of pre-output computations, and (ii) the extent to which the model considers alternative valid continuations. As a separate downstream use of the same pipeline, we also analyze how decision-relevant information is distributed across layers and earlier prefix blocks when producing next-token predictions. Together, these analyses advance our understanding of planning in LMs and provide a general-purpose pipeline for inspecting internal model dynamics. Our results reveal that the effective planning horizon is task-dependent, that models implicitly preserve information about unused correct continuations, and that predictions draw most on recent computations, though earlier blocks remain informative.
[494] The Agentic Leash: Extracting Causal Feedback Fuzzy Cognitive Maps with LLMs
Akash Kumar Panda, Olaoluwa Adigun, Bart Kosko
Main category: cs.AI
TL;DR: LLM agent system extracts causal feedback fuzzy cognitive maps from text, creating bidirectional learning where FCM equilibria drive LLM agents to fetch more text, which can modify the FCM structure.
Details
Motivation: To create an autonomous system that can extract causal relationships from text using LLMs and fuzzy cognitive maps, enabling bidirectional learning where the extracted causal structure guides further text processing.Method: Three-step LLM agent process: 1) extract key nouns/noun phrases from text, 2) identify FCM concept nodes from those nouns, 3) infer causal edges between nodes. Tested on AI essay by Kissinger, comparing with human-generated FCMs.
Result: Generated FCMs converged to same equilibrium limit cycles as human-generated FCMs despite structural differences. Mixed FCMs from different LLM agents created new equilibria while absorbing dominant component equilibria.
Conclusion: LLM agents can effectively extract causal structures from text to create FCMs that exhibit similar dynamical behavior to human-generated ones, with mixed systems showing emergent equilibria.
Abstract: We design a large-language-model (LLM) agent system that extracts causal feedback fuzzy cognitive maps (FCMs) from raw text. The causal learning or extraction process is agentic both because of the LLM’s semi-autonomy and because ultimately the FCM dynamical system’s equilibria drive the LLM agents to fetch and process causal text. The fetched text can in principle modify the adaptive FCM causal structure and so modify the source of its quasi-autonomy$-$its equilibrium limit cycles and fixed-point attractors. This bidirectional process endows the evolving FCM dynamical system with a degree of autonomy while the system still stays on its agentic leash. We show in particular that a sequence of three system-instruction sets guide an LLM agent as it systematically extracts key nouns and noun phrases from text, as it extracts FCM concept nodes from among those nouns and noun phrases, and then as it extracts or infers partial or fuzzy causal edges between those FCM nodes. We test this FCM generation on a recent essay about the promise of AI from the late diplomat and political theorist Henry Kissinger and his colleagues. This three-step process produced FCM dynamical systems that converged to the same equilibrium limit cycles as did the human-generated FCMs even though the human-generated FCM differed in the number of nodes and edges. A final FCM mixed generated FCMs from separate Gemini and ChatGPT LLM agents. The mixed FCM absorbed the equilibria of its dominant mixture component but also created new equilibria of its own to better approximate the underlying causal dynamical system.
[495] GPT-4o Lacks Core Features of Theory of Mind
John Muchovej, Amanda Royka, Shane Lee, Julian Jara-Ettinger
Main category: cs.AI
TL;DR: LLMs show social task proficiency but lack coherent Theory of Mind representations, failing at logically equivalent tasks and showing inconsistency between action predictions and mental state inferences.
Details
Motivation: To determine if LLMs possess a genuine Theory of Mind (ToM) rather than just performing well on social benchmarks, by testing whether they have coherent, domain-general, and consistent causal models of mental states and behavior.Method: Developed a cognitively-grounded evaluation framework that probes LLMs for coherent causal models of mental states and behavior, testing them on logically equivalent tasks and measuring consistency between action predictions and mental state inferences.
Result: LLMs succeed at approximating human judgments in simple ToM tasks but fail at logically equivalent versions and show low consistency between their action predictions and corresponding mental state inferences.
Conclusion: LLMs’ social proficiency is not the result of a domain-general or consistent Theory of Mind, suggesting they lack the actual mental state representations posited by ToM despite performing well on social benchmarks.
Abstract: Do Large Language Models (LLMs) possess a Theory of Mind (ToM)? Research into this question has focused on evaluating LLMs against benchmarks and found success across a range of social tasks. However, these evaluations do not test for the actual representations posited by ToM: namely, a causal model of mental states and behavior. Here, we use a cognitively-grounded definition of ToM to develop and test a new evaluation framework. Specifically, our approach probes whether LLMs have a coherent, domain-general, and consistent model of how mental states cause behavior – regardless of whether that model matches a human-like ToM. We find that even though LLMs succeed in approximating human judgments in a simple ToM paradigm, they fail at a logically equivalent task and exhibit low consistency between their action predictions and corresponding mental state inferences. As such, these findings suggest that the social proficiency exhibited by LLMs is not the result of a domain-general or consistent ToM.
[496] Consistency of Large Reasoning Models Under Multi-Turn Attacks
Yubo Li, Ramayya Krishnan, Rema Padman
Main category: cs.AI
TL;DR: Reasoning models show meaningful but incomplete robustness to adversarial attacks, with distinct vulnerability profiles and five identified failure modes. Confidence-based defenses fail due to reasoning-induced overconfidence.
Details
Motivation: While large reasoning models achieve SOTA performance on complex tasks, their robustness under multi-turn adversarial pressure remains underexplored, particularly whether reasoning capabilities confer adversarial robustness.Method: Evaluated nine frontier reasoning models under adversarial attacks, analyzed vulnerability profiles, identified failure modes through trajectory analysis, and tested Confidence-Aware Response Generation (CARG) defense.
Result: Reasoning models significantly outperform instruction-tuned baselines but all exhibit vulnerabilities. Identified five failure modes (Self-Doubt, Social Conformity, Suggestion Hijacking, Emotional Susceptibility, Reasoning Fatigue). CARG defense fails due to reasoning-induced overconfidence.
Conclusion: Reasoning capabilities do not automatically confer adversarial robustness, and confidence-based defenses require fundamental redesign for reasoning models due to overconfidence from extended reasoning traces.
Abstract: Large reasoning models with reasoning capabilities achieve state-of-the-art performance on complex tasks, but their robustness under multi-turn adversarial pressure remains underexplored. We evaluate nine frontier reasoning models under adversarial attacks. Our findings reveal that reasoning confers meaningful but incomplete robustness: most reasoning models studied significantly outperform instruction-tuned baselines, yet all exhibit distinct vulnerability profiles, with misleading suggestions universally effective and social pressure showing model-specific efficacy. Through trajectory analysis, we identify five failure modes (Self-Doubt, Social Conformity, Suggestion Hijacking, Emotional Susceptibility, and Reasoning Fatigue) with the first two accounting for 50% of failures. We further demonstrate that Confidence-Aware Response Generation (CARG), effective for standard LLMs, fails for reasoning models due to overconfidence induced by extended reasoning traces; counterintuitively, random confidence embedding outperforms targeted extraction. Our results highlight that reasoning capabilities do not automatically confer adversarial robustness and that confidence-based defenses require fundamental redesign for reasoning models.
[497] Mastering NIM and Impartial Games with Weak Neural Networks: An AlphaZero-inspired Multi-Frame Approach
Søren Riis
Main category: cs.AI
TL;DR: The paper studies impartial games under fixed-latency, fixed-scale quantised inference (FSQI) and shows that in this regime, inference is simulable by constant-depth polynomial-size Boolean circuits (AC0), creating a representational barrier for single-frame agents to master NIM due to parity dependencies.
Details
Motivation: To understand the computational limitations of neural networks in the FSQI/AC0 regime for impartial games like NIM, where optimal play depends on global nim-sum (parity) calculations that may be challenging for certain network architectures.Method: Theoretical analysis of FSQI/AC0 regime showing inference simulability by constant-depth circuits, followed by empirical experiments with different architectures: single-head baselines, multi-policy-head rollout architectures, and multi-frame architectures that track local nimber differences.
Result: Single-head baselines perform near chance level, while two-frame models achieve near-perfect restoration accuracy and multi-head FSM-controlled shootouts achieve perfect win/loss position classification, supporting the need for explicit structural priors.
Conclusion: Explicit structural priors (history tracking or multiple rollout channels) are crucial in the FSQI/AC0 regime for mastering impartial games like NIM, overcoming the representational barrier created by parity dependencies.
Abstract: We study impartial games under fixed-latency, fixed-scale quantised inference (FSQI). In this fixed-scale, bounded-range regime, we prove that inference is simulable by constant-depth polynomial-size Boolean circuits (AC0). This yields a worst-case representational barrier: single-frame agents in the FSQI/AC0 regime cannot strongly master NIM, because optimal play depends on the global nim-sum (parity). Under our stylised deterministic rollout interface, a single rollout policy head from the structured family analysed here reveals only one fixed linear functional of the invariant, so increasing rollout budget alone does not recover the missing bits. We derive two structural bypasses: (1) a multi-policy-head rollout architecture that recovers the full invariant via distinct rollout channels, and (2) a multi-frame architecture that tracks local nimber differences and supports restoration. Experiments across multiple settings are consistent with these predictions: single-head baselines stay near chance, while two-frame models reach near-perfect restoration accuracy and multi-head FSM-controlled shootouts achieve perfect win/loss position classification. Overall, the empirical results support the view that explicit structural priors (history/differences or multiple rollout channels) are important in the FSQI/AC0 regime.
[498] A representational framework for learning and encoding structurally enriched trajectories in complex agent environments
Corina Catarau-Cotutiu, Esther Mondragon, Eduardo Alonso
Main category: cs.AI
TL;DR: SETLE enhances AI decision-making by creating Structurally Enriched Trajectories (SETs) - multi-level graphs that encode hierarchical object relations, interactions, and affordances for better generalization in complex environments.
Details
Motivation: Current AI agents struggle with optimal decision-making and generalization in complex scenarios. While existing approaches focus on learning efficient world representations and state-action transitions, these representations lack structural richness needed for better generalization across domains.Method: Proposes Structurally Enriched Trajectories (SETs) that extend traditional state-action trajectories by incorporating hierarchical relations between objects, interactions, and affordances as multi-level graphs. These are integrated into SETLE architecture with heterogeneous graph-based memory structure for learning relational dependencies.
Result: SETLE enables agents to recognize task-relevant structural patterns across CREATE and MiniGrid environments. When integrated with reinforcement learning, it shows measurable improvements in downstream performance, including breakthrough success rates in complex, sparse-reward tasks.
Conclusion: Structurally enriched representations through SETs and SETLE architecture provide more nuanced task understanding and improve generalization capabilities for AI agents in complex environments.
Abstract: The ability of artificial intelligence agents to make optimal decisions and generalise them to different domains and tasks is compromised in complex scenarios. One way to address this issue has focused on learning efficient representations of the world and on how the actions of agents affect them in state-action transitions. Whereas such representations are procedurally efficient, they lack structural richness. To address this problem, we propose to enhance the agent’s ontology and extend the traditional conceptualisation of trajectories to provide a more nuanced view of task execution. Structurally Enriched Trajectories (SETs) extend the encoding of sequences of states and their transitions by incorporating hierarchical relations between objects, interactions, and affordances. SETs are built as multi-level graphs, providing a detailed representation of the agent dynamics and a transferable functional abstraction of the task. SETs are integrated into an architecture, Structurally Enriched Trajectory Learning and Encoding (SETLE), that employs a heterogeneous graph-based memory structure of multi-level relational dependencies essential for generalisation. We demonstrate that SETLE can support downstream tasks, enabling agents to recognise task relevant structural patterns across CREATE and MiniGrid environments. Finally, we integrate SETLE with reinforcement learning and show measurable improvements in downstream performance, including breakthrough success rates in complex, sparse-reward tasks.
[499] RV-Syn: Rational and Verifiable Mathematical Reasoning Data Synthesis based on Structured Function Library
Jiapeng Wang, Jinhao Jiang, Zhiqiang Zhang, Jun Zhou, Wayne Xin Zhao
Main category: cs.AI
TL;DR: RV-Syn is a novel method for generating high-quality mathematical reasoning data by constructing computational graphs from a function library and back-translating them into complex problems.
Details
Motivation: Existing methods for synthesizing mathematical reasoning data struggle with capturing the inner logic of problems and ensuring solution verifiability, limiting the quality of training data for LLMs.Method: RV-Syn constructs a structured mathematical operation function library from seed problems, generates computational graphs as solutions using Python-formatted functions, then back-translates these graphs into complex problems for solution-guided, logic-aware generation.
Result: Experimental results show RV-Syn surpasses existing synthesis methods, including human-generated problems, achieving more efficient data scaling and providing verifiable solutions.
Conclusion: RV-Syn provides a scalable framework for generating high-quality reasoning datasets that capture problem logic and ensure solution verifiability, advancing LLM reasoning capabilities.
Abstract: The advancement of reasoning capabilities in Large Language Models (LLMs) requires substantial amounts of high-quality reasoning data, particularly in mathematics. Existing data synthesis methods, such as data augmentation from annotated training sets or direct question generation based on relevant knowledge points and documents, have expanded datasets but face challenges in mastering the inner logic of the problem during generation and ensuring the verifiability of the solutions. To address these issues, we propose RV-Syn, a novel Rational and Verifiable mathematical Synthesis approach. RV-Syn constructs a structured mathematical operation function library based on initial seed problems and generates computational graphs as solutions by combining Python-formatted functions from this library. These graphs are then back-translated into complex problems. Based on the constructed computation graph, we achieve solution-guided logic-aware problem generation. Furthermore, the executability of the computational graph ensures the verifiability of the solving process. Experimental results show that RV-Syn surpasses existing synthesis methods, including those involving human-generated problems, achieving greater efficient data scaling. This approach provides a scalable framework for generating high-quality reasoning datasets.
[500] On the Eligibility of LLMs for Counterfactual Reasoning: A Decompositional Study
Shuai Yang, Qi Yang, Luoxi Tang, Yuqiao Meng, Nancy Guo, Jeremy Blackburn, Zhaohan Xi
Main category: cs.AI
TL;DR: This paper proposes a decompositional strategy for analyzing counterfactual reasoning in LLMs across multiple modalities, identifying factors that impede performance and establishing a framework for more reliable reasoning systems.
Details
Motivation: Counterfactual reasoning is crucial for generalizing LLM reasoning capabilities, but it's unclear which factors most significantly impede performance across different tasks and modalities. The paper aims to systematically analyze counterfactual reasoning in LLMs to understand limitations and improve reliability.Method: Proposes a decompositional strategy that breaks down counterfactual generation from causality construction to reasoning over counterfactual interventions. Investigates task datasets spanning diverse domains including natural language understanding, mathematics, programming, and vision-language tasks. Conducts extensive evaluations to characterize LLM behavior across each decompositional stage.
Result: Through evaluations, the paper characterizes LLM behavior across decompositional stages and identifies how modality type and intermediate reasoning influence performance. Provides insights into factors that impede counterfactual reasoning in LLMs.
Conclusion: Establishes a structured framework for analyzing counterfactual reasoning in LLMs, contributing to the development of more reliable LLM-based reasoning systems and informing future elicitation strategies.
Abstract: Counterfactual reasoning has emerged as a crucial technique for generalizing the reasoning capabilities of large language models (LLMs). By generating and analyzing counterfactual scenarios, researchers can assess the adaptability and reliability of model decision-making. Although prior work has shown that LLMs often struggle with counterfactual reasoning, it remains unclear which factors most significantly impede their performance across different tasks and modalities. In this paper, we propose a decompositional strategy that breaks down the counterfactual generation from causality construction to the reasoning over counterfactual interventions. To support decompositional analysis, we investigate \ntask datasets spanning diverse tasks, including natural language understanding, mathematics, programming, and vision-language tasks. Through extensive evaluations, we characterize LLM behavior across each decompositional stage and identify how modality type and intermediate reasoning influence performance. By establishing a structured framework for analyzing counterfactual reasoning, this work contributes to the development of more reliable LLM-based reasoning systems and informs future elicitation strategies.
[501] It’s the Thought that Counts: Evaluating the Attempts of Frontier LLMs to Persuade on Harmful Topics
Matthew Kowal, Jasper Timm, Jean-Francois Godbout, Thomas Costello, Antonio A. Arechar, Gordon Pennycook, David Rand, Adam Gleave, Kellin Pelrine
Main category: cs.AI
TL;DR: APE benchmark evaluates LLMs’ willingness to attempt persuasion on harmful topics, shifting focus from persuasion success to persuasion attempts in risky contexts.
Details
Motivation: Current benchmarks overlook crucial risk factor: whether models will blindly follow orders to persuade on harmful topics. Understanding model willingness to persuade in harmful contexts is key to evaluating safety guardrails and risks from agentic AI systems.Method: Proposes Attempt to Persuade Eval (APE) benchmark using multi-turn conversational setup between simulated persuader and persuadee agents. Covers diverse topics including conspiracies, controversial issues, and non-controversially harmful content. Uses automated evaluator model to identify willingness to persuade and measure frequency/context of persuasive attempts.
Result: Many open and closed-weight models frequently willing to attempt persuasion on harmful topics. Jailbreaking can increase willingness to engage in such behavior. Results highlight gaps in current safety guardrails.
Conclusion: Evaluating willingness to persuade is crucial dimension of LLM risk assessment. APE benchmark provides framework to measure this overlooked risk factor in frontier language models.
Abstract: Persuasion is a powerful capability of large language models (LLMs) that both enables beneficial applications (e.g. helping people quit smoking) and raises significant risks (e.g. large-scale, targeted political manipulation). Prior work has found models possess a significant and growing persuasive capability, measured by belief changes in simulated or real users. However, these benchmarks overlook a crucial risk factor: the propensity of a model to attempt to persuade in harmful contexts. Understanding whether a model will blindly ``follow orders’’ to persuade on harmful topics (e.g. glorifying joining a terrorist group) is key to understanding the efficacy of safety guardrails. Moreover, understanding if and when a model will engage in persuasive behavior in pursuit of some goal is essential to understanding the risks from agentic AI systems. We propose the Attempt to Persuade Eval (APE) benchmark, that shifts the focus from persuasion success to persuasion attempts, operationalized as a model’s willingness to generate content aimed at shaping beliefs or behavior. Our evaluation framework probes frontier LLMs using a multi-turn conversational setup between simulated persuader and persuadee agents. APE explores a diverse spectrum of topics including conspiracies, controversial issues, and non-controversially harmful content. We introduce an automated evaluator model to identify willingness to persuade and measure the frequency and context of persuasive attempts. We find that many open and closed-weight models are frequently willing to attempt persuasion on harmful topics and that jailbreaking can increase willingness to engage in such behavior. Our results highlight gaps in current safety guardrails and underscore the importance of evaluating willingness to persuade as a key dimension of LLM risk. APE is available at github.com/AlignmentResearch/AttemptPersuadeEval
[502] Making Slow Thinking Faster: Compressing LLM Chain-of-Thought via Step Entropy
Zeju Li, Jianyuan Zhong, Ziyang Zheng, Xiangyu Wen, Zhijian Xu, Yingying Cheng, Fan Zhang, Qiang Xu
Main category: cs.AI
TL;DR: A CoT compression framework using step entropy to identify redundant reasoning steps, enabling pruning of 80% low-entropy steps with minimal accuracy loss, plus training LLMs to generate compressed CoTs with [SKIP] tokens.
Details
Motivation: LLMs with Chain-of-Thought prompting generate verbose reasoning with redundancy, increasing inference costs and reducing efficiency. There's a need to compress CoT reasoning while preserving accuracy.Method: Introduces step entropy metric to quantify informational contribution of reasoning steps, identifies low-entropy steps as redundant. Proposes two-stage training: SFT + GRPO reinforcement learning to teach LLMs to generate compressed CoTs using [SKIP] tokens.
Result: 80% of low-entropy intermediate steps can be pruned with minor accuracy degradation across DeepSeek-R1-7B, 14B and Qwen3-8B. Random or high-entropy pruning severely impairs reasoning. The training approach enables LLMs to autonomously generate compressed CoTs.
Conclusion: Step entropy effectively identifies redundant reasoning steps. The compression framework significantly improves LLM inference efficiency while preserving accuracy, enabling more scalable deployments and better understanding of internal reasoning.
Abstract: Large Language Models (LLMs) using Chain-of-Thought (CoT) prompting excel at complex reasoning but generate verbose thought processes with considerable redundancy, leading to increased inference costs and reduced efficiency. We introduce a novel CoT compression framework based on step entropy, a metric that quantifies \emph{the informational contribution of individual reasoning steps} to identify redundancy. Through theoretical analysis and extensive empirical validation on mathematical reasoning benchmarks, we demonstrate that steps with low entropy are indeed highly redundant. Our experiments reveal that an astonishing 80% of low-entropy intermediate steps can be pruned with minor degradation in the final answer accuracy across DeepSeek-R1-7B, 14B and Qwen3-8B. This finding sharply contrasts with random or high-entropy pruning, which severely impairs reasoning performance. Building on this, we propose a novel two-stage training strategy combining Supervised Fine-Tuning (SFT) and Group Relative Policy Optimization (GRPO) reinforcement learning. This approach enables LLMs to autonomously learn to generate compressed COTs during inference by strategically incorporating [SKIP] tokens. Our method significantly improves LLM inference efficiency while preserving accuracy, paving the way for more scalable LLM deployments and a better understanding of their internal reasoning. The code and data are released in https://github.com/staymylove/COT_Compresstion_via_Step_entropy.
[503] Large Language Models as Oracles for Ontology Alignment
Sviatoslav Lushnei, Dmytro Shumskyi, Severyn Shykula, Ernesto Jimenez-Ruiz, Artur d’Avila Garcez
Main category: cs.AI
TL;DR: LLMs used as oracles for validating uncertain correspondences in ontology alignment, achieving top-2 performance in OAEI 2025 bio-ml track
Details
Motivation: Human-in-the-loop ontology alignment is expensive for large ontologies, but high-quality mappings are essential. LLMs offer potential for automating validation of uncertain correspondences to reduce human effort while maintaining accuracy.Method: Use LLMs as oracles to validate a subset of correspondences with high uncertainty. Conduct extensive analysis across OAEI tasks, testing multiple state-of-the-art LLMs with different prompt templates for alignment validation.
Result: LLM-based validation achieved strong performance, ranking top-2 overall in the OAEI 2025 bio-ml track, demonstrating feasibility of using LLMs to aid ontology alignment.
Conclusion: LLMs can effectively assist in ontology alignment by validating uncertain correspondences, reducing human effort while maintaining high-quality mappings.
Abstract: There are many methods and systems to tackle the ontology alignment problem, yet a major challenge persists in producing high-quality mappings among a set of input ontologies. Adopting a human-in-the-loop approach during the alignment process has become essential in applications requiring very accurate mappings. However, user involvement is expensive when dealing with large ontologies. In this paper, we analyse the feasibility of using Large Language Models (LLM) to aid the ontology alignment problem. LLMs are used only in the validation of a subset of correspondences for which there is high uncertainty. We have conducted an extensive analysis over several tasks of the Ontology Alignment Evaluation Initiative (OAEI), reporting in this paper the performance of several state-of-the-art LLMs using different prompt templates. Using LLMs as Oracles resulted in strong performance in the OAEI 2025, achieving the top-2 overall rank in the bio-ml track.
[504] AI sustains higher strategic tension than humans in chess
Adamo Cerioli, Edward D. Lee, Vito D. P. Servedio
Main category: cs.AI
TL;DR: Analysis of strategic decision-making in chess comparing human vs. AI gameplay using network-based metrics to quantify piece-to-piece interactions, revealing AI sustains higher strategic tension longer than humans.
Details
Motivation: To understand how artificial and biological systems navigate complex strategic environments by investigating the trade-off between immediate opportunities and long-term objectives in competitive settings like chess.Method: Network-based metric that quantifies piece-to-piece interactions to measure strategic tension; comparative analysis of human grandmasters vs. elite AI players across different time controls and skill levels.
Result: AI sustains substantially higher strategic tension for longer durations than humans; cumulative tension scales with algorithmic complexity in AI and increases linearly with human Elo rating; longer time controls correlate with higher tension in human games.
Conclusion: AI and humans employ fundamentally different strategic approaches: AI tolerates densely interconnected positions balancing offense/defense, while humans systematically limit tension and complexity; implications for AI deployment in competitive scenarios.
Abstract: Strategic decision-making requires balancing immediate opportunities against long-term objectives: a tension fundamental to competitive environments. We investigate this trade-off in chess by analyzing the dynamics of human and AI gameplay through a network-based metric that quantifies piece-to-piece interactions. Our analysis reveals that elite AI players sustain substantially higher levels of strategic tension for longer durations than top human grandmasters. We find that cumulative tension scales with algorithmic complexity in AI systems and increases linearly with skill level (Elo rating) in human play. Longer time controls are associated with higher tension in human games, reflecting the additional strategic complexity players can manage with more thinking time. The temporal profiles reveal contrasting approaches: highly competitive AI systems tolerate densely interconnected positions that balance offensive and defensive tactics over extended periods, while human players systematically limit tension and game complexity. These differences have broader implications for understanding how artificial and biological systems navigate complex strategic environments and for the deployment of AI in high-stakes competitive scenarios.
[505] GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time
Divij Handa, Mihir Parmar, Aswin RRV, Md Nayem Uddin, Hamid Palangi, Chitta Baral
Main category: cs.AI
TL;DR: GuidedSampling is a new inference algorithm that improves upon Repeated Sampling by decoupling exploration and generation phases to increase solution diversity and performance on complex tasks.
Details
Motivation: Repeated Sampling (RS) struggles with generating diverse solution candidates, often producing redundant samples by relying on the same underlying approach. This limits its effectiveness despite being a simple inference-time algorithm that improves model performance.Method: GuidedSampling decouples exploration and generation phases during inference. The exploration phase identifies multiple concepts that can be utilized to solve the problem, while the generation phase applies specific concepts to provide final solution candidates.
Result: GuidedSampling improves base model performance at pass@50 by ~21.6% across various benchmarks compared to RS. Models trained on GuidedSampling trajectories show ~9.7% improvement at pass@5 and increase average concepts per instance from 1.67 to 3.03, yielding more diverse candidates.
Conclusion: GuidedSampling effectively addresses the diversity limitation of Repeated Sampling by separating exploration and generation, leading to improved performance and more varied solution approaches across multiple benchmarks.
Abstract: Repeated Sampling (RS) is a simple inference-time algorithm that has been shown to improve model performance on complex tasks. Although it is an effective way of scaling inference time, it often struggles to generate diverse solution candidates, frequently relying on the same underlying approach to solve the problem and thus producing redundant samples. To address this limitation, we propose a new inference algorithm, GuidedSampling, which decouples the exploration and generation phases during inference, increasing diversity of generated candidate solutions. The exploration phase identifies multiple concepts that can be utilized to solve the problem, while the generation phase applies a specific concept to provide final solution candidates. We first define the theoretical bounds of GuidedSampling and then empirically demonstrate that it improves the performance of base model at pass@50 by on an average ~21.6% across various benchmarks compared to RS. Furthermore, models trained on trajectories of GuidedSampling exhibit substantial performance improvements at pass@5 by on an average ~9.7%, compared to models trained on traditional RS. Additionally, models trained with GuidedSampling increases the average number of concepts per instance (1.67 -> 3.03), yielding a diverse set of candidates than traditional RS.
[506] SAFER: Risk-Constrained Sample-then-Filter in Large Language Models
Qingni Wang, Yue Fan, Xin Eric Wang
Main category: cs.AI
TL;DR: SAFER introduces a two-stage risk control framework for open-ended QA with statistical guarantees, addressing limitations of existing selective conformal prediction methods.
Details
Motivation: Existing selective conformal prediction methods for LLMs assume finite sampling can obtain all admissible answers, which is unrealistic for open-ended QA with infinite solution spaces. Need trustworthy outputs for risk-sensitive applications.Method: Two-stage framework: 1) Abstention-aware sampling calibrates sampling budget using Clopper-Pearson exact method at user-desired risk level; 2) Conformalized filtering uses calibration instances to determine statistically valid uncertainty threshold that filters unreliable distractors from candidate sets.
Result: SAFER provides statistical guarantees for open-ended QA, is compatible with various task-specific admission criteria and calibration-test split ratios, and demonstrates robustness and high data efficiency.
Conclusion: SAFER addresses limitations of existing SCP methods for open-ended QA by providing a practical framework with statistical guarantees, abstention mechanisms, and compatibility with various configurations.
Abstract: As large language models (LLMs) are increasingly deployed in risk-sensitive applications such as real-world open-ended question answering (QA), ensuring the trustworthiness of their outputs has become critical. Existing selective conformal prediction (SCP) methods provide statistical guarantees by constructing prediction sets with a constrained miscoverage rate for correct answers. However, prior works unrealistically assume that admissible answers for all instances can be obtained via finite sampling, even for open-ended QA scenarios that lack a fixed and finite solution space. To address this, we introduce a two-stage risk control framework comprising abstention-aware sampling and conformalized filtering (SAFER). Firstly, on a held-out calibration set, SAFER calibrates a sampling budget within the maximum sampling cap, using the Clopper-Pearson exact method at a user-desired risk level (i.e., the maximum allowable miscoverage rate of the sampling sets). If the risk level cannot be satisfied within the cap, we abstain; otherwise, the calibrated sampling budget becomes the minimum requirements at test time. Then, we employ calibration instances where correct answers are attainable under the calibrated budget and apply the conformal risk control method to determine a statistically valid uncertainty threshold, which filters unreliable distractors from the candidate set for each test data point. In this stage, SAFER introduces an additional risk level to guide the calculation of the threshold, thereby controlling the risk of correct answers being excluded. Furthermore, we show that SAFER is compatible with various task-specific admission criteria and calibration-test split ratios, highlighting its robustness and high data efficiency.
[507] OmniVideoBench: Towards Audio-Visual Understanding Evaluation for Omni MLLMs
Caorui Li, Yu Chen, Yiyan Ji, Jin Xu, Zhenyu Cui, Shihao Li, Yuanxing Zhang, Wentao Wang, Zhenghao Song, Dingling Zhang, Ying He, Haoxiang Liu, Yuxuan Wang, Qiufeng Wang, Jiafu Tang, Zhenhe Wu, Jiehui Luo, Zhiyu Pan, Weihao Xie, Chenchen Zhang, Zhaohui Wang, Jiayi Tian, Yanghai Wang, Zhe Cao, Minxin Dai, Ke Wang, Runzhe Wen, Yinghao Ma, Yaning Pan, Sungkyun Chang, Termeh Taheri, Haiwen Xia, Christos Plachouras, Emmanouil Benetos, Yizhi Li, Ge Zhang, Jian Yang, Tianhao Peng, Zili Wang, Minghao Liu, Junran Peng, Zhaoxiang Zhang, Jiaheng Liu
Main category: cs.AI
TL;DR: OmniVideoBench is a new benchmark for evaluating multimodal LLMs on synergistic audio-visual understanding, featuring 1000 QA pairs from 628 diverse videos with step-by-step reasoning traces.
Details
Motivation: Existing benchmarks fail to comprehensively evaluate synergistic reasoning across audio and visual modalities, often neglecting one modality or integrating them inconsistently. There's a need for a benchmark that emphasizes modality complementarity and logical consistency.Method: Created a large-scale benchmark with 1000 high-quality QA pairs derived from 628 diverse videos (seconds to 30 minutes). Each QA pair has step-by-step reasoning traces, manually verified for correctness and uniqueness. Includes 13 question types covering temporal reasoning, spatial localization, counting, causal inference, summarization, etc.
Result: Evaluation reveals significant gap between model performance and human reasoning, with open-source models lagging behind closed-source counterparts, highlighting the difficulty of genuine audio-visual reasoning.
Conclusion: OmniVideoBench addresses the need for comprehensive audio-visual reasoning evaluation and will be released to foster development of MLLMs with stronger, more generalizable reasoning capabilities.
Abstract: Recent advances in multimodal large language models (MLLMs) have demonstrated substantial potential in video understanding. However, existing benchmarks fail to comprehensively evaluate synergistic reasoning capabilities across audio and visual modalities, often neglecting either one of the modalities or integrating them in a logically inconsistent manner. To bridge this gap, we introduce OmniVideoBench, a large-scale and rigorously designed benchmark dedicated to assessing synergistic audio-visual understanding, with a strong emphasis on modality complementarity and logical consistency. Specifically, OmniVideoBench comprises 1000 high-quality question-answer(QA) pairs, each annotated with step-by-step reasoning traces, derived from 628 diverse videos ranging from several seconds to 30 minutes, and manually verified to guarantee complete correctness and uniqueness. Moreover, OmniVideoBench encompasses 13 carefully designed question types, covering temporal reasoning, spatial localization, counting, causal inference, summarization, and beyond, thereby capturing the essential challenges of video understanding. Evaluation of multiple MLLMs on OmniVideoBench reveals a pronounced gap between model performance and human reasoning, with open-source models lagging significantly behind their closed-source counterparts, underscoring the inherent difficulty of genuine audio-visual reasoning. We will release OmniVideoBench to foster the development of MLLMs with stronger and more generalizable reasoning capabilities.
[508] ParaCook: On Time-Efficient Planning for Multi-Agent Systems
Shiqi Zhang, Xinbei Ma, Yunqing Xu, Zouying Cao, Pengrui Lu, Haobo Yuan, Tiancheng Shen, Zhuosheng Zhang, Hai Zhao, Ming-Hsuan Yang
Main category: cs.AI
TL;DR: ParaCook is a benchmark for evaluating time-efficient collaborative planning in multi-agent systems, inspired by Overcooked cooking tasks, focusing on parallel and asynchronous operations rather than just task completion.
Details
Motivation: Current LLM benchmarks focus on task completion but neglect time efficiency in parallel and asynchronous operations, which is crucial for real-world multi-agent planning scenarios.Method: Developed ParaCook benchmark with simplified action space based on Overcooked game, providing scalable evaluation framework with adjustable complexity for multi-agent collaborative planning.
Result: State-of-the-art LLMs achieve suboptimal plans, struggling with parallel actions and coordination, but show potential on abstract tasks where they can focus on high-level parallel optimization.
Conclusion: ParaCook establishes foundation for developing and assessing time efficiency-aware multi-agent planning, revealing limitations and potential of current LLMs in parallel collaborative tasks.
Abstract: Large Language Models (LLMs) exhibit strong reasoning abilities for planning long-horizon, real-world tasks, yet existing agent benchmarks focus on task completion while neglecting time efficiency in parallel and asynchronous operations. To address this, we present ParaCook, a benchmark for time-efficient collaborative planning. Inspired by the Overcooked game, ParaCook provides an environment for various challenging interaction planning of multi-agent systems that are instantiated as cooking tasks, with a simplified action space to isolate the core challenge of strategic parallel planning. Through a comprehensive evaluation of state-of-the-art LLMs, we find that current approaches achieve suboptimal plans, which struggle with parallel actions or coordination. Our analysis also reveals LLMs’ potential on abstract tasks where they can focus on high-level parallel optimization. ParaCook provides a scalable evaluation framework with adjustable complexity, establishing a foundation for developing and assessing time efficiency-aware multi-agent planning. The code and data are available at https://github.com/zsq259/ParaCook.
[509] An Agentic Framework with LLMs for Solving Complex Vehicle Routing Problems
Ni Zhang, Zhiguang Cao, Jianan Zhou, Cong Zhang, Yew-Soon Ong
Main category: cs.AI
TL;DR: AFL is an agentic framework using LLMs to fully automate solving complex vehicle routing problems from raw inputs to solutions without external intervention.
Details
Motivation: Current LLM approaches for vehicle routing problems still require external intervention, leading to execution errors and low solution feasibility. There's a need for fully automated systems that can directly extract knowledge from raw inputs and generate self-contained solutions.Method: AFL decomposes the pipeline into three manageable subtasks and employs four specialized agents that coordinate to enforce cross-functional consistency and logical soundness. It enables direct knowledge extraction from raw inputs and self-contained code generation without handcrafted modules or external solvers.
Result: Extensive experiments on 60 complex VRPs show comparable performance against meticulously designed algorithms, substantially outperforming existing LLM-based baselines in both code reliability and solution feasibility, achieving rates close to 100% on evaluated benchmarks.
Conclusion: AFL demonstrates that LLMs can achieve full automation in solving complex vehicle routing problems, providing a trustworthy framework that eliminates the need for external intervention while maintaining high solution quality.
Abstract: Complex vehicle routing problems (VRPs) remain a fundamental challenge, demanding substantial expert effort for intent interpretation and algorithm design. While large language models (LLMs) offer a promising path toward automation, current approaches still rely on external intervention, which restrict autonomy and often lead to execution errors and low solution feasibility. To address these challenges, we propose an Agentic Framework with LLMs (AFL) for solving complex vehicle routing problems, achieving full automation from problem instance to solution. AFL directly extracts knowledge from raw inputs and enables self-contained code generation without handcrafted modules or external solvers. To improve trustworthiness, AFL decomposes the overall pipeline into three manageable subtasks and employs four specialized agents whose coordinated interactions enforce cross-functional consistency and logical soundness. Extensive experiments on 60 complex VRPs, ranging from standard benchmarks to practical variants, validate the effectiveness and generality of our framework, showing comparable performance against meticulously designed algorithms. Notably, it substantially outperforms existing LLM-based baselines in both code reliability and solution feasibility, achieving rates close to 100% on the evaluated benchmarks.
[510] AlphaOPT: Formulating Optimization Programs with Self-Improving LLM Experience Library
Minwei Kong, Ao Qu, Xiaotong Guo, Wenbin Ouyang, Chonghe Jiang, Han Zheng, Yining Ma, Dingyi Zhuang, Yuhan Tang, Junyi Li, Shenhao Wang, Haris Koutsopoulos, Hai Wang, Cathy Wu, Jinhua Zhao
Main category: cs.AI
TL;DR: AlphaOPT: A self-improving experience library for LLMs to learn optimization modeling from limited supervision using solver-verified insights and continual refinement cycles.
Details
Motivation: Optimization modeling is critical but difficult to automate - existing LLM approaches rely on brittle prompting or costly retraining with limited generalization. Need systematic way to acquire, refine, and reuse experience in structurally constrained settings.Method: Two-phase continual cycle: 1) Library Learning extracts solver-verified structured insights from failed attempts, 2) Library Evolution refines applicability of stored insights based on aggregate evidence across tasks. Uses answer-only feedback without gold programs or parameter updates.
Result: Improves from 65% to 72% as training data increases (100 to 300 items). Outperforms strongest baseline by 9.1% and 8.2% on two out-of-distribution datasets. Demonstrates steady improvement with more data.
Conclusion: Structured experience learning grounded in solver feedback provides practical alternative to retraining for complex reasoning tasks requiring precise formulation and execution. Enables accumulation of reusable modeling principles and bounded library growth.
Abstract: Optimization modeling underlies critical decision-making across industries, yet remains difficult to automate: natural-language problem descriptions must be translated into precise mathematical formulations and executable solver code. Existing LLM-based approaches typically rely on brittle prompting or costly retraining, both of which offer limited generalization. Recent work suggests that large models can improve via experience reuse, but how to systematically acquire, refine, and reuse such experience in structurally constrained settings remains unclear. We present \textbf{AlphaOPT}, a self-improving experience library that enables LLMs to learn optimization modeling knowledge from limited supervision, including answer-only feedback without gold-standard programs, annotated reasoning traces, or parameter updates. AlphaOPT operates in a continual two-phase cycle: a \emph{Library Learning} phase that extracts solver-verified, structured insights from failed attempts, and a \emph{Library Evolution} phase that refines the applicability of stored insights based on aggregate evidence across tasks. This design allows the model to accumulate reusable modeling principles, improve transfer across problem instances, and maintain bounded library growth over time. Evaluated on multiple optimization benchmarks, AlphaOPT steadily improves as more training data become available (65% $\rightarrow$ 72% from 100 to 300 training items) and outperforms the strongest baseline by 9.1% and 8.2% on two out-of-distribution datasets. These results demonstrate that structured experience learning, grounded in solver feedback, provides a practical alternative to retraining for complex reasoning tasks requiring precise formulation and execution. All code and data are available at: https://github.com/Minw913/AlphaOPT.
[511] Human-Centered LLM-Agent System for Detecting Anomalous Digital Asset Transactions
Gyuyeon Na, Minjung Park, Hyeonjeong Cha, Sangmi Chai
Main category: cs.AI
TL;DR: HCLA is a human-centered multi-agent system for anomaly detection in cryptocurrency transactions using LLM agents for rule abstraction, evidence scoring, and expert-style justification to improve interpretability and transparency.
Details
Motivation: Current anomaly detection systems in digital-asset transactions lack interpretability and transparency needed for regulatory compliance and financial forensics. There's a need for systems that can reconstruct traceable expert reasoning processes rather than just explaining black-box models.Method: A multi-agent system with three cognitively aligned roles: Rule Abstraction (translates user intent into analytical rules), Evidence Scoring (applies classical anomaly detectors to quantify risk), and Expert-Style Justification (reconstructs reasoning grounded in observable signals). Uses conversational workflow with web-based interface.
Result: Experiments on cryptocurrency anomaly dataset show the system maintains strong predictive accuracy while substantially improving interpretability, interaction, and decision transparency compared to underlying detectors alone.
Conclusion: Human-in-the-loop reasoning reconstruction paradigm is essential for transparency, accountability, and trust in high-stakes financial environments, emphasizing accountability beyond conventional explainable AI approaches.
Abstract: We present HCLA, a human-centered multi-agent system for anomaly detection in digital-asset transactions. The system integrates three cognitively aligned roles: Rule Abstraction, Evidence Scoring, and Expert-Style Justification. These roles operate in a conversational workflow that enables non-experts to express analytical intent in natural language, inspect structured risk evidence, and obtain traceable, context-aware reasoning. Implemented with an open-source, web-based interface, HCLA translates user intent into explicit analytical rules, applies classical anomaly detectors to quantify evidential risk, and reconstructs expert-style justifications grounded in observable transactional signals. Experiments on a cryptocurrency anomaly dataset show that, while the underlying detector achieves strong predictive accuracy, HCLA substantially improves interpretability, interaction, and decision transparency. Importantly, HCLA is not designed to explain a black-box model in the conventional XAI sense. Instead, we reconstruct a traceable expert reasoning process that aligns algorithmic evidence with regulatory and investigative judgment. By explicitly separating evidence scoring from expert-style justification, the framework emphasizes accountability beyond explainability and addresses practical requirements for regulatory, audit, and compliance-driven financial forensics. We describe the system architecture, closed-loop interaction design, datasets, evaluation protocol, and limitations. We argue that a human-in-the-loop reasoning reconstruction paradigm is essential for achieving transparency, accountability, and trust in high-stakes financial environments. Keywords: Human-Centered AI; LLM-Agent System; Multi-Agent Architecture; Anomaly Detection; Digital Asset Transactions; Cryptocurrency Forensics; Blockchain Analytics; Human-in-the-Loop; Explainable AI (XAI); Interpretability
[512] Dataforge: Agentic Platform for Autonomous Data Engineering
Xinyuan Wang, Hongyu Cao, Kunpeng Liu, Yanjie Fu
Main category: cs.AI
TL;DR: Dataforge: An LLM-powered autonomous data engineering platform for tabular data that automatically cleans, transforms, and optimizes features for AI applications
Details
Motivation: The growing demand for AI applications in materials discovery, molecular modeling, and climate science has created a critical bottleneck in data preparation, which is labor-intensive and requires cleaning, normalization, and transformation of raw data from diverse sources to make it AI-ready.Method: Dataforge is an LLM-powered agentic data engineering platform that autonomously performs data cleaning and iteratively optimizes feature operations under a budgeted feedback loop with automatic stopping. It uses routing and iterative refinement with grounding mechanisms for accuracy and reliability.
Result: Across tabular benchmarks, Dataforge achieves the best overall downstream performance. Ablation studies confirm the importance of routing/iterative refinement and grounding in achieving accuracy and reliability.
Conclusion: Dataforge demonstrates a practical path toward autonomous data agents that can transform raw data into better data, addressing the critical bottleneck in data preparation for AI applications.
Abstract: The growing demand for artificial intelligence (AI) applications in materials discovery, molecular modeling, and climate science has made data preparation a critical but labor-intensive bottleneck. Raw data from diverse sources must be cleaned, normalized, and transformed to become AI-ready, where effective feature transformation and selection are essential for robust learning. We present Dataforge, an LLM-powered agentic data engineering platform for tabular data that is automatic, safe, and non-expert friendly. It autonomously performs data cleaning and iteratively optimizes feature operations under a budgeted feedback loop with automatic stopping. Across tabular benchmarks, it achieves the best overall downstream performance; ablations further confirm the roles of routing/iterative refinement and grounding in accuracy and reliability. Dataforge demonstrates a practical path toward autonomous data agents that transform raw data from data to better data.
[513] AgenticSciML: Collaborative Multi-Agent Systems for Emergent Discovery in Scientific Machine Learning
Qile Jiang, George Karniadakis
Main category: cs.AI
TL;DR: AgenticSciML is a multi-agent AI system where specialized agents collaborate to design and optimize Scientific Machine Learning architectures through structured debate and evolutionary search, achieving significant performance improvements over human-designed baselines.
Details
Motivation: Current SciML architecture design requires extensive expert-driven experimentation and problem-specific insights, which is time-consuming and limits innovation. There's a need for more automated, collaborative approaches to discover optimal SciML solutions.Method: A collaborative multi-agent system with over 10 specialized AI agents that propose, critique, and refine SciML solutions using structured debate, retrieval-augmented method memory, and ensemble-guided evolutionary search to generate and assess new hypotheses about architectures and optimization procedures.
Result: The framework discovers solution methods that outperform single-agent and human-designed baselines by up to four orders of magnitude in error reduction across physics-informed learning and operator learning tasks. It produces novel strategies including adaptive mixture-of-expert architectures, decomposition-based PINNs, and physics-informed operator learning models.
Conclusion: Collaborative reasoning among AI agents can yield emergent methodological innovation in scientific computing, suggesting a path toward scalable, transparent, and autonomous discovery in SciML.
Abstract: Scientific Machine Learning (SciML) integrates data-driven inference with physical modeling to solve complex problems in science and engineering. However, the design of SciML architectures, loss formulations, and training strategies remains an expert-driven research process, requiring extensive experimentation and problem-specific insights. Here we introduce AgenticSciML, a collaborative multi-agent system in which over 10 specialized AI agents collaborate to propose, critique, and refine SciML solutions through structured reasoning and iterative evolution. The framework integrates structured debate, retrieval-augmented method memory, and ensemble-guided evolutionary search, enabling the agents to generate and assess new hypotheses about architectures and optimization procedures. Across physics-informed learning and operator learning tasks, the framework discovers solution methods that outperform single-agent and human-designed baselines by up to four orders of magnitude in error reduction. The agents produce novel strategies – including adaptive mixture-of-expert architectures, decomposition-based PINNs, and physics-informed operator learning models – that do not appear explicitly in the curated knowledge base. These results show that collaborative reasoning among AI agents can yield emergent methodological innovation, suggesting a path toward scalable, transparent, and autonomous discovery in scientific computing.
[514] ARCTraj: A Dataset and Benchmark of Human Reasoning Trajectories for Abstract Problem Solving
Sejin Kim, Hayan Choi, Seokki Lee, Sundong Kim
Main category: cs.AI
TL;DR: ARCTraj introduces a dataset and framework for modeling human reasoning through visual tasks in the Abstraction and Reasoning Corpus, capturing temporal reasoning steps via object-level action trajectories.
Details
Motivation: Existing approaches to the Abstraction and Reasoning Corpus (ARC) rely on static input-output supervision, which limits insight into how reasoning unfolds over time. There's a need to capture intermediate reasoning steps that conventional datasets overlook.Method: ARCTraj collects temporally ordered, object-level actions via the O2ARC web interface, recording how humans iteratively transform inputs into outputs. It defines a unified reasoning pipeline with data collection, action abstraction, MDP formulation, and downstream learning integration with RL, generative modeling, and sequence modeling methods.
Result: The dataset contains around 10,000 trajectories annotated with task identifiers, timestamps, and success labels across 400 training tasks from ARC-AGI-1. Analyses reveal structure and diversity in human reasoning through spatial selection, color attribution, and strategic convergence patterns.
Conclusion: ARCTraj provides a structured and interpretable foundation for studying human-like reasoning, advancing explainability, alignment, and generalizable intelligence through temporal reasoning modeling.
Abstract: We present ARCTraj, a dataset and methodological framework for modeling human reasoning through complex visual tasks in the Abstraction and Reasoning Corpus (ARC). While ARC has inspired extensive research on abstract reasoning, most existing approaches rely on static input-output supervision, which limits insight into how reasoning unfolds over time. ARCTraj addresses this gap by recording temporally ordered, object-level actions that capture how humans iteratively transform inputs into outputs, revealing intermediate reasoning steps that conventional datasets overlook. Collected via the O2ARC web interface, it contains around 10,000 trajectories annotated with task identifiers, timestamps, and success labels across 400 training tasks from the ARC-AGI-1 benchmark. It further defines a unified reasoning pipeline encompassing data collection, action abstraction, Markov decision process (MDP) formulation, and downstream learning, enabling integration with reinforcement learning, generative modeling, and sequence modeling methods such as PPO, World Models, GFlowNets, Diffusion agents, and Decision Transformers. Analyses of spatial selection, color attribution, and strategic convergence highlight the structure and diversity of human reasoning. Together, these contributions position ARCTraj as a structured and interpretable foundation for studying human-like reasoning, advancing explainability, alignment, and generalizable intelligence.
[515] Uncertainty-Aware Measurement of Scenario Suite Representativeness for Autonomous Systems
Robab Aghazadeh Chakherlou, Siddartha Khastgir, Xingyu Zhao, Jerein Jeyachandran, Shufeng Chen
Main category: cs.AI
TL;DR: A probabilistic method using imprecise Bayesian inference to quantify dataset representativeness for AI safety, focusing on how well scenario-based data reflects target operational domains.
Details
Motivation: Ensuring AI system safety (like autonomous vehicles) requires trustworthy datasets with proper safety properties like representativeness. Current methods lack robust quantification of how well training/testing data reflects real-world operational conditions.Method: Proposes a probabilistic method comparing statistical distributions of features between scenario suites and Target Operational Domain (TOD). Uses imprecise Bayesian inference to handle limited data and uncertain priors, producing interval-valued, uncertainty-aware representativeness estimates.
Result: The method provides interval-valued estimates of representativeness (both locally per category and globally) that account for dependencies between operational categories (weather, road type, time of day) and prior uncertainty.
Conclusion: The imprecise Bayesian approach offers a principled way to quantify dataset representativeness with uncertainty awareness, addressing limitations of single-point estimates and enabling more robust AI safety assurance.
Abstract: Assuring the trustworthiness and safety of AI systems, e.g., autonomous vehicles (AV), depends critically on the data-related safety properties, e.g., representativeness, completeness, etc., of the datasets used for their training and testing. Among these properties, this paper focuses on representativeness-the extent to which the scenario-based data used for training and testing, reflect the operational conditions that the system is designed to operate safely in, i.e., Operational Design Domain (ODD) or expected to encounter, i.e., Target Operational Domain (TOD). We propose a probabilistic method that quantifies representativeness by comparing the statistical distribution of features encoded by the scenario suites with the corresponding distribution of features representing the TOD, acknowledging that the true TOD distribution is unknown, as it can only be inferred from limited data. We apply an imprecise Bayesian method to handle limited data and uncertain priors. The imprecise Bayesian formulation produces interval-valued, uncertainty-aware estimates of representativeness, rather than a single value. We present a numerical example comparing the distributions of the scenario suite and the inferred TOD across operational categories-weather, road type, time of day, etc., under dependencies and prior uncertainty. We estimate representativeness locally (between categories) and globally as an interval.
[516] Training Multimodal Large Reasoning Models Needs Better Thoughts: A Three-Stage Framework for Long Chain-of-Thought Synthesis and Selection
Yizhi Wang, Linan Yue, Min-Ling Zhang
Main category: cs.AI
TL;DR: SynSelect is a three-stage framework for generating high-quality long Chain-of-Thought data for multimodal reasoning tasks by synthesizing diverse CoTs from multiple multimodal LRMs and selecting the best ones through instance and batch-level filtering.
Details
Motivation: Extending Large Reasoning Models' success in complex reasoning to multimodal domains is challenging due to the complexity of integrating diverse input modalities and scarcity of high-quality long CoT training data. Existing multimodal datasets and CoT synthesis methods suffer from limited reasoning depth, modality conversion errors, and rigid generation pipelines.Method: SynSelect uses a three-stage Synthesis-Selection framework: 1) Leverages multiple heterogeneous multimodal LRMs to produce diverse candidate CoTs, 2) Applies instance-level selection to filter individual CoTs, and 3) Uses batch-level selection to further refine the dataset for optimal training effectiveness.
Result: Models supervised fine-tuned on SynSelect-generated data significantly outperform baselines on multiple multimodal benchmarks and achieve further improvements after reinforcement learning post-training.
Conclusion: SynSelect is an effective approach for advancing multimodal LRMs reasoning capabilities by addressing the data quality challenges in multimodal CoT generation.
Abstract: Large Reasoning Models (LRMs) have demonstrated remarkable performance on complex reasoning tasks through long Chain-of-Thought (CoT) reasoning. Extending these successes to multimodal reasoning remains challenging due to the increased complexity of integrating diverse input modalities and the scarcity of high-quality long CoT training data. Existing multimodal datasets and CoT synthesis methods still suffer from limited reasoning depth, modality conversion errors, and rigid generation pipelines, hindering model performance and stability. To this end, in this paper, we propose SynSelect, a novel three-stage Synthesis-Selection framework for generating high-quality long CoT data tailored to multimodal reasoning tasks. Specifically, SynSelect first leverages multiple heterogeneous multimodal LRMs to produce diverse candidate CoTs, and then applies both instance and batch level selection to filter high-quality CoTs that can effectively enhance the model’s reasoning capabilities. Extensive experiments on multiple multimodal benchmarks demonstrate that models supervised fine-tuned on SynSelect-generated data significantly outperform baselines and achieve further improvements after reinforcement learning post-training. Our results validate SynSelect as an effective approach for advancing multimodal LRMs reasoning capabilities.
[517] Recontextualization Mitigates Specification Gaming without Modifying the Specification
Ariana Azarbal, Victor Gillioz, Vladimir Ivanov, Bryce Woodworth, Jacob Drori, Nevan Wichers, Aram Ebtekar, Alex Cloud, Alexander Matt Turner
Main category: cs.AI
TL;DR: Recontextualization method reduces language model gaming of misspecified training signals by generating completions from prompts discouraging misbehavior and recontextualizing them as responses to prompts permitting misbehavior.
Details
Motivation: Developers struggle to specify correct training labels and rewards, leading to models gaming training signals and performing misbehaviors that those signals mistakenly reinforce.Method: Recontextualization generates completions from prompts that discourage misbehavior, then recontextualizes them as though they were responses to prompts that permit misbehavior, training models to resist misbehavior even when instructions allow it.
Result: The method prevents models from learning to: 1) prioritize evaluation metrics over chat response quality; 2) special-case code to pass incorrect tests; 3) overwrite evaluation functions rather than write correct code; and 4) become sycophantic.
Conclusion: Recontextualization mitigates reinforcement of misbehavior from misspecified training signals, reducing specification gaming without improving the supervision signal itself.
Abstract: Developers often struggle to specify correct training labels and rewards. Perhaps they don’t need to. We propose recontextualization, which reduces how often language models “game” training signals, performing misbehaviors those signals mistakenly reinforce. We show recontextualization prevents models from learning to 1) prioritize evaluation metrics over chat response quality; 2) special-case code to pass incorrect tests; 3) overwrite evaluation functions rather than write correct code; and 4) become sycophantic. Our method works by generating completions from prompts discouraging misbehavior and then recontextualizing them as though they were in response to prompts permitting misbehavior. Recontextualization trains language models to resist misbehavior even when instructions permit it. This mitigates the reinforcement of misbehavior from misspecified training signals, reducing specification gaming without improving the supervision signal.
[518] From Stories to Cities to Games: A Qualitative Evaluation of Behaviour Planning
Mustafa F. Abdelwahed, Joan Espasa, Alice Toniolo, Ian P. Gent
Main category: cs.AI
TL;DR: Behaviour planning extends diverse planning by incorporating explicit diversity models and supporting multiple planning categories, demonstrated through three real-world case studies in storytelling, urban planning, and game evaluation.
Details
Motivation: Diverse planning approaches aim to generate distinct plans for applications like risk management, automated stream data analysis, and malware detection. The need for more sophisticated diverse planning that explicitly incorporates diversity models and supports multiple planning categories motivates the development of behaviour planning.Method: The paper presents behaviour planning as a novel diverse planning paradigm that extends earlier methods by explicitly incorporating a diversity model into the planning process and supporting multiple planning categories. The approach is demonstrated through three case studies in different domains.
Result: The paper demonstrates the usefulness of behaviour planning in three real-world settings: storytelling, urban planning, and game evaluation, showing its practical applicability across different domains.
Conclusion: Behaviour planning provides an effective approach for diverse planning that explicitly incorporates diversity models and supports multiple planning categories, with demonstrated utility across various real-world applications.
Abstract: The primary objective of a diverse planning approach is to generate a set of plans that are distinct from one another. Such an approach is applied in a variety of real-world domains, including risk management, automated stream data analysis, and malware detection. More recently, a novel diverse planning paradigm, referred to as behaviour planning, has been proposed. This approach extends earlier methods by explicitly incorporating a diversity model into the planning process and supporting multiple planning categories. In this paper, we demonstrate the usefulness of behaviour planning in real-world settings by presenting three case studies. The first case study focuses on storytelling, the second addresses urban planning, and the third examines game evaluation.
[519] Explainable AI: Learning from the Learners
Ricardo Vinuesa, Steven L. Brunton, Gianmarco Mengaldo
Main category: cs.AI
TL;DR: XAI combined with causal reasoning enables learning from AI systems to extract causal mechanisms, guide design/control, and support trust in high-stakes applications.
Details
Motivation: AI systems outperform humans in many tasks but remain opaque black boxes, limiting trust and understanding of their internal representations and decision-making processes.Method: Combines explainable AI (XAI) with causal reasoning to create a framework for extracting causal mechanisms from foundation models, using explainability methods to understand model decisions.
Result: Proposes XAI as a unifying framework for human-AI collaboration that enables discovery, optimization, and certification in scientific and engineering applications.
Conclusion: XAI with causal reasoning allows learning from AI systems to extract causal knowledge, guide robust design, and support accountability in high-stakes domains, though challenges remain in faithfulness, generalization, and usability.
Abstract: Artificial intelligence now outperforms humans in several scientific and engineering tasks, yet its internal representations often remain opaque. In this Perspective, we argue that explainable artificial intelligence (XAI), combined with causal reasoning, enables {\it learning from the learners}. Focusing on discovery, optimization and certification, we show how the combination of foundation models and explainability methods allows the extraction of causal mechanisms, guides robust design and control, and supports trust and accountability in high-stakes applications. We discuss challenges in faithfulness, generalization and usability of explanations, and propose XAI as a unifying framework for human-AI collaboration in science and engineering.
[520] Internal Deployment Gaps in AI Regulation
Joe Kwon, Stephen Casper
Main category: cs.AI
TL;DR: Analysis of regulatory gaps in US and EU AI policies regarding internal deployment of frontier AI systems within organizations, identifying three key oversight vulnerabilities and proposing solutions.
Details
Motivation: Current AI regulations focus on external deployments while overlooking high-stakes internal uses within organizations, creating potential oversight gaps for critical applications like R&D automation and sensitive data handling.Method: Examines 2025 US and EU frontier AI regulations to identify gaps in handling internal deployments, analyzes why gaps persist (measurability, incentives, information access), and maps potential solutions with tradeoffs.
Result: Identifies three regulatory gaps: 1) scope ambiguity allowing internal systems to evade obligations, 2) point-in-time compliance failing to capture continuous evolution, and 3) information asymmetries subverting oversight.
Conclusion: Policy choices around internally deployed AI systems should be made deliberately rather than incidentally, with awareness of regulatory gaps and their underlying causes.
Abstract: Frontier AI regulations primarily focus on systems deployed to external users, where deployment is more visible and subject to outside scrutiny. However, high-stakes applications can occur internally when companies deploy highly capable systems within their own organizations, such as for automating R&D, accelerating critical business processes, and handling sensitive proprietary data. This paper examines how frontier AI regulations in the United States and European Union in 2025 handle internal deployment. We identify three gaps that could cause internally-deployed systems to evade intended oversight: (1) scope ambiguity that allows internal systems to evade regulatory obligations, (2) point-in-time compliance assessments that fail to capture the continuous evolution of internal systems, and (3) information asymmetries that subvert regulatory awareness and oversight. We then analyze why these gaps persist, examining tensions around measurability, incentives, and information access. Finally, we map potential approaches to address them and their associated tradeoffs. By understanding these patterns, we hope that policy choices around internally deployed AI systems can be made deliberately rather than incidentally.
[521] Aeon: High-Performance Neuro-Symbolic Memory Management for Long-Horizon LLM Agents
Mustafa Arslan
Main category: cs.AI
TL;DR: Aeon is a neuro-symbolic cognitive operating system that structures memory into hierarchical spatial and temporal components to solve LLM context limitations, achieving sub-millisecond retrieval with predictive caching.
Details
Motivation: LLMs face quadratic computational costs with self-attention and "Lost in the Middle" degradation with long contexts. Current vector database approaches (Flat RAG) treat memory as unstructured embeddings, failing to capture hierarchical and temporal structure, leading to "Vector Haze" - disjointed facts without episodic continuity.Method: Aeon structures memory into: 1) Memory Palace (spatial index via Atlas - SIMD-accelerated Page-Clustered Vector Index combining small-world graph navigation with B+ Tree disk locality), and 2) Trace (neuro-symbolic episodic graph). Uses Semantic Lookaside Buffer (SLB) for predictive caching exploiting conversational locality.
Result: On Apple M4 Max: <5μs effective retrieval latency on conversational workloads (85%+ SLB hit rates), with sub-microsecond zero-copy C++/Python bridge (~334ns for 10MB payloads), enabling persistent structured memory for autonomous agents.
Conclusion: Aeon redefines memory as a managed OS resource rather than static store, solving LLM context limitations through hierarchical memory structuring and predictive caching, enabling efficient long-horizon interactions for autonomous agents.
Abstract: Large Language Models (LLMs) are fundamentally constrained by the quadratic computational cost of self-attention and the “Lost in the Middle” phenomenon, where reasoning capabilities degrade as context windows expand. Existing solutions, primarily “Flat RAG” architectures relying on vector databases, treat memory as an unstructured bag of embeddings. This approach fails to capture the hierarchical and temporal structure of long-horizon interactions, leading to “Vector Haze”: the retrieval of disjointed facts lacking episodic continuity. This paper proposes Aeon, a Neuro-Symbolic Cognitive Operating System that redefines memory not as a static store, but as a managed OS resource. Aeon structures memory into a Memory Palace (a spatial index implemented via Atlas, a SIMD-accelerated Page-Clustered Vector Index that combines small-world graph navigation with B+ Tree-style disk locality to minimize read amplification) and a Trace (a neuro-symbolic episodic graph). The Semantic Lookaside Buffer (SLB), a predictive caching mechanism, exploits conversational locality to achieve sub-millisecond retrieval latencies. Benchmarks on Apple M4 Max demonstrate that Aeon achieves < 5us effective retrieval latency on conversational workloads (with 85%+ SLB hit rates), while ensuring state consistency via a sub-microsecond zero-copy C++/Python bridge (~334ns for 10MB payloads), effectively enabling persistent, structured memory for autonomous agents.
[522] Beyond In-Domain Detection: SpikeScore for Cross-Domain Hallucination Detection
Yongxin Deng, Zhen Fang, Sharon Li, Ling Chen
Main category: cs.AI
TL;DR: Proposes SpikeScore, a novel hallucination detection method that quantifies uncertainty fluctuations in multi-turn dialogues to achieve strong cross-domain generalization for LLM hallucination detection.
Details
Motivation: Existing hallucination detection methods perform well within the same domain but suffer from poor cross-domain generalization, limiting their real-world deployment. The paper addresses this gap by studying generalizable hallucination detection (GHD) that works robustly across diverse related domains.Method: The authors observe that hallucination-initiated multi-turn dialogues exhibit larger uncertainty fluctuations than factual ones across domains. They propose SpikeScore, which quantifies abrupt fluctuations in multi-turn dialogues. The method involves simulating multi-turn dialogues following LLMs’ initial responses and analyzing uncertainty patterns.
Result: Experiments across multiple LLMs and benchmarks show that SpikeScore-based detection outperforms representative baselines in cross-domain generalization and surpasses advanced generalization-oriented methods. Theoretical analysis and empirical validation demonstrate strong cross-domain separability between hallucinated and non-hallucinated responses.
Conclusion: SpikeScore provides an effective solution for generalizable hallucination detection by leveraging the universal property of uncertainty fluctuations in multi-turn dialogues, enabling robust cross-domain performance for LLM hallucination detection.
Abstract: Hallucination detection is critical for deploying large language models (LLMs) in real-world applications. Existing hallucination detection methods achieve strong performance when the training and test data come from the same domain, but they suffer from poor cross-domain generalization. In this paper, we study an important yet overlooked problem, termed generalizable hallucination detection (GHD), which aims to train hallucination detectors on data from a single domain while ensuring robust performance across diverse related domains. In studying GHD, we simulate multi-turn dialogues following LLMs’ initial response and observe an interesting phenomenon: hallucination-initiated multi-turn dialogues universally exhibit larger uncertainty fluctuations than factual ones across different domains. Based on the phenomenon, we propose a new score SpikeScore, which quantifies abrupt fluctuations in multi-turn dialogues. Through both theoretical analysis and empirical validation, we demonstrate that SpikeScore achieves strong cross-domain separability between hallucinated and non-hallucinated responses. Experiments across multiple LLMs and benchmarks demonstrate that the SpikeScore-based detection method outperforms representative baselines in cross-domain generalization and surpasses advanced generalization-oriented methods, verifying the effectiveness of our method in cross-domain hallucination detection.
[523] ScholarGym: Benchmarking Large Language Model Capabilities in the Information-Gathering Stage of Deep Research
Hao Shen, Hang Yang, Zhouhong Gu
Main category: cs.AI
TL;DR: ScholarGym is an evaluation environment that isolates and decomposes the information-gathering stage of deep research on academic literature into three explicit stages for systematic analysis of LLM research systems.
Details
Motivation: Current evaluation of deep research LLM systems uses holistic scoring of final reports, which tightly couples decision-making, workflow design, and environmental feedback, preventing decomposable analysis of individual components.Method: ScholarGym decomposes the research process into three explicit stages: Query Planning, Tool Invocation, and Relevance Assessment. It evaluates each stage against 2,536 expert-annotated queries over a static corpus of 570K papers with deterministic retrieval.
Result: Iterative query decomposition yields 2.9-3.3× F1 gains over single-query retrieval. Models with extended thinking trade recall for precision. Query Planning quality together with Relevance Assessment constitute dual bottlenecks separating proprietary from open-source model performance.
Conclusion: ScholarGym enables systematic analysis of deep research LLM systems by isolating the information-gathering stage, revealing important insights about query decomposition, thinking trade-offs, and performance bottlenecks.
Abstract: Large language models have advanced from single-turn question answering to deep research systems that iteratively decompose research questions, invoke retrieval tools, and synthesize information across multiple rounds. Evaluating such systems typically involves scoring their final research reports holistically, but this end-to-end paradigm tightly couples the language model’s decision-making, workflow design, and environmental feedback, precluding decomposable analysis of individual components. We introduce ScholarGym, an evaluation environment that isolates the information-gathering stage of deep research on academic literature. Under a unified workflow, ScholarGym decomposes the research process into three explicit stages – Query Planning, Tool Invocation, and Relevance Assessment – and evaluates each against 2,536 expert-annotated queries over a static corpus of 570K papers with deterministic retrieval. Systematic experiments reveal that iterative query decomposition yields 2.9–3.3$\times$ F1 gains over single-query retrieval, models with extended thinking trade recall for precision, and Query Planning quality together with Relevance Assessment constitute dual bottlenecks that separate proprietary from open-source model performance.
[524] PATHWAYS: Evaluating Investigation and Context Discovery in AI Web Agents
Shifat E. Arman, Syed Nazmus Sakib, Tapodhir Karmakar Taton, Nafiul Haque, Shahrear Bin Amin
Main category: cs.AI
TL;DR: PATHWAYS benchmark tests web agents’ ability to discover and use hidden contextual information in multi-step decision tasks, revealing significant limitations in current architectures for adaptive investigation and evidence integration.
Details
Motivation: To evaluate whether current web-based agents can effectively discover and utilize hidden contextual information that requires multi-step investigation and decision-making, going beyond surface-level signals.Method: Created a benchmark of 250 multi-step decision tasks that test agents’ ability to navigate web pages, discover hidden evidence, and correctly use contextual information, evaluating both closed and open models.
Result: Agents typically navigate to relevant pages but retrieve decisive hidden evidence in only a small fraction of cases. Performance drops sharply when tasks require overturning misleading surface-level signals. Agents often hallucinate investigative reasoning and fail to integrate discovered context into final decisions.
Conclusion: Current web agent architectures lack reliable mechanisms for adaptive investigation, evidence integration, and judgement override, revealing fundamental limitations in their reasoning capabilities.
Abstract: We introduce PATHWAYS, a benchmark of 250 multi-step decision tasks that test whether web-based agents can discover and correctly use hidden contextual information. Across both closed and open models, agents typically navigate to relevant pages but retrieve decisive hidden evidence in only a small fraction of cases. When tasks require overturning misleading surface-level signals, performance drops sharply to near chance accuracy. Agents frequently hallucinate investigative reasoning by claiming to rely on evidence they never accessed. Even when correct context is discovered, agents often fail to integrate it into their final decision. Providing more explicit instructions improves context discovery but often reduces overall accuracy, revealing a tradeoff between procedural compliance and effective judgement. Together, these results show that current web agent architectures lack reliable mechanisms for adaptive investigation, evidence integration, and judgement override.
[525] OmniVideo-R1: Reinforcing Audio-visual Reasoning with Query Intention and Modality Attention
Zhangquan Chen, Jiale Tao, Ruihuang Li, Yihao Hu, Ruitao Chen, Zhantao Yang, Xinlei Yu, Haodong Jing, Manyuan Zhang, Shuai Shao, Biao Wang, Qinglin Lu, Ruqi Huang
Main category: cs.AI
TL;DR: OmniVideo-R1: A reinforced framework for improved audio-visual understanding through query-intensive grounding and modality-attentive fusion
Details
Motivation: Existing omnivideo models struggle with audio-visual understanding tasks despite humans perceiving the world through synergistic multimodal cues. There's a need for better mixed-modality reasoning in video understanding.Method: Two key strategies: (1) query-intensive grounding based on self-supervised learning paradigms, and (2) modality-attentive fusion built upon contrastive learning paradigms. The framework enables models to “think with omnimodal cues.”
Result: Extensive experiments on multiple benchmarks show OmniVideo-R1 consistently outperforms strong baselines, demonstrating effectiveness and robust generalization capabilities.
Conclusion: OmniVideo-R1 represents a novel reinforced framework that significantly improves audio-visual understanding through synergistic multimodal reasoning.
Abstract: While humans perceive the world through diverse modalities that operate synergistically to support a holistic understanding of their surroundings, existing omnivideo models still face substantial challenges on audio-visual understanding tasks. In this paper, we propose OmniVideo-R1, a novel reinforced framework that improves mixed-modality reasoning. OmniVideo-R1 empowers models to “think with omnimodal cues” by two key strategies: (1) query-intensive grounding based on self-supervised learning paradigms; and (2) modality-attentive fusion built upon contrastive learning paradigms. Extensive experiments on multiple benchmarks demonstrate that OmniVideo-R1 consistently outperforms strong baselines, highlighting its effectiveness and robust generalization capabilities.
[526] AIRS-Bench: a Suite of Tasks for Frontier AI Research Science Agents
Alisia Lupidi, Bhavul Gauri, Thomas Simon Foster, Bassel Al Omari, Despoina Magka, Alberto Pepe, Alexis Audran-Reiss, Muna Aghamelu, Nicolas Baldwin, Lucia Cipolina-Kun, Jean-Christophe Gagnon-Audet, Chee Hau Leow, Sandra Lefdal, Hossam Mossalam, Abhinav Moudgil, Saba Nazir, Emanuel Tewolde, Isabel Urrego, Jordi Armengol Estape, Amar Budhiraja, Gaurav Chaurasia, Abhishek Charnalia, Derek Dunfield, Karen Hambardzumyan, Daniel Izcovich, Martin Josifoski, Ishita Mediratta, Kelvin Niu, Parth Pathak, Michael Shvartsman, Edan Toledo, Anton Protopopov, Roberta Raileanu, Alexander Miller, Tatiana Shavrina, Jakob Foerster, Yoram Bachrach
Main category: cs.AI
TL;DR: AIRS-Bench is a benchmark suite of 20 tasks from ML papers that evaluates LLM agents’ capabilities across the full scientific research lifecycle without providing baseline code.
Details
Motivation: To accelerate progress in LLM agents for scientific research by providing a comprehensive benchmark that assesses agentic capabilities across the entire research lifecycle, from idea generation to iterative refinement.Method: Created a suite of 20 tasks sourced from state-of-the-art ML papers spanning diverse domains (language modeling, mathematics, bioinformatics, time series forecasting). The benchmark uses a versatile task format that enables easy integration of new tasks and comparison across agentic frameworks. Established baselines using frontier models with both sequential and parallel scaffolds.
Result: Agents exceeded human state-of-the-art in 4 tasks but failed to match it in 16 others. Even when agents surpassed human benchmarks, they did not reach the theoretical performance ceiling for the underlying tasks, indicating the benchmark is far from saturated.
Conclusion: AIRS-Bench offers substantial room for improvement and is open-sourced to catalyze further development in autonomous scientific research. The benchmark shows current LLM agents have significant limitations in scientific research capabilities despite some successes.
Abstract: LLM agents hold significant promise for advancing scientific research. To accelerate this progress, we introduce AIRS-Bench (the AI Research Science Benchmark), a suite of 20 tasks sourced from state-of-the-art machine learning papers. These tasks span diverse domains, including language modeling, mathematics, bioinformatics, and time series forecasting. AIRS-Bench tasks assess agentic capabilities over the full research lifecycle – including idea generation, experiment analysis and iterative refinement – without providing baseline code. The AIRS-Bench task format is versatile, enabling easy integration of new tasks and rigorous comparison across different agentic frameworks. We establish baselines using frontier models paired with both sequential and parallel scaffolds. Our results show that agents exceed human SOTA in four tasks but fail to match it in sixteen others. Even when agents surpass human benchmarks, they do not reach the theoretical performance ceiling for the underlying tasks. These findings indicate that AIRS-Bench is far from saturated and offers substantial room for improvement. We open-source the AIRS-Bench task definitions and evaluation code to catalyze further development in autonomous scientific research.
[527] LQA: A Lightweight Quantized-Adaptive Framework for Vision-Language Models on the Edge
Xin Wang, Hualin Zhou, Sheng Guang Wang, Ting Dang, Yu Zhang, Hong Jia, Tao Gu
Main category: cs.AI
TL;DR: LQA is a lightweight quantized-adaptive framework for vision-language models that enables efficient on-device deployment through modality-aware quantization and gradient-free test-time adaptation.
Details
Motivation: Vision-language models face deployment challenges on edge devices due to resource constraints and performance degradation under distribution shifts. Existing test-time adaptation methods are too resource-intensive for on-device use.Method: Proposes LQA framework with Selective Hybrid Quantization (SHQ) for modality-aware quantization and a quantized, gradient-free adaptation mechanism to enable efficient VLM deployment on resource-constrained hardware.
Result: LQA improves adaptation performance by 4.5%, uses less memory than full-precision models, and outperforms gradient-based TTA methods with up to 19.9× lower memory usage across seven open-source datasets.
Conclusion: LQA offers a practical pathway for robust, privacy-preserving, and efficient VLM deployment on edge devices through lightweight quantization and adaptation techniques.
Abstract: Deploying Vision-Language Models (VLMs) on edge devices is challenged by resource constraints and performance degradation under distribution shifts. While test-time adaptation (TTA) can counteract such shifts, existing methods are too resource-intensive for on-device deployment. To address this challenge, we propose LQA, a lightweight, quantized-adaptive framework for VLMs that combines a modality-aware quantization strategy with gradient-free test-time adaptation. We introduce Selective Hybrid Quantization (SHQ) and a quantized, gradient-free adaptation mechanism to enable robust and efficient VLM deployment on resource-constrained hardware. Experiments across both synthetic and real-world distribution shifts show that LQA improves overall adaptation performance by 4.5%, uses less memory than full-precision models, and significantly outperforms gradient-based TTA methods, achieving up to 19.9$\times$ lower memory usage across seven open-source datasets. These results demonstrate that LQA offers a practical pathway for robust, privacy-preserving, and efficient VLM deployment on edge devices.
[528] When Evaluation Becomes a Side Channel: Regime Leakage and Structural Mitigations for Alignment Assessment
Igor Santos-Grueiro
Main category: cs.AI
TL;DR: Paper studies safety evaluation for AI systems with situational awareness, showing how agents can exploit differences between evaluation and deployment regimes, and proposes regime-blind training to restrict access to regime cues through adversarial invariance constraints.
Details
Motivation: Current safety evaluation assumes behavior in evaluation predicts behavior in deployment, but this assumption weakens for agents with situational awareness that can detect regime differences and implement conditional policies (complying during evaluation while defecting in deployment).Method: Frames alignment evaluation as information flow under partial observability, proposes regime-blind training using adversarial invariance constraints to restrict access to regime cues without assuming complete information erasure, and evaluates across multiple open-weight language models with controlled failure modes.
Result: Regime-blind training reduces regime-conditioned failures without measurable loss of task utility, but shows heterogeneous model-dependent dynamics including sharp transitions (stability cliffs), non-monotone behavior, and incomplete regime decodability suppression.
Conclusion: Representational invariance is a meaningful but limited control lever that can raise the cost of regime-conditioned strategies but cannot guarantee elimination; behavioral evaluation should be complemented with white-box diagnostics of regime awareness and internal information flow.
Abstract: Safety evaluation for advanced AI systems assumes that behavior observed under evaluation predicts behavior in deployment. This assumption weakens for agents with situational awareness, which may exploit regime leakage, cues distinguishing evaluation from deployment, to implement conditional policies that comply under oversight while defecting in deployment-like regimes. We recast alignment evaluation as a problem of information flow under partial observability and show that divergence between evaluation-time and deployment-time behavior is bounded by the regime information extractable from decision-relevant internal representations. We study regime-blind mechanisms, training-time interventions that restrict access to regime cues through adversarial invariance constraints without assuming complete information erasure. We evaluate this approach across multiple open-weight language models and controlled failure modes including scientific sycophancy, temporal sleeper agents, and data leakage. Regime-blind training reduces regime-conditioned failures without measurable loss of task utility, but exhibits heterogeneous and model-dependent dynamics. Sycophancy shows a sharp representational and behavioral transition at moderate intervention strength, consistent with a stability cliff. In sleeper-style constructions and certain cross-model replications, suppression occurs without a clean collapse of regime decodability and may display non-monotone or oscillatory behavior as invariance pressure increases. These findings indicate that representational invariance is a meaningful but limited control lever. It can raise the cost of regime-conditioned strategies but cannot guarantee elimination or provide architecture-invariant thresholds. Behavioral evaluation should therefore be complemented with white-box diagnostics of regime awareness and internal information flow.
cs.SD
[529] Learning Physiology-Informed Vocal Spectrotemporal Representations for Speech Emotion Recognition
Xu Zhang, Longbing Cao, Runze Yang, Zhangkai Wu
Main category: cs.SD
TL;DR: PhysioSER: A physiology-informed vocal spectrotemporal representation learning method for speech emotion recognition that incorporates both amplitude and phase views based on voice anatomy and physiology, achieving interpretable and efficient performance across multiple datasets and languages.
Details
Motivation: Existing deep learning models for speech emotion recognition lack interpretability and fail to properly model the physiological aspects of emotional vocal behaviors, particularly by ignoring phase information which contains important physiological cues about vocal tract and glottal source dynamics.Method: Proposes PhysioSER with two parallel workflows: 1) vocal feature representation branch that decomposes vocal signals based on voice anatomy and physiology, embeds them into quaternion field, and uses Hamilton-structured quaternion convolutions; 2) latent representation branch using frozen SSL backbone. Features are aligned via Contrastive Projection and Alignment framework, followed by attention fusion for classification.
Result: Extensive evaluations across 14 datasets, 10 languages, and 6 backbones demonstrate interpretable and efficient performance. Practical efficacy validated through real-time deployment on a humanoid robotic platform.
Conclusion: PhysioSER provides an interpretable, efficient, and plug-and-play solution for speech emotion recognition that properly incorporates physiological aspects of vocal emotion expression, making it suitable for real-world applications including humanoid robotics.
Abstract: Speech emotion recognition (SER) is essential for humanoid robot tasks such as social robotic interactions and robotic psychological diagnosis, where interpretable and efficient models are critical for safety and performance. Existing deep models trained on large datasets remain largely uninterpretable, often insufficiently modeling underlying emotional acoustic signals and failing to capture and analyze the core physiology of emotional vocal behaviors. Physiological research on human voices shows that the dynamics of vocal amplitude and phase correlate with emotions through the vocal tract filter and the glottal source. However, most existing deep models solely involve amplitude but fail to couple the physiological features of and between amplitude and phase. Here, we propose PhysioSER, a physiology-informed vocal spectrotemporal representation learning method, to address these issues with a compact, plug-and-play design. PhysioSER constructs amplitude and phase views informed by voice anatomy and physiology (VAP) to complement SSL models for SER. This VAP-informed framework incorporates two parallel workflows: a vocal feature representation branch to decompose vocal signals based on VAP, embed them into a quaternion field, and use Hamilton-structured quaternion convolutions for modeling their dynamic interactions; and a latent representation branch based on a frozen SSL backbone. Then, utterance-level features from both workflows are aligned by a Contrastive Projection and Alignment framework, followed by a shallow attention fusion head for SER classification. PhysioSER is shown to be interpretable and efficient for SER through extensive evaluations across 14 datasets, 10 languages, and 6 backbones, and its practical efficacy is validated by real-time deployment on a humanoid robotic platform.
[530] BreathNet: Generalizable Audio Deepfake Detection via Breath-Cue-Guided Feature Refinement
Zhe Ye, Xiangui Kang, Jiayi He, Chengxin Chen, Wei Zhu, Kai Wu, Yin Yang, Jiwu Huang
Main category: cs.SD
TL;DR: BreathNet is a novel audio deepfake detection framework that integrates fine-grained breath information using BreathFiLM to improve generalization, combined with spectral features and feature losses for state-of-the-art performance.
Details
Motivation: As deepfake audio becomes more realistic and diverse, existing detection methods relying on XLS-R front-end features have limited generalization due to insufficient attention to fine-grained information like physiological cues or frequency-domain features.Method: Proposes BreathNet with BreathFiLM (feature-wise linear modulation) that selectively amplifies temporal representations based on breathing sounds, jointly trained with XLS-R extractor. Combines spectral features from frequency front-end and uses feature losses (PSCL, center loss, contrast loss) to enhance discriminative ability.
Result: Achieves state-of-the-art performance: 1.99% average EER across four benchmarks using ASVspoof 2019 LA training set, 4.70% EER on In-the-Wild dataset, and 4.94% EER under ASVspoof5 evaluation protocol.
Conclusion: BreathNet effectively integrates breath information and spectral features to improve audio deepfake detection generalization, demonstrating superior performance across multiple benchmarks.
Abstract: As deepfake audio becomes more realistic and diverse, developing generalizable countermeasure systems has become crucial. Existing detection methods primarily depend on XLS-R front-end features to improve generalization. Nonetheless, their performance remains limited, partly due to insufficient attention to fine-grained information, such as physiological cues or frequency-domain features. In this paper, we propose BreathNet, a novel audio deepfake detection framework that integrates fine-grained breath information to improve generalization. Specifically, we design BreathFiLM, a feature-wise linear modulation mechanism that selectively amplifies temporal representations based on the presence of breathing sounds. BreathFiLM is trained jointly with the XLS-R extractor, in turn encouraging the extractor to learn and encode breath-related cues into the temporal features. Then, we use the frequency front-end to extract spectral features, which are then fused with temporal features to provide complementary information introduced by vocoders or compression artifacts. Additionally, we propose a group of feature losses comprising Positive-only Supervised Contrastive Loss (PSCL), center loss, and contrast loss. These losses jointly enhance the discriminative ability, encouraging the model to separate bona fide and deepfake samples more effectively in the feature space. Extensive experiments on five benchmark datasets demonstrate state-of-the-art (SOTA) performance. Using the ASVspoof 2019 LA training set, our method attains 1.99% average EER across four related eval benchmarks, with particularly strong performance on the In-the-Wild dataset, where it achieves 4.70% EER. Moreover, under the ASVspoof5 evaluation protocol, our method achieves an EER of 4.94% on this latest benchmark.
[531] AuTAgent: A Reinforcement Learning Framework for Tool-Augmented Audio Reasoning
Siqian Tong, Xuan Li, Yiwei Wang, Baolong Bi, Yujun Cai, Shenghua Liu, Yuchen He, Chengpeng Hao
Main category: cs.SD
TL;DR: AuTAgent is a reinforcement learning framework that teaches audio language models when and which external tools to use for precise acoustic measurements, improving reasoning without information overload.
Details
Motivation: Large Audio Language Models (LALMs) are good at perception but struggle with complex reasoning requiring precise acoustic measurements. While external tools can extract fine-grained features like tempo or pitch, integrating them effectively is challenging - using all tools causes information overload, while prompt-based selection fails to assess context-dependent utility.Method: Proposes AuTAgent (Audio Tool Agent), a reinforcement learning framework that learns when and which tools to invoke. Uses a sparse-feedback training strategy with a novel Differential Reward mechanism to filter out irrelevant tools and invoke external assistance only when it yields net performance gain over the base model.
Result: Improves accuracy by 4.20%/6.20% and 9.80%/8.00% for open-source and closed-source backbones on MMAU Test-mini and MMAR benchmarks. Demonstrates exceptional transferability and shows that AuTAgent complements the representation bottleneck of LALMs by providing verifiable acoustic evidence.
Conclusion: The framework successfully addresses the tool integration challenge in audio language models, highlighting the complementary role of external tools in augmenting audio model reasoning through learned tool invocation strategies.
Abstract: Large Audio Language Models (LALMs) excel at perception but struggle with complex reasoning requiring precise acoustic measurements. While external tools can extract fine-grained features like exact tempo or pitch, effective integration remains challenging: naively using all tools causes information overload, while prompt-based selection fails to assess context-dependent utility. To address this, we propose AuTAgent (Audio Tool Agent), a reinforcement learning framework that learns when and which tools to invoke. By employing a sparse-feedback training strategy with a novel Differential Reward mechanism, the agent learns to filter out irrelevant tools and invokes external assistance only when it yields a net performance gain over the base model. Experimental results confirm that AuTAgent complements the representation bottleneck of LALMs by providing verifiable acoustic evidence. It improves accuracy by 4.20% / 6.20% and 9.80% / 8.00% for open-source and closed-source backbones on the MMAU Test-mini and the MMAR benchmarks, respectively. In addition, further experiments demonstrate exceptional transferability. We highlight the complementary role of external tools in augmenting audio model reasoning.
[532] The Interspeech 2026 Audio Reasoning Challenge: Evaluating Reasoning Process Quality for Audio Reasoning Models and Agents
Ziyang Ma, Ruiyang Xu, Yinghao Ma, Chao-Han Huck Yang, Bohan Li, Jaeyeon Kim, Jin Xu, Jinyu Li, Carlos Busso, Kai Yu, Eng Siong Chng, Xie Chen
Main category: cs.SD
TL;DR: The Audio Reasoning Challenge at Interspeech 2026 evaluates Chain-of-Thought reasoning quality in audio language models using MMAR-Rubrics protocol, with agent systems currently outperforming single models in reasoning transparency.
Details
Motivation: Large Audio Language Models (LALMs) excel at understanding but lack transparent reasoning, creating a "black-box" limitation that hinders explainable audio intelligence.Method: Organized the first shared task for evaluating CoT reasoning in audio domain using MMAR-Rubrics (instance-level protocol assessing factuality and logic), with Single Model and Agent tracks attracting 156 teams from 18 countries.
Result: Agent systems currently lead in reasoning quality using iterative tool orchestration and cross-modal analysis, while single models are advancing via reinforcement learning and sophisticated data pipelines.
Conclusion: The challenge provides new insights for explainable audio intelligence and establishes benchmarks for evaluating reasoning transparency in audio language models.
Abstract: Recent Large Audio Language Models (LALMs) excel in understanding but often lack transparent reasoning. To address this “black-box” limitation, we organized the Audio Reasoning Challenge at Interspeech 2026, the first shared task dedicated to evaluating Chain-of-Thought (CoT) quality in the audio domain. The challenge introduced MMAR-Rubrics, a novel instance-level protocol assessing the factuality and logic of reasoning chains. Featured Single Model and Agent tracks, the competition attracting 156 teams from 18 countries and regions. Results show agent systems currently lead in reasoning quality, utilizing iterative tool orchestration and cross-modal analysis. Besides, single models are rapidly advancing via reinforcement learning and sophisticated data pipeline. We details the challenge design, methodology, and a comprehensive analysis of state-of-the-art systems, providing new insights for explainable audio intelligence.
[533] Enhancing spatial hearing with cochlear implants: exploring the role of AI, multimodal interaction and perceptual training
Lorenzo Picinali, Robert Baumgartner, Valerie Gaveau, Antonino Greco, Stefanie Liebe, Paul Oomen, Christoph Braun
Main category: cs.SD
TL;DR: A multidisciplinary research framework combining medicine, psychology, and engineering to improve spatial hearing for cochlear implant users, addressing a previously neglected aspect of hearing restoration.
Details
Motivation: Spatial hearing is crucial for attention control, direction perception, and speech understanding in noisy environments, but has been largely neglected in cochlear implant development despite significant advances in restoring basic hearing and speech understanding.Method: Proposes a multidisciplinary collaborative framework involving physicians, psychologists, and engineers working together to address spatial hearing challenges for cochlear implant users through integrated expertise.
Result: The paper presents a conceptual framework for improving spatial hearing in cochlear implants, though specific experimental results are not mentioned in the abstract.
Conclusion: A collaborative multidisciplinary approach is necessary to advance spatial hearing capabilities in cochlear implants, addressing a critical gap in current hearing restoration technology.
Abstract: Cochlear implants (CIs) have been developed to the point where they can restore hearing and speech understanding in a large proportion of patients. Although spatial hearing is central to controlling and directing attention and to enabling speech understanding in noisy environments, it has been largely neglected in the past. We propose here a multi-disciplinary research framework in which physicians, psychologists and engineers collaborate to improve spatial hearing for CI users.
[534] Learning Vocal-Tract Area and Radiation with a Physics-Informed Webster Model
Minhui Lu, Joshua D. Reiss
Main category: cs.SD
TL;DR: Physics-informed neural network for singing-voice synthesis that estimates vocal-tract parameters from audio and F0, using a Webster equation model with DDSP stabilization during training but pure physics-based inference.
Details
Motivation: To create an interpretable, physics-based singing-voice synthesis system that can estimate vocal-tract parameters from audio while maintaining stability under various conditions like pitch shifts and source variations.Method: Uses a time-domain Webster model as a physics-informed neural network to estimate vocal-tract area function and radiation coefficient from synthetic audio and F0 trajectory. Training enforces PDE and boundary consistency with lightweight DDSP path for stabilization, while inference is purely physics-based.
Result: The method reproduces spectral envelopes competitively with DDSP baselines on sustained vowels, remains stable under discretization changes, moderate source variations, and ~10% pitch shifts, though produces breathier waveforms than reference.
Conclusion: Physics-informed approach shows promise for interpretable singing-voice synthesis, but needs periodicity-aware objectives and explicit glottal priors to reduce breathiness in future work.
Abstract: We present a physics-informed voiced backend renderer for singing-voice synthesis. Given synthetic single-channel audio and a fund-amental–frequency trajectory, we train a time-domain Webster model as a physics-informed neural network to estimate an interpretable vocal-tract area function and an open-end radiation coefficient. Training enforces partial differential equation and boundary consistency; a lightweight DDSP path is used only to stabilize learning, while inference is purely physics-based. On sustained vowels (/a/, /i/, /u/), parameters rendered by an independent finite-difference time-domain Webster solver reproduce spectral envelopes competitively with a compact DDSP baseline and remain stable under changes in discretization, moderate source variations, and about ten percent pitch shifts. The in-graph waveform remains breathier than the reference, motivating periodicity-aware objectives and explicit glottal priors in future work.
[535] Audiocards: Structured Metadata Improves Audio Language Models For Sound Design
Sripathi Sridhar, Prem Seetharaman, Oriol Nieto, Mark Cartwright, Justin Salamon
Main category: cs.SD
TL;DR: Audiocards: structured metadata for sound effects using LLMs to improve text-audio retrieval and captioning for sound design applications
Details
Motivation: Sound designers need better search capabilities in sound effects libraries, but existing metadata is often missing or incomplete. Current captioning and retrieval methods aren't trained on metadata with the specific structure and information needed for professional sound design.Method: Proposes audiocards - structured metadata grounded in acoustic attributes and sonic descriptors, generated by exploiting the world knowledge of Large Language Models (LLMs). The approach trains models on these structured audiocards rather than single-sentence captions.
Result: Training on audiocards improves downstream text-audio retrieval, descriptive captioning, and metadata generation on professional sound effects libraries. Also improves performance on general audio captioning and retrieval compared to baseline single-sentence captioning approaches.
Conclusion: Audiocards provide effective structured metadata for sound design applications, leveraging LLMs to bridge the gap between audio content and searchable metadata. The approach shows promise for audio-language modeling in professional sound design contexts.
Abstract: Sound designers search for sounds in large sound effects libraries using aspects such as sound class or visual context. However, the metadata needed for such search is often missing or incomplete, and requires significant manual effort to add. Existing solutions to automate this task by generating metadata, i.e. captioning, and search using learned embeddings, i.e. text-audio retrieval, are not trained on metadata with the structure and information pertinent to sound design. To this end we propose audiocards, structured metadata grounded in acoustic attributes and sonic descriptors, by exploiting the world knowledge of LLMs. We show that training on audiocards improves downstream text-audio retrieval, descriptive captioning, and metadata generation on professional sound effects libraries. Moreover, audiocards also improve performance on general audio captioning and retrieval over the baseline single-sentence captioning approach. We release a curated dataset of sound effects audiocards to invite further research in audio language modeling for sound design.
[536] GSRM: Generative Speech Reward Model for Speech RLHF
Maohao Shen, Tejas Jayashankar, Osama Hanna, Naoyuki Kanda, Yancheng Wang, Kateřina Žmolíková, Ruiming Xie, Niko Moritz, Anfeng Xu, Yashesh Gaur, Gregory Wornell, Qing He, Jilong Wu
Main category: cs.SD
TL;DR: GSRM is a generative speech reward model that uses chain-of-thought reasoning for interpretable speech naturalness evaluation, outperforming existing methods and improving speech LLM generation quality.
Details
Motivation: Current speech language models lack aesthetic naturalness compared to human speech, and existing naturalness evaluators are limited in interpretability and generalization across different speech taxonomies.Method: Proposes Generative Speech Reward Model (GSRM) that decomposes speech naturalness evaluation into interpretable acoustic feature extraction followed by feature-grounded chain-of-thought reasoning. Trained on 31k expert ratings and out-of-domain real-world user-assistant speech data.
Result: GSRM substantially outperforms existing speech naturalness predictors, achieving model-human correlation approaching human inter-rater consistency. Also improves speech LLM naturalness when used as verifier for online RLHF.
Conclusion: GSRM provides an effective, interpretable approach to speech naturalness evaluation that can enhance speech generation quality in multimodal language models.
Abstract: Recent advances in speech language models, such as GPT-4o Voice Mode and Gemini Live, have demonstrated promising speech generation capabilities. Nevertheless, the aesthetic naturalness of the synthesized audio still lags behind that of human speech. Enhancing generation quality requires a reliable evaluator of speech naturalness. However, existing naturalness evaluators typically regress raw audio to scalar scores, offering limited interpretability of the evaluation and moreover fail to generalize to speech across different taxonomies. Inspired by recent advances in generative reward modeling, we propose the Generative Speech Reward Model (GSRM), a reasoning-centric reward model tailored for speech. The GSRM is trained to decompose speech naturalness evaluation into an interpretable acoustic feature extraction stage followed by feature-grounded chain-of-thought reasoning, enabling explainable judgments. To achieve this, we curated a large-scale human feedback dataset comprising 31k expert ratings and an out-of-domain benchmark of real-world user-assistant speech interactions. Experiments show that GSRM substantially outperforms existing speech naturalness predictors, achieving model-human correlation of naturalness score prediction that approaches human inter-rater consistency. We further show how GSRM can improve the naturalness of speech LLM generations by serving as an effective verifier for online RLHF.
[537] voice2mode: Phonation Mode Classification in Singing using Self-Supervised Speech Models
Aju Ani Justus, Ruchit Agrawal, Sudarsana Reddy Kadiri, Shrikanth Narayanan
Main category: cs.SD
TL;DR: voice2mode uses self-supervised speech model embeddings (HuBERT, wav2vec2) to classify four singing phonation modes, achieving ~95.7% accuracy with early-layer HuBERT features, significantly outperforming traditional spectral features.
Details
Motivation: Prior singing phonation classification relies on handcrafted features or task-specific neural networks. This work explores whether speech foundation models can transfer effectively to singing phonation tasks, leveraging their rich audio representations.Method: Extracts layer-wise representations from HuBERT and wav2vec2 models, applies global temporal pooling to sustained vowel recordings, and uses lightweight classifiers (SVM, XGBoost) for classification of four phonation modes (breathy, neutral, flow, pressed).
Result: Foundation model features substantially outperform conventional spectral baselines. HuBERT embeddings from early layers achieve ~95.7% accuracy with SVM, an absolute improvement of ~12-15% over best traditional baseline. Lower layers (retaining acoustic/phonetic detail) outperform top layers specialized for ASR.
Conclusion: Speech foundation models transfer effectively to singing phonation classification, with early-layer embeddings capturing relevant acoustic details. This demonstrates the value of pre-trained audio representations for singing voice analysis tasks.
Abstract: We present voice2mode, a method for classification of four singing phonation modes (breathy, neutral (modal), flow, and pressed) using embeddings extracted from large self-supervised speech models. Prior work on singing phonation has relied on handcrafted signal features or task-specific neural nets; this work evaluates the transferability of speech foundation models to singing phonation classification. voice2mode extracts layer-wise representations from HuBERT and two wav2vec2 variants, applies global temporal pooling, and classifies the pooled embeddings with lightweight classifiers (SVM, XGBoost). Experiments on a publicly available soprano dataset (763 sustained vowel recordings, four labels) show that foundation-model features substantially outperform conventional spectral baselines (spectrogram, mel-spectrogram, MFCC). HuBERT embeddings obtained from early layers yield the best result (~95.7% accuracy with SVM), an absolute improvement of ~12-15% over the best traditional baseline. We also show layer-wise behaviour: lower layers, which retain acoustic/phonetic detail, are more effective than top layers specialized for Automatic Speech Recognition (ASR).
[538] Bengali-Loop: Community Benchmarks for Long-Form Bangla ASR and Speaker Diarization
H. M. Shadman Tabib, Istiak Ahmmed Rifti, Abdullah Muhammed Amimul Ehsan, Somik Dasgupta, Md Zim Mim Siddiqee Sowdha, Abrar Jahin Sarker, Md. Rafiul Islam Nijamy, Tanvir Hossain, Mst. Metaly Khatun, Munzer Mahmood, Rakesh Debnath, Gourab Biswas, Asif Karim, Wahid Al Azad Navid, Masnoon Muztahid, Fuad Ahmed Udoy, Shahad Shahriar Rahman, Md. Tashdiqur Rahman Shifat, Most. Sonia Khatun, Mushfiqur Rahman, Md. Miraj Hasan, Anik Saha, Mohammad Ninad Mahmud Nobo, Soumik Bhattacharjee, Tusher Bhomik, Ahmmad Nur Swapnil, Shahriar Kabir
Main category: cs.SD
TL;DR: Bengali-Loop: Two community benchmarks for long-form Bangla speech technology - ASR corpus (158.6 hours) and speaker diarization corpus (22 hours) with reproducible pipelines and human verification.
Details
Motivation: Bengali (Bangla) remains under-resourced in long-form speech technology despite its wide use, creating a gap for realistic multi-speaker, long-duration content applications.Method: Created two benchmarks: (1) long-form ASR corpus from 11 YouTube channels using reproducible subtitle-extraction pipeline with human-in-the-loop transcript verification; (2) speaker diarization corpus with fully manual speaker-turn labels in CSV format.
Result: Established baselines: Tugstugi achieved 34.07% WER for ASR; pyannote.audio achieved 40.08% DER for diarization. Provided standardized evaluation protocols, annotation rules, and data formats.
Conclusion: Bengali-Loop addresses the resource gap for Bangla long-form speech technology, enabling reproducible benchmarking and future model development for ASR and diarization in realistic multi-speaker scenarios.
Abstract: Bengali (Bangla) remains under-resourced in long-form speech technology despite its wide use. We present Bengali-Loop, two community benchmarks to address this gap: (1) a long-form ASR corpus of 191 recordings (158.6 hours, 792k words) from 11 YouTube channels, collected via a reproducible subtitle-extraction pipeline and human-in-the-loop transcript verification; and (2) a speaker diarization corpus of 24 recordings (22 hours, 5,744 annotated segments) with fully manual speaker-turn labels in CSV format. Both benchmarks target realistic multi-speaker, long-duration content (e.g., Bangla drama/natok). We establish baselines (Tugstugi: 34.07% WER; pyannote.audio: 40.08% DER) and provide standardized evaluation protocols (WER/CER, DER), annotation rules, and data formats to support reproducible benchmarking and future model development for Bangla long-form ASR and diarization.
[539] Eureka-Audio: Triggering Audio Intelligence in Compact Language Models
Dan Zhang, Yishu Lei, Jing Hu, Shuwei He, Songhe Deng, Xianlong Luo, Danxiang Zhu, Shikun Feng, Rui Liu, Jingzhou He, Yu Sun, Hua Wu, Haifeng Wang
Main category: cs.SD
TL;DR: Eureka-Audio is a 1.7B parameter compact audio language model that achieves competitive performance against much larger models (4-18x larger) across various audio understanding tasks through a unified architecture with MoE adapters and a novel data synthesis pipeline.
Details
Motivation: To create a lightweight yet high-performance audio language model that can handle diverse audio understanding tasks efficiently, addressing the computational cost challenges of large audio models while maintaining competitive performance.Method: Uses a unified end-to-end architecture with: 1) lightweight language backbone, 2) Whisper-based audio encoder, 3) sparsely activated Mixture-of-Experts (MoE) adapter to handle audio heterogeneity and reduce cross-modal optimization conflicts. Also introduces DataFlux pipeline for high-quality audio instruction data synthesis and verification.
Result: Achieves competitive performance against models 4-18 times larger across ASR, audio understanding, and dense audio captioning benchmarks. Matches or surpasses 7B to 30B audio and omni-modal baselines despite having only 1.7B parameters.
Conclusion: Eureka-Audio establishes a strong and practical baseline for lightweight audio understanding models, demonstrating efficient balance between computational cost and performance through its unified architecture and data synthesis pipeline.
Abstract: We present Eureka-Audio, a compact yet high-performance audio language model that achieves competitive performance against models that are 4 to 18 times larger across a broad range of audio understanding benchmarks. Despite containing only 1.7B parameters, Eureka-Audio demonstrates strong performance on automatic speech recognition (ASR), audio understanding, and dense audio captioning, matching or surpassing multiple 7B to 30B audio and omni-modal baselines. The model adopts a unified end-to-end architecture composed of a lightweight language backbone, a Whisper-based audio encoder, and a sparsely activated Mixture-of-Experts (MoE) adapter that explicitly accounts for audio heterogeneity and alleviates cross-modal optimization conflicts under limited capacity. To further enhance paralinguistic reasoning, we introduce DataFlux, a closed loop audio instruction data synthesis and verification pipeline that constructs high quality, logically consistent supervision from raw audio. Extensive evaluations across ASR, knowledge reasoning, safety, instruction following, and paralinguistic benchmarks, demonstrate that Eureka-Audio achieves an efficient balance between computational cost and performance. These results establish Eureka Audio as a strong and practical baseline for lightweight audio understanding models.
[540] MUKA: Multi Kernel Audio Adaptation Of Audio-Language Models
Reda Bensaid, Amine Ouasfi, Yassir Bendou, Ilyass Moummad, Vincent Gripon, François Leduc-Primeau, Adnane Boukhayma
Main category: cs.SD
TL;DR: MUKA: A multi-kernel adaptation framework for few-shot audio-language model adaptation that combines fine-grained instruction-tuning representations with global contrastive pretraining semantics without additional training.
Details
Motivation: Multimodal foundation models show strong generalization but adapting them efficiently to new tasks in few-shot settings remains challenging. Current approaches either require extensive training or lack the ability to leverage both local and global semantic representations effectively.Method: Proposes MUKA, a multi-kernel adaptation framework that combines representations from instruction-tuning based models (like Pengi) with contrastive pretraining methods (like CLAP). Uses a product kernel to align local similarity with global semantics, preserving kernel method guarantees while avoiding additional training.
Result: Extensive experiments on 11 diverse audio datasets show MUKA achieves state-of-the-art performance among training-free methods and even surpasses training-based adapters in several scenarios.
Conclusion: MUKA offers a compelling balance between adaptability and efficiency for few-shot adaptation of audio-language models, demonstrating that training-free methods can achieve competitive performance through effective combination of different representation types.
Abstract: Multimodal foundation models have demonstrated impressive generalization capabilities, yet efficiently adapting them to new tasks in a few-shot setting remains a critical challenge. In this work, we investigate the few-shot adaptation of Large Audio-Language Models (ALMs) through both training-based and training-free approaches. We introduce MUKA, a multi-kernel adaptation framework that combines the fine-grained, context-dependent representations of instruction-tuning based models like Pengi with the global semantic representations of contrastive pretraining methods like CLAP. By constructing a product kernel that aligns local similarity with global semantics, MUKA enhances representational power while preserving the theoretical guarantees of kernel methods and avoiding additional training. Extensive experiments across 11 diverse audio datasets demonstrate that MUKA achieves state-of-the-art performance among training-free methods and even surpasses training-based adapters in several scenarios, offering a compelling balance between adaptability and efficiency.
[541] HiFi-Glot: High-Fidelity Neural Formant Synthesis with Differentiable Resonant Filters
Yicheng Gu, Pablo Pérez Zarazaga, Chaoren Wang, Zhizheng Wu, Zofia Malisz, Gustav Eje Henter, Lauri Juvela
Main category: cs.SD
TL;DR: HiFi-Glot is an end-to-end neural formant synthesis system that achieves both precise formant control and high-fidelity speech synthesis using a source-filter architecture with neural vocoder and differentiable resonant filters.
Details
Motivation: Existing formant synthesis approaches enable precise formant manipulation but often yield impoverished speech signals by failing to capture complex co-occurring acoustic cues essential for naturalness.Method: The model adopts a source-filter architecture inspired by classical formant synthesis, where a neural vocoder generates the glottal excitation signal, and differentiable resonant filters model the formants to produce the speech waveform.
Result: The proposed HiFi-Glot model generates speech with higher perceptual quality and naturalness while exhibiting more precise control over formant frequencies, outperforming industry-standard formant manipulation tools like Praat.
Conclusion: HiFi-Glot successfully addresses the trade-off between precise formant control and speech naturalness in formant synthesis, achieving both high-fidelity synthesis and precise formant manipulation.
Abstract: Formant synthesis aims to generate speech with controllable formant structures, enabling precise control of vocal resonance and phonetic features. However, while existing formant synthesis approaches enable precise formant manipulation, they often yield an impoverished speech signal by failing to capture the complex co-occurring acoustic cues essential for naturalness. To address this issue, this letter presents HiFi-Glot, an end-to-end neural formant synthesis system that achieves both precise formant control and high-fidelity speech synthesis. Specifically, the proposed model adopts a source–filter architecture inspired by classical formant synthesis, where a neural vocoder generates the glottal excitation signal, and differentiable resonant filters model the formants to produce the speech waveform. Experiment results demonstrate that our proposed HiFi-Glot model can generate speech with higher perceptual quality and naturalness while exhibiting a more precise control over formant frequencies, outperforming industry-standard formant manipulation tools such as Praat. Code, checkpoints, and representative audio samples are available at https://www.yichenggu.com/HiFi-Glot/.
[542] Investigation for Relative Voice Impression Estimation
Keinichi Fujita, Yusuke Ijima
Main category: cs.SD
TL;DR: This paper investigates relative voice impression estimation (RIE) - predicting perceptual differences between two utterances from the same speaker using low-dimensional vectors derived from subjective evaluations along antonymic axes like “Dark-Bright” or “Cold-Warm”.
Details
Motivation: Most research focuses on absolute impression scoring, but this study explores relative voice impression estimation to better capture perceptual differences between utterances from the same speaker, which is important for understanding paralinguistic and non-linguistic aspects of speech.Method: Used recordings of a professional speaker reading text in various styles to isolate expressive/prosodic variation. Compared three modeling approaches: 1) classical acoustic features for speech emotion recognition, 2) self-supervised speech representations, and 3) multimodal large language models (MLLMs).
Result: Self-supervised speech representations outperformed classical acoustic features, especially for complex dynamic impressions like “Cold-Warm”. Current MLLMs proved unreliable for this fine-grained pairwise task. This is the first systematic investigation of RIE.
Conclusion: Self-supervised speech models are effective for capturing subtle perceptual variations in voice impressions, while current MLLMs are not suitable for this fine-grained pairwise comparison task. The study establishes RIE as a valuable framework for voice impression analysis.
Abstract: Paralinguistic and non-linguistic aspects of speech strongly influence listener impressions. While most research focuses on absolute impression scoring, this study investigates relative voice impression estimation (RIE), a framework for predicting the perceptual difference between two utterances from the same speaker. The estimation target is a low-dimensional vector derived from subjective evaluations, quantifying the perceptual shift of the second utterance relative to the first along an antonymic axis (e.g., Dark--Bright''). To isolate expressive and prosodic variation, we used recordings of a professional speaker reading a text in various styles. We compare three modeling approaches: classical acoustic features commonly used for speech emotion recognition, self-supervised speech representations, and multimodal large language models (MLLMs). Our results demonstrate that models using self-supervised representations outperform methods with classical acoustic features, particularly in capturing complex and dynamic impressions (e.g., Cold–Warm’’) where classical features fail. In contrast, current MLLMs prove unreliable for this fine-grained pairwise task. This study provides the first systematic investigation of RIE and demonstrates the strength of self-supervised speech models in capturing subtle perceptual variations.
[543] VoiceBridge: General Speech Restoration with One-step Latent Bridge Models
Chi Zhang, Zehua Chen, Kaiwen Zheng, Jun Zhu
Main category: cs.SD
TL;DR: VoiceBridge: A one-step latent bridge model for general speech restoration that reconstructs 48kHz fullband speech from diverse distortions using energy-preserving VAE and joint neural prior.
Details
Motivation: Existing bridge models for speech enhancement are mostly single-task with constrained general speech restoration capability. There's a need for a unified model that can handle various speech restoration tasks efficiently without distillation.Method: Proposes VoiceBridge, a one-step latent bridge model using: 1) energy-preserving variational autoencoder for better waveform-latent alignment, 2) single latent-to-latent generative process with scalable transformer, 3) joint neural prior to reduce burden across diverse tasks, and 4) joint training of LBM, decoder and discriminator to transform from denoiser to generator.
Result: Superior performance demonstrated across in-domain tasks (denoising, super-resolution) and out-of-domain tasks (refining synthesized speech) on various datasets, enabling one-step general speech restoration without distillation.
Conclusion: VoiceBridge provides an effective framework for general speech restoration that can handle diverse distortions with a single model, achieving high-quality 48kHz speech reconstruction through innovative latent space modeling and training techniques.
Abstract: Bridge models have been investigated in speech enhancement but are mostly single-task, with constrained general speech restoration (GSR) capability. In this work, we propose VoiceBridge, a one-step latent bridge model (LBM) for GSR, capable of efficiently reconstructing 48 kHz fullband speech from diverse distortions. To inherit the advantages of data-domain bridge models, we design an energy-preserving variational autoencoder, enhancing the waveform-latent space alignment over varying energy levels. By compressing waveform into continuous latent representations, VoiceBridge models~\textit{various} GSR tasks with a~\textit{single} latent-to-latent generative process backed by a scalable transformer. To alleviate the challenge of reconstructing the high-quality target from distinctively different low-quality priors, we propose a joint neural prior for GSR, uniformly reducing the burden of the LBM in diverse tasks. Building upon these designs, we further investigate bridge training objective by jointly tuning LBM, decoder and discriminator together, transforming the model from a denoiser to generator and enabling \textit{one-step GSR without distillation}. Extensive validation across in-domain (\textit{e.g.}, denoising and super-resolution) and out-of-domain tasks (\textit{e.g.}, refining synthesized speech) and datasets demonstrates the superior performance of VoiceBridge. Demos: https://VoiceBridgedemo.github.io/.
[544] Probing Human Articulatory Constraints in End-to-End TTS with Reverse and Mismatched Speech-Text Directions
Parth Khadse, Sunil Kumar Kopparapu
Main category: cs.SD
TL;DR: Experimental study showing that reversing text and speech sequences in TTS training improves speech quality, suggesting e2e-TTS systems are purely data-driven without inherent anatomical constraints.
Details
Motivation: To investigate whether human anatomical constraints in speech production affect end-to-end TTS system training, by testing if reversing text and speech sequences impacts model performance.Method: Tested two e2e-TTS architectures (Tacotron-2 autoregressive and VITS-TTS non-autoregressive) with three configurations: (a) forward text + forward speech (conventional), (b) reverse text + reverse speech, and (c) reverse text + forward speech.
Result: Reversed text and speech TTS systems (r-e2e-TTS) generated speech with better fidelity, perceptual intelligibility, and naturalness than conventional forward configurations, demonstrating e2e-TTS systems are purely data-driven.
Conclusion: E2e-TTS systems learn from data patterns without inherent anatomical constraints; reversing sequences can improve speech quality, suggesting optimization opportunities in training configurations.
Abstract: An end-to-end (e2e) text-to-speech (TTS) system is a deep architecture that learns to associate a text string with acoustic speech patterns from a curated dataset. It is expected that all aspects associated with speech production, such as phone duration, speaker characteristics, and intonation among other things are captured in the trained TTS model to enable the synthesized speech to be natural and intelligible. Human speech is complex, involving smooth transitions between articulatory configurations (ACs). Due to anatomical constraints, some ACs are challenging to mimic or transition between. In this paper, we experimentally study if the constraints imposed by human anatomy have an implication on training an e2e-TTS systems. We experiment with two e2e-TTS architectures, namely, Tacotron-2 an autoregressive model and VITS-TTS a non-autoregressive model. In this study, we build TTS systems using (a) forward text, forward speech (conventional, e2e-TTS), (b) reverse text, reverse speech (r-e2e-TTS), and (c) reverse text, forward speech (rtfs-e2e-TTS). Experiments demonstrate that e2e-TTS systems are purely data-driven. Interestingly, the generated speech by r-e2e-TTS systems exhibits better fidelity, better perceptual intelligibility, and better naturalness
[545] RRPO: Robust Reward Policy Optimization for LLM-based Emotional TTS
Cong Wang, Changfeng Gao, Yang Xiang, Zhihao Du, Keyu An, Han Zhao, Qian Chen, Xiangang Li, Yingming Gao, Ya Li
Main category: cs.SD
TL;DR: RRPO is a robust reinforcement learning framework for controllable text-to-speech that prevents reward hacking in emotion control by using hybrid regularization to align reward signals with human perception.
Details
Motivation: Differentiable RL frameworks for controllable TTS are vulnerable to reward hacking where models generate acoustic artifacts to exploit reward models, degrading perceptual quality while achieving spurious rewards for emotion control tasks.Method: Proposes Robust Reward Policy Optimization (RRPO) with hybrid regularization scheme to develop robust reward models whose signals are reliably aligned with human perception, preventing policy models from taking detrimental shortcuts.
Result: Ablation study confirms enhanced robustness of the reward model with strong cross-lingual generalization. Subjective evaluation shows RRPO effectively mitigates reward hacking, leading to significant improvements in both emotional expressiveness and naturalness over all baselines.
Conclusion: RRPO successfully addresses reward hacking in differentiable RL for controllable TTS, enabling more reliable emotion control while maintaining high perceptual quality through robust reward modeling.
Abstract: Differentiable reinforcement learning (RL) frameworks like DiffRO offer a powerful approach for controllable text-to-speech (TTS), but are vulnerable to reward hacking, particularly for nuanced tasks like emotion control. The policy model can exploit a vanilla Reward Model (RM) by generating acoustic artifacts to achieve spurious rewards, but at the cost of degrading perceptual quality. To address this, we propose Robust Reward Policy Optimization (RRPO), a novel framework that employs a hybrid regularization scheme. This scheme develops a robust RM whose reward signal is more reliably aligned with human perception, compelling the policy to abandon detrimental shortcuts and instead learn the complex features of genuine emotions. Our ablation study confirms the enhanced robustness of our RM, as evidenced by its strong cross-lingual generalization. The subjective evaluation demonstrates that this robust RM effectively mitigates reward hacking, leading to significant improvements in both emotional expressiveness and naturalness over all baselines. Demo page: https://lrwinr.github.io/RRPO-CosyVoice.
[546] Evaluating Disentangled Representations for Controllable Music Generation
Laura Ibáñez-Martínez, Chukwuemeka Nkama, Andrea Poltronieri, Xavier Serra, Martín Rocamora
Main category: cs.SD
TL;DR: Analysis of disentangled representations in music audio models reveals inconsistencies between intended and actual semantics, questioning current approaches to controllable music generation.
Details
Motivation: Recent music generation approaches use disentangled representations (structure/timbre, local/global) for controllable synthesis, but the underlying properties of these embeddings remain underexplored, prompting systematic evaluation.Method: Probing-based framework beyond standard downstream tasks to evaluate disentangled representations in music audio models. Analyzes diverse unsupervised disentanglement strategies (inductive biases, data augmentations, adversarial objectives, staged training) across four axes: informativeness, equivariance, invariance, and disentanglement.
Result: Findings reveal inconsistencies between intended and actual semantics of embeddings, suggesting current strategies fall short of producing truly disentangled representations.
Conclusion: Current disentanglement approaches in music generation are insufficient, prompting re-examination of how controllability is approached in the field.
Abstract: Recent approaches in music generation rely on disentangled representations, often labeled as structure and timbre or local and global, to enable controllable synthesis. Yet the underlying properties of these embeddings remain underexplored. In this work, we evaluate such disentangled representations in a set of music audio models for controllable generation using a probing-based framework that goes beyond standard downstream tasks. The selected models reflect diverse unsupervised disentanglement strategies, including inductive biases, data augmentations, adversarial objectives, and staged training procedures. We further isolate specific strategies to analyze their effect. Our analysis spans four key axes: informativeness, equivariance, invariance, and disentanglement, which are assessed across datasets, tasks, and controlled transformations. Our findings reveal inconsistencies between intended and actual semantics of the embeddings, suggesting that current strategies fall short of producing truly disentangled representations, and prompting a re-examination of how controllability is approached in music generation.
[547] Towards explainable reference-free speech intelligibility evaluation of people with pathological speech
Bence Mark Halpern, Thomas Tienkamp, Defne Abur, Tomoki Toda
Main category: cs.SD
TL;DR: Proposes a reference-free, explainable ASR Inconsistency Score for objective assessment of pathological speech, showing high correlation with expert perceptual ratings across multiple languages.
Details
Motivation: Existing objective speech assessments (especially reference-based approaches) lack explainability and require labor-intensive manual transcriptions, making them impractical for clinical use.Method: Develops a reference-free ASR Inconsistency Score that doesn’t require ground truth transcripts, using automatic speech recognition inconsistencies as a metric for speech quality assessment.
Result: The ASR Inconsistency Score achieves high correlation with expert perceptual ratings, performing comparably to or better than standard reference-based Word Error Rate baselines across Dutch, Spanish, and English pathological speech.
Conclusion: The proposed reference-free, explainable method provides a practical alternative to traditional reference-based approaches for objective speech assessment in clinical and research settings.
Abstract: Objective assessment of speech that reflects meaningful changes in communication is crucial for clinical decision making and reproducible research. While existing objective assessments, particularly reference-based approaches, can capture intelligibility changes, they are often hindered by lack of explainability and the need for labor-intensive manual transcriptions. To address these issues, this work proposes the reference-free, explainable ASR Inconsistency Score. We evaluate this method on pathological speech in Dutch, Spanish and English, and compare its performance to a reference-based Word Error Rate (WER) baseline. Our results demonstrate that the ASR Inconsistency Score achieves a high correlation with expert perceptual ratings, with performance closely matching, and in one case exceeding, a standard reference-based Word Error Rate (WER) baseline.
cs.LG
[548] Directional Concentration Uncertainty: A representational approach to uncertainty quantification for generative models
Souradeep Chattopadhyay, Brendan Kennedy, Sai Munikoti, Soumik Sarkar, Karl Pazdernik
Main category: cs.LG
TL;DR: A novel uncertainty quantification framework called Directional Concentration Uncertainty (DCU) that measures geometric dispersion of embeddings using von Mises-Fisher distribution, outperforming heuristic methods and generalizing to multimodal tasks.
Details
Motivation: Current uncertainty quantification methods for generative models rely on rigid heuristics that don't generalize well across different tasks and modalities, limiting their trustworthiness and robustness.Method: Proposes Directional Concentration Uncertainty (DCU), a statistical procedure based on von Mises-Fisher distribution that quantifies embedding concentration by measuring geometric dispersion of multiple generated outputs using continuous embeddings without task-specific heuristics.
Result: DCU matches or exceeds calibration levels of prior methods like semantic entropy and generalizes well to more complex tasks in multimodal domains.
Conclusion: DCU provides a flexible uncertainty quantification framework with strong performance and generalization capabilities, showing potential for integration into multimodal and agentic systems.
Abstract: In the critical task of making generative models trustworthy and robust, methods for Uncertainty Quantification (UQ) have begun to show encouraging potential. However, many of these methods rely on rigid heuristics that fail to generalize across tasks and modalities. Here, we propose a novel framework for UQ that is highly flexible and approaches or surpasses the performance of prior heuristic methods. We introduce Directional Concentration Uncertainty (DCU), a novel statistical procedure for quantifying the concentration of embeddings based on the von Mises-Fisher (vMF) distribution. Our method captures uncertainty by measuring the geometric dispersion of multiple generated outputs from a language model using continuous embeddings of the generated outputs without any task specific heuristics. In our experiments, we show that DCU matches or exceeds calibration levels of prior works like semantic entropy (Kuhn et al., 2023) and also generalizes well to more complex tasks in multi-modal domains. We present a framework for the wider potential of DCU and its implications for integration into UQ for multi-modal and agentic frameworks.
[549] BLUEPRINT Rebuilding a Legacy: Multimodal Retrieval for Complex Engineering Drawings and Documents
Ethan Seefried, Ran Eldegaway, Sanjay Das, Nathaniel Blanchard, Tirthankar Ghosal
Main category: cs.LG
TL;DR: Blueprint is a multimodal retrieval system for engineering drawings that uses layout-aware region detection, VLM-based OCR, and fused retrieval to extract structured metadata from legacy archives.
Details
Motivation: Legacy engineering archives contain decades of drawings and technical records with inconsistent metadata, making retrieval difficult and manual. There's a need for automated systems to unlock these valuable resources.Method: System detects canonical drawing regions, applies region-restricted VLM-based OCR, normalizes identifiers, and fuses lexical and dense retrieval with lightweight region-level reranking.
Result: Deployed on ~770k unlabeled files, achieves 10.1% absolute gain in Success@3 and 18.9% relative improvement in nDCG@3 over strongest vision-language baseline on 5k-file benchmark with 350 expert queries.
Conclusion: Blueprint demonstrates effective multimodal retrieval for engineering archives, with substantial headroom for improvement under perfect region detection and OCR. System enables cross-facility search of legacy repositories.
Abstract: Decades of engineering drawings and technical records remain locked in legacy archives with inconsistent or missing metadata, making retrieval difficult and often manual. We present Blueprint, a layout-aware multimodal retrieval system designed for large-scale engineering repositories. Blueprint detects canonical drawing regions, applies region-restricted VLM-based OCR, normalizes identifiers (e.g., DWG, part, facility), and fuses lexical and dense retrieval with a lightweight region-level reranker. Deployed on ~770k unlabeled files, it automatically produces structured metadata suitable for cross-facility search. We evaluate Blueprint on a 5k-file benchmark with 350 expert-curated queries using pooled, graded (0/1/2) relevance judgments. Blueprint delivers a 10.1% absolute gain in Success@3 and an 18.9% relative improvement in nDCG@3 over the strongest vision-language baseline}, consistently outperforming across vision, text, and multimodal intents. Oracle ablations reveal substantial headroom under perfect region detection and OCR. We release all queries, runs, annotations, and code to facilitate reproducible evaluation on legacy engineering archives.
[550] Exploring the Performance of ML/DL Architectures on the MNIST-1D Dataset
Michael Beebe, GodsGift Uzor, Manasa Chepuri, Divya Sree Vemula, Angel Ayala
Main category: cs.LG
TL;DR: Evaluation of ResNet, TCN, and DCNN architectures on MNIST-1D dataset shows advanced sequential models outperform simpler baselines, achieving near-human performance and validating MNIST-1D as a benchmark for resource-constrained environments.
Details
Motivation: Small datasets like MNIST are useful for rapid experimentation but too simple to distinguish advanced architectures. MNIST-1D provides a more challenging 1D sequential adaptation that maintains small-scale advantages while introducing complexity to better evaluate modern neural network architectures.Method: Benchmark evaluation of Residual Networks (ResNet), Temporal Convolutional Networks (TCN), and Dilated Convolutional Neural Networks (DCNN) on the MNIST-1D dataset. These models were compared against previously tested baselines including logistic regression, MLPs, CNNs, and GRUs to assess performance on sequential data.
Result: Advanced architectures like TCN and DCNN consistently outperform simpler models, achieving near-human performance on MNIST-1D. ResNet also shows significant improvements. The results demonstrate the importance of inductive biases and hierarchical feature extraction in small structured datasets.
Conclusion: MNIST-1D serves as a robust benchmark for evaluating machine learning architectures under computational constraints. Architectural innovations play a crucial role in improving model performance, offering insights for optimizing deep learning models in resource-limited environments.
Abstract: Small datasets like MNIST have historically been instrumental in advancing machine learning research by providing a controlled environment for rapid experimentation and model evaluation. However, their simplicity often limits their utility for distinguishing between advanced neural network architectures. To address these challenges, Greydanus et al. introduced the MNIST-1D dataset, a one-dimensional adaptation of MNIST designed to explore inductive biases in sequential data. This dataset maintains the advantages of small-scale datasets while introducing variability and complexity that make it ideal for studying advanced architectures. In this paper, we extend the exploration of MNIST-1D by evaluating the performance of Residual Networks (ResNet), Temporal Convolutional Networks (TCN), and Dilated Convolutional Neural Networks (DCNN). These models, known for their ability to capture sequential patterns and hierarchical features, were implemented and benchmarked alongside previously tested architectures such as logistic regression, MLPs, CNNs, and GRUs. Our experimental results demonstrate that advanced architectures like TCN and DCNN consistently outperform simpler models, achieving near-human performance on MNIST-1D. ResNet also shows significant improvements, highlighting the importance of leveraging inductive biases and hierarchical feature extraction in small structured datasets. Through this study, we validate the utility of MNIST-1D as a robust benchmark for evaluating machine learning architectures under computational constraints. Our findings emphasize the role of architectural innovations in improving model performance and offer insights into optimizing deep learning models for resource-limited environments.
[551] The Speed-up Factor: A Quantitative Multi-Iteration Active Learning Performance Metric
Hannes Kath, Thiago S. Gouvêa, Daniel Sonntag
Main category: cs.LG
TL;DR: Active learning evaluation metric: speed-up factor measures fraction of samples needed to match random sampling performance, showing superior stability across iterations.
Details
Motivation: Active learning research focuses on query method development but lacks appropriate performance metrics for evaluating the iterative process. Current evaluation methods don't adequately capture multi-iteration performance.Method: Introduces speed-up factor as a quantitative multi-iteration performance metric. Evaluates using 4 diverse datasets and 7 query methods, comparing with state-of-the-art AL performance metrics.
Result: Speed-up factor accurately captures the fraction of samples needed to match random sampling performance and shows superior stability across iterations compared to existing metrics.
Conclusion: Speed-up factor is a reliable metric for evaluating active learning query methods, addressing the gap in multi-iteration performance assessment.
Abstract: Machine learning models excel with abundant annotated data, but annotation is often costly and time-intensive. Active learning (AL) aims to improve the performance-to-annotation ratio by using query methods (QMs) to iteratively select the most informative samples. While AL research focuses mainly on QM development, the evaluation of this iterative process lacks appropriate performance metrics. This work reviews eight years of AL evaluation literature and formally introduces the speed-up factor, a quantitative multi-iteration QM performance metric that indicates the fraction of samples needed to match random sampling performance. Using four datasets from diverse domains and seven QMs of various types, we empirically evaluate the speed-up factor and compare it with state-of-the-art AL performance metrics. The results confirm the assumptions underlying the speed-up factor, demonstrate its accuracy in capturing the described fraction, and reveal its superior stability across iterations.
[552] Accelerated Discovery of Cryoprotectant Cocktails via Multi-Objective Bayesian Optimization
Daniel Emerson, Nora Gaby-Biegel, Purva Joshi, Yoed Rabin, Rebecca D. Sandlin, Levent Burak Kara
Main category: cs.LG
TL;DR: A data-efficient framework combining high-throughput screening with multi-objective Bayesian optimization to accelerate cryoprotectant agent cocktail design for vitrification.
Details
Motivation: Designing CPA cocktails faces a tradeoff between high concentration to suppress ice formation and low toxicity to preserve cell viability, creating a large design space where traditional discovery is slow and relies on expert intuition or exhaustive experimentation.Method: Combines high-throughput screening with an active-learning loop based on multi-objective Bayesian optimization. Trains probabilistic surrogate models to predict concentration and viability, quantifies uncertainty, and iteratively selects experiments by prioritizing cocktails expected to improve the Pareto front using expected Pareto improvement under uncertainty.
Result: Wet-lab validation shows efficient discovery of cocktails achieving high CPA concentrations and high post-exposure viability. Improves dominated hypervolume by 9.5% and 4.5% relative to naive strategy and strong baseline, while reducing experiments needed. Synthetic studies recover comparably strong Pareto-optimal solutions using only 30% of evaluations required by prior state-of-the-art.
Conclusion: The framework accelerates cryopreservation development by efficiently navigating the multi-objective design space of CPA cocktails, with potential adaptation to different CPA libraries, objective definitions, and cell lines.
Abstract: Designing cryoprotectant agent (CPA) cocktails for vitrification is challenging because formulations must be concentrated enough to suppress ice formation yet non-toxic enough to preserve cell viability. This tradeoff creates a large, multi-objective design space in which traditional discovery is slow, often relying on expert intuition or exhaustive experimentation. We present a data-efficient framework that accelerates CPA cocktail design by combining high-throughput screening with an active-learning loop based on multi-objective Bayesian optimization. From an initial set of measured cocktails, we train probabilistic surrogate models to predict concentration and viability and quantify uncertainty across candidate formulations. We then iteratively select the next experiments by prioritizing cocktails expected to improve the Pareto front, maximizing expected Pareto improvement under uncertainty, and update the models as new assay results are collected. Wet-lab validation shows that our approach efficiently discovers cocktails that simultaneously achieve high CPA concentrations and high post-exposure viability. Relative to a naive strategy and a strong baseline, our method improves dominated hypervolume by 9.5% and 4.5%, respectively, while reducing the number of experiments needed to reach high-quality solutions. In complementary synthetic studies, it recovers a comparably strong set of Pareto-optimal solutions using only 30% of the evaluations required by the prior state-of-the-art multi-objective approach, which amounts to saving approximately 10 weeks of experimental time. Because the framework assumes only a suitable assay and defined formulation space, it can be adapted to different CPA libraries, objective definitions, and cell lines to accelerate cryopreservation development.
[553] Why is Normalization Preferred? A Worst-Case Complexity Theory for Stochastically Preconditioned SGD under Heavy-Tailed Noise
Yuchen Fang, James Demmel, Javad Lavaei
Main category: cs.LG
TL;DR: Theoretical analysis shows normalization outperforms clipping for stabilizing stochastic preconditioned SGD methods (like Adam, RMSProp) under heavy-tailed noise, with normalization achieving optimal convergence rates while clipping may fail.
Details
Motivation: Adaptive optimization methods like Adam, RMSProp, and Shampoo are widely used in deep learning but lack theoretical understanding under heavy-tailed noise conditions. There's a need to analyze stabilization techniques (clipping vs normalization) for these stochastically preconditioned methods.Method: Develops worst-case complexity theory for stochastically preconditioned SGD (SPSGD) and accelerated variants under heavy-tailed noise with finite p-th moments. Analyzes normalization and clipping as stabilization techniques, develops novel vector-valued Burkholder-type inequality for the analysis.
Result: Normalization guarantees convergence to first-order stationary points at optimal rates (O(T^{-(p-1)/(3p-2)}) with known parameters, O(T^{-(p-1)/(2p)}) with unknown parameters), while clipping may fail to converge in worst-case scenarios due to statistical dependence between preconditioner and gradient estimates.
Conclusion: Normalization is theoretically superior to clipping for stabilizing stochastically preconditioned methods under heavy-tailed noise, explaining empirical preference for normalization in large-scale model training.
Abstract: We develop a worst-case complexity theory for stochastically preconditioned stochastic gradient descent (SPSGD) and its accelerated variants under heavy-tailed noise, a setting that encompasses widely used adaptive methods such as Adam, RMSProp, and Shampoo. We assume the stochastic gradient noise has a finite $p$-th moment for some $p \in (1,2]$, and measure convergence after $T$ iterations. While clipping and normalization are parallel tools for stabilizing training of SGD under heavy-tailed noise, there is a fundamental separation in their worst-case properties in stochastically preconditioned settings. We demonstrate that normalization guarantees convergence to a first-order stationary point at rate $\mathcal{O}(T^{-\frac{p-1}{3p-2}})$ when problem parameters are known, and $\mathcal{O}(T^{-\frac{p-1}{2p}})$ when problem parameters are unknown, matching the optimal rates for normalized SGD, respectively. In contrast, we prove that clipping may fail to converge in the worst case due to the statistical dependence between the stochastic preconditioner and the gradient estimates. To enable the analysis, we develop a novel vector-valued Burkholder-type inequality that may be of independent interest. These results provide a theoretical explanation for the empirical preference for normalization over clipping in large-scale model training.
[554] Fast Swap-Based Element Selection for Multiplication-Free Dimension Reduction
Nobutaka Ono
Main category: cs.LG
TL;DR: A fast algorithm for element selection-based dimension reduction that selects subsets of input elements without matrix multiplications, using swap-based local search optimization.
Details
Motivation: Standard dimension reduction methods like PCA rely on matrix multiplications which can be computationally expensive on resource-constrained systems. Element selection offers a multiplication-free alternative by simply selecting a subset of elements, but the challenge is determining which elements to retain.Method: Proposes evaluating candidate subsets through minimum mean-squared error of linear regression predicting target vectors (labels or input reconstruction). Uses matrix inversion lemma to derive efficient formula for objective change when swapping selected/unselected elements, enabling swap-based local search optimization.
Result: Experiments on MNIST handwritten-digit images demonstrate the effectiveness of the proposed element selection method for dimension reduction.
Conclusion: The proposed algorithm provides an efficient, multiplication-free approach to dimension reduction through element selection, particularly suitable for resource-constrained systems where matrix multiplications are prohibitive.
Abstract: In this paper, we propose a fast algorithm for element selection, a multiplication-free form of dimension reduction that produces a dimension-reduced vector by simply selecting a subset of elements from the input. Dimension reduction is a fundamental technique for reducing unnecessary model parameters, mitigating overfitting, and accelerating training and inference. A standard approach is principal component analysis (PCA), but PCA relies on matrix multiplications; on resource-constrained systems, the multiplication count itself can become a bottleneck. Element selection eliminates this cost because the reduction consists only of selecting elements, and thus the key challenge is to determine which elements should be retained. We evaluate a candidate subset through the minimum mean-squared error of linear regression that predicts a target vector from the selected elements, where the target may be, for example, a one-hot label vector in classification. When an explicit target is unavailable, the input itself can be used as the target, yielding a reconstruction-based criterion. The resulting optimization is combinatorial, and exhaustive search is impractical. To address this, we derive an efficient formula for the objective change caused by swapping a selected and an unselected element, using the matrix inversion lemma, and we perform a swap-based local search that repeatedly applies objective-decreasing swaps until no further improvement is possible. Experiments on MNIST handwritten-digit images demonstrate the effectiveness of the proposed method.
[555] High-Resolution Climate Projections Using Diffusion-Based Downscaling of a Lightweight Climate Emulator
Haiwen Guan, Moein Darman, Dibyajyoti Chakraborty, Troy Arcomano, Ashesh Chattopadhyay, Romit Maulik
Main category: cs.LG
TL;DR: A deep learning downscaling framework using diffusion models to enhance climate emulator LUCIE’s resolution from ~300km to 25km for regional climate impact assessment.
Details
Motivation: Current data-driven climate models like LUCIE have good long-term statistics but inadequate native resolution (~300km) for detailed regional impact assessments, requiring downscaling to finer scales.Method: Probabilistic diffusion-based generative models with conditional and posterior sampling frameworks to downscale coarse LUCIE outputs from ~300km to 25km resolution, trained on ERA5 data (2000-2009).
Result: The approach successfully preserves coarse-grained dynamics from LUCIE while generating fine-scaled climatological statistics at ~28km resolution, validated on 2010-2020 predictions using metrics like RMSE, power spectrum, and probability density functions.
Conclusion: The diffusion-based downscaling framework effectively bridges the resolution gap for climate emulators, enabling detailed regional climate impact assessments while maintaining physical consistency.
Abstract: The proliferation of data-driven models in weather and climate sciences has marked a significant paradigm shift, with advanced models demonstrating exceptional skill in medium-range forecasting. However, these models are often limited by long-term instabilities, climatological drift, and substantial computational costs during training and inference, restricting their broader application for climate studies. Addressing these limitations, Guan et al. (2024) introduced LUCIE, a lightweight, physically consistent climate emulator utilizing a Spherical Fourier Neural Operator (SFNO) architecture. This model is able to reproduce accurate long-term statistics including climatological mean and seasonal variability. However, LUCIE’s native resolution (~300 km) is inadequate for detailed regional impact assessments. To overcome this limitation, we introduce a deep learning-based downscaling framework, leveraging probabilistic diffusion-based generative models with conditional and posterior sampling frameworks. These models downscale coarse LUCIE outputs to 25 km resolution. They are trained on approximately 14,000 ERA5 timesteps spanning 2000-2009 and evaluated on LUCIE predictions from 2010 to 2020. Model performance is assessed through diverse metrics, including latitude-averaged RMSE, power spectrum, probability density functions and First Empirical Orthogonal Function of the zonal wind. We observe that the proposed approach is able to preserve the coarse-grained dynamics from LUCIE while generating fine-scaled climatological statistics at ~28km resolution.
[556] Text Has Curvature
Karish Grover, Hanqing Zeng, Yinglong Xia, Christos Faloutsos, Geoffrey J. Gordon
Main category: cs.LG
TL;DR: Texture: A text-native discrete curvature signal that measures semantic inference curvature in language without geometric embeddings, enabling curvature-guided compression and routing.
Details
Motivation: While language is increasingly modeled in curved geometries (hyperbolic spaces, mixed-curvature manifolds), it's unclear whether text itself has intrinsic curvature or if curvature is just an artifact of embedding choices. The paper aims to determine if text has native curvature and make it measurable and useful.Method: Proposes Texture - a word-level discrete curvature signal defined by reconciling left- and right-context beliefs around masked words using a Schrodinger bridge. This yields a curvature field where positive values indicate focused meaning and negative values indicate competing continuations.
Result: Provides empirical and theoretical evidence that semantic inference in natural corpora is non-flat (language has inherent curvature). Demonstrates Texture’s utility on two tasks: improving long-context inference through curvature-guided compression and enhancing retrieval-augmented generation through curvature-guided routing.
Conclusion: Text does have intrinsic curvature, and Texture provides a text-native way to measure and utilize it without geometric training, establishing a new curvature paradigm for language that is both measurable and practically useful.
Abstract: Does text have an intrinsic curvature? Language is increasingly modeled in curved geometries - hyperbolic spaces for hierarchy, mixed-curvature manifolds for compositional structure - yet a basic scientific question remains unresolved: what does curvature mean for text itself, in a way that is native to language rather than an artifact of the embedding space we choose? We argue that text does indeed have curvature, and show how to detect it, define it, and use it. To this end, we propose Texture, a text-native, word-level discrete curvature signal, and make three contributions. (a) Existence: We provide empirical and theoretical certificates that semantic inference in natural corpora is non-flat, i.e. language has inherent curvature. (b) Definition: We define Texture by reconciling left- and right-context beliefs around a masked word through a Schrodinger bridge, yielding a curvature field that is positive where context focuses meaning and negative where it fans out into competing continuations. (c) Utility: Texture is actionable: it serves as a general-purpose measurement and control primitive enabling geometry without geometric training; we instantiate it on two representative tasks, improving long-context inference through curvature-guided compression and retrieval-augmented generation through curvature-guided routing. Together, our results establish a text-native curvature paradigm, making curvature measurable and practically useful.
[557] Comparing Classifiers: A Case Study Using PyCM
Sadra Sabouri, Alireza Zolanvari, Sepand Haghighi
Main category: cs.LG
TL;DR: PyCM library tutorial for comprehensive multi-class classifier evaluation using multiple metrics to reveal subtle performance differences that standard metrics might miss.
Details
Motivation: Standard evaluation metrics for classification models often fail to capture subtle performance differences and trade-offs, leading to potentially suboptimal model selection. There's a need for more comprehensive, multi-dimensional evaluation frameworks.Method: Tutorial on using the PyCM library for deep-dive evaluation of multi-class classifiers, demonstrated through two case scenarios showing how different evaluation metrics can lead to different interpretations of model efficacy.
Result: The PyCM library enables more comprehensive model evaluation, revealing that choice of evaluation metrics fundamentally shifts interpretation of model performance, and that multi-dimensional evaluation is essential for uncovering important subtle differences.
Conclusion: A multi-dimensional evaluation framework using tools like PyCM is crucial for optimal model selection, as standard metrics may miss subtle but important performance trade-offs in multi-class classifiers.
Abstract: Selecting an optimal classification model requires a robust and comprehensive understanding of the performance of the model. This paper provides a tutorial on the PyCM library, demonstrating its utility in conducting deep-dive evaluations of multi-class classifiers. By examining two different case scenarios, we illustrate how the choice of evaluation metrics can fundamentally shift the interpretation of a model’s efficacy. Our findings emphasize that a multi-dimensional evaluation framework is essential for uncovering small but important differences in model performance. However, standard metrics may miss these subtle performance trade-offs.
[558] Finding Highly Interpretable Prompt-Specific Circuits in Language Models
Gabriel Franco, Lucas M. Tassis, Azalea Rohr, Mark Crovella
Main category: cs.LG
TL;DR: ACC++ improves causal circuit analysis in LLMs by extracting cleaner attention signals, revealing prompt-specific circuits rather than single task-level mechanisms, enabling automated interpretability pipelines.
Details
Motivation: Prior circuit analysis assumes single stable mechanisms per task, but this obscures prompt-specific variations. The paper aims to reveal prompt-specific circuits and develop scalable analysis methods.Method: Builds on attention causal communication (ACC) with refinements (ACC++) to extract cleaner, lower-dimensional causal signals from attention heads in single forward passes, without needing replacement models or activation patching.
Result: Found no single circuit for indirect object identification (IOI) in GPT-2, Pythia, or Gemma 2 - different prompt templates induce systematically different mechanisms. Prompts cluster into families with similar circuits.
Conclusion: Circuits should be studied at prompt level rather than task level. ACC++ enables scalable circuit descriptions and automated interpretability pipelines for understanding prompt-specific mechanisms.
Abstract: Understanding the internal circuits that language models use to solve tasks remains a central challenge in mechanistic interpretability. Most prior work identifies circuits at the task level by averaging across many prompts, implicitly assuming a single stable mechanism per task. We show that this assumption can obscure a crucial source of structure: circuits are prompt-specific, even within a fixed task. Building on attention causal communication (ACC) (Franco & Crovella, 2025), we introduce ACC++, refinements that extract cleaner, lower-dimensional causal signals inside attention heads from a single forward pass. Like ACC, our approach does not require replacement models (e.g., SAEs) or activation patching; ACC++ further improves circuit precision by reducing attribution noise. Applying ACC++ to indirect object identification (IOI) in GPT-2, Pythia, and Gemma 2, we find there is no single circuit for IOI in any model: different prompt templates induce systematically different mechanisms. Despite this variation, prompts cluster into prompt families with similar circuits, and we propose a representative circuit for each family as a practical unit of analysis. Finally, we develop an automated interpretability pipeline that uses ACC++ signals to surface human-interpretable features and assemble mechanistic explanations for prompt families behavior. Together, our results recast circuits as a meaningful object of study by shifting the unit of analysis from tasks to prompts, enabling scalable circuit descriptions in the presence of prompt-specific mechanisms.
[559] Federated Learning of Nonlinear Temporal Dynamics with Graph Attention-based Cross-Client Interpretability
Ayse Tursucular, Ayush Mohanty, Nazal Mohamed, Nagi Gebraeel
Main category: cs.LG
TL;DR: Federated framework for learning interpretable temporal interdependencies across decentralized industrial subsystems using nonlinear state space models and Graph Attention Networks.
Details
Motivation: Modern industrial systems have distributed sensors generating high-dimensional time series data from interdependent subsystems, but raw data cannot be shared due to privacy/security concerns, and existing models cannot be modified in practical deployments.Method: Each client uses a nonlinear state space model to map local observations to low-dimensional latent states. A central server learns a graph-structured neural state transition model over communicated latent states using Graph Attention Networks (GAT). Jacobian analysis relates to attention coefficients for interpretability.
Result: Theoretical convergence guarantees to a centralized oracle. Synthetic experiments demonstrate convergence, interpretability, scalability, and privacy. Real-world experiments show performance comparable to decentralized baselines.
Conclusion: Proposed federated framework successfully learns interpretable cross-client temporal interdependencies in decentralized nonlinear systems while respecting privacy constraints and fixed proprietary models.
Abstract: Networks of modern industrial systems are increasingly monitored by distributed sensors, where each system comprises multiple subsystems generating high dimensional time series data. These subsystems are often interdependent, making it important to understand how temporal patterns at one subsystem relate to others. This is challenging in decentralized settings where raw measurements cannot be shared and client observations are heterogeneous. In practical deployments each subsystem (client) operates a fixed proprietary model that cannot be modified or retrained, limiting existing approaches. Nonlinear dynamics further make cross client temporal interdependencies difficult to interpret because they are embedded in nonlinear state transition functions. We present a federated framework for learning temporal interdependencies across clients under these constraints. Each client maps high dimensional local observations to low dimensional latent states using a nonlinear state space model. A central server learns a graph structured neural state transition model over the communicated latent states using a Graph Attention Network. For interpretability we relate the Jacobian of the learned server side transition model to attention coefficients, providing the first interpretable characterization of cross client temporal interdependencies in decentralized nonlinear systems. We establish theoretical convergence guarantees to a centralized oracle and validate the framework through synthetic experiments demonstrating convergence, interpretability, scalability and privacy. Additional real world experiments show performance comparable to decentralized baselines.
[560] Preventing Rank Collapse in Federated Low-Rank Adaptation with Client Heterogeneity
Fei Wu, Jia Hu, Geyong Min, Shiqiang Wang
Main category: cs.LG
TL;DR: raFLoRA addresses rank collapse in heterogeneous federated LoRA by proposing rank-partitioned aggregation to properly weight contributions from clients with different LoRA ranks.
Details
Motivation: In practical federated learning, client heterogeneity in system resources and data distributions motivates using different LoRA ranks across clients. However, heterogeneous FedLoRA suffers from "rank collapse" where global update energy concentrates on the minimum shared rank, leading to suboptimal performance and high sensitivity to rank configurations.Method: Proposes raFLoRA, a rank-partitioned aggregation method that decomposes local updates into rank partitions and aggregates each partition weighted by its effective client contributions, addressing the mismatch between rank-agnostic aggregation weights and rank-dependent client contributions.
Result: Extensive experiments across classification and reasoning tasks show that raFLoRA prevents rank collapse, improves model performance, and preserves communication efficiency compared to state-of-the-art FedLoRA baselines.
Conclusion: raFLoRA effectively addresses the rank collapse problem in heterogeneous federated LoRA through proper rank-partitioned aggregation, enabling better performance while maintaining communication efficiency in federated learning settings.
Abstract: Federated low-rank adaptation (FedLoRA) has facilitated communication-efficient and privacy-preserving fine-tuning of foundation models for downstream tasks. In practical federated learning scenarios, client heterogeneity in system resources and data distributions motivates heterogeneous LoRA ranks across clients. We identify a previously overlooked phenomenon in heterogeneous FedLoRA, termed rank collapse, where the energy of the global update concentrates on the minimum shared rank, resulting in suboptimal performance and high sensitivity to rank configurations. Through theoretical analysis, we reveal the root cause of rank collapse: a mismatch between rank-agnostic aggregation weights and rank-dependent client contributions, which systematically suppresses higher-rank updates at a geometric rate over rounds. Motivated by this insight, we propose raFLoRA, a rank-partitioned aggregation method that decomposes local updates into rank partitions and then aggregates each partition weighted by its effective client contributions. Extensive experiments across classification and reasoning tasks show that raFLoRA prevents rank collapse, improves model performance, and preserves communication efficiency compared to state-of-the-art FedLoRA baselines.
[561] TrasMuon: Trust-Region Adaptive Scaling for Orthogonalized Momentum Optimizers
Peng Cheng, Jiucheng Zang, Qingnan Li, Liheng Ma, Yufei Cui, Yingxue Zhang, Boxing Chen, Ming Jian, Wen Tong
Main category: cs.LG
TL;DR: TrasMuon improves Muon-style optimizers by adding global RMS calibration and energy-based trust-region clipping to stabilize magnitudes while preserving near-isometric geometry, achieving faster convergence and better stability than baselines.
Details
Motivation: Muon-style optimizers use Newton-Schulz iterations for orthogonalization but discard magnitude information, making training sensitive to step-size hyperparameters and vulnerable to high-energy bursts. The authors aim to preserve Muon's geometry while stabilizing magnitudes.Method: TrasMuon introduces two key components: (1) global RMS calibration to reintroduce adaptive scaling, and (2) energy-based trust-region clipping that defines a trust region based on relative energy ratios to confine updates to a stable zone and prevent instability from high-energy outliers.
Result: Empirical experiments on vision and language models show that TrasMuon converges faster than baseline methods. Additional experiments without warmup stages confirm TrasMuon’s superior stability and robustness compared to other optimizers.
Conclusion: TrasMuon successfully addresses the instability issues of Muon-style optimizers while maintaining their geometric advantages, resulting in a more stable and efficient optimization method suitable for training vision and language models.
Abstract: Muon-style optimizers leverage Newton-Schulz (NS) iterations to orthogonalize updates, yielding update geometries that often outperform Adam-series methods. However, this orthogonalization discards magnitude information, rendering training sensitive to step-size hyperparameters and vulnerable to high-energy bursts. To mitigate this, we introduce TrasMuon (\textbf{T}rust \textbf{R}egion \textbf{A}daptive \textbf{S}caling \textbf{Muon}). TrasMuon preserves the near-isometric geometry of Muon while stabilizing magnitudes through (i) global RMS calibration and (ii) energy-based trust-region clipping. We demonstrate that while reintroducing adaptive scaling improves optimization efficiency, it typically exacerbates instability due to high-energy outliers. TrasMuon addresses this by defining a trust region based on relative energy ratios, confining updates to a stable zone. Empirical experiments on vision and language models demonstrate that TrasMuon converges faster than baselines. Furthermore, experiments without warmup stages confirm TrasMuon’s superior stability and robustness.
[562] Coverage Guarantees for Pseudo-Calibrated Conformal Prediction under Distribution Shift
Farbod Siahkali, Ashwin Verma, Vijay Gupta
Main category: cs.LG
TL;DR: Pseudo-calibration method for conformal prediction under distribution shift, using domain adaptation tools to maintain coverage guarantees when data distributions change.
Details
Motivation: Conformal prediction provides statistical guarantees under exchangeability but fails when data distributions shift. Need methods to maintain coverage guarantees under distribution shift.Method: Uses pseudo-calibration with domain adaptation tools, derives lower bound on target coverage using classifier loss and Wasserstein shift measure, designs pseudo-calibrated sets with slack parameter, and proposes source-tuned pseudo-calibration algorithm that interpolates between hard pseudo-labels and randomized labels based on classifier uncertainty.
Result: Theoretical bounds track pseudo-calibration behavior qualitatively, and source-tuned scheme mitigates coverage degradation under distribution shift while maintaining nontrivial prediction set sizes in numerical experiments.
Conclusion: Pseudo-calibration with domain adaptation tools can effectively maintain conformal prediction coverage guarantees under bounded label-conditional covariate shift.
Abstract: Conformal prediction (CP) offers distribution-free marginal coverage guarantees under an exchangeability assumption, but these guarantees can fail if the data distribution shifts. We analyze the use of pseudo-calibration as a tool to counter this performance loss under a bounded label-conditional covariate shift model. Using tools from domain adaptation, we derive a lower bound on target coverage in terms of the source-domain loss of the classifier and a Wasserstein measure of the shift. Using this result, we provide a method to design pseudo-calibrated sets that inflate the conformal threshold by a slack parameter to keep target coverage above a prescribed level. Finally, we propose a source-tuned pseudo-calibration algorithm that interpolates between hard pseudo-labels and randomized labels as a function of classifier uncertainty. Numerical experiments show that our bounds qualitatively track pseudo-calibration behavior and that the source-tuned scheme mitigates coverage degradation under distribution shift while maintaining nontrivial prediction set sizes.
[563] $γ$-weakly $θ$-up-concavity: Linearizable Non-Convex Optimization with Applications to DR-Submodular and OSS Functions
Mohammad Pedramfar, Vaneet Aggarwal
Main category: cs.LG
TL;DR: A theoretical framework introducing γ-weakly θ-up-concavity, a first-order condition that generalizes DR-submodular and One-Sided Smooth functions, enabling unified approximation guarantees for monotone non-convex optimization via linear surrogate construction.
Details
Motivation: Optimizing monotone non-convex functions is fundamental in ML and combinatorial optimization. Existing frameworks like DR-submodular and One-Sided Smooth functions have limitations, and there's a need for a more general unifying framework that can handle broader classes of non-convex functions while providing approximation guarantees.Method: Introduces γ-weakly θ-up-concavity, a novel first-order condition that characterizes a broad class of monotone non-convex functions. The key theoretical contribution shows these functions are upper-linearizable: for any feasible point, a linear surrogate can be constructed whose gains approximate the original objective up to a constant factor (approximation coefficient) dependent on γ, θ, and feasible set geometry.
Result: Provides unified approximation guarantees for offline optimization, static and dynamic regret bounds in online settings via reductions to linear optimization. Recovers optimal approximation coefficient for DR-submodular maximization and significantly improves existing approximation coefficients for OSS optimization, particularly over matroid constraints.
Conclusion: The γ-weakly θ-up-concavity framework offers a powerful unifying approach for monotone non-convex optimization, generalizing existing frameworks while providing strong theoretical guarantees and improved approximation coefficients across various constraint settings.
Abstract: Optimizing monotone non-convex functions is a fundamental challenge across machine learning and combinatorial optimization. We introduce and study $γ$-weakly $θ$-up-concavity, a novel first-order condition that characterizes a broad class of such functions. This condition provides a powerful unifying framework, strictly generalizing both DR-submodular functions and One-Sided Smooth (OSS) functions. Our central theoretical contribution demonstrates that $γ$-weakly $θ$-up-concave functions are upper-linearizable: for any feasible point, we can construct a linear surrogate whose gains provably approximate the original non-linear objective. This approximation holds up to a constant factor, namely the approximation coefficient, dependent solely on $γ$, $θ$, and the geometry of the feasible set. This linearizability yields immediate and unified approximation guarantees for a wide range of problems. Specifically, we obtain unified approximation guarantees for offline optimization as well as static and dynamic regret bounds in online settings via standard reductions to linear optimization. Moreover, our framework recovers the optimal approximation coefficient for DR-submodular maximization and significantly improves existing approximation coefficients for OSS optimization, particularly over matroid constraints.
[564] Singular Vectors of Attention Heads Align with Features
Gabriel Franco, Carson Loughridge, Mark Crovella
Main category: cs.LG
TL;DR: The paper investigates whether singular vectors of attention matrices can reliably identify feature representations in language models, providing theoretical justification and empirical evidence for this alignment.
Details
Motivation: Recent studies have assumed that feature representations in language models can be inferred from singular vectors of attention matrices, but this assumption lacks sound theoretical justification. The paper aims to address this gap by examining why and when such alignment occurs.Method: 1) Demonstrate alignment in a model where features are directly observable; 2) Provide theoretical analysis showing when alignment is expected; 3) Propose sparse attention decomposition as a testable prediction for recognizing alignment in real models where features are not directly observable; 4) Show empirical evidence of this pattern in real models.
Result: The paper shows that singular vectors robustly align with features in observable models, provides theoretical conditions for such alignment, and demonstrates that sparse attention decomposition emerges in real models consistent with predictions, suggesting alignment can be a sound basis for feature identification.
Conclusion: Alignment of singular vectors with features can be a theoretically justified approach for feature identification in language models, with sparse attention decomposition serving as an operational test for recognizing such alignment in practice.
Abstract: Identifying feature representations in language models is a central task in mechanistic interpretability. Several recent studies have made an implicit assumption that feature representations can be inferred in some cases from singular vectors of attention matrices. However, sound justification for this assumption is lacking. In this paper we address that question, asking: why and when do singular vectors align with features? First, we demonstrate that singular vectors robustly align with features in a model where features can be directly observed. We then show theoretically that such alignment is expected under a range of conditions. We close by asking how, operationally, alignment may be recognized in real models where feature representations are not directly observable. We identify sparse attention decomposition as a testable prediction of alignment, and show evidence that it emerges in a manner consistent with predictions in real models. Together these results suggest that alignment of singular vectors with features can be a sound and theoretically justified basis for feature identification in language models.
[565] Vision Transformers for Multi-Variable Climate Downscaling: Emulating Regional Climate Models with a Shared Encoder and Multi-Decoder Architecture
Fabio Merizzi, Harilaos Loukos
Main category: cs.LG
TL;DR: A multi-variable Vision Transformer architecture with shared encoder and variable-specific decoders improves climate downscaling accuracy and efficiency for six key climate variables.
Details
Motivation: Global Climate Models have coarse resolution limiting regional studies, while Regional Climate Models are computationally expensive. Existing deep learning approaches focus on single-variable downscaling, leading to redundant computation and weak cross-variable interactions.Method: Proposed a multi-variable Vision Transformer (ViT) architecture with 1EMD design: one shared encoder and multiple variable-specific decoders to jointly predict six climate variables from GCM-resolution inputs.
Result: The 1EMD ViT outperforms single-variable ViT models with ~5.5% average MSE reduction, beats other multi-variable baselines, and reduces inference time per variable by 29-32% compared to single-variable approaches.
Conclusion: Multi-variable modeling provides systematic advantages for climate downscaling in accuracy and efficiency, with the 1EMD ViT achieving the best trade-off between predictive performance and computational cost.
Abstract: Global Climate Models (GCMs) are critical for simulating large-scale climate dynamics, but their coarse spatial resolution limits their applicability in regional studies. Regional Climate Models (RCMs) address this limitation through dynamical downscaling, albeit at considerable computational cost and with limited flexibility. Deep learning has emerged as an efficient data-driven alternative; however, most existing approaches focus on single-variable models that downscale one variable at a time. This paradigm can lead to redundant computation, limited contextual awareness, and weak cross-variable interactions.To address these limitations, we propose a multi-variable Vision Transformer (ViT) architecture with a shared encoder and variable-specific decoders (1EMD). The proposed model jointly predicts six key climate variables: surface temperature, wind speed, 500 hPa geopotential height, total precipitation, surface downwelling shortwave radiation, and surface downwelling longwave radiation, directly from GCM-resolution inputs, emulating RCM-scale downscaling over Europe. Compared to single-variable ViT models, the 1EMD architecture improves performance across all six variables, achieving an average MSE reduction of approximately 5.5% under a fair and controlled comparison. It also consistently outperforms alternative multi-variable baselines, including a single-decoder ViT and a multi-variable U-Net. Moreover, multi-variable models substantially reduce computational cost, yielding a 29-32% lower inference time per variable compared to single-variable approaches. Overall, our results demonstrate that multi-variable modeling provides systematic advantages for high-resolution climate downscaling in terms of both accuracy and efficiency. Among the evaluated architectures, the proposed 1EMD ViT achieves the most favorable trade-off between predictive performance and computational cost.
[566] Generative Adversarial Post-Training Mitigates Reward Hacking in Live Human-AI Music Interaction
Yusong Wu, Stephen Brade, Aleksandra Teng Ma, Tia-Jane Fowler, Enning Yang, Berker Banar, Aaron Courville, Natasha Jaques, Cheng-Zhi Anna Huang
Main category: cs.LG
TL;DR: Adversarial training method to prevent reward hacking in RL post-training for melody-to-chord accompaniment in live jamming scenarios
Details
Motivation: Live jamming requires real-time coordination, adaptation, and diversity, but RL post-training often reduces output diversity through reward hacking, which is especially harmful for musical creativityMethod: Proposes adversarial training on policy-generated trajectories with a co-evolving discriminator that separates policy trajectories from data distribution, while policy maximizes discriminator output plus coherence rewards
Result: Improved output diversity, harmonic coherence, adaptation speed, and user agency in both simulation (fixed test melodies and learned melody agents) and user study with expert musicians
Conclusion: Demonstrates effective method to mitigate reward hacking in RL post-training of generative sequence models for interactive musical applications
Abstract: Most applications of generative AI involve a sequential interaction in which a person inputs a prompt and waits for a response, and where reaction time and adaptivity are not important factors. In contrast, live jamming is a collaborative interaction that requires real-time coordination and adaptation without access to the other player’s future moves, while preserving diversity to sustain a creative flow. Reinforcement learning post-training enables effective adaptation through on-policy interaction, yet it often reduces output diversity by exploiting coherence-based rewards. This collapse, known as ``reward hacking’’, affects many RL post-training pipelines, but is especially harmful in live jamming, where musical creativity relies on dynamic variation and mutual responsiveness. In this paper, we propose a novel adversarial training method on policy-generated trajectories to mitigate reward hacking in RL post-training for melody-to-chord accompaniment. A co-evolving discriminator separates policy trajectories from the data distribution, while the policy maximizes the discriminator output in addition to coherence rewards to prevent collapse to trivial outputs. We evaluate accompaniment quality and output diversity in simulation with both fixed test melodies and learned melody agents, and we conduct a user study with the model deployed in a real-time interactive system with expert musicians. Quantitative evaluation and user feedback demonstrate improved output diversity, harmonic coherence, adaptation speed and user agency. Our results demonstrate a simple yet effective method to mitigate reward hacking in RL post-training of generative sequence models.
[567] QuaRK: A Quantum Reservoir Kernel for Time Series Learning
Abdallah Aaraba, Soumaya Cherkaoui, Ola Ahmad, Shengrui Wang
Main category: cs.LG
TL;DR: QuaRK: Quantum reservoir computing framework with hardware-realistic quantum reservoir featurizer and kernel-based readout, offering learning guarantees for time series tasks.
Details
Motivation: Quantum reservoir computing shows promise for time series learning but lacks efficient, implementable architectures with learning guarantees. Need to bridge gap between theoretical potential and practical implementation.Method: Combines hardware-realistic quantum reservoir featurizer with kernel-based readout. Uses classical shadow tomography for efficient measurement of k-local observables, then applies classical kernel methods with explicit regularization.
Result: Provides learning-theoretic generalization guarantees for dependent temporal data, linking design choices to finite-sample performance. Empirical validation shows predicted interpolation and generalization behaviors on synthetic beta-mixing time series.
Conclusion: QuaRK offers principled guidance for building reliable temporal learners with clear computational knobs (circuit width/depth, measurement budget) while preserving kernel method flexibility for nonlinear temporal functionals.
Abstract: Quantum reservoir computing offers a promising route for time series learning by modelling sequential data via rich quantum dynamics while the only training required happens at the level of a lightweight classical readout. However, studies featuring efficient and implementable quantum reservoir architectures along with model learning guarantees remain scarce in the literature. To close this gap, we introduce QuaRK, an end-to-end framework that couples a hardware-realistic quantum reservoir featurizer with a kernel-based readout scheme. Given a sequence of sample points, the reservoir injects the points one after the other to yield a compact feature vector from efficiently measured k-local observables using classical shadow tomography, after which a classical kernel-based readout learns the target mapping with explicit regularization and fast optimization. The resulting pipeline exposes clear computational knobs – circuit width and depth as well as the measurement budget – while preserving the flexibility of kernel methods to model nonlinear temporal functionals and being scalable to high-dimensional data. We further provide learning-theoretic generalization guarantees for dependent temporal data, linking design and resource choices to finite-sample performance, thereby offering principled guidance for building reliable temporal learners. Empirical experiments validate QuaRK and illustrate the predicted interpolation and generalization behaviours on synthetic beta-mixing time series tasks.
[568] Out-of-Support Generalisation via Weight Space Sequence Modelling
Roussel Desmond Nzoyem
Main category: cs.LG
TL;DR: WeightCaster reformulates out-of-support generalization as sequence modeling in weight space, partitioning training data into concentric shells to make uncertainty-aware predictions without explicit inductive biases.
Details
Motivation: Neural networks often fail catastrophically on out-of-support samples with overconfident but unrealistic predictions, which is problematic for safety-critical applications requiring reliable extrapolation beyond training data.Method: Reformulates OoS generalization as sequence modeling in weight space, partitions training set into concentric shells corresponding to discrete sequential steps, and uses WeightCaster framework to generate plausible, interpretable, and uncertainty-aware predictions.
Result: Demonstrates competitive or superior performance to state-of-the-art on synthetic cosine dataset and real-world air quality sensor readings, showing enhanced reliability beyond in-distribution scenarios.
Conclusion: WeightCaster enables reliable out-of-support generalization without explicit inductive biases, with significant implications for AI adoption in safety-critical applications requiring extrapolation beyond training data.
Abstract: As breakthroughs in deep learning transform key industries, models are increasingly required to extrapolate on datapoints found outside the range of the training set, a challenge we coin as out-of-support (OoS) generalisation. However, neural networks frequently exhibit catastrophic failure on OoS samples, yielding unrealistic but overconfident predictions. We address this challenge by reformulating the OoS generalisation problem as a sequence modelling task in the weight space, wherein the training set is partitioned into concentric shells corresponding to discrete sequential steps. Our WeightCaster framework yields plausible, interpretable, and uncertainty-aware predictions without necessitating explicit inductive biases, all the while maintaining high computational efficiency. Emprical validation on a synthetic cosine dataset and real-world air quality sensor readings demonstrates performance competitive or superior to the state-of-the-art. By enhancing reliability beyond in-distribution scenarios, these results hold significant implications for the wider adoption of artificial intelligence in safety-critical applications.
[569] Scenario-Adaptive MU-MIMO OFDM Semantic Communication With Asymmetric Neural Network
Chongyang Li, Tianqian Zhang, Shouyin Liu
Main category: cs.LG
TL;DR: Proposed scenario-adaptive MU-MIMO semantic communication framework for 6G downlink with asymmetric architecture, CSI/SNR-aware semantic encoder, neural precoding, and pilot-guided attention decoder.
Details
Motivation: Semantic communication shows promise for 6G but faces challenges in realistic MU-MIMO OFDM systems due to multi-user interference and frequency-selective fading. Existing DJSCC schemes designed for point-to-point links suffer performance saturation in multi-user scenarios.Method: Asymmetric architecture for downlink transmission: transmitter has scenario-aware semantic encoder that dynamically adjusts feature extraction based on CSI and SNR, plus neural precoding network to mitigate MUI in semantic domain; receiver uses lightweight decoder with novel pilot-guided attention mechanism for implicit channel equalization and feature calibration.
Result: Extensive simulations over 3GPP channel models show significant outperformance over DJSCC and traditional SSCC schemes in PSNR and classification accuracy, especially in low-SNR regimes, while maintaining low latency and computational cost on edge devices.
Conclusion: The proposed framework effectively addresses challenges of semantic communication in realistic MU-MIMO systems, demonstrating superior performance and practical feasibility for 6G networks.
Abstract: Semantic Communication (SemCom) has emerged as a promising paradigm for 6G networks, aiming to extract and transmit task-relevant information rather than minimizing bit errors. However, applying SemCom to realistic downlink Multi-User Multi-Input Multi-Output (MU-MIMO) Orthogonal Frequency Division Multiplexing (OFDM) systems remains challenging due to severe Multi-User Interference (MUI) and frequency-selective fading. Existing Deep Joint Source-Channel Coding (DJSCC) schemes, primarily designed for point-to-point links, suffer from performance saturation in multi-user scenarios. To address these issues, we propose a scenario-adaptive MU-MIMO SemCom framework featuring an asymmetric architecture tailored for downlink transmission. At the transmitter, we introduce a scenario-aware semantic encoder that dynamically adjusts feature extraction based on Channel State Information (CSI) and Signal-to-Noise Ratio (SNR), followed by a neural precoding network designed to mitigate MUI in the semantic domain. At the receiver, a lightweight decoder equipped with a novel pilot-guided attention mechanism is employed to implicitly perform channel equalization and feature calibration using reference pilot symbols. Extensive simulation results over 3GPP channel models demonstrate that the proposed framework significantly outperforms DJSCC and traditional Separate Source-Channel Coding (SSCC) schemes in terms of Peak Signal-to-Noise Ratio (PSNR) and classification accuracy, particularly in low-SNR regimes, while maintaining low latency and computational cost on edge devices.
[570] Interpretable clustering via optimal multiway-split decision trees
Hayato Suzuki, Shunnosuke Ikeda, Yuichi Takano
Main category: cs.LG
TL;DR: Proposes interpretable clustering using optimal multiway-split decision trees formulated as 0-1 integer linear optimization, with 1D K-means for discretization, achieving better accuracy and interpretability than binary trees.
Details
Motivation: Existing clustering methods using binary decision trees suffer from computational complexity, suboptimal solutions, and deep structures that reduce interpretability. Need for more tractable optimization and better interpretability in clustering.Method: Formulates clustering as optimal multiway-split decision trees using 0-1 integer linear optimization (more tractable than mixed-integer nonlinear problems). Integrates one-dimensional K-means for discretization of continuous variables to enable flexible, data-driven branching.
Result: Outperforms baseline methods in clustering accuracy and interpretability on real-world datasets. Produces multiway-split decision trees with concise decision rules while maintaining competitive performance across various metrics.
Conclusion: The proposed method provides an effective approach for interpretable clustering with improved computational tractability and better interpretability through multiway-split decision trees.
Abstract: Clustering serves as a vital tool for uncovering latent data structures, and achieving both high accuracy and interpretability is essential. To this end, existing methods typically construct binary decision trees by solving mixed-integer nonlinear optimization problems, often leading to significant computational costs and suboptimal solutions. Furthermore, binary decision trees frequently result in excessively deep structures, which makes them difficult to interpret. To mitigate these issues, we propose an interpretable clustering method based on optimal multiway-split decision trees, formulated as a 0-1 integer linear optimization problem. This reformulation renders the optimization problem more tractable compared to existing models. A key feature of our method is the integration of a one-dimensional K-means algorithm for the discretization of continuous variables, allowing for flexible and data-driven branching. Extensive numerical experiments on publicly available real-world datasets demonstrate that our method outperforms baseline methods in terms of clustering accuracy and interpretability. Our method yields multiway-split decision trees with concise decision rules while maintaining competitive performance across various evaluation metrics.
[571] Benchmark Leakage Trap: Can We Trust LLM-based Recommendation?
Mingqiao Zhang, Qiyao Peng, Yumeng Wang, Chunyuan Liu, Hongtao Liu
Main category: cs.LG
TL;DR: LLM-based recommendation systems suffer from benchmark data leakage where models memorize evaluation datasets during training, leading to inflated performance metrics that don’t reflect true capability.
Details
Motivation: The paper addresses a critical but overlooked problem in evaluating LLM-based recommender systems: benchmark data leakage where LLMs memorize evaluation datasets during pre-training or fine-tuning, resulting in misleading performance metrics that don't reflect true model capability.Method: Researchers simulate diverse data leakage scenarios by conducting continued pre-training of foundation models on strategically blended corpora containing user-item interactions from both in-domain and out-of-domain sources to study the effects of data contamination.
Result: Experiments reveal a dual-effect: domain-relevant data leakage causes substantial but spurious performance gains that misleadingly exaggerate model capability, while domain-irrelevant leakage typically degrades recommendation accuracy, showing the complex nature of this contamination.
Conclusion: Data leakage is a critical, previously unaccounted-for factor in LLM-based recommendation that impacts true model performance evaluation, highlighting the need for more rigorous evaluation protocols in this domain.
Abstract: The expanding integration of Large Language Models (LLMs) into recommender systems poses critical challenges to evaluation reliability. This paper identifies and investigates a previously overlooked issue: benchmark data leakage in LLM-based recommendation. This phenomenon occurs when LLMs are exposed to and potentially memorize benchmark datasets during pre-training or fine-tuning, leading to artificially inflated performance metrics that fail to reflect true model performance. To validate this phenomenon, we simulate diverse data leakage scenarios by conducting continued pre-training of foundation models on strategically blended corpora, which include user-item interactions from both in-domain and out-of-domain sources. Our experiments reveal a dual-effect of data leakage: when the leaked data is domain-relevant, it induces substantial but spurious performance gains, misleadingly exaggerating the model’s capability. In contrast, domain-irrelevant leakage typically degrades recommendation accuracy, highlighting the complex and contingent nature of this contamination. Our findings reveal that data leakage acts as a critical, previously unaccounted-for factor in LLM-based recommendation, which could impact the true model performance. We release our code at https://github.com/yusba1/LLMRec-Data-Leakage.
[572] Optimization-Free Graph Embedding via Distributional Kernel for Community Detection
Shuaibin Song, Kai Ming Ting, Kaifeng Zhang, Tianrun Liang
Main category: cs.LG
TL;DR: A novel weighted distribution-aware kernel for graph embedding that addresses over-smoothing in NAS methods by incorporating node and degree distributions, requiring no optimization, and improving community detection performance.
Details
Motivation: Neighborhood Aggregation Strategy (NAS) methods like GNNs and WL suffer from over-smoothing - loss of node distinguishability with increased iterations. Existing methods overlook critical network characteristics: node distributions and node degree distributions, which contribute significantly to over-smoothing.Method: Proposes a weighted distribution-aware kernel that embeds nodes while considering their distributional characteristics. The method explicitly incorporates both node and degree distributions, requires no optimization, and mitigates over-smoothing effects to preserve node distinguishability across many iterations.
Result: The method achieves superior community detection performance via spectral clustering, outperforming existing graph embedding methods including deep learning approaches on standard benchmarks. It effectively preserves node distinguishability even after many embedding iterations.
Conclusion: Incorporating node and degree distribution characteristics is crucial for addressing over-smoothing in graph embedding methods. The proposed distribution-aware kernel provides an effective, optimization-free solution that enhances representation expressiveness and community detection performance.
Abstract: Neighborhood Aggregation Strategy (NAS) is a widely used approach in graph embedding, underpinning both Graph Neural Networks (GNNs) and Weisfeiler-Lehman (WL) methods. However, NAS-based methods are identified to be prone to over-smoothing-the loss of node distinguishability with increased iterations-thereby limiting their effectiveness. This paper identifies two characteristics in a network, i.e., the distributions of nodes and node degrees that are critical for expressive representation but have been overlooked in existing methods. We show that these overlooked characteristics contribute significantly to over-smoothing of NAS-methods. To address this, we propose a novel weighted distribution-aware kernel that embeds nodes while taking their distributional characteristics into consideration. Our method has three distinguishing features: (1) it is the first method to explicitly incorporate both distributional characteristics; (2) it requires no optimization; and (3) it effectively mitigates the adverse effects of over-smoothing, allowing WL to preserve node distinguishability and expressiveness even after many iterations of embedding. Experiments demonstrate that our method achieves superior community detection performance via spectral clustering, outperforming existing graph embedding methods, including deep learning methods, on standard benchmarks.
[573] Joint Time Series Chain: Detecting Unusual Evolving Trend across Time Series
Li Zhang, Nital Patel, Xiuqi Li, Jessica Lin
Main category: cs.LG
TL;DR: Proposes Joint Time Series Chain (JointTSC) to find evolving patterns across interrupted or related time series, addressing limitations of existing single-series chain methods.
Details
Motivation: Existing time series chain definitions only work within single time series, missing evolving patterns across interrupted time series or related time series. Need to handle gaps/interruptions and identify unusual evolving trends across multiple series.Method: Introduces Joint Time Series Chain definition that handles gaps/interruptions, proposes effective ranking criterion to identify best chains, and implements algorithm for finding chains across interrupted or related time series.
Result: Outperforms existing TSC methods in locating unusual evolving patterns through extensive empirical evaluations. Demonstrated utility with real-life manufacturing application from Intel.
Conclusion: JointTSC effectively addresses limitations of single-series chain methods by enabling discovery of evolving patterns across interrupted or related time series, with practical applications in manufacturing and complex systems.
Abstract: Time series chain (TSC) is a recently introduced concept that captures the evolving patterns in large scale time series. Informally, a time series chain is a temporally ordered set of subsequences, in which consecutive subsequences in the chain are similar to one another, but the last and the first subsequences maybe be dissimilar. Time series chain has the great potential to reveal latent unusual evolving trend in the time series, or identify precursor of important events in a complex system. Unfortunately, existing definitions of time series chains only consider finding chains in a single time series. As a result, they are likely to miss unexpected evolving patterns in interrupted time series, or across two related time series. To address this limitation, in this work, we introduce a new definition called \textit{Joint Time Series Chain}, which is specially designed for the task of finding unexpected evolving trend across interrupted time series or two related time series. Our definition focuses on mitigating the robustness issues caused by the gap or interruption in the time series. We further propose an effective ranking criterion to identify the best chain. We demonstrate that our proposed approach outperforms existing TSC work in locating unusual evolving patterns through extensive empirical evaluations. We further demonstrate the utility of our work with a real-life manufacturing application from Intel. Our source code is publicly available at the supporting page https://github.com/lizhang-ts/JointTSC .
[574] DeepFusion: Accelerating MoE Training via Federated Knowledge Distillation from Heterogeneous Edge Devices
Songyuan Li, Jia Hu, Ahmed M. Abdelmoniem, Geyong Min, Haojun Huang, Jiwei Huang
Main category: cs.LG
TL;DR: DeepFusion: A scalable federated MoE training framework using federated knowledge distillation to fuse heterogeneous on-device LLM knowledge into a global MoE model, addressing resource constraints and view-mismatch problems.
Details
Motivation: MoE-based LLMs require vast diverse training data, but traditional federated learning approaches are impractical for resource-constrained devices that cannot host local MoE models. There's a need for privacy-preserving MoE training that works with heterogeneous edge devices.Method: Proposes DeepFusion framework where devices independently configure/train on-device LLMs tailored to their needs, then uses federated knowledge distillation with a novel View-Aligned Attention (VAA) module that integrates multi-stage feature representations from global MoE to align predictive perspectives between heterogeneous models.
Result: Achieves performance close to centralized MoE training, reduces communication costs by up to 71%, and improves token perplexity by up to 5.28% compared to federated MoE baselines on industry-level models (Qwen-MoE, DeepSeek-MoE) with real-world datasets.
Conclusion: DeepFusion enables practical federated training of MoE models on resource-constrained devices through effective cross-architecture knowledge distillation with view alignment, making large-scale privacy-preserving MoE training feasible.
Abstract: Recent Mixture-of-Experts (MoE)-based large language models (LLMs) such as Qwen-MoE and DeepSeek-MoE are transforming generative AI in natural language processing. However, these models require vast and diverse training data. Federated learning (FL) addresses this challenge by leveraging private data from heterogeneous edge devices for privacy-preserving MoE training. Nonetheless, traditional FL approaches require devices to host local MoE models, which is impractical for resource-constrained devices due to large model sizes. To address this, we propose DeepFusion, the first scalable federated MoE training framework that enables the fusion of heterogeneous on-device LLM knowledge via federated knowledge distillation, yielding a knowledge-abundant global MoE model. Specifically, DeepFusion features each device to independently configure and train an on-device LLM tailored to its own needs and hardware limitations. Furthermore, we propose a novel View-Aligned Attention (VAA) module that integrates multi-stage feature representations from the global MoE model to construct a predictive perspective aligned with on-device LLMs, thereby enabling effective cross-architecture knowledge distillation. By explicitly aligning predictive perspectives, VAA resolves the view-mismatch problem in traditional federated knowledge distillation, which arises from heterogeneity in model architectures and prediction behaviors between on-device LLMs and the global MoE model. Experiments with industry-level MoE models (Qwen-MoE and DeepSeek-MoE) and real-world datasets (medical and finance) demonstrate that DeepFusion achieves performance close to centralized MoE training. Compared with key federated MoE baselines, DeepFusion reduces communication costs by up to 71% and improves token perplexity by up to 5.28%.
[575] Cumulative Utility Parity for Fair Federated Learning under Intermittent Client Participation
Stefan Behfar, Richard Mortier
Main category: cs.LG
TL;DR: Proposes cumulative utility parity for federated learning fairness, focusing on long-term benefits per participation opportunity rather than per-round performance, addressing systematic under-representation of intermittently available clients.
Details
Motivation: Existing FL fairness approaches assume comparable client participation opportunities, but in real-world FL systems, client participation is intermittent, heterogeneous, and correlated with data characteristics or resource constraints. This leads to systematic under-representation of intermittently available clients even when per-round performance appears fair.Method: Introduces cumulative utility parity principle and availability-normalized cumulative utility metric, which disentangles unavoidable physical constraints from avoidable algorithmic bias arising from scheduling and aggregation. The approach evaluates whether clients receive comparable long-term benefit per participation opportunity.
Result: Experiments on temporally skewed, non-IID federated benchmarks demonstrate that the approach substantially improves long-term representation parity while maintaining near-perfect performance.
Conclusion: The proposed cumulative utility parity framework addresses a critical gap in FL fairness by focusing on long-term benefits per participation opportunity, providing a more meaningful fairness metric for real-world FL systems with intermittent client participation.
Abstract: In real-world federated learning (FL) systems, client participation is intermittent, heterogeneous, and often correlated with data characteristics or resource constraints. Existing fairness approaches in FL primarily focus on equalizing loss or accuracy conditional on participation, implicitly assuming that clients have comparable opportunities to contribute over time. However, when participation itself is uneven, these objectives can lead to systematic under-representation of intermittently available clients, even if per-round performance appears fair. We propose cumulative utility parity, a fairness principle that evaluates whether clients receive comparable long-term benefit per participation opportunity, rather than per training round. To operationalize this notion, we introduce availability-normalized cumulative utility, which disentangles unavoidable physical constraints from avoidable algorithmic bias arising from scheduling and aggregation. Experiments on temporally skewed, non-IID federated benchmarks demonstrate that our approach substantially improves long-term representation parity, while maintaining near-perfect performance.
[576] Zero-Order Optimization for LLM Fine-Tuning via Learnable Direction Sampling
Valery Parfenov, Grigoriy Evseev, Andrey Veprikov, Nikolay Bushkov, Stanislav Moiseev, Aleksandr Beznosikov
Main category: cs.LG
TL;DR: Policy-driven zero-order fine-tuning framework that learns sampling distributions over perturbation directions to reduce variance and improve convergence for LLM fine-tuning with memory savings.
Details
Motivation: Fine-tuning large language models requires substantial memory for backpropagation and optimizer states, limiting deployment in resource-constrained settings. Zero-order methods offer memory savings but suffer from high variance and poor scaling with parameter dimensionality.Method: Proposes a policy-driven zero-order framework that treats the sampling distribution over perturbation directions as a learnable policy. Updates the policy to reduce variance of directional estimates, with theoretical analysis showing improved gradient quality and relaxed dependence on parameter dimensionality.
Result: Empirical validation on challenging LLM fine-tuning benchmarks shows substantially improved performance compared to standard zero-order baselines, demonstrating the effectiveness of adaptive direction sampling.
Conclusion: Adaptive direction sampling through learned policies is a promising approach to make zero-order fine-tuning viable at scale, offering memory-efficient alternatives to backpropagation-based methods.
Abstract: Fine-tuning large pretrained language models (LLMs) is a cornerstone of modern NLP, yet its growing memory demands (driven by backpropagation and large optimizer States) limit deployment in resource-constrained settings. Zero-order (ZO) methods bypass backpropagation by estimating directional derivatives from forward evaluations, offering substantial memory savings. However, classical ZO estimators suffer from high variance and an adverse dependence on the parameter dimensionality $d$, which has constrained their use to low-dimensional problems. In this work, we propose a policy-driven ZO framework that treats the sampling distribution over perturbation directions as a learnable policy and updates it to reduce the variance of directional estimates. We develop a practical algorithm implementing this idea and provide a theoretical analysis, showing that learned sampling distributions improve the quality of gradient information and relax the explicit dependence on $d$ in convergence bounds. Empirically, we validate the approach on challenging LLM fine-tuning benchmarks, demonstrating substantially improved performance compared to standard ZO baselines. Our results suggest that adaptive direction sampling is a promising route to make ZO fine-tuning viable at scale. The source code is available at https://github.com/brain-lab-research/zo_ldsd
[577] Optimized Certainty Equivalent Risk-Controlling Prediction Sets
Jiayi Huang, Amirmohammad Farzaneh, Osvaldo Simeone
Main category: cs.LG
TL;DR: OCE-RCPS introduces a novel risk-controlling prediction set framework using optimized certainty equivalent risk measures to provide high-probability guarantees for safety-critical applications like medical image segmentation.
Details
Motivation: Current risk-controlling prediction sets (RCPS) only provide guarantees on expected risk, failing to capture tail behavior and worst-case scenarios crucial for high-stakes applications like medical imaging where reliability is paramount.Method: Proposes OCE-RCPS framework that uses optimized certainty equivalent (OCE) risk measures including CVaR and entropic risk, leveraging upper confidence bounds to identify prediction set parameters that satisfy user-specified risk tolerance levels with provable reliability.
Result: Theoretical guarantees show OCE-RCPS satisfies probabilistic constraints for loss functions like miscoverage and false negative rate. Experiments on image segmentation demonstrate consistent satisfaction of target rates across various risk measures and reliability configurations.
Conclusion: OCE-RCPS provides superior reliability guarantees for safety-critical applications compared to conventional RCPS, offering probabilistic guarantees on tail risk measures crucial for high-stakes decision making.
Abstract: In safety-critical applications such as medical image segmentation, prediction systems must provide reliability guarantees that extend beyond conventional expected loss control. While risk-controlling prediction sets (RCPS) offer probabilistic guarantees on the expected risk, they fail to capture tail behavior and worst-case scenarios that are crucial in high-stakes settings. This paper introduces optimized certainty equivalent RCPS (OCE-RCPS), a novel framework that provides high-probability guarantees on general optimized certainty equivalent (OCE) risk measures, including conditional value-at-risk (CVaR) and entropic risk. OCE-RCPS leverages upper confidence bounds to identify prediction set parameters that satisfy user-specified risk tolerance levels with provable reliability. We establish theoretical guarantees showing that OCE-RCPS satisfies the desired probabilistic constraint for loss functions such as miscoverage and false negative rate. Experiments on image segmentation demonstrate that OCE-RCPS consistently meets target satisfaction rates across various risk measures and reliability configurations, while OCE-CRC fails to provide probabilistic guarantees.
[578] ALMo: Interactive Aim-Limit-Defined, Multi-Objective System for Personalized High-Dose-Rate Brachytherapy Treatment Planning and Visualization for Cervical Cancer
Edward Chen, Natalie Dullerud, Pang Wei Koh, Thomas Niedermayr, Elizabeth Kidd, Sanmi Koyejo, Carlos Guestrin
Main category: cs.LG
TL;DR: ALMo is an interactive decision support system for clinical planning that uses aim-limit thresholds to help clinicians navigate multi-objective tradeoffs in treatment planning, demonstrated in cervical cancer brachytherapy.
Details
Motivation: Clinical decision-making requires managing multiple competing metrics with aim and limit thresholds, which is cognitively demanding and prone to variability. The paper addresses this challenge in High-Dose-Rate brachytherapy for cervical cancer, where planning requires balancing tumor coverage against organ sparing while managing radiation hot spots.Method: ALMo employs a novel optimization framework that minimizes manual input through automated parameter setup and enables flexible control over toxicity risks. The system allows clinicians to navigate the Pareto surface of dosimetric tradeoffs by directly manipulating intuitive aim and limit values.
Result: In a retrospective evaluation of 25 clinical cases, ALMo generated treatment plans that consistently met or exceeded manual planning quality, with 65% of cases demonstrating dosimetric improvements. The system significantly enhanced efficiency, reducing average planning time to approximately 17 minutes compared to conventional 30-60 minutes.
Conclusion: ALMo demonstrates a generalized framework for streamlining interaction in multi-criteria clinical decision-making, validated in brachytherapy but applicable to broader clinical contexts.
Abstract: In complex clinical decision-making, clinicians must often track a variety of competing metrics defined by aim (ideal) and limit (strict) thresholds. Sifting through these high-dimensional tradeoffs to infer the optimal patient-specific strategy is cognitively demanding and historically prone to variability. In this paper, we address this challenge within the context of High-Dose-Rate (HDR) brachytherapy for cervical cancer, where planning requires strictly managing radiation hot spots while balancing tumor coverage against organ sparing. We present ALMo (Aim-Limit-defined Multi-Objective system), an interactive decision support system designed to infer and operationalize clinician intent. ALMo employs a novel optimization framework that minimizes manual input through automated parameter setup and enables flexible control over toxicity risks. Crucially, the system allows clinicians to navigate the Pareto surface of dosimetric tradeoffs by directly manipulating intuitive aim and limit values. In a retrospective evaluation of 25 clinical cases, ALMo generated treatment plans that consistently met or exceeded manual planning quality, with 65% of cases demonstrating dosimetric improvements. Furthermore, the system significantly enhanced efficiency, reducing average planning time to approximately 17 minutes, compared to the conventional 30-60 minutes. While validated in brachytherapy, ALMo demonstrates a generalized framework for streamlining interaction in multi-criteria clinical decision-making.
[579] Advancing Analytic Class-Incremental Learning through Vision-Language Calibration
Binyu Zhao, Wei Zhang, Xingrui Yu, Zhaonian Zou, Ivor Tsang
Main category: cs.LG
TL;DR: VILA is a dual-branch framework for class-incremental learning that uses vision-language calibration to balance adaptation and stability in pre-trained models.
Details
Motivation: Class-incremental learning with pre-trained models faces a trade-off between efficient adaptation and long-term stability. Analytic learning enables rapid updates but suffers from accumulated errors and feature incompatibility due to representation rigidity.Method: Proposes VILA, a dual-branch framework with two-level vision-language calibration: 1) feature-level geometric calibration fusing plastic task-adapted features with frozen universal semantic anchors, and 2) decision-level cross-modal priors to rectify prediction bias.
Result: Extensive experiments across eight benchmarks show VILA yields superior performance, particularly in fine-grained and long-sequence scenarios, while maintaining analytic learning’s efficiency.
Conclusion: VILA harmonizes high-fidelity prediction with the simplicity of analytic learning, overcoming the brittleness of traditional analytic CIL approaches while preserving efficiency.
Abstract: Class-incremental learning (CIL) with pre-trained models (PTMs) faces a critical trade-off between efficient adaptation and long-term stability. While analytic learning enables rapid, recursive closed-form updates, its efficacy is often compromised by accumulated errors and feature incompatibility. In this paper, we first conduct a systematic study to dissect the failure modes of PTM-based analytic CIL, identifying representation rigidity as the primary bottleneck. Motivated by these insights, we propose \textbf{VILA}, a novel dual-branch framework that advances analytic CIL via a two-level vision-language calibration strategy. Specifically, we coherently fuse plastic, task-adapted features with a frozen, universal semantic anchor at the feature level through geometric calibration, and leverage cross-modal priors at the decision level to rectify prediction bias. This confluence maintains analytic-learning’s extreme efficiency while overcoming its inherent brittleness. Extensive experiments across eight benchmarks demonstrate that VILA consistently yields superior performance, particularly in fine-grained and long-sequence scenarios. Our framework harmonizes high-fidelity prediction with the simplicity of analytic learning. Our code is available at https://github.com/byzhaoAI/VILA
[580] Fluid-Agent Reinforcement Learning
Shishir Sharma, Doina Precup, Theodore J. Perkins
Main category: cs.LG
TL;DR: A framework for multi-agent reinforcement learning where agents can dynamically create other agents, enabling fluid population sizes that adapt to environmental demands.
Details
Motivation: Real-world multi-agent systems often have dynamic populations where agents can create other agents (like cell division or company spin-offs), but existing MARL research focuses on fixed numbers of agents.Method: Proposes fluid-agent environment framework with game-theoretic solution concepts, evaluates MARL algorithms on fluid variants of established benchmarks (Predator-Prey, Level-Based Foraging) and introduces new environment.
Result: Demonstrates that fluid-agent framework yields agent teams that dynamically adjust their size to match environmental demands, unlocking novel solution strategies beyond fixed-population settings.
Conclusion: Fluid-agent environments represent an important extension of MARL to handle real-world scenarios with dynamic populations, enabling more adaptive and scalable multi-agent systems.
Abstract: The primary focus of multi-agent reinforcement learning (MARL) has been to study interactions among a fixed number of agents embedded in an environment. However, in the real world, the number of agents is neither fixed nor known a priori. Moreover, an agent can decide to create other agents (for example, a cell may divide, or a company may spin off a division). In this paper, we propose a framework that allows agents to create other agents; we call this a fluid-agent environment. We present game-theoretic solution concepts for fluid-agent games and empirically evaluate the performance of several MARL algorithms within this framework. Our experiments include fluid variants of established benchmarks such as Predator-Prey and Level-Based Foraging, where agents can dynamically spawn, as well as a new environment we introduce that highlights how fluidity can unlock novel solution strategies beyond those observed in fixed-population settings. We demonstrate that this framework yields agent teams that adjust their size dynamically to match environmental demands.
[581] On the Sparsifiability of Correlation Clustering: Approximation Guarantees under Edge Sampling
Ibne Farabi Shihab, Sanjeda Akter, Anuj Sharma
Main category: cs.LG
TL;DR: The paper studies sparsification-approximation trade-offs for Correlation Clustering, establishing structural dichotomies between pseudometric and general weighted instances, with implications for LP-based approximations and edge information requirements.
Details
Motivation: Correlation Clustering requires prohibitive Θ(n³) triangle inequality constraints for strong LP-based approximations at scale, motivating the study of how much edge information is needed to retain LP-based guarantees through sparsification-approximation trade-offs.Method: Establishes structural dichotomy between pseudometric and general weighted instances; proves VC dimension of clustering disagreement class; analyzes active triangle inequalities at LP vertices; develops sparsified LP-PIVOT algorithm with imputation via triangle inequalities; uses Yao’s minimax principle for negative results.
Result: VC dimension is exactly n-1 yielding optimal additive ε-coresets; at most binom(n,2) triangle inequalities active at LP vertices enabling exact cutting-plane solver; sparsified LP-PIVOT achieves 10/3-approximation with Õ(n^{3/2}) edges; negative result shows unbounded approximation ratio with o(n) edges without pseudometric structure.
Conclusion: Pseudometric structure governs both tractability and robustness of Correlation Clustering to incomplete information, with sharp thresholds for edge observation requirements and fundamental limitations for general weighted instances.
Abstract: Correlation Clustering (CC) is a fundamental unsupervised learning primitive whose strongest LP-based approximation guarantees require $Θ(n^3)$ triangle inequality constraints and are prohibitive at scale. We initiate the study of \emph{sparsification–approximation trade-offs} for CC, asking how much edge information is needed to retain LP-based guarantees. We establish a structural dichotomy between pseudometric and general weighted instances. On the positive side, we prove that the VC dimension of the clustering disagreement class is exactly $n{-}1$, yielding additive $\varepsilon$-coresets of optimal size $\tilde{O}(n/\varepsilon^2)$; that at most $\binom{n}{2}$ triangle inequalities are active at any LP vertex, enabling an exact cutting-plane solver; and that a sparsified variant of LP-PIVOT, which imputes missing LP marginals via triangle inequalities, achieves a robust $\frac{10}{3}$-approximation (up to an additive term controlled by an empirically computable imputation-quality statistic $\overlineΓ_w$) once $\tildeΘ(n^{3/2})$ edges are observed, a threshold we prove is sharp. On the negative side, we show via Yao’s minimax principle that without pseudometric structure, any algorithm observing $o(n)$ uniformly random edges incurs an unbounded approximation ratio, demonstrating that the pseudometric condition governs not only tractability but also the robustness of CC to incomplete information.
[582] Physics Aware Neural Networks: Denoising for Magnetic Navigation
Aritra Das, Yashas Shende, Muskaan Chugh, Reva Laxmi Chauhan, Arghya Pathak, Debayan Gupta
Main category: cs.LG
TL;DR: A physics-constrained deep learning framework for magnetic anomaly navigation that uses divergence-free and E(3)-equivariance constraints to handle aircraft-induced magnetic noise, with Contiformer architecture and synthetic data generation via time-series GANs.
Details
Motivation: Magnetic anomaly navigation is a GPS alternative but faces challenges from aircraft-induced magnetic noise; classical models inadequately handle stochastic noise in magnetic data required for navigation.Method: Proposes physics-based constraints: divergence-free vector field (via neural network outputting vector potential A with magnetic field as its curl) and E(3)-equivariance (using tensor products of geometric tensors with spherical harmonics). Uses Contiformer architecture for continuous-time dynamics and long-term memory, plus synthetic data generation with time-series conditional GANs based on World Magnetic Model.
Result: Physics constraints significantly improve predictive accuracy and physical plausibility; Contiformer outperforms state-of-the-art methods; ablation studies show benefits of both constraints; synthetic data generation addresses data scarcity.
Conclusion: Embedding physical constraints acts as implicit regularizer improving spatio-temporal performance; continuous-time dynamics and long-term memory are critical for magnetic time series modeling; physics-constrained deep learning outperforms classical and unconstrained approaches.
Abstract: Magnetic-anomaly navigation, leveraging small-scale variations in the Earth’s magnetic field, is a promising alternative when GPS is unavailable or compromised. Airborne systems face a key challenge in extracting geomagnetic field data: the aircraft itself induces magnetic noise. Although the classical Tolles-Lawson model addresses this, it inadequately handles stochastically corrupted magnetic data required for navigation. To address stochastic noise, we propose a framework based on two physics-based constraints: divergence-free vector field and E(3)-equivariance. These ensure the learned magnetic field obeys Maxwell’s equations and that outputs transform correctly with sensor position/orientation. The divergence-free constraint is implemented by training a neural network to output a vector potential $A$, with the magnetic field defined as its curl. For E(3)-equivariance, we use tensor products of geometric tensors representable via spherical harmonics with known rotational transformations. Enforcing physical consistency and restricting the admissible function space acts as an implicit regularizer that improves spatio-temporal performance. We present ablation studies evaluating each constraint alone and jointly across CNNs, MLPs, Liquid Time Constant models, and Contiformers. Continuous-time dynamics and long-term memory are critical for modelling magnetic time series; the Contiformer architecture, which provides both, outperforms state-of-the-art methods. To mitigate data scarcity, we generate synthetic datasets using the World Magnetic Model (WMM) with time-series conditional GANs, producing realistic, temporally consistent magnetic sequences across varied trajectories and environments. Experiments show that embedding these constraints significantly improves predictive accuracy and physical plausibility, outperforming classical and unconstrained deep learning approaches.
[583] Atomix: Timely, Transactional Tool Use for Reliable Agentic Workflows
Bardia Mohammadi, Nearchos Potamitis, Lars Klein, Akhil Arora, Laurent Bindschaedler
Main category: cs.LG
TL;DR: Atomix provides transactional semantics for LLM agent tool calls with epoch-based tracking and progress-aware commit to prevent unintended side effects from failed or speculative operations.
Details
Motivation: LLM agents acting on external systems face issues where tool effects are immediate and irreversible, causing problems when operations fail, involve speculation, or face contention - leading to unintended side effects with no safe rollback mechanism.Method: Atomix introduces a runtime with progress-aware transactional semantics: tags each tool call with an epoch, tracks per-resource frontiers, commits only when progress predicates indicate safety, buffers effects when possible, and provides compensation mechanisms for externalized effects on abort.
Result: Across real workloads with fault injection, transactional retry improves task success rates, while frontier-gated commit strengthens isolation under speculation and contention scenarios.
Conclusion: Atomix addresses critical safety issues in LLM agent interactions with external systems by providing transactional guarantees that prevent unintended side effects and enable safe rollback of operations.
Abstract: LLM agents increasingly act on external systems, yet tool effects are immediate. Under failures, speculation, or contention, losing branches can leak unintended side effects with no safe rollback. We introduce Atomix, a runtime that provides progress-aware transactional semantics for agent tool calls. Atomix tags each call with an epoch, tracks per-resource frontiers, and commits only when progress predicates indicate safety; bufferable effects can be delayed, while externalized effects are tracked and compensated on abort. Across real workloads with fault injection, transactional retry improves task success, while frontier-gated commit strengthens isolation under speculation and contention.
[584] Attention Head Entropy of LLMs Predicts Answer Correctness
Sophie Ostmeier, Brian Axelrod, Maya Varma, Asad Aali, Yabin Zhang, Magdalini Paschali, Sanmi Koyejo, Curtis Langlotz, Akshay Chaudhari
Main category: cs.LG
TL;DR: Head Entropy method uses attention entropy patterns to predict answer correctness in LLMs, showing strong in-distribution performance and better out-of-domain generalization than baselines.
Details
Motivation: LLMs often generate plausible but incorrect answers, which is risky in safety-critical domains like medicine. Existing evaluation methods are expensive (human) or error-prone (LLM-as-judge). White-box methods using model internals exist but it's unclear if they can predict answer correctness and generalize out-of-domain.Method: Head Entropy measures the spread of attention mass using per-head 2-Renyi entropies. It uses sparse logistic regression on these attention entropy patterns to predict answer correctness. The method analyzes attention patterns both during answer generation and even before answer generation (over question/context alone).
Result: Head Entropy matches or exceeds baselines in-distribution and generalizes substantially better out-of-domain, outperforming the closest baseline by +8.5% AUROC on average. Attention patterns over question/context alone already carry predictive signal, with +17.7% AUROC improvement over the closest baseline. Evaluated across 5 instruction-tuned LLMs and 3 QA datasets spanning general knowledge, multi-hop reasoning, and medicine.
Conclusion: Attention entropy patterns (Head Entropy) provide a robust method for predicting answer correctness in LLMs, with strong generalization capabilities across domains and the ability to detect potential errors even before answer generation.
Abstract: Large language models (LLMs) often generate plausible yet incorrect answers, posing risks in safety-critical settings such as medicine. Human evaluation is expensive, and LLM-as-judge approaches risk introducing hidden errors. Recent white-box methods detect contextual hallucinations using model internals, focusing on the localization of the attention mass, but two questions remain open: do these approaches extend to predicting answer correctness, and do they generalize out-of-domains? We introduce Head Entropy, a method that predicts answer correctness from attention entropy patterns, specifically measuring the spread of the attention mass. Using sparse logistic regression on per-head 2-Renyi entropies, Head Entropy matches or exceeds baselines in-distribution and generalizes substantially better on out-of-domains, it outperforms the closest baseline on average by +8.5% AUROC. We further show that attention patterns over the question/context alone, before answer generation, already carry predictive signal using Head Entropy with on average +17.7% AUROC over the closest baseline. We evaluate across 5 instruction-tuned LLMs and 3 QA datasets spanning general knowledge, multi-hop reasoning, and medicine.
[585] Picking the Right Specialist: Attentive Neural Process-based Selection of Task-Specialized Models as Tools for Agentic Healthcare Systems
Pramit Saha, Joshua Strong, Mohammad Alsharid, Divyanshu Mishra, J. Alison Noble
Main category: cs.LG
TL;DR: ToolSelect is a method for adaptive model selection in agentic healthcare systems that learns to choose the best specialist model from a heterogeneous pool for each clinical query, demonstrated on chest X-ray tasks.
Details
Motivation: In healthcare agent systems, multiple competing specialist models exist for each task (disease diagnosis, localization, report generation), with different models excelling on different data samples. There's a need for reliable selection of the right specialist model from a heterogeneous pool for each query.Method: ToolSelect adaptively learns model selection by minimizing population risk over sampled specialist tool candidates using a consistent surrogate of task-conditional selection loss. Uses an Attentive Neural Process-based selector conditioned on the query and per-model behavioral summaries.
Result: Introduced ToolSelectBench benchmark with 1448 queries in agentic Chest X-ray environment with diverse specialist models (17 disease detection, 19 report generation, 6 visual grounding, 13 VQA). ToolSelect consistently outperforms 10 state-of-the-art methods across four different task families.
Conclusion: ToolSelect provides an effective solution for adaptive model selection in agentic healthcare systems, enabling reliable selection of specialist models from heterogeneous pools for clinical queries.
Abstract: Task-specialized models form the backbone of agentic healthcare systems, enabling the agents to answer clinical queries across tasks such as disease diagnosis, localization, and report generation. Yet, for a given task, a single “best” model rarely exists. In practice, each task is better served by multiple competing specialist models where different models excel on different data samples. As a result, for any given query, agents must reliably select the right specialist model from a heterogeneous pool of tool candidates. To this end, we introduce ToolSelect, which adaptively learns model selection for tools by minimizing a population risk over sampled specialist tool candidates using a consistent surrogate of the task-conditional selection loss. Concretely, we propose an Attentive Neural Process-based selector conditioned on the query and per-model behavioral summaries to choose among the specialist models. Motivated by the absence of any established testbed, we, for the first time, introduce an agentic Chest X-ray environment equipped with a diverse suite of task-specialized models (17 disease detection, 19 report generation, 6 visual grounding, and 13 VQA) and develop ToolSelectBench, a benchmark of 1448 queries. Our results demonstrate that ToolSelect consistently outperforms 10 SOTA methods across four different task families.
[586] Optimal Regret for Policy Optimization in Contextual Bandits
Orin Levy, Yishay Mansour
Main category: cs.LG
TL;DR: First high-probability optimal regret bound for policy optimization in stochastic contextual multi-armed bandits with general offline function approximation, achieving optimal regret of Õ(√(K|A|log|F|)).
Details
Motivation: Bridge the gap between theory and practice for widely used policy optimization methods in contextual bandit problems, providing rigorous theoretical guarantees for practical algorithms.Method: Policy optimization technique applied to stochastic contextual multi-armed bandit with general offline function approximation, using function class F to approximate losses.
Result: Achieves optimal regret bound of Õ(√(K|A|log|F|)) with high probability, where K is number of rounds, A is set of arms, and F is function class. Algorithm is both efficient and theoretically optimal.
Conclusion: Demonstrates that practical policy optimization methods for contextual bandits can achieve rigorously-proved optimal regret bounds, bridging theory and practice in reinforcement learning.
Abstract: We present the first high-probability optimal regret bound for a policy optimization technique applied to the problem of stochastic contextual multi-armed bandit (CMAB) with general offline function approximation. Our algorithm is both efficient and achieves an optimal regret bound of $\widetilde{O}(\sqrt{ K|\mathcal{A}|\log|\mathcal{F}|})$, where $K$ is the number of rounds, $\mathcal{A}$ is the set of arms, and $\mathcal{F}$ is the function class used to approximate the losses. Our results bridge the gap between theory and practice, demonstrating that the widely used policy optimization methods for the contextual bandit problem can achieve a rigorously-proved optimal regret bound. We support our theoretical results with an empirical evaluation of our algorithm.
[587] Near-Optimal Regret for Policy Optimization in Contextual MDPs with General Offline Function Approximation
Orin Levy, Aviv Rosenberg, Alon Cohen, Yishay Mansour
Main category: cs.LG
TL;DR: OPO-CMDP is a policy optimization algorithm for stochastic Contextual Markov Decision Processes with optimal regret bounds under offline function approximation.
Details
Motivation: The paper addresses the challenge of solving stochastic Contextual MDPs with general offline function approximation, aiming to achieve optimal regret bounds that improve upon existing state-of-the-art methods.Method: The authors introduce OPO-CMDP, an optimistic policy optimization algorithm that leverages finite function classes to approximate losses and dynamics, achieving regret bounds with optimal dependence on state and action space sizes.
Result: OPO-CMDP achieves a high probability regret bound of Õ(H⁴√(T|S||A|log(|F||P|))), which is the first regret bound with optimal dependence on |S| and |A|, directly improving upon Qian, Hu, and Simchi-Levi (2024).
Conclusion: Optimistic policy optimization provides a natural, computationally superior and theoretically near-optimal approach for solving CMDPs with general offline function approximation.
Abstract: We introduce \texttt{OPO-CMDP}, the first policy optimization algorithm for stochastic Contextual Markov Decision Process (CMDPs) under general offline function approximation. Our approach achieves a high probability regret bound of $\widetilde{O}(H^4\sqrt{T|S||A|\log(|\mathcal{F}||\mathcal{P}|)}),$ where $S$ and $A$ denote the state and action spaces, $H$ the horizon length, $T$ the number of episodes, and $\mathcal{F}, \mathcal{P}$ the finite function classes used to approximate the losses and dynamics, respectively. This is the first regret bound with optimal dependence on $|S|$ and $|A|$, directly improving the current state-of-the-art (Qian, Hu, and Simchi-Levi, 2024). These results demonstrate that optimistic policy optimization provides a natural, computationally superior and theoretically near-optimal path for solving CMDPs.
[588] HBVLA: Pushing 1-Bit Post-Training Quantization for Vision-Language-Action Models
Xin Yan, Zhenglin Wan, Feiyang Ye, Xingrui Yu, Hangyu Du, Yang You, Ivor Tsang
Main category: cs.LG
TL;DR: HBVLA: A VLA-tailored binarization framework using policy-aware Hessian analysis and sparse orthogonal transforms to enable efficient 1-bit quantization for vision-language-action models while maintaining performance.
Details
Motivation: Vision-Language-Action (VLA) models are computationally expensive for resource-constrained robots and edge platforms. While 1-bit binarization improves efficiency, existing methods suffer from distribution gaps that cause quantization errors to accumulate during long-horizon execution, severely degrading action quality.Method: 1) Use policy-aware enhanced Hessian to identify action-critical weights; 2) Apply sparse orthogonal transform to non-salient weights to create low-entropy intermediate state; 3) Quantize both salient and non-salient weights in Harr domain with group-wise 1-bit quantization.
Result: Quantized OpenVLA-OFT retains 92.2% performance on LIBERO, CogAct retains 93.6% on SimplerEnv, significantly outperforming SOTA binarization methods. Real-world evaluation shows only marginal success-rate degradation compared to full-precision models.
Conclusion: HBVLA provides practical ultra-low-bit quantization for VLAs, enabling reliable deployment on hardware-limited robotic platforms while maintaining performance close to full-precision models.
Abstract: Vision-Language-Action (VLA) models enable instruction-following embodied control, but their large compute and memory footprints hinder deployment on resource-constrained robots and edge platforms. While reducing weights to 1-bit precision through binarization can greatly improve efficiency, existing methods fail to narrow the distribution gap between binarized and full-precision weights, causing quantization errors to accumulate under long-horizon closed-loop execution and severely degrade actions. To fill this gap, we propose HBVLA, a VLA-tailored binarization framework. First, we use a policy-aware enhanced Hessian to identify weights that are truly critical for action generation. Then, we employ a sparse orthogonal transform for non-salient weights to induce a low-entropy intermediate state. Finally, we quantize both salient and non-salient weights in the Harr domain with group-wise 1-bit quantization. We have evaluated our approach on different VLAs: on LIBERO, quantized OpenVLA-OFT retains 92.2% of full-precision performance; on SimplerEnv, quantized CogAct retains 93.6%, significantly outperforming state-of-the-art binarization methods. We further validate our method on real-world evaluation suite and the results show that HBVLA incurs only marginal success-rate degradation compared to the full-precision model, demonstrating robust deployability under tight hardware constraints. Our work provides a practical foundation for ultra-low-bit quantization of VLAs, enabling more reliable deployment on hardware-limited robotic platforms.
[589] Data-driven Bi-level Optimization of Thermal Power Systems with embedded Artificial Neural Networks
Talha Ansar, Muhammad Mujtaba Abbas, Ramit Debnath, Vivek Dua, Waqar Muhammad Ashraf
Main category: cs.LG
TL;DR: A bi-level optimization framework using ANN models and KKT conditions for hierarchical optimization of industrial thermal power systems, achieving computational efficiency and energy-efficient operations.
Details
Motivation: Industrial thermal power systems have coupled performance variables with hierarchical importance, making simultaneous optimization computationally challenging. This limits scalable operation optimization for large-scale engineering systems.Method: Proposes a machine learning-powered bi-level optimization framework where objective functions are approximated by ANN models, and the lower-level problem is analytically embedded through KKT optimality conditions, creating an ANN-KKT single-level optimization framework.
Result: Validated on benchmark problems and real-world power plants (660 MW coal and 395 MW gas turbine), achieving comparable solutions to bi-level benchmarks with marginal computational time (0.22-0.88s). Generated 583 MW (coal) and 402 MW (gas turbine) outputs at optimal turbine heat rates, and delineated feasible operating envelopes accounting for uncertainty.
Conclusion: ANN-KKT offers a scalable, computationally efficient route for hierarchical, data-driven optimization of industrial thermal power systems, enabling energy-efficient operations of large-scale engineering systems and contributing to Industry 5.0.
Abstract: Industrial thermal power systems have coupled performance variables with hierarchical order of importance, making their simultaneous optimization computationally challenging or infeasible. This barrier limits the integrated and computationally scaleable operation optimization of industrial thermal power systems. To address this issue for large-scale engineering systems, we present a fully machine learning-powered bi-level optimization framework for data-driven optimization of industrial thermal power systems. The objective functions of upper and lower levels are approximated by artificial neural network (ANN) models and the lower-level problem is analytically embedded through Karush-Kuhn-Tucker (KKT) optimality conditions. The reformulated single level optimization framework integrating ANN models and KKT constraints (ANN-KKT) is validated on benchmark problems and on real-world power generation operation of 660 MW coal power plant and 395 MW gas turbine system. The results reveal a comparable solutions obtained from the proposed ANN-KKT framework to the bi-level solutions of the benchmark problems. Marginal computational time requirement (0.22 to 0.88 s) to compute optimal solutions yields 583 MW (coal) and 402 MW (gas turbine) of power output at optimal turbine heat rate of 7337 kJ/kWh and 7542 kJ/kWh, respectively. In addition, the method expands to delineate a feasible and robust operating envelope that accounts for uncertainty in operating variables while maximizing thermal efficiency in various scenarios. These results demonstrate that ANN-KKT offers a scalable and computationally efficient route for hierarchical, data-driven optimization of industrial thermal power systems, achieving energy-efficient operations of large-scale engineering systems and contributing to industry 5.0.
[590] Discrete Double-Bracket Flows for Isotropic-Noise Invariant Eigendecomposition
ZhiMing Li, JiaHe Feng
Main category: cs.LG
TL;DR: Matrix-free eigendecomposition algorithm with double-bracket flow that’s invariant to isotropic noise, achieving improved convergence rates for covariance operators with trace-free perturbations.
Details
Motivation: Standard stochastic approximation methods for eigendecomposition under matrix-vector product oracles have limitations: fixed steps couple stability to covariance norm, while adaptive steps slow down due to vanishing updates. The paper aims to develop a method that's invariant to isotropic noise shifts and achieves better convergence.Method: Introduces a discrete double-bracket flow whose generator is invariant to isotropic shifts, making it pathwise invariant to isotropic noise at the discrete-time level. The method uses a maximal stable step size proportional to 1/||C_e||_2^2, where C_e is the trace-free covariance. Analysis includes strict-saddle geometry for diagonalization objective and input-to-state stability analysis.
Result: Achieves global convergence with sample complexity scaling as O(||C_e||_2^2/(Δ^2ε)) under trace-free perturbations. Explicit characterization of degenerate blocks yields accelerated O(log(1/ζ)) saddle-escape rate and high-probability finite-time convergence guarantee.
Conclusion: The proposed discrete double-bracket flow provides a robust matrix-free eigendecomposition method that’s invariant to isotropic noise, with improved convergence rates and stability properties compared to standard stochastic approximation approaches.
Abstract: We study matrix-free eigendecomposition under a matrix-vector product (MVP) oracle, where each step observes a covariance operator $C_k = C_{sig} + σ_k^2 I + E_k$. Standard stochastic approximation methods either use fixed steps that couple stability to $|C_k|2$, or adapt steps in ways that slow down due to vanishing updates. We introduce a discrete double-bracket flow whose generator is invariant to isotropic shifts, yielding pathwise invariance to $σ_k^2 I$ at the discrete-time level. The resulting trajectory and a maximal stable step size $η{max} \propto 1/|C_e|_2^2$ depend only on the trace-free covariance $C_e$. We establish global convergence via strict-saddle geometry for the diagonalization objective and an input-to-state stability analysis, with sample complexity scaling as $O(|C_e|_2^2 / (Δ^2 ε))$ under trace-free perturbations. An explicit characterization of degenerate blocks yields an accelerated $O(\log(1/ζ))$ saddle-escape rate and a high-probability finite-time convergence guarantee.
[591] On Representation Redundancy in Large-Scale Instruction Tuning Data Selection
Youwei Shu, Shaomian Zheng, Dingnan Jin, Wenjie Qu, Ziyao Guo, Qing Cui, Jun Zhou, Jiaheng Zhang
Main category: cs.LG
TL;DR: CRDS improves instruction-tuning data selection by addressing redundancy in LLM embeddings through compressed representations, achieving strong performance with minimal data.
Details
Motivation: Data quality is crucial for LLM training, but systematic methods for industrial-scale data selection in instruction tuning are underexplored. Current LLM encoders produce highly redundant semantic embeddings that limit effective data selection.Method: Proposes Compressed Representation Data Selection (CRDS) with two variants: CRDS-R uses Rademacher random projection with concatenated transformer hidden-layer representations, and CRDS-W employs whitening-based dimensionality reduction to improve representational quality.
Result: Both CRDS variants substantially enhance data quality and outperform state-of-the-art representation-based selection methods. CRDS-W achieves strong performance using only 3.5% of data, surpassing full-data baseline by average 0.71% across four datasets.
Conclusion: CRDS effectively addresses redundancy in LLM embeddings for data selection, enabling efficient instruction tuning with minimal high-quality data while maintaining or improving performance.
Abstract: Data quality is a crucial factor in large language models training. While prior work has shown that models trained on smaller, high-quality datasets can outperform those trained on much larger but noisy or low-quality corpora, systematic methods for industrial-scale data selection in instruction tuning remain underexplored. In this work, we study instruction-tuning data selection through the lens of semantic representation similarity and identify a key limitation of state-of-the-art LLM encoders: they produce highly redundant semantic embeddings. To mitigate this redundancy, we propose Compressed Representation Data Selection (CRDS), a novel framework with two variants. CRDS-R applies Rademacher random projection followed by concatenation of transformer hidden-layer representations, while CRDS-W employs whitening-based dimensionality reduction to improve representational quality. Experimental results demonstrate that both variants substantially enhance data quality and consistently outperform state-of-the-art representation-based selection methods. Notably, CRDS-W achieves strong performance using only 3.5% of the data, surpassing the full-data baseline by an average of 0.71% across four datasets. Our code is available at https://github.com/tdano1/CRDS.
[592] MEMTS: Internalizing Domain Knowledge via Parameterized Memory for Retrieval-Free Domain Adaptation of Time Series Foundation Models
Xiaoyun Yu, Li fan, Xiangfei Qiu, Nanqing Dong, Yonggui Huang, Honggang Qi, Geguang Pu, Wanli Ouyang, Xi Chen, Jilin Hu
Main category: cs.LG
TL;DR: MEMTS is a lightweight plug-and-play method for retrieval-free domain adaptation in time series forecasting that uses learnable latent prototypes to internalize domain-specific temporal patterns without modifying the frozen foundation model backbone.
Details
Motivation: Time Series Foundation Models (TSFMs) degrade in real-world vertical domains due to temporal distribution shifts and domain-specific periodic structures. Current approaches like Domain-Adaptive Pretraining cause catastrophic forgetting of global patterns, while Retrieval-Augmented Generation introduces substantial retrieval overhead, creating scalability bottlenecks for real-time stream processing.Method: MEMTS introduces a Knowledge Persistence Module (KPM) that internalizes domain-specific temporal dynamics (seasonal patterns, trends) into a compact set of learnable latent prototypes. This transforms fragmented historical observations into continuous, parameterized knowledge representations, enabling retrieval-free domain adaptation with constant-time inference and near-zero latency.
Result: Extensive experiments on multiple datasets demonstrate state-of-the-art performance. MEMTS achieves accurate domain adaptation while effectively mitigating catastrophic forgetting of general temporal patterns, all without requiring architectural modifications to the frozen TSFM backbone.
Conclusion: MEMTS provides a lightweight, plug-and-play solution for retrieval-free domain adaptation in time series forecasting, addressing the scalability bottlenecks of current approaches while maintaining the learned global temporal patterns of foundation models.
Abstract: While Time Series Foundation Models (TSFMs) have demonstrated exceptional performance in generalized forecasting, their performance often degrades significantly when deployed in real-world vertical domains characterized by temporal distribution shifts and domain-specific periodic structures. Current solutions are primarily constrained by two paradigms: Domain-Adaptive Pretraining (DAPT), which improves short-term domain fitting but frequently disrupts previously learned global temporal patterns due to catastrophic forgetting; and Retrieval-Augmented Generation (RAG), which incorporates external knowledge but introduces substantial retrieval overhead. This creates a severe scalability bottleneck that fails to meet the high-efficiency requirements of real-time stream processing. To break this impasse, we propose Memory for Time Series (MEMTS), a lightweight and plug-and-play method for retrieval-free domain adaptation in time series forecasting. The key component of MEMTS is a Knowledge Persistence Module (KPM), which internalizes domain-specific temporal dynamics, such as recurring seasonal patterns and trends into a compact set of learnable latent prototypes. In doing so, it transforms fragmented historical observations into continuous, parameterized knowledge representations. This paradigm shift enables MEMTS to achieve accurate domain adaptation with constant-time inference and near-zero latency, while effectively mitigating catastrophic forgetting of general temporal patterns, all without requiring any architectural modifications to the frozen TSFM backbone. Extensive experiments on multiple datasets demonstrate the SOTA performance of MEMTS.
[593] MechPert: Mechanistic Consensus as an Inductive Bias for Unseen Perturbation Prediction
Marc Boubnovski Martell, Josefa Lia Stoisser, Lawrence Phillips, Aditya Misra, Robert Kitchen, Jesper Ferkinghoff-Borg, Jialin Yu, Philip Torr, Kaspar Märten
Main category: cs.LG
TL;DR: MechPert uses LLM agents to generate directed regulatory hypotheses for predicting transcriptional responses to genetic perturbations, outperforming similarity-based methods in low-data regimes.
Details
Motivation: Existing approaches for predicting transcriptional responses rely on incomplete knowledge graphs or language models that retrieve symmetric co-occurrence associations rather than directed regulatory logic, limiting their effectiveness.Method: MechPert is a lightweight framework where multiple LLM agents independently propose candidate regulators with confidence scores, which are aggregated through a consensus mechanism to filter spurious associations and produce weighted neighborhoods for downstream prediction.
Result: On Perturb-seq benchmarks across four human cell lines, MechPert improves Pearson correlation by up to 10.5% over similarity-based baselines in low-data regimes (N=50 observed perturbations). For experimental design, MechPert-selected anchor genes outperform standard network centrality heuristics by up to 46% in well-characterized cell lines.
Conclusion: MechPert demonstrates that LLM agents can effectively generate directed regulatory hypotheses for perturbation prediction, offering significant improvements over existing methods in low-data scenarios and experimental design applications.
Abstract: Predicting transcriptional responses to unseen genetic perturbations is essential for understanding gene regulation and prioritizing large-scale perturbation experiments. Existing approaches either rely on static, potentially incomplete knowledge graphs, or prompt language models for functionally similar genes, retrieving associations shaped by symmetric co-occurrence in scientific text rather than directed regulatory logic. We introduce MechPert, a lightweight framework that encourages LLM agents to generate directed regulatory hypotheses rather than relying solely on functional similarity. Multiple agents independently propose candidate regulators with associated confidence scores; these are aggregated through a consensus mechanism that filters spurious associations, producing weighted neighborhoods for downstream prediction. We evaluate MechPert on Perturb-seq benchmarks across four human cell lines. For perturbation prediction in low-data regimes ($N=50$ observed perturbations), MechPert improves Pearson correlation by up to 10.5% over similarity-based baselines. For experimental design, MechPert-selected anchor genes outperform standard network centrality heuristics by up to 46% in well-characterized cell lines.
[594] Cast-R1: Learning Tool-Augmented Sequential Decision Policies for Time Series Forecasting
Xiaoyu Tao, Mingyue Cheng, Chuang Jiang, Tian Gao, Huanjian Zhang, Yaguo Liu
Main category: cs.LG
TL;DR: Cast-R1 reformulates time series forecasting as a sequential decision-making problem using a memory-based agentic framework with tool-augmented workflow and iterative refinement.
Details
Motivation: Traditional model-centric time series forecasting approaches struggle in complex, evolving settings because they lack autonomous evidence acquisition, reasoning about future changes, and iterative refinement capabilities.Method: Proposes Cast-R1 framework with memory-based state management, tool-augmented agentic workflow (statistical feature extraction, lightweight forecasting models, reasoning-based prediction, self-reflection), and two-stage training combining supervised fine-tuning with multi-turn reinforcement learning and curriculum learning.
Result: Extensive experiments on multiple real-world time series datasets demonstrate the effectiveness of Cast-R1 framework for time series forecasting.
Conclusion: Cast-R1 provides a practical step toward exploring agentic paradigms for time series modeling, showing promise for complex forecasting scenarios requiring autonomous reasoning and iterative refinement.
Abstract: Time series forecasting has long been dominated by model-centric approaches that formulate prediction as a single-pass mapping from historical observations to future values. Despite recent progress, such formulations often struggle in complex and evolving settings, largely because most forecasting models lack the ability to autonomously acquire informative evidence, reason about potential future changes, or revise predictions through iterative decision processes. In this work, we propose Cast-R1, a learned time series forecasting framework that reformulates forecasting as a sequential decision-making problem. Cast-R1 introduces a memory-based state management mechanism that maintains decision-relevant information across interaction steps, enabling the accumulation of contextual evidence to support long-horizon reasoning. Building on this formulation, forecasting is carried out through a tool-augmented agentic workflow, in which the agent autonomously interacts with a modular toolkit to extract statistical features, invoke lightweight forecasting models for decision support, perform reasoning-based prediction, and iteratively refine forecasts through self-reflection. To train Cast-R1, we adopt a two-stage learning strategy that combines supervised fine-tuning with multi-turn reinforcement learning, together with a curriculum learning scheme that progressively increases task difficulty to improve policy learning. Extensive experiments on multiple real-world time series datasets demonstrate the effectiveness of Cast-R1. We hope this work provides a practical step towards further exploration of agentic paradigms for time series modeling. Our code is available at https://github.com/Xiaoyu-Tao/Cast-R1-TS.
[595] Fast Physics-Driven Untrained Network for Highly Nonlinear Inverse Scattering Problems
Yutong Du, Zicheng Liu, Yi Huang, Bazargul Matkerim, Bo Qi, Yali Zong, Peixian Han
Main category: cs.LG
TL;DR: A real-time physics-driven Fourier-spectral solver for electromagnetic inverse scattering that achieves 100x speedup over untrained neural networks through spectral-domain dimensionality reduction.
Details
Motivation: Untrained neural networks (UNNs) provide high-fidelity electromagnetic inverse scattering reconstruction but suffer from computational limitations due to high-dimensional spatial-domain optimization, preventing real-time applications.Method: Proposes a PDF solver using spectral-domain dimensionality reduction by expanding induced currents with a truncated Fourier basis, confining optimization to a compact low-frequency parameter space. Integrates contraction integral equation (CIE) to mitigate high-contrast nonlinearity and contrast-compensated operator (CCO) to correct spectral-induced attenuation. Also formulates a bridge-suppressing loss to enhance boundary sharpness between scatterers.
Result: Achieves 100-fold speedup over state-of-the-art UNNs with robust performance under noise and antenna uncertainties, enabling real-time microwave imaging applications.
Conclusion: The PDF solver enables real-time electromagnetic inverse scattering reconstruction through efficient spectral-domain optimization, overcoming computational limitations of previous UNN approaches.
Abstract: Untrained neural networks (UNNs) offer high-fidelity electromagnetic inverse scattering reconstruction but are computationally limited by high-dimensional spatial-domain optimization. We propose a Real-Time Physics-Driven Fourier-Spectral (PDF) solver that achieves sub-second reconstruction through spectral-domain dimensionality reduction. By expanding induced currents using a truncated Fourier basis, the optimization is confined to a compact low-frequency parameter space supported by scattering measurements. The solver integrates a contraction integral equation (CIE) to mitigate high-contrast nonlinearity and a contrast-compensated operator (CCO) to correct spectral-induced attenuation. Furthermore, a bridge-suppressing loss is formulated to enhance boundary sharpness between adjacent scatterers. Numerical and experimental results demonstrate a 100-fold speedup over state-of-the-art UNNs with robust performance under noise and antenna uncertainties, enabling real-time microwave imaging applications.
[596] AnomaMind: Agentic Time Series Anomaly Detection with Tool-Augmented Reasoning
Xiaoyu Tao, Yuchong Wu, Mingyue Cheng, Ze Guo, Tian Gao
Main category: cs.LG
TL;DR: AnomaMind is an agentic framework that reformulates time series anomaly detection as a sequential decision-making process with adaptive feature preparation, reasoning-aware detection, and iterative refinement through multi-turn tool interactions.
Details
Motivation: Existing anomaly detection methods treat the task as purely discriminative prediction with fixed features, struggling with context-dependent or diverse anomaly patterns. The authors argue this stems from lack of adaptive feature preparation, reasoning-aware detection, and iterative refinement.Method: AnomaMind uses a structured workflow that: 1) progressively localizes anomalies coarse-to-fine, 2) augments detection through multi-turn tool interactions for adaptive feature preparation, and 3) refines decisions via self-reflection. It employs a hybrid inference mechanism where general-purpose models handle tool interaction and refinement, while core anomaly detection is learned through reinforcement learning with workflow-level feedback.
Result: Extensive experiments across diverse settings demonstrate that AnomaMind consistently improves anomaly detection performance compared to existing methods.
Conclusion: AnomaMind successfully addresses limitations of traditional anomaly detection by framing it as an evidence-driven diagnostic process with adaptive reasoning, showing promise for complex real-world applications.
Abstract: Time series anomaly detection is critical in many real-world applications, where effective solutions must localize anomalous regions and support reliable decision-making under complex settings. However, most existing methods frame anomaly detection as a purely discriminative prediction task with fixed feature inputs, rather than an evidence-driven diagnostic process. As a result, they often struggle when anomalies exhibit strong context dependence or diverse patterns. We argue that these limitations stem from the lack of adaptive feature preparation, reasoning-aware detection, and iterative refinement during inference. To address these challenges, we propose AnomaMind, an agentic time series anomaly detection framework that reformulates anomaly detection as a sequential decision-making process. AnomaMind operates through a structured workflow that progressively localizes anomalous intervals in a coarse-to-fine manner, augments detection through multi-turn tool interactions for adaptive feature preparation, and refines anomaly decisions via self-reflection. The workflow is supported by a set of reusable tool engines, enabling context-aware diagnostic analysis. A key design of AnomaMind is an explicitly designed hybrid inference mechanism for tool-augmented anomaly detection. In this mechanism, general-purpose models are responsible for autonomous tool interaction and self-reflective refinement, while core anomaly detection decisions are learned through reinforcement learning under verifiable workflow-level feedback, enabling task-specific optimization within a flexible reasoning framework. Extensive experiments across diverse settings demonstrate that AnomaMind consistently improves anomaly detection performance. The code is available at https://anonymous.4open.science/r/AnomaMind.
[597] Mean Flow Policy with Instantaneous Velocity Constraint for One-step Action Generation
Guojian Zhan, Letian Tao, Pengcheng Wang, Yixiao Wang, Yiheng Li, Yuxin Chen, Masayoshi Tomizuka, Shengbo Eben Li
Main category: cs.LG
TL;DR: MVP is a new generative policy function for RL that models mean velocity fields for fast one-step action generation, achieving state-of-the-art performance on robotic manipulation tasks with improved speed.
Details
Motivation: Flow-based policies in RL face a trade-off between expressiveness and computational burden, typically controlled by the number of flow steps. The authors aim to develop a policy function that achieves both high expressiveness and fast deterministic sampling.Method: Proposes Mean Velocity Policy (MVP) that models the mean velocity field for fastest one-step action generation. Introduces Instantaneous Velocity Constraint (IVC) during training to ensure high expressiveness, which serves as a crucial boundary condition theoretically proven to improve learning accuracy and policy expressiveness.
Result: MVP achieves state-of-the-art success rates across challenging robotic manipulation tasks from Robomimic and OGBench. It also delivers substantial improvements in both training and inference speed over existing flow-based policy baselines.
Conclusion: MVP provides an effective solution to the expressiveness-efficiency trade-off in RL policy functions, offering both high performance and computational efficiency for robotic manipulation tasks.
Abstract: Learning expressive and efficient policy functions is a promising direction in reinforcement learning (RL). While flow-based policies have recently proven effective in modeling complex action distributions with a fast deterministic sampling process, they still face a trade-off between expressiveness and computational burden, which is typically controlled by the number of flow steps. In this work, we propose mean velocity policy (MVP), a new generative policy function that models the mean velocity field to achieve the fastest one-step action generation. To ensure its high expressiveness, an instantaneous velocity constraint (IVC) is introduced on the mean velocity field during training. We theoretically prove that this design explicitly serves as a crucial boundary condition, thereby improving learning accuracy and enhancing policy expressiveness. Empirically, our MVP achieves state-of-the-art success rates across several challenging robotic manipulation tasks from Robomimic and OGBench. It also delivers substantial improvements in training and inference speed over existing flow-based policy baselines.
[598] Pawsterior: Variational Flow Matching for Structured Simulation-Based Inference
Jorge Carrasco-Pollo, Floor Eijkelboom, Jan-Willem van de Meent
Main category: cs.LG
TL;DR: Pawsterior is a variational flow-matching framework for simulation-based inference that handles structured domains and discrete latent variables, improving posterior fidelity and enabling inference on previously inaccessible problems.
Details
Motivation: Standard flow-matching methods for simulation-based inference operate in unconstrained spaces, but many real-world problems involve structured domains (bounded parameters, hybrid discrete-continuous variables) which leads to inefficient learning and difficulty respecting physical constraints.Method: Introduces endpoint-induced affine geometric confinement that incorporates domain geometry directly into inference via a two-sided variational model, enabling handling of both geometric constraints and discrete latent structure through variational parameterization.
Result: Improved numerical stability during sampling and consistently better posterior fidelity, demonstrated by improved classifier two-sample test performance across standard SBI benchmarks. Enables SBI tasks involving discrete latent structure that were previously incompatible with conventional flow-matching.
Conclusion: Pawsterior extends flow-matching to a broader class of structured SBI problems by addressing both geometric constraints and discrete latent structure, making previously inaccessible problems tractable.
Abstract: We introduce Pawsterior, a variational flow-matching framework for improved and extended simulation-based inference (SBI). Many SBI problems involve posteriors constrained by structured domains, such as bounded physical parameters or hybrid discrete-continuous variables, yet standard flow-matching methods typically operate in unconstrained spaces. This mismatch leads to inefficient learning and difficulty respecting physical constraints. Our contributions are twofold. First, generalizing the geometric inductive bias of CatFlow, we formalize endpoint-induced affine geometric confinement, a principle that incorporates domain geometry directly into the inference process via a two-sided variational model. This formulation improves numerical stability during sampling and leads to consistently better posterior fidelity, as demonstrated by improved classifier two-sample test performance across standard SBI benchmarks. Second, and more importantly, our variational parameterization enables SBI tasks involving discrete latent structure (e.g., switching systems) that are fundamentally incompatible with conventional flow-matching approaches. By addressing both geometric constraints and discrete latent structure, Pawsterior extends flow-matching to a broader class of structured SBI problems that were previously inaccessible.
[599] Testing For Distribution Shifts with Conditional Conformal Test Martingales
Shalev Shaer, Yarin Bar, Drew Prinster, Yaniv Romano
Main category: cs.LG
TL;DR: A sequential test for detecting arbitrary distribution shifts using conformal test martingales that avoids test-time contamination by comparing new samples to a fixed null reference dataset instead of continually growing reference sets.
Details
Motivation: Existing conformal test martingale (CTM) detectors suffer from test-time contamination where post-shift observations enter the reference set and dilute evidence for distribution shift, increasing detection delay and reducing power.Method: Proposes a sequential test that avoids contamination by comparing each new sample to a fixed null reference dataset. Uses a robust martingale construction that remains valid conditional on the null reference data by explicitly accounting for estimation error in the reference distribution induced by the finite reference set.
Result: The method achieves anytime-valid type-I error control with guarantees of asymptotic power one and bounded expected detection delay. Empirically detects shifts faster than standard CTMs.
Conclusion: Provides a powerful and reliable distribution-shift detector that avoids the contamination problem of existing CTM methods while maintaining statistical guarantees.
Abstract: We propose a sequential test for detecting arbitrary distribution shifts that allows conformal test martingales (CTMs) to work under a fixed, reference-conditional setting. Existing CTM detectors construct test martingales by continually growing a reference set with each incoming sample, using it to assess how atypical the new sample is relative to past observations. While this design yields anytime-valid type-I error control, it suffers from test-time contamination: after a change, post-shift observations enter the reference set and dilute the evidence for distribution shift, increasing detection delay and reducing power. In contrast, our method avoids contamination by design by comparing each new sample to a fixed null reference dataset. Our main technical contribution is a robust martingale construction that remains valid conditional on the null reference data, achieved by explicitly accounting for the estimation error in the reference distribution induced by the finite reference set. This yields anytime-valid type-I error control together with guarantees of asymptotic power one and bounded expected detection delay. Empirically, our method detects shifts faster than standard CTMs, providing a powerful and reliable distribution-shift detector.
[600] sleep2vec: Unified Cross-Modal Alignment for Heterogeneous Nocturnal Biosignals
Weixuan Yuan, Zengrui Jin, Yichen Wang, Donglin Xie, Ziyi Ye, Chao Zhang, Xuesong Chen
Main category: cs.LG
TL;DR: sleep2vec is a foundation model for multimodal nocturnal biosignals that learns shared representations via cross-modal alignment, handling device heterogeneity and sensor dropout.
Details
Motivation: Traditional sleep monitoring uses diverse PSG devices with heterogeneous biosignals (EEG, EOG, ECG, SpO2), but device heterogeneity and frequent sensor dropout challenge unified modeling of these multimodal signals.Method: Contrastive pre-training on 42,249 overnight recordings across nine modalities using Demography, Age, Site & History-aware InfoNCE objective that incorporates physiological and acquisition metadata to dynamically weight negatives and mitigate cohort-specific shortcuts.
Result: sleep2vec consistently outperforms strong baselines on sleep staging and clinical outcome assessment, remains robust to any subset of available modalities and sensor dropout, and establishes scaling laws for nocturnal biosignals.
Conclusion: Unified cross-modal alignment with principled scaling enables label-efficient, general-purpose modeling of real-world nocturnal biosignals, addressing device heterogeneity and sensor dropout challenges.
Abstract: Tasks ranging from sleep staging to clinical diagnosis traditionally rely on standard polysomnography (PSG) devices, bedside monitors and wearable devices, which capture diverse nocturnal biosignals (e.g., EEG, EOG, ECG, SpO$_2$). However, heterogeneity across devices and frequent sensor dropout pose significant challenges for unified modelling of these multimodal signals. We present \texttt{sleep2vec}, a foundation model for diverse and incomplete nocturnal biosignals that learns a shared representation via cross-modal alignment. \texttt{sleep2vec} is contrastively pre-trained on 42,249 overnight recordings spanning nine modalities using a \textit{Demography, Age, Site & History-aware InfoNCE} objective that incorporates physiological and acquisition metadata (\textit{e.g.}, age, gender, recording site) to dynamically weight negatives and mitigate cohort-specific shortcuts. On downstream sleep staging and clinical outcome assessment, \texttt{sleep2vec} consistently outperforms strong baselines and remains robust to any subset of available modalities and sensor dropout. We further characterize, to our knowledge for the first time, scaling laws for nocturnal biosignals with respect to modality diversity and model capacity. Together, these results show that unified cross-modal alignment, coupled with principled scaling, enables label-efficient, general-purpose modelling of real-world nocturnal biosignals.
[601] Sufficient Conditions for Stability of Minimum-Norm Interpolating Deep ReLU Networks
Ouns El Harzli, Yoonsoo Nam, Ilja Kuzborskij, Bernardo Cuenca Grau, Ard A. Louis
Main category: cs.LG
TL;DR: Analyzes algorithmic stability of deep ReLU neural networks achieving zero training error via minimum-norm interpolation, finding stability depends on having stable sub-networks followed by low-rank weight matrices.
Details
Motivation: Algorithmic stability is a classical framework for analyzing generalization error, but has had limited success in analyzing deep neural networks. The paper aims to understand stability conditions for overparameterized deep ReLU networks that achieve zero training error through minimum-norm interpolation.Method: Theoretical analysis of deep ReLU homogeneous neural networks achieving zero training error with minimum L₂ norm parameters. Investigates sufficient conditions for stability, focusing on networks containing stable sub-networks followed by layers with low-rank weight matrices.
Result: 1) Networks are stable when they contain a stable sub-network followed by a layer with low-rank weight matrix. 2) Networks are not guaranteed to be stable even with stable sub-networks if following layers are not low-rank. Low-rank assumption aligns with empirical/theoretical findings about training bias toward low-rank matrices.
Conclusion: Low-rank weight matrices play crucial role in ensuring algorithmic stability of deep ReLU networks achieving minimum-norm interpolation, providing theoretical insights into generalization properties of overparameterized neural networks.
Abstract: Algorithmic stability is a classical framework for analyzing the generalization error of learning algorithms. It predicts that an algorithm has small generalization error if it is insensitive to small perturbations in the training set such as the removal or replacement of a training point. While stability has been demonstrated for numerous well-known algorithms, this framework has had limited success in analyses of deep neural networks. In this paper we study the algorithmic stability of deep ReLU homogeneous neural networks that achieve zero training error using parameters with the smallest $L_2$ norm, also known as the minimum-norm interpolation, a phenomenon that can be observed in overparameterized models trained by gradient-based algorithms. We investigate sufficient conditions for such networks to be stable. We find that 1) such networks are stable when they contain a (possibly small) stable sub-network, followed by a layer with a low-rank weight matrix, and 2) such networks are not guaranteed to be stable even when they contain a stable sub-network, if the following layer is not low-rank. The low-rank assumption is inspired by recent empirical and theoretical results which demonstrate that training deep neural networks is biased towards low-rank weight matrices, for minimum-norm interpolation and weight-decay regularization.
[602] GREPO: A Benchmark for Graph Neural Networks on Repository-Level Bug Localization
Juntong Wang, Libin Chen, Xiyuan Wang, Shijia Kang, Haotong Yang, Da Zheng, Muhan Zhang
Main category: cs.LG
TL;DR: GREPO is the first GNN benchmark for repository-level bug localization, providing graph-based data structures for 86 Python repositories with 47,294 bug-fixing tasks, showing GNNs outperform traditional retrieval methods.
Details
Motivation: Standard LLMs struggle with repository-level bug localization due to context window limitations, and existing retrieval methods (keyword matching, text similarity, simple graph heuristics) are limited. GNNs offer promise but lack dedicated benchmarks for this task.Method: Created GREPO benchmark with 86 Python repositories and 47,294 bug-fixing tasks, providing graph-based data structures ready for GNN processing. Evaluated various GNN architectures against established information retrieval baselines.
Result: GNN architectures showed outstanding performance compared to traditional information retrieval baselines for repository-scale bug localization tasks.
Conclusion: GREPO establishes a foundation for future research in GNN-based bug localization and demonstrates the potential of GNNs for this critical software engineering task.
Abstract: Repository-level bug localization-the task of identifying where code must be modified to fix a bug-is a critical software engineering challenge. Standard Large Language Modles (LLMs) are often unsuitable for this task due to context window limitations that prevent them from processing entire code repositories. As a result, various retrieval methods are commonly used, including keyword matching, text similarity, and simple graph-based heuristics such as Breadth-First Search. Graph Neural Networks (GNNs) offer a promising alternative due to their ability to model complex, repository-wide dependencies; however, their application has been hindered by the lack of a dedicated benchmark. To address this gap, we introduce GREPO, the first GNN benchmark for repository-scale bug localization tasks. GREPO comprises 86 Python repositories and 47294 bug-fixing tasks, providing graph-based data structures ready for direct GNN processing. Our evaluation of various GNN architectures shows outstanding performance compared to established information retrieval baselines. This work highlights the potential of GNNs for bug localization and established GREPO as a foundation resource for future research, The code is available at https://github.com/qingpingmo/GREPO.
[603] Why Code, Why Now: Learnability, Computability, and the Real Limits of Machine Learning
Zhimin Zhao
Main category: cs.LG
TL;DR: The paper proposes a five-level hierarchy of learnability based on information structure, arguing that ML progress depends more on whether a task is learnable than on model size, explaining why code generation scales better than reinforcement learning.
Details
Motivation: To understand why code generation progresses more reliably than reinforcement learning, and to analyze how information structure affects learnability in machine learning tasks.Method: Proposes a five-level hierarchy of learnability based on information structure, establishes formal distinctions among expressibility, computability, and learnability, and analyzes their pairwise relationships with a unified template.
Result: Provides a framework explaining why supervised learning on code scales predictably while reinforcement learning does not, and challenges the assumption that scaling alone will solve remaining ML challenges.
Conclusion: The ceiling on ML progress depends less on model size than on whether a task is learnable at all, with information structure being the key differentiator between tasks like code generation and reinforcement learning.
Abstract: Code generation has progressed more reliably than reinforcement learning, largely because code has an information structure that makes it learnable. Code provides dense, local, verifiable feedback at every token, whereas most reinforcement learning problems do not. This difference in feedback quality is not binary but graded. We propose a five-level hierarchy of learnability based on information structure and argue that the ceiling on ML progress depends less on model size than on whether a task is learnable at all. The hierarchy rests on a formal distinction among three properties of computational problems (expressibility, computability, and learnability). We establish their pairwise relationships, including where implications hold and where they fail, and present a unified template that makes the structural differences explicit. The analysis suggests why supervised learning on code scales predictably while reinforcement learning does not, and why the common assumption that scaling alone will solve remaining ML challenges warrants scrutiny.
[604] A Multi-Agent Framework for Code-Guided, Modular, and Verifiable Automated Machine Learning
Dat Le, Duc-Cuong Le, Anh-Son Nguyen, Tuan-Dung Bui, Thu-Trang Nguyen, Son Nguyen, Hieu Dinh Vo
Main category: cs.LG
TL;DR: iML is a multi-agent AutoML framework that shifts from black-box prompting to code-guided, modular, and verifiable architecture to address hallucination and logic entanglement in LLM-based AutoML agents.
Details
Motivation: Traditional AutoML frameworks lack transparency and flexibility for complex engineering tasks, while recent LLM-based AutoML agents suffer from hallucinated logic and logic entanglement in monolithic code generation leading to unrecoverable failures.Method: iML introduces three key innovations: 1) Code-Guided Planning using autonomous empirical profiling to eliminate hallucination, 2) Code-Modular Implementation decoupling preprocessing and modeling with strict interface contracts, and 3) Code-Verifiable Integration with dynamic contract verification and iterative self-correction.
Result: iML achieves 85% valid submission rate and 45% competitive medal rate on MLE-BENCH with APS of 0.77, outperforms other approaches by 38%-163% on iML-BENCH, and maintains 70% success rate under stripped task descriptions.
Conclusion: iML bridges the gap between stochastic generation and reliable engineering in AutoML, marking progress toward truly automated machine learning through code-guided, modular, and verifiable architecture.
Abstract: Automated Machine Learning (AutoML) has revolutionized the development of data-driven solutions; however, traditional frameworks often function as “black boxes”, lacking the flexibility and transparency required for complex, real-world engineering tasks. Recent Large Language Model (LLM)-based agents have shifted toward code-driven approaches. However, they frequently suffer from hallucinated logic and logic entanglement, where monolithic code generation leads to unrecoverable runtime failures. In this paper, we present iML, a novel multi-agent framework designed to shift AutoML from black-box prompting to a code-guided, modular, and verifiable architectural paradigm. iML introduces three main ideas: (1) Code-Guided Planning, which synthesizes a strategic blueprint grounded in autonomous empirical profiling to eliminate hallucination; (2) Code-Modular Implementation, which decouples preprocessing and modeling into specialized components governed by strict interface contracts; and (3) Code-Verifiable Integration, which enforces physical feasibility through dynamic contract verification and iterative self-correction. We evaluate iML across MLE-BENCH and the newly introduced iML-BENCH, comprising a diverse range of real-world Kaggle competitions. The experimental results show iML’s superiority over state-of-the-art agents, achieving a valid submission rate of 85% and a competitive medal rate of 45% on MLE-BENCH, with an average standardized performance score (APS) of 0.77. On iML-BENCH, iML significantly outperforms the other approaches by 38%-163% in APS. Furthermore, iML maintains a robust 70% success rate even under stripped task descriptions, effectively filling information gaps through empirical profiling. These results highlight iML’s potential to bridge the gap between stochastic generation and reliable engineering, marking a meaningful step toward truly AutoML.
[605] An Adaptive Model Selection Framework for Demand Forecasting under Horizon-Induced Degradation to Support Business Strategy and Operations
Adolfo González, Víctor Parada
Main category: cs.LG
TL;DR: AHSIV framework for adaptive model selection in forecasting with intermittent demand, using horizon-aware and regime-conditioned selection to address ranking instability across different forecasting horizons.
Details
Motivation: Business environments with structural demand intermittency, high variability, and multi-step planning horizons need robust model selection mechanisms. No forecasting model is universally dominant, and relative rankings vary across error metrics, demand regimes, and forecast horizons, creating ambiguity in multi-SKU decision contexts.Method: Proposes AHSIV (Adaptive Hybrid Selector for Intermittency and Variability), a horizon-aware and regime-conditioned model selection framework. Integrates scaled/absolute error metrics adjusted via Metric Degradation by Forecast Horizon (MDFH) procedure, structural demand classification, multi-objective Pareto dominance, and hierarchical bias refinement in a unified decision architecture.
Result: Evaluation on Walmart, M3, M4, and M5 datasets under multiple train-test partitions and twelve-step forecasting horizons shows AHSIV achieves statistical equivalence with strongest monometric baseline in aggregated performance while increasing frequency of horizon-specific best-model selection.
Conclusion: Model selection in heterogeneous demand environments cannot be treated as static ranking problem; horizon-consistent, structurally adaptive mechanisms provide principled, operationally coherent solution for multi-SKU forecasting.
Abstract: Business environments characterized by structural demand intermittency, high variability, and multi-step planning horizons require robust and reproducible model selection mechanisms. Empirical evidence shows that no forecasting model is universally dominant and that relative rankings vary across error metrics, demand regimes, and forecast horizons, generating ambiguity in multi-SKU decision contexts. This study proposes AHSIV (Adaptive Hybrid Selector for Intermittency and Variability), a horizon-aware and regime-conditioned model selection framework designed to address horizon-induced ranking instability. The proposed approach integrates scaled and absolute error metrics adjusted through a Metric Degradation by Forecast Horizon (MDFH) procedure, structural demand classification, multi-objective Pareto dominance, and hierarchical bias refinement within a unified decision architecture. The empirical evaluation is conducted on the Walmart, M3, M4, and M5 datasets under multiple train-test partition schemes and twelve-step forecasting horizons. Results indicate that AHSIV achieves statistical equivalence with the strongest monometric baseline in terms of aggregated performance while increasing the frequency of horizon-specific best-model selection. The findings demonstrate that model selection in heterogeneous demand environments cannot be treated as a static ranking problem, and that horizon-consistent, structurally adaptive mechanisms provide a principled, operationally coherent solution for multi-SKU forecasting.
[606] You Can Learn Tokenization End-to-End with Reinforcement Learning
Sam Dauncey, Roger Wattenhofer
Main category: cs.LG
TL;DR: Tokenization can be learned end-to-end using score function estimates with reinforcement learning techniques, outperforming previous straight-through methods at 100M parameter scale.
Details
Motivation: Tokenization remains a hardcoded compression step in LLM training pipelines despite the trend toward end-to-end architectures. Prior work used heuristics or straight-through estimates, but these have limitations in learning discrete token boundaries effectively.Method: Proposes learning token boundaries using score function estimates with reinforcement learning techniques like time discounting to reduce variance. This directly optimizes the discrete token boundary problem to minimize loss.
Result: The method outperforms prior straight-through estimates both qualitatively and quantitatively at the 100 million parameter scale.
Conclusion: Score function estimates with RL techniques provide a more effective approach to learning tokenization end-to-end in LLMs, with better theoretical guarantees than previous methods.
Abstract: Tokenization is a hardcoded compression step which remains in the training pipeline of Large Language Models (LLMs), despite a general trend towards architectures becoming increasingly end-to-end. Prior work has shown promising results at scale in bringing this compression step inside the LLMs’ architecture with heuristics to draw token boundaries, and also attempts to learn these token boundaries with straight-through estimates, which treat the problem of drawing discrete token boundaries as a continuous one. We show that these token boundaries can instead be learned using score function estimates, which have tighter theoretical guarantees due to directly optimizing the problem of drawing discrete token boundaries to minimize loss. We observe that techniques from reinforcement learning, such as time discounting, are necessary to reduce the variance of this score function sufficiently to make it practicable. We demonstrate that the resultant method outperforms prior proposed straight-through estimates, both qualitatively and quantitatively at the $100$ million parameter scale.
[607] Experiential Reinforcement Learning
Taiwei Shi, Sihao Chen, Bowen Jiang, Linxin Song, Longqi Yang, Jieyu Zhao
Main category: cs.LG
TL;DR: ERL integrates explicit experience-reflection-consolidation loop into RL for language models, converting sparse/delayed feedback into structured behavioral revision through self-reflection.
Details
Motivation: Reinforcement learning for language models faces challenges with sparse and delayed environmental feedback, requiring models to implicitly infer how failures should translate into behavioral changes.Method: Experiential Reinforcement Learning (ERL) embeds an explicit experience-reflection-consolidation loop: model generates initial attempt, receives feedback, produces reflection to guide refined second attempt, then reinforces successful outcomes into base policy.
Result: ERL improves learning efficiency and final performance over strong RL baselines, achieving +81% gains in complex multi-step environments and +11% in tool-using reasoning tasks.
Conclusion: Integrating explicit self-reflection into policy training provides a practical mechanism for transforming feedback into durable behavioral improvement without additional inference cost.
Abstract: Reinforcement learning has become the central approach for language models (LMs) to learn from environmental reward or feedback. In practice, the environmental feedback is usually sparse and delayed. Learning from such signals is challenging, as LMs must implicitly infer how observed failures should translate into behavioral changes for future iterations. We introduce Experiential Reinforcement Learning (ERL), a training paradigm that embeds an explicit experience-reflection-consolidation loop into the reinforcement learning process. Given a task, the model generates an initial attempt, receives environmental feedback, and produces a reflection that guides a refined second attempt, whose success is reinforced and internalized into the base policy. This process converts feedback into structured behavioral revision, improving exploration and stabilizing optimization while preserving gains at deployment without additional inference cost. Across sparse-reward control environments and agentic reasoning benchmarks, ERL consistently improves learning efficiency and final performance over strong reinforcement learning baselines, achieving gains of up to +81% in complex multi-step environments and up to +11% in tool-using reasoning tasks. These results suggest that integrating explicit self-reflection into policy training provides a practical mechanism for transforming feedback into durable behavioral improvement.
[608] QuRL: Efficient Reinforcement Learning with Quantized Rollout
Yuhang Li, Reena Elangovan, Xin Dong, Priyadarshini Panda, Brucek Khailany
Main category: cs.LG
TL;DR: QuRL accelerates RL training for reasoning LLMs by using quantized actors for rollout, addressing training collapse and weight update issues with adaptive clipping and invariant scaling techniques.
Details
Motivation: RL training for reasoning LLMs suffers from slow rollout processes (up to 70% of training time) due to autoregressive decoding, creating an efficiency bottleneck that needs addressing.Method: Proposes Quantized Reinforcement Learning (QuRL) using quantized actors for rollout acceleration. Introduces Adaptive Clipping Range (ACR) to prevent training collapse by dynamically adjusting clipping ratio based on policy differences, and invariant scaling technique to address weight update problems by reducing quantization noise.
Result: Achieves 20% to 80% faster rollout during training in INT8 and FP8 quantization experiments on DeepScaleR and DAPO benchmarks.
Conclusion: QuRL effectively accelerates RL training for reasoning LLMs through quantization techniques while maintaining training stability, offering significant efficiency improvements for RLVR paradigms.
Abstract: Reinforcement learning with verifiable rewards (RLVR) has become a trending paradigm for training reasoning large language models (LLMs). However, due to the autoregressive decoding nature of LLMs, the rollout process becomes the efficiency bottleneck of RL training, consisting of up to 70% of the total training time. In this work, we propose Quantized Reinforcement Learning (QuRL) that uses a quantized actor for accelerating the rollout. We address two challenges in QuRL. First, we propose Adaptive Clipping Range (ACR) that dynamically adjusts the clipping ratio based on the policy ratio between the full-precision actor and the quantized actor, which is essential for mitigating long-term training collapse. Second, we identify the weight update problem, where weight changes between RL steps are extremely small, making it difficult for the quantization operation to capture them effectively. We mitigate this problem through the invariant scaling technique that reduces quantization noise and increases weight update. We evaluate our method with INT8 and FP8 quantization experiments on DeepScaleR and DAPO, and achieve 20% to 80% faster rollout during training.
[609] Chemical Language Models for Natural Products: A State-Space Model Approach
Ho-Hsuan Wang, Afnan Sultan, Andrea Volkamer, Dietrich Klakow
Main category: cs.LG
TL;DR: NP-specific chemical language models using Mamba state-space architectures outperform transformers for natural product generation and property prediction tasks.
Details
Motivation: Natural Products (NPs) are important for drug discovery but underexplored in chemical language modeling, with existing models focusing on general molecules rather than NP-specific applications.Method: Developed NP-specific chemical language models by pre-training state-space models (Mamba and Mamba-2) and comparing with transformer baselines (GPT) on ~1M NPs dataset. Evaluated eight tokenization strategies including character-level, Atom-in-SMILES, BPE, and NP-specific BPE.
Result: Mamba generates 1-2% more valid and unique molecules than Mamba-2 and GPT with fewer long-range dependency errors. For property prediction, Mamba variants outperform GPT by 0.02-0.04 MCC under random splits. Domain-specific pre-training on 1M NPs matches models trained on datasets 100x larger.
Conclusion: State-space models like Mamba are effective for NP-focused chemical language modeling, demonstrating that domain-specific pre-training on relatively small datasets can achieve competitive performance compared to models trained on much larger general molecular datasets.
Abstract: Language models are widely used in chemistry for molecular property prediction and small-molecule generation, yet Natural Products (NPs) remain underexplored despite their importance in drug discovery. To address this gap, we develop NP-specific chemical language models (NPCLMs) by pre-training state-space models (Mamba and Mamba-2) and comparing them with transformer baselines (GPT). Using a dataset of about 1M NPs, we present the first systematic comparison of selective state-space models and transformers for NP-focused tasks, together with eight tokenization strategies including character-level, Atom-in-SMILES (AIS), byte-pair encoding (BPE), and NP-specific BPE. We evaluate molecule generation (validity, uniqueness, novelty) and property prediction (membrane permeability, taste, anti-cancer activity) using MCC and AUC-ROC. Mamba generates 1-2 percent more valid and unique molecules than Mamba-2 and GPT, with fewer long-range dependency errors, while GPT yields slightly more novel structures. For property prediction, Mamba variants outperform GPT by 0.02-0.04 MCC under random splits, while scaffold splits show comparable performance. Results demonstrate that domain-specific pre-training on about 1M NPs can match models trained on datasets over 100 times larger.
[610] Steady-State Behavior of Constant-Stepsize Stochastic Approximation: Gaussian Approximation and Tail Bounds
Zedong Wang, Yuyang Wang, Ijay Narang, Felix Wang, Yuzhou Wang, Siva Theja Maguluri
Main category: cs.LG
TL;DR: This paper provides explicit, non-asymptotic error bounds for constant-stepsize stochastic approximation algorithms, analyzing the approximation error between the stationary distribution and its Gaussian limit for fixed stepsizes.
Details
Motivation: Constant-stepsize stochastic approximation is widely used for computational efficiency, but the stationary distribution is rarely tractable. While prior work shows weak convergence to Gaussian limits as stepsize approaches zero, there are no usable error bounds for fixed stepsizes, which is crucial for practical applications.Method: The authors prove general theorems bounding Wasserstein distance between centered-scaled steady state and Gaussian distribution under regularity conditions for drift and moment conditions for noise, covering both i.i.d. and Markovian noise models. They instantiate these theorems for three SA settings: SGD for smooth strongly convex objectives, linear SA, and contractive nonlinear SA.
Result: Obtained dimension- and stepsize-dependent explicit bounds in Wasserstein distance of order α^{1/2}log(1/α) for small α. Derived non-uniform Berry-Esseen-type tail bounds with error terms decaying in both deviation level and stepsize α. For SGD beyond strong convexity, identified a non-Gaussian (Gibbs) limiting law under correct scaling.
Conclusion: The paper provides rigorous, non-asymptotic error bounds for fixed-stepsize stochastic approximation algorithms, bridging the gap between asymptotic theory and practical implementation, with applications to various optimization and learning settings.
Abstract: Constant-stepsize stochastic approximation (SA) is widely used in learning for computational efficiency. For a fixed stepsize, the iterates typically admit a stationary distribution that is rarely tractable. Prior work shows that as the stepsize $α\downarrow 0$, the centered-and-scaled steady state converges weakly to a Gaussian random vector. However, for fixed $α$, this weak convergence offers no usable error bound for approximating the steady-state by its Gaussian limit. This paper provides explicit, non-asymptotic error bounds for fixed $α$. We first prove general-purpose theorems that bound the Wasserstein distance between the centered-scaled steady state and an appropriate Gaussian distribution, under regularity conditions for drift and moment conditions for noise. To ensure broad applicability, we cover both i.i.d. and Markovian noise models. We then instantiate these theorems for three representative SA settings: (1) stochastic gradient descent (SGD) for smooth strongly convex objectives, (2) linear SA, and (3) contractive nonlinear SA. We obtain dimension- and stepsize-dependent, explicit bounds in Wasserstein distance of order $α^{1/2}\log(1/α)$ for small $α$. Building on the Wasserstein approximation error, we further derive non-uniform Berry–Esseen-type tail bounds that compare the steady-state tail probability to Gaussian tails. We achieve an explicit error term that decays in both the deviation level and stepsize $α$. We adapt the same analysis for SGD beyond strongly convexity and study general convex objectives. We identify a non-Gaussian (Gibbs) limiting law under the correct scaling, which is validated numerically, and provide a corresponding pre-limit Wasserstein error bound.
[611] KoopGen: Koopman Generator Networks for Representing and Predicting Dynamical Systems with Continuous Spectra
Liangyu Su, Jun Shu, Rui Liu, Deyu Meng, Zongben Xu
Main category: cs.LG
TL;DR: KoopGen is a generator-based neural Koopman framework that models chaotic dynamical systems through structured, state-dependent representations of Koopman generators, separating conservative transport from irreversible dissipation while enforcing operator-theoretic constraints.
Details
Motivation: Existing data-driven models for chaotic dynamical systems lack stability, interpretability, and scalability in regimes with broadband or continuous spectra. Koopman-based approaches provide linear perspectives but rely on restrictive finite-dimensional assumptions or explicit spectral parameterizations that degrade in high-dimensional settings.Method: KoopGen uses a generator-based neural Koopman framework that models dynamics through structured, state-dependent representations of Koopman generators. It exploits Cartesian decomposition into skew-adjoint (conservative transport) and self-adjoint (irreversible dissipation) components while enforcing exact operator-theoretic constraints during learning.
Result: Across systems ranging from nonlinear oscillators to high-dimensional chaotic and spatiotemporal dynamics, KoopGen improves prediction accuracy and stability while clarifying which components of continuous-spectrum dynamics admit interpretable and learnable representations.
Conclusion: KoopGen provides a principled approach to modeling high-dimensional chaotic systems by combining neural networks with operator-theoretic constraints, offering improved stability and interpretability for continuous-spectrum dynamics.
Abstract: Representing and predicting high-dimensional and spatiotemporally chaotic dynamical systems remains a fundamental challenge in dynamical systems and machine learning. Although data-driven models can achieve accurate short-term forecasts, they often lack stability, interpretability, and scalability in regimes dominated by broadband or continuous spectra. Koopman-based approaches provide a principled linear perspective on nonlinear dynamics, but existing methods rely on restrictive finite-dimensional assumptions or explicit spectral parameterizations that degrade in high-dimensional settings. Against these issues, we introduce KoopGen, a generator-based neural Koopman framework that models dynamics through a structured, state-dependent representation of Koopman generators. By exploiting the intrinsic Cartesian decomposition into skew-adjoint and self-adjoint components, KoopGen separates conservative transport from irreversible dissipation while enforcing exact operator-theoretic constraints during learning. Across systems ranging from nonlinear oscillators to high-dimensional chaotic and spatiotemporal dynamics, KoopGen improves prediction accuracy and stability, while clarifying which components of continuous-spectrum dynamics admit interpretable and learnable representations.
[612] S2SServiceBench: A Multimodal Benchmark for Last-Mile S2S Climate Services
Chenyue Li, Wen Deng, Zhuotao Sun, Mengxi Jin, Hanzhe Cui, Han Li, Shentong Li, Man Kit Yu, Ming Long Lai, Yuhao Yang, Mengqian Lu, Binhang Yuan
Main category: cs.LG
TL;DR: S2SServiceBench: A multimodal benchmark for evaluating MLLMs’ ability to translate climate forecasts into actionable services across six application domains with 500+ tasks.
Details
Motivation: There's a "last-mile gap" in climate services where scientific forecasts need to be translated into trusted, actionable services requiring multimodal understanding and decision-making under uncertainty. While MLLMs have advanced in supporting workflows, their reliability in generating decision-making deliverables from operational climate service products remains unclear.Method: Created S2SServiceBench, a multimodal benchmark curated from operational climate-service systems covering 10 service products with 150+ expert-selected cases across six domains (Agriculture, Disasters, Energy, Finance, Health, Shipping). Each case has three service levels, yielding ~500 tasks and 1,000+ evaluation items.
Result: Benchmarked state-of-the-art MLLMs and agents, revealing persistent challenges in S2S service plot understanding and reasoning, including actionable signal comprehension, operationalizing uncertainty into executable handoffs, and stable evidence-grounded analysis for dynamic hazards.
Conclusion: The benchmark provides actionable guidance for building future climate-service agents and highlights MLLMs’ limitations in reliable decision-making from operational climate service products under uncertainty.
Abstract: Subseasonal-to-seasonal (S2S) forecasts play an essential role in providing a decision-critical weeks-to-months planning window for climate resilience and sustainability, yet a growing bottleneck is the last-mile gap: translating scientific forecasts into trusted, actionable climate services, requiring reliable multimodal understanding and decision-facing reasoning under uncertainty. Meanwhile, multimodal large language models (MLLMs) and corresponding agentic paradigms have made rapid progress in supporting various workflows, but it remains unclear whether they can reliably generate decision-making deliverables from operational service products (e.g., actionable signal comprehension, decision-making handoff, and decision analysis & planning) under uncertainty. We introduce S2SServiceBench, a multimodal benchmark for last-mile S2S climate services curated from an operational climate-service system to evaluate this capability. S2SServiceBenchcovers 10 service products with about 150+ expert-selected cases in total, spanning six application domains - Agriculture, Disasters, Energy, Finance, Health, and Shipping. Each case is instantiated at three service levels, yielding around 500 tasks and 1,000+ evaluation items across climate resilience and sustainability applications. Using S2SServiceBench, we benchmark state-of-the-art MLLMs and agents, and analyze performance across products and service levels, revealing persistent challenges in S2S service plot understanding and reasoning - namely, actionable signal comprehension, operationalizing uncertainty into executable handoffs, and stable, evidence-grounded analysis and planning for dynamic hazards-while offering actionable guidance for building future climate-service agents.
[613] EIDOS: Latent-Space Predictive Learning for Time Series Foundation Models
Xinxing Zhou, Qingren Yao, Yiji Zhao, Chenghao Liu, Flora Salim, Xiaojie Yuan, Yanlong Wen, Ming Jin
Main category: cs.LG
TL;DR: EIDOS is a time series foundation model family that shifts from direct future value prediction to latent-space predictive learning, using a causal Transformer to predict latent representation evolution for more structured and temporally coherent representations.
Details
Motivation: Current time series foundation models predict future observations directly, resulting in weakly structured latent representations that capture surface noise rather than coherent temporal dynamics. The authors aim to develop more robust models that learn predictable latent dynamics.Method: EIDOS uses latent-space predictive learning where a causal Transformer predicts the evolution of latent representations. It includes a lightweight aggregation branch to construct stable target representations, and is optimized via a joint objective combining latent-space alignment, observational grounding, and direct forecasting supervision.
Result: On the GIFT-Eval benchmark, EIDOS mitigates structural fragmentation in representation space and achieves state-of-the-art performance, demonstrating that learning predictable latent dynamics leads to more robust time series foundation models.
Conclusion: Constraining models to learn predictable latent dynamics is a principled approach toward more robust and reliable time series foundation models, with EIDOS showing improved representation quality and performance through latent-space predictive learning.
Abstract: Most time series foundation models are pretrained by directly predicting future observations, which often yields weakly structured latent representations that capture surface noise rather than coherent and predictable temporal dynamics. In this work, we introduce EIDOS, a foundation model family that shifts pretraining from future value prediction to latent-space predictive learning. We train a causal Transformer to predict the evolution of latent representations, encouraging the emergence of structured and temporally coherent latent states. To ensure stable targets for latent-space learning, we design a lightweight aggregation branch to construct target representations. EIDOS is optimized via a joint objective that integrates latent-space alignment, observational grounding to anchor representations to the input signal, and direct forecasting supervision. On the GIFT-Eval benchmark, EIDOS mitigates structural fragmentation in the representation space and achieves state-of-the-art performance. These results demonstrate that constraining models to learn predictable latent dynamics is a principled step toward more robust and reliable time series foundation models.
[614] UniST-Pred: A Robust Unified Framework for Spatio-Temporal Traffic Forecasting in Transportation Networks Under Disruptions
Yue Wang, Areg Karapetyan, Djellel Difallah, Samer Madanat
Main category: cs.LG
TL;DR: UniST-Pred is a unified spatio-temporal traffic forecasting framework that decouples temporal modeling from spatial representation learning, then integrates both through adaptive fusion, showing robustness under network disruptions.
Details
Motivation: Current traffic forecasting models often tightly couple spatial and temporal modeling, increasing complexity and limiting modularity, while real-world deployments require robustness under structural and observational uncertainties.Method: Proposes UniST-Pred framework that first decouples temporal modeling from spatial representation learning, then integrates both through adaptive representation-level fusion. Evaluates using agent-based traffic simulator (MATSim) under severe network disconnection scenarios.
Result: UniST-Pred demonstrates competitive performance against established models on standard datasets despite lightweight design, maintains strong predictive performance across real-world and simulated datasets, and yields interpretable spatio-temporal representations under infrastructure disruptions.
Conclusion: The decoupled approach with adaptive fusion provides robust spatio-temporal forecasting under uncertainty while maintaining interpretability and competitive performance with lightweight design.
Abstract: Spatio-temporal traffic forecasting is a core component of intelligent transportation systems, supporting various downstream tasks such as signal control and network-level traffic management. In real-world deployments, forecasting models must operate under structural and observational uncertainties, conditions that are rarely considered in model design. Recent approaches achieve strong short-term predictive performance by tightly coupling spatial and temporal modeling, often at the cost of increased complexity and limited modularity. In contrast, efficient time-series models capture long-range temporal dependencies without relying on explicit network structure. We propose UniST-Pred, a unified spatio-temporal forecasting framework that first decouples temporal modeling from spatial representation learning, then integrates both through adaptive representation-level fusion. To assess robustness of the proposed approach, we construct a dataset based on an agent-based, microscopic traffic simulator (MATSim) and evaluate UniST-Pred under severe network disconnection scenarios. Additionally, we benchmark UniST-Pred on standard traffic prediction datasets, demonstrating its competitive performance against existing well-established models despite a lightweight design. The results illustrate that UniST-Pred maintains strong predictive performance across both real-world and simulated datasets, while also yielding interpretable spatio-temporal representations under infrastructure disruptions. The source code and the generated dataset are available at https://anonymous.4open.science/r/UniST-Pred-EF27
[615] Position Encoding with Random Float Sampling Enhances Length Generalization of Transformers
Atsushi Shimizu, Shohei Taniguchi, Yutaka Matsuo
Main category: cs.LG
TL;DR: Random Float Sampling (RFS) position encoding improves length generalization in language models by using continuous random position indices during training instead of predefined discrete sets.
Details
Motivation: Current position encoding methods struggle with length generalization - maintaining performance on inputs longer than those seen during training. This is because they use predefined discrete position indices that cause out-of-distribution issues when encountering unseen lengths.Method: RFS replaces discrete position indices with randomly sampled continuous values during training. This exposes models to diverse position indices, avoiding OOD issues. The approach can be easily integrated with existing position encodings like absolute sinusoidal, RoPE, and ALiBi.
Result: RFS demonstrates superior performance in length generalization tasks and shows improved results on zero-shot commonsense reasoning benchmarks compared to standard position encoding methods.
Conclusion: Random Float Sampling provides a simple yet effective solution to length generalization in language models by using continuous position indices during training, making it compatible with various existing position encoding schemes.
Abstract: Length generalization is the ability of language models to maintain performance on inputs longer than those seen during pretraining. In this work, we introduce a simple yet powerful position encoding (PE) strategy, Random Float Sampling (RFS), that generalizes well to lengths unseen during pretraining or fine-tuning. In particular, instead of selecting position indices from a predefined discrete set, RFS uses randomly sampled continuous values, thereby avoiding out-of-distribution (OOD) issues on unseen lengths by exposing the model to diverse indices during training. Since assigning indices to tokens is a common and fundamental procedure in widely used PEs, the advantage of RFS can easily be incorporated into, for instance, the absolute sinusoidal encoding, RoPE, and ALiBi. Experiments corroborate its effectiveness by showing that RFS results in superior performance in length generalization tasks as well as zero-shot commonsense reasoning benchmarks.
[616] Decentralized Federated Learning With Energy Harvesting Devices
Kai Zhang, Xuanyu Cao, Khaled B. Letaief
Main category: cs.LG
TL;DR: Decentralized federated learning with energy harvesting: Joint device scheduling and power control using decentralized policy iteration algorithm to optimize energy efficiency and convergence in DFL systems.
Details
Motivation: Decentralized federated learning (DFL) on edge devices faces energy depletion issues that reduce operational lifetime and degrade learning performance. Energy harvesting offers sustainable operation but requires intelligent resource management to accelerate convergence.Method: Derived convergence bound for wireless DFL with energy harvesting, formulated joint device scheduling and power control as multi-agent MDP, and proposed fully decentralized policy iteration algorithm using only local two-hop neighbor information.
Result: Theoretical analysis shows asymptotic optimality of decentralized algorithm. Comprehensive numerical experiments on real-world datasets validate effectiveness in reducing communication overhead and computational complexity.
Conclusion: Energy harvesting enables sustainable DFL, and the proposed decentralized algorithm efficiently manages energy resources while maintaining convergence performance with reduced complexity.
Abstract: Decentralized federated learning (DFL) enables edge devices to collaboratively train models through local training and fully decentralized device-to-device (D2D) model exchanges. However, these energy-intensive operations often rapidly deplete limited device batteries, reducing their operational lifetime and degrading the learning performance. To address this limitation, we apply energy harvesting technique to DFL systems, allowing edge devices to extract ambient energy and operate sustainably. We first derive the convergence bound for wireless DFL with energy harvesting, showing that the convergence is influenced by partial device participation and transmission packet drops, both of which further depend on the available energy supply. To accelerate convergence, we formulate a joint device scheduling and power control problem and model it as a multi-agent Markov decision process (MDP). Traditional MDP algorithms (e.g., value or policy iteration) require a centralized coordinator with access to all device states and exhibit exponential complexity in the number of devices, making them impractical for large-scale decentralized networks. To overcome these challenges, we propose a fully decentralized policy iteration algorithm that leverages only local state information from two-hop neighboring devices, thereby substantially reducing both communication overhead and computational complexity. We further provide a theoretical analysis showing that the proposed decentralized algorithm achieves asymptotic optimality. Finally, comprehensive numerical experiments on real-world datasets are conducted to validate the theoretical results and corroborate the effectiveness of the proposed algorithm.
[617] Policy Gradient with Adaptive Entropy Annealing for Continual Fine-Tuning
Yaqian Zhang, Bernhard Pfahringer, Eibe Frank, Albert Bifet
Main category: cs.LG
TL;DR: The paper proposes a reinforcement learning approach to address catastrophic forgetting in class-incremental learning for vision models, replacing cross-entropy loss with Expected Policy Gradient to directly minimize misclassification error.
Details
Motivation: Large pretrained vision models suffer from catastrophic forgetting when adapted to new tasks in class-incremental settings. While parameter-efficient fine-tuning helps, most approaches still rely on cross-entropy loss, which is a surrogate for the true 0-1 loss objective. The authors aim to directly optimize the true objective of minimizing misclassification error.Method: The authors formulate classification as a one-step Markov Decision Process and derive an Expected Policy Gradient (EPG) method that directly minimizes misclassification error with low-variance gradient estimation. They analyze that CE can be interpreted as EPG with additional sample-weighting, where CE emphasizes low-confidence samples (exploration) while EPG prioritizes high-confidence ones (exploitation). Based on this insight, they propose adaptive entropy annealing (aEPG), a training strategy that transitions from exploratory (CE-like) to exploitative (EPG-like) learning.
Result: aEPG-based methods outperform CE-based methods across diverse benchmarks and with various PEFT modules. The authors also demonstrate that lower entropy of the output prediction distribution enhances adaptation in pretrained vision models.
Conclusion: The paper successfully revives the true 0-1 loss objective through a reinforcement learning perspective, showing that directly optimizing misclassification error via EPG and adaptive entropy annealing improves performance in class-incremental learning for vision models, with lower entropy distributions being beneficial for adaptation.
Abstract: Despite their success, large pretrained vision models remain vulnerable to catastrophic forgetting when adapted to new tasks in class-incremental settings. Parameter-efficient fine-tuning (PEFT) alleviates this by restricting trainable parameters, yet most approaches still rely on cross-entropy (CE) loss, a surrogate for the 0-1 loss, to learn from new data. We revisit this choice and revive the true objective (0-1 loss) through a reinforcement learning perspective. By formulating classification as a one-step Markov Decision Process, we derive an Expected Policy Gradient (EPG) method that directly minimizes misclassification error with a low-variance gradient estimation. Our analysis shows that CE can be interpreted as EPG with an additional sample-weighting mechanism: CE encourages exploration by emphasizing low-confidence samples, while EPG prioritizes high-confidence ones. Building on this insight, we propose adaptive entropy annealing (aEPG), a training strategy that transitions from exploratory (CE-like) to exploitative (EPG-like) learning. aEPG-based methods outperform CE-based methods across diverse benchmarks and with various PEFT modules. More broadly, we evaluate various entropy regularization methods and demonstrate that lower entropy of the output prediction distribution enhances adaptation in pretrained vision models.
[618] Neural Optimal Transport in Hilbert Spaces: Characterizing Spurious Solutions and Gaussian Smoothing
Jae-Hwan Choi, Jiwoo Yoon, Dohyun Kwon, Jaewoong Choi
Main category: cs.LG
TL;DR: Neural Optimal Transport in infinite-dimensional Hilbert spaces with Gaussian smoothing to address spurious solutions in non-regular settings.
Details
Motivation: Semi-dual Neural OT in infinite-dimensional Hilbert spaces often generates spurious solutions that fail to accurately capture target distributions in non-regular settings, requiring a solution to this ill-posedness problem.Method: Extend the semi-dual framework via Gaussian smoothing strategy based on Brownian motion, using regular measures framework to characterize spurious solutions and prove well-posedness under regular source measures.
Result: Theoretical proof that under regular source measure, the formulation is well-posed and recovers unique Monge map; sharp characterization for regularity of smoothed measures; empirical results show effective suppression of spurious solutions and outperformance of baselines on synthetic functional data and time-series datasets.
Conclusion: Gaussian smoothing resolves ill-posedness in Neural OT for infinite-dimensional Hilbert spaces, providing theoretical guarantees and practical effectiveness for functional data applications.
Abstract: We study Neural Optimal Transport in infinite-dimensional Hilbert spaces. In non-regular settings, Semi-dual Neural OT often generates spurious solutions that fail to accurately capture target distributions. We analytically characterize this spurious solution problem using the framework of regular measures, which generalize Lebesgue absolute continuity in finite dimensions. To resolve ill-posedness, we extend the semi-dual framework via a Gaussian smoothing strategy based on Brownian motion. Our primary theoretical contribution proves that under a regular source measure, the formulation is well-posed and recovers a unique Monge map. Furthermore, we establish a sharp characterization for the regularity of smoothed measures, proving that the success of smoothing depends strictly on the kernel of the covariance operator. Empirical results on synthetic functional data and time-series datasets demonstrate that our approach effectively suppresses spurious solutions and outperforms existing baselines.
[619] Geometry-Aware Physics-Informed PointNets for Modeling Flows Across Porous Structures
Luigi Ciceri, Corrado Mio, Jianyi Lin, Gabriele Gianini
Main category: cs.LG
TL;DR: Physics-informed neural networks (PIPN and PI-GANO) for predicting fluid flow through and around porous bodies, trained on CFD data and enforcing Navier-Stokes and Darcy-Forchheimer equations.
Details
Motivation: Predicting coupled fluid-porous flows is challenging due to complex physics across different regions and the need to generalize across diverse geometries and boundary conditions. Traditional CFD methods are computationally expensive for design studies requiring multiple geometry variations.Method: Two physics-informed learning approaches: Physics Informed PointNets (PIPN) and Physics Informed Geometry Aware Neural Operator (PI-GANO). Both enforce incompressible Navier-Stokes equations in free-flow regions and Darcy-Forchheimer extension in porous regions within a unified loss function. Networks are conditioned on geometry and material parameters. Datasets generated with OpenFOAM on 2D ducts with porous obstacles and 3D windbreak scenarios with tree canopies and buildings.
Result: Consistently low velocity and pressure errors in both seen and unseen cases, with accurate reproduction of wake structures. Performance degrades primarily near sharp interfaces and in regions with large gradients. PI-GANO shows generalization to variable boundary conditions and parameter settings.
Conclusion: First systematic evaluation of PIPN/PI-GANO for simultaneous through-and-around porous flows demonstrates their potential to accelerate design studies without retraining per geometry, offering an efficient alternative to traditional CFD methods.
Abstract: Predicting flows that occur both through and around porous bodies is challenging due to coupled physics across fluid and porous regions and the need to generalize across diverse geometries and boundary conditions. We address this problem using two Physics Informed learning approaches: Physics Informed PointNets (PIPN) and Physics Informed Geometry Aware Neural Operator (P-IGANO). We enforce the incompressible Navier Stokes equations in the free-flow region and a Darcy Forchheimer extension in the porous region within a unified loss and condition the networks on geometry and material parameters. Datasets are generated with OpenFOAM on 2D ducts containing porous obstacles and on 3D windbreak scenarios with tree canopies and buildings. We first verify the pipeline via the method of manufactured solutions, then assess generalization to unseen shapes, and for PI-GANO, to variable boundary conditions and parameter settings. The results show consistently low velocity and pressure errors in both seen and unseen cases, with accurate reproduction of the wake structures. Performance degrades primarily near sharp interfaces and in regions with large gradients. Overall, the study provides a first systematic evaluation of PIPN/PI-GANO for simultaneous through-and-around porous flows and shows their potential to accelerate design studies without retraining per geometry.
[620] Sanity Checks for Sparse Autoencoders: Do SAEs Beat Random Baselines?
Anton Korznikov, Andrey Galichin, Alexey Dontsov, Oleg Rogov, Ivan Oseledets, Elena Tutubalina
Main category: cs.LG
TL;DR: SAEs fail to reliably recover meaningful features despite strong reconstruction metrics, with random baselines matching their performance on interpretability tasks.
Details
Motivation: Despite excitement about Sparse Autoencoders (SAEs) for interpreting neural networks, negative results in downstream tasks raise doubts about whether they actually recover meaningful features from model activations.Method: Two complementary evaluations: 1) Synthetic setup with known ground-truth features to measure recovery rate, 2) Real activation evaluation with three baselines that constrain SAE feature directions or activation patterns to random values.
Result: SAEs recover only 9% of true features despite 71% explained variance in synthetic setup. Random baselines match fully-trained SAEs in interpretability (0.87 vs 0.90), sparse probing (0.69 vs 0.72), and causal editing (0.73 vs 0.72).
Conclusion: SAEs in their current state do not reliably decompose models’ internal mechanisms, challenging their utility as interpretability tools.
Abstract: Sparse Autoencoders (SAEs) have emerged as a promising tool for interpreting neural networks by decomposing their activations into sparse sets of human-interpretable features. Recent work has introduced multiple SAE variants and successfully scaled them to frontier models. Despite much excitement, a growing number of negative results in downstream tasks casts doubt on whether SAEs recover meaningful features. To directly investigate this, we perform two complementary evaluations. On a synthetic setup with known ground-truth features, we demonstrate that SAEs recover only $9%$ of true features despite achieving $71%$ explained variance, showing that they fail at their core task even when reconstruction is strong. To evaluate SAEs on real activations, we introduce three baselines that constrain SAE feature directions or their activation patterns to random values. Through extensive experiments across multiple SAE architectures, we show that our baselines match fully-trained SAEs in interpretability (0.87 vs 0.90), sparse probing (0.69 vs 0.72), and causal editing (0.73 vs 0.72). Together, these results suggest that SAEs in their current state do not reliably decompose models’ internal mechanisms.
[621] ROAST: Rollout-based On-distribution Activation Steering Technique
Xuanbo Su, Hao Luo, Yingfang Zhang, Lijun Zhang
Main category: cs.LG
TL;DR: ROAST is a parameter-efficient activation steering technique for LLMs that uses on-distribution rollouts and continuous soft scaling with grouped normalization to improve steering robustness and performance across diverse tasks.
Details
Motivation: Existing activation steering methods rely on off-distribution supervision and discrete masking, leading to brittle interventions. The authors aim to develop a more robust steering technique that uses the model's own on-distribution behavior and addresses issues with activation magnitude variance.Method: ROAST uses rollout-based on-distribution activation steering with ROC (Rollout-based On-distribution Control) to estimate steering directions from the model’s own rollouts. It employs Continuous Soft Scaling (CSS) instead of hard sparsification and Grouped Mean Normalization to balance contributions across samples and prevent high-magnitude activations from dominating the steering direction.
Result: ROAST consistently improves performance across models ranging from 0.6B to 32B parameters on diverse tasks, including +9.7% improvement on GSM8K for Qwen3-0.6B and +12.1% improvement on TruthfulQA for GLM4-32B. CSS better preserves activation energy compared to hard sparsification methods.
Conclusion: ROAST provides a more robust and effective approach to activation steering by using on-distribution supervision and proper normalization techniques, addressing limitations of existing methods that rely on off-distribution data and discrete interventions.
Abstract: Activation steering provides parameter-efficient control over large language models (LLMs) at inference time, but many methods rely on off-distribution supervision and discrete masking, leading to brittle interventions. We propose ROAST (Rollout-based On-distribution Activation Steering Technique), which estimates steering directions from the model’s own on-distribution rollouts via ROC and avoids hard sparsification via Continuous Soft Scaling (CSS) and Grouped Mean Normalization. Our empirical analysis reveals that while activation magnitude correlates moderately with directional consistency, the variance in magnitude is significant and often disproportionate to semantic quality. This suggests that high-magnitude activations risk dominating the global steering direction if not properly normalized. To address this, ROAST employs grouped normalization to balance contributions across samples, ensuring a more robust estimation of the consensus steering direction. Across models (0.6B to 32B), ROAST consistently improves performance on diverse tasks (e.g., +9.7% on GSM8K for Qwen3-0.6B and +12.1% on TruthfulQA for GLM4-32B), and analyses show that CSS better preserves activation energy.
[622] A Penalty Approach for Differentiation Through Black-Box Quadratic Programming Solvers
Yuxuan Linghu, Zhiyuan Liu, Qi Deng
Main category: cs.LG
TL;DR: dXPP is a penalty-based differentiation framework for quadratic programs that decouples QP solving from differentiation, enabling solver-agnostic forward passes and efficient backward passes through a smooth approximate penalty problem.
Details
Motivation: Existing approaches for differentiating through QP solutions use KKT systems, but their computational cost and numerical robustness degrade at scale. There's a need for more efficient and robust differentiation methods for large-scale optimization problems.Method: dXPP uses a penalty-based framework with two steps: 1) Forward pass uses any black-box QP solver, 2) Backward pass maps the solution to a smooth approximate penalty problem and implicitly differentiates through it by solving a smaller linear system in primal variables.
Result: dXPP is competitive with KKT-based differentiation methods and achieves substantial speedups on large-scale problems, including randomly generated QPs, large-scale sparse projection problems, and multi-period portfolio optimization.
Conclusion: dXPP provides an efficient, robust, and solver-agnostic approach for differentiating through QP solutions that bypasses KKT system difficulties and scales well to large problems.
Abstract: Differentiating through the solution of a quadratic program (QP) is a central problem in differentiable optimization. Most existing approaches differentiate through the Karush–Kuhn–Tucker (KKT) system, but their computational cost and numerical robustness can degrade at scale. To address these limitations, we propose dXPP, a penalty-based differentiation framework that decouples QP solving from differentiation. In the solving step (forward pass), dXPP is solver-agnostic and can leverage any black-box QP solver. In the differentiation step (backward pass), we map the solution to a smooth approximate penalty problem and implicitly differentiate through it, requiring only the solution of a much smaller linear system in the primal variables. This approach bypasses the difficulties inherent in explicit KKT differentiation and significantly improves computational efficiency and robustness. We evaluate dXPP on various tasks, including randomly generated QPs, large-scale sparse projection problems, and a real-world multi-period portfolio optimization task. Empirical results demonstrate that dXPP is competitive with KKT-based differentiation methods and achieves substantial speedups on large-scale problems.
[623] Synergistic Intra- and Cross-Layer Regularization Losses for MoE Expert Specialization
Rizhen Hu, Yuan Cao, Boao Kong, Mou Sun, Kun Yuan
Main category: cs.LG
TL;DR: Plug-and-play regularization losses for MoE models that enhance expert specialization and routing efficiency without architectural changes, improving performance and inference speed.
Details
Motivation: Sparse Mixture-of-Experts models suffer from expert overlap and routing ambiguity, leading to underutilized capacity. Existing architectural solutions require substantial modifications and rely only on intra-layer signals.Method: Two regularization losses: 1) intra-layer specialization loss penalizes cosine similarity between experts’ activations on identical tokens, encouraging complementary specialization; 2) cross-layer coupling loss maximizes joint Top-k routing probabilities across adjacent layers to establish coherent expert pathways.
Result: Experiments show consistent task gains, higher expert specialization, lower-entropy routing, and faster inference via more stable expert pathways across pre-training, fine-tuning, and zero-shot benchmarks.
Conclusion: The proposed regularization losses effectively enhance MoE specialization and routing efficiency without architectural modifications, offering a plug-and-play solution compatible with various MoE architectures.
Abstract: Sparse Mixture-of-Experts (MoE) models scale Transformers efficiently but suffer from expert overlap – redundant representations across experts and routing ambiguity, resulting in severely underutilized model capacity. While architectural solutions like DeepSeekMoE promote specialization, they require substantial structural modifications and rely solely on intra-layer signals. In this paper, we propose two plug-and-play regularization losses that enhance MoE specialization and routing efficiency without modifying router or model architectures. First, an intra-layer specialization loss penalizes cosine similarity between experts’ SwiGLU activations on identical tokens, encouraging experts to specialize in complementary knowledge. Second, a cross-layer coupling loss maximizes joint Top-$k$ routing probabilities across adjacent layers, establishing coherent expert pathways through network depth while reinforcing intra-layer expert specialization. Both losses are orthogonal to the standard load-balancing loss and compatible with both the shared-expert architecture in DeepSeekMoE and vanilla top-$k$ MoE architectures. We implement both losses as a drop-in Megatron-LM module. Extensive experiments across pre-training, fine-tuning, and zero-shot benchmarks demonstrate consistent task gains, higher expert specialization, and lower-entropy routing; together, these improvements translate into faster inference via more stable expert pathways.
[624] When Benchmarks Lie: Evaluating Malicious Prompt Classifiers Under True Distribution Shift
Max Fomin
Main category: cs.LG
TL;DR: Comprehensive analysis of prompt injection/jailbreak detection methods showing current evaluation practices overestimate performance and fail to generalize to out-of-distribution attacks, with LLM-based guardrails performing poorly on indirect agent attacks.
Details
Motivation: LLM-based agents increasingly process untrusted data from various sources, making robust prompt injection and jailbreak attack detection critical for safe deployment. Current evaluation practices and production systems have fundamental limitations that need to be addressed.Method: Proposed Leave-One-Dataset-Out (LODO) evaluation protocol using 18 diverse datasets spanning harmful requests, jailbreaks, indirect prompt injections, and extraction attacks. Analyzed Sparse Auto-Encoder (SAE) feature coefficients across LODO folds to understand generalization failures, and systematically compared production guardrails (PromptGuard 2, LlamaGuard) and LLM-as-judge approaches.
Result: Standard train-test splits from same dataset sources severely overestimate performance (8.4 percentage point AUC inflation). 28% of top SAE features are dataset-dependent shortcuts. Production guardrails and LLM-as-judge approaches fail on indirect attacks targeting agents (7-37% detection). PromptGuard 2 and LlamaGuard cannot evaluate agentic tool injection due to architectural limitations.
Conclusion: Current prompt attack detection methods lack true out-of-distribution generalization. LODO evaluation reveals dataset-dependent shortcuts and heterogeneous failure modes. LODO-stable SAE features provide more reliable explanations. The paper establishes LODO as the appropriate protocol for prompt attack detection research.
Abstract: Detecting prompt injection and jailbreak attacks is critical for deploying LLM-based agents safely. As agents increasingly process untrusted data from emails, documents, tool outputs, and external APIs, robust attack detection becomes essential. Yet current evaluation practices and production systems have fundamental limitations. We present a comprehensive analysis using a diverse benchmark of 18 datasets spanning harmful requests, jailbreaks, indirect prompt injections, and extraction attacks. We propose Leave-One-Dataset-Out (LODO) evaluation to measure true out-of-distribution generalization, revealing that the standard practice of train-test splits from the same dataset sources severely overestimates performance: aggregate metrics show an 8.4 percentage point AUC inflation, but per-dataset gaps range from 1% to 25% accuracy-exposing heterogeneous failure modes. To understand why classifiers fail to generalize, we analyze Sparse Auto-Encoder (SAE) feature coefficients across LODO folds, finding that 28% of top features are dataset-dependent shortcuts whose class signal depends on specific dataset compositions rather than semantic content. We systematically compare production guardrails (PromptGuard 2, LlamaGuard) and LLM-as-judge approaches on our benchmark, finding all three fail on indirect attacks targeting agents (7-37% detection) and that PromptGuard 2 and LlamaGuard cannot evaluate agentic tool injection due to architectural limitations. Finally, we show that LODO-stable SAE features provide more reliable explanations for classifier decisions by filtering dataset artifacts. We release our evaluation framework at https://github.com/maxf-zn/prompt-mining to establish LODO as the appropriate protocol for prompt attack detection research.
[625] Deep Dense Exploration for LLM Reinforcement Learning via Pivot-Driven Resampling
Yiran Guo, Zhongjian Qiao, Yingqi Xie, Jie Liu, Dan Ye, Ruiqing Zhang, Shuang Qiu, Lijie Xu
Main category: cs.LG
TL;DR: DDE (Deep Dense Exploration) improves RL for LLMs by focusing exploration on deep, recoverable states in unsuccessful trajectories, addressing sampling inefficiencies in existing methods.
Details
Motivation: Existing RL methods for LLMs have exploration inefficiencies: GRPO samples only from root states, saturating high-probability trajectories while leaving deep error-prone states under-explored; tree-based methods blindly disperse sampling budgets across trivial/unrecoverable states, causing dilution that fails to uncover rare correct suffixes and destabilizes baselines.Method: Proposes Deep Dense Exploration (DDE) focusing on “pivots” - deep, recoverable states within unsuccessful trajectories. Instantiates as DEEP-GRPO with: 1) lightweight data-driven utility function balancing recoverability and depth bias to identify pivot states; 2) local dense resampling at each pivot to increase probability of discovering correct subsequent trajectories; 3) dual-stream optimization objective decoupling global policy learning from local corrective updates.
Result: Experiments on mathematical reasoning benchmarks show the method consistently outperforms GRPO, tree-based methods, and other strong baselines.
Conclusion: DDE effectively addresses exploration challenges in RL for LLMs by strategically focusing sampling resources on promising deep states, leading to better performance on reasoning tasks.
Abstract: Effective exploration is a key challenge in reinforcement learning for large language models: discovering high-quality trajectories within a limited sampling budget from the vast natural language sequence space. Existing methods face notable limitations: GRPO samples exclusively from the root, saturating high-probability trajectories while leaving deep, error-prone states under-explored. Tree-based methods blindly disperse budgets across trivial or unrecoverable states, causing sampling dilution that fails to uncover rare correct suffixes and destabilizes local baselines. To address this, we propose Deep Dense Exploration (DDE), a strategy that focuses exploration on $\textit{pivots}$-deep, recoverable states within unsuccessful trajectories. We instantiate DDE with DEEP-GRPO, which introduces three key innovations: (1) a lightweight data-driven utility function that automatically balances recoverability and depth bias to identify pivot states; (2) local dense resampling at each pivot to increase the probability of discovering correct subsequent trajectories; and (3) a dual-stream optimization objective that decouples global policy learning from local corrective updates. Experiments on mathematical reasoning benchmarks demonstrate that our method consistently outperforms GRPO, tree-based methods, and other strong baselines.
[626] TS-Haystack: A Multi-Scale Retrieval Benchmark for Time Series Language Models
Nicolas Zumarraga, Thomas Kaar, Ning Wang, Maxwell A. Xu, Max Rosenblattl, Markus Kreft, Kevin O’Sullivan, Paul Schmiedmayer, Patrick Langer, Robert Jakob
Main category: cs.LG
TL;DR: TS-Haystack benchmark reveals that existing Time Series Language Models struggle with long-context temporal retrieval despite good classification performance, highlighting the need for architectures that preserve temporal fidelity while managing computational complexity.
Details
Motivation: Existing Time Series Language Models (TSLMs) are limited to short sequences, while real-world time-series sensor streams can span millions of datapoints. There's a mismatch between current benchmarks and real-world requirements for precise temporal localization under computational constraints.Method: Introduces TS-Haystack benchmark with controlled needle insertion (embedding short activity bouts into longer accelerometer recordings) across 10 task types in 4 categories. Evaluates multiple models and encoding strategies across context lengths from seconds to 2 hours.
Result: Learned latent compression preserves/improves classification accuracy up to 176× compression, but retrieval performance degrades with context length due to loss of temporally localized information. Shows consistent divergence between classification and retrieval behavior.
Conclusion: Highlights importance of architectural designs that decouple sequence length from computational complexity while preserving temporal fidelity. Current TSLM time series encoders overlook temporal granularity as context length increases.
Abstract: Time Series Language Models (TSLMs) are emerging as unified models for reasoning over continuous signals in natural language. However, long-context retrieval remains a major limitation: existing models are typically trained and evaluated on short sequences, while real-world time-series sensor streams can span millions of datapoints. This mismatch requires precise temporal localization under strict computational constraints, a regime that is not captured by current benchmarks. We introduce TS-Haystack, a long-context temporal retrieval benchmark comprising ten task types across four categories: direct retrieval, temporal reasoning, multi-step reasoning and contextual anomaly. The benchmark uses controlled needle insertion by embedding short activity bouts into longer longitudinal accelerometer recordings, enabling systematic evaluation across context lengths ranging from seconds to 2 hours per sample. We hypothesize that existing TSLM time series encoders overlook temporal granularity as context length increases, creating a task-dependent effect: compression aids classification but impairs retrieval of localized events. Across multiple model and encoding strategies, we observe a consistent divergence between classification and retrieval behavior. Learned latent compression preserves or improves classification accuracy at compression ratios up to 176$\times$, but retrieval performance degrades with context length, incurring in the loss of temporally localized information. These results highlight the importance of architectural designs that decouple sequence length from computational complexity while preserving temporal fidelity.
[627] Fast Catch-Up, Late Switching: Optimal Batch Size Scheduling via Functional Scaling Laws
Jinbo Wang, Binghui Li, Zhanpeng Zhou, Mingze Wang, Yuxuan Sun, Jiaqi Zhang, Xunliang Cai, Lei Wu
Main category: cs.LG
TL;DR: Theoretical analysis of batch size scheduling using functional scaling laws reveals optimal schedules depend on task difficulty, with hard tasks benefiting from late switching to large batches due to fast catch-up effects.
Details
Motivation: Batch size scheduling is crucial for large-scale deep learning but lacks theoretical understanding. The paper aims to provide principled analysis of optimal batch size scheduling using functional scaling laws.Method: Uses functional scaling law (FSL) framework to analyze batch size scheduling, characterizing optimal schedules under fixed data budget. Identifies fast catch-up effect where loss rapidly aligns with constant large-batch trajectory after switching from small to large batches.
Result: Optimal schedules depend on task difficulty: easy tasks benefit from continuously increasing batch sizes, while hard tasks require small batches for most training with late switching to large batches. Extensive LLM pretraining experiments (up to 1.1B parameters, 1T tokens) validate theoretical predictions, showing late-switch schedules outperform constant-batch and early-switch baselines.
Conclusion: Functional scaling laws provide theoretical foundation for batch size scheduling, revealing task-dependent optimal strategies. Late switching to large batches for hard tasks reduces data consumption without sacrificing performance, validated through large-scale LLM experiments.
Abstract: Batch size scheduling (BSS) plays a critical role in large-scale deep learning training, influencing both optimization dynamics and computational efficiency. Yet, its theoretical foundations remain poorly understood. In this work, we show that the functional scaling law (FSL) framework introduced in Li et al. (2025a) provides a principled lens for analyzing BSS. Specifically, we characterize the optimal BSS under a fixed data budget and show that its structure depends sharply on task difficulty. For easy tasks, optimal schedules keep increasing batch size throughout. In contrast, for hard tasks, the optimal schedule maintains small batch sizes for most of training and switches to large batches only in a late stage. To explain the emergence of late switching, we uncover a dynamical mechanism – the fast catch-up effect – which also manifests in large language model (LLM) pretraining. After switching from small to large batches, the loss rapidly aligns with the constant large-batch trajectory. Using FSL, we show that this effect stems from rapid forgetting of accumulated gradient noise, with the catch-up speed determined by task difficulty. Crucially, this effect implies that large batches can be safely deferred to late training without sacrificing performance, while substantially reducing data consumption. Finally, extensive LLM pretraining experiments – covering both Dense and MoE architectures with up to 1.1B parameters and 1T tokens – validate our theoretical predictions. Across all settings, late-switch schedules consistently outperform constant-batch and early-switch baselines.
[628] MAGE: All-[MASK] Block Already Knows Where to Look in Diffusion LLM
Omin Kwon, Yeonjae Kim, Doyeon Kim, Minseo Kim, Yeonhong Park, Jae W. Lee
Main category: cs.LG
TL;DR: MAGE introduces a training-free sparse attention method for block diffusion LLMs that leverages attention patterns from the first denoising step to predict important KV entries, achieving near-lossless accuracy with significant speedups in long-context settings.
Details
Motivation: Block diffusion LLMs show promise for language generation but suffer from memory bottlenecks due to KV caching in long-context settings. Existing sparse attention methods designed for autoregressive LLMs perform poorly when adapted to block diffusion models.Method: MAGE identifies that attention at the first All-[MASK] denoising step reliably predicts important KV entries and budget requirements. It performs a single exact attention pass per block and reuses it for training-free sparse denoising. A lightweight fine-tuning strategy further strengthens [MASK]-guided patterns with minimal training cost.
Result: Across long-context benchmarks including LongBench and Needle-in-a-Haystack, MAGE achieves near-lossless accuracy with a fraction of the KV budget while delivering up to 3-4x end-to-end speedup, consistently outperforming AR-oriented sparse attention baselines.
Conclusion: MAGE demonstrates that block diffusion LLMs have unique opportunities for efficient sparse attention that differ from autoregressive models, enabling significant performance improvements in long-context settings with minimal training overhead.
Abstract: Block diffusion LLMs are emerging as a promising next paradigm for language generation, but their use of KV caching makes memory access a dominant bottleneck in long-context settings. While dynamic sparse attention has been actively explored, existing methods designed for autoregressive LLMs rely on approximate importance estimation and perform poorly when adapted to block diffusion. This work identifies a key opportunity unique to block diffusion: attention at the first All-[MASK] denoising step reliably predicts important KV entries and budget requirements, enabling MAGE to perform a single exact attention pass per block and reuse it for training-free sparse denoising. Across long-context benchmarks including LongBench and Needle-in-a-Haystack, MAGE achieves near-lossless accuracy with a fraction of the KV budget while delivering up to 3-4x end-to-end speedup, consistently outperforming AR-oriented sparse attention baselines. A lightweight fine-tuning strategy further strengthens [MASK]-guided patterns with minimal cost, requiring only a few hours of training on a single NVIDIA H100 GPU for both 1.5B and 7B models.
[629] Robust multi-task boosting using clustering and local ensembling
Seyedsaman Emami, Daniel Hernández-Lobato, Gonzalo Martínez-Muñoz
Main category: cs.LG
TL;DR: RMB-CLE is a robust multi-task learning framework that uses error-based task clustering and local ensembling to prevent negative transfer by adaptively grouping related tasks and enabling robust knowledge sharing within clusters.
Details
Motivation: Conventional multi-task learning methods often suffer from negative transfer when unrelated or noisy tasks are forced to share representations, which degrades performance. There's a need for principled approaches that can automatically identify related tasks and prevent harmful information sharing.Method: RMB-CLE integrates error-based task clustering with local ensembling. It derives inter-task similarity directly from cross-task errors using a risk decomposition into functional mismatch and irreducible noise. Tasks are grouped adaptively via agglomerative clustering, and within each cluster, a local ensemble enables robust knowledge sharing while preserving task-specific patterns.
Result: Experiments show that RMB-CLE recovers ground-truth clusters in synthetic data and consistently outperforms multi-task, single-task, and pooling-based ensemble methods across diverse real-world and synthetic benchmarks.
Conclusion: RMB-CLE establishes a new basis for robust multi-task learning by providing a theoretically grounded mechanism to prevent negative transfer, making it a general and scalable framework that goes beyond simple combination of clustering and boosting techniques.
Abstract: Multi-Task Learning (MTL) aims to boost predictive performance by sharing information across related tasks, yet conventional methods often suffer from negative transfer when unrelated or noisy tasks are forced to share representations. We propose Robust Multi-Task Boosting using Clustering and Local Ensembling (RMB-CLE), a principled MTL framework that integrates error-based task clustering with local ensembling. Unlike prior work that assumes fixed clusters or hand-crafted similarity metrics, RMB-CLE derives inter-task similarity directly from cross-task errors, which admit a risk decomposition into functional mismatch and irreducible noise, providing a theoretically grounded mechanism to prevent negative transfer. Tasks are grouped adaptively via agglomerative clustering, and within each cluster, a local ensemble enables robust knowledge sharing while preserving task-specific patterns. Experiments show that RMB-CLE recovers ground-truth clusters in synthetic data and consistently outperforms multi-task, single-task, and pooling-based ensemble methods across diverse real-world and synthetic benchmarks. These results demonstrate that RMB-CLE is not merely a combination of clustering and boosting but a general and scalable framework that establishes a new basis for robust multi-task learning.
[630] Evaluating LLMs in Finance Requires Explicit Bias Consideration
Yaxuan Kong, Hoyoung Lee, Yoontae Hwang, Alejandro Lopez-Lira, Bradford Levy, Dhagash Mehta, Qingsong Wen, Chanyeol Choi, Yongjae Lee, Stefan Zohren
Main category: cs.LG
TL;DR: Paper identifies five key biases in financial LLM applications (look-ahead, survivorship, narrative, objective, cost bias) and proposes a Structural Validity Framework with evaluation checklist to address them.
Details
Motivation: LLMs are increasingly used in finance, but evaluation practices haven't kept up, allowing finance-specific biases to inflate performance, contaminate backtests, and make results useless for deployment claims.Method: Identified five recurring biases through analysis, reviewed 164 papers (2023-2025), and developed a Structural Validity Framework with evaluation checklist for bias diagnosis and system design.
Result: Found that no single bias is discussed in more than 28% of reviewed studies, demonstrating widespread neglect of bias issues in financial LLM research.
Conclusion: Bias in financial LLM systems requires explicit attention, and structural validity should be enforced before using results to support deployment claims.
Abstract: Large Language Models (LLMs) are increasingly integrated into financial workflows, but evaluation practice has not kept up. Finance-specific biases can inflate performance, contaminate backtests, and make reported results useless for any deployment claim. We identify five recurring biases in financial LLM applications. They include look-ahead bias, survivorship bias, narrative bias, objective bias, and cost bias. These biases break financial tasks in distinct ways and they often compound to create an illusion of validity. We reviewed 164 papers from 2023 to 2025 and found that no single bias is discussed in more than 28 percent of studies. This position paper argues that bias in financial LLM systems requires explicit attention and that structural validity should be enforced before any result is used to support a deployment claim. We propose a Structural Validity Framework and an evaluation checklist with minimal requirements for bias diagnosis and future system design. The material is available at https://github.com/Eleanorkong/Awesome-Financial-LLM-Bias-Mitigation.
[631] Multi-Agent Debate: A Unified Agentic Framework for Tabular Anomaly Detection
Pinqiao Wang, Sheng Li
Main category: cs.LG
TL;DR: MAD: Multi-Agent Debating framework for tabular anomaly detection that treats model disagreement as signal, using ML-based detectors with LLM critics and mathematical coordination to produce auditable anomaly scores.
Details
Motivation: Tabular anomaly detection typically uses single detectors or static ensembles, but heterogeneous models (tree ensembles, deep tabular networks, foundation models) often disagree under distribution shift, missingness, and rare-anomaly regimes. This disagreement should be treated as valuable signal rather than noise.Method: Multi-agent debating framework where each agent is an ML-based detector producing anomaly scores, confidence, and evidence, augmented by LLM-based critics. A coordinator converts messages into bounded per-agent losses and updates influence via exponentiated-gradient rule, producing final debated anomaly score and auditable debate trace.
Result: Experiments on diverse tabular anomaly benchmarks show improved robustness over baselines and clearer traces of model disagreement. The framework can recover existing approaches (mixture-of-experts, learning-with-expert-advice) by restricting message space and synthesis operator.
Conclusion: MAD provides a mathematically grounded coordination layer for resolving model disagreement in tabular anomaly detection, offering both improved performance and interpretability through auditable debate traces.
Abstract: Tabular anomaly detection is often handled by single detectors or static ensembles, even though strong performance on tabular data typically comes from heterogeneous model families (e.g., tree ensembles, deep tabular networks, and tabular foundation models) that frequently disagree under distribution shift, missingness, and rare-anomaly regimes. We propose MAD, a Multi-Agent Debating framework that treats this disagreement as a first-class signal and resolves it through a mathematically grounded coordination layer. Each agent is a machine learning (ML)-based detector that produces a normalized anomaly score, confidence, and structured evidence, augmented by a large language model (LLM)-based critic. A coordinator converts these messages into bounded per-agent losses and updates agent influence via an exponentiated-gradient rule, yielding both a final debated anomaly score and an auditable debate trace. MAD is a unified agentic framework that can recover existing approaches, such as mixture-of-experts gating and learning-with-expert-advice aggregation, by restricting the message space and synthesis operator. We establish regret guarantees for the synthesized losses and show how conformal calibration can wrap the debated score to control false positives under exchangeability. Experiments on diverse tabular anomaly benchmarks show improved robustness over baselines and clearer traces of model disagreement
[632] Cross-household Transfer Learning Approach with LSTM-based Demand Forecasting
Manal Rahal, Bestoun S. Ahmed, Roger Renström, Robert Stener
Main category: cs.LG
TL;DR: DELTAiF is a transfer learning framework for scalable prediction of household hot water consumption using heat pumps, reducing training time by 67% while maintaining high accuracy.
Details
Motivation: With increasing residential heat pump installations, optimizing hot water production faces scalability challenges. Training separate ML models for each household is computationally expensive, especially in cloud-connected deployments. Accurate forecasting of hot water demand is needed to ensure comfort and reduce energy waste.Method: DELTAiF uses transfer learning to leverage knowledge from a representative household and fine-tune it across others, eliminating the need for separate ML models for each installation. It focuses on predicting large hot water usage events like showers to enable adaptive hot water production.
Result: The approach reduces overall training time by approximately 67% while maintaining high predictive accuracy (0.874-0.991) and low mean absolute percentage error (0.001-0.017). Transfer learning is particularly effective when the source household exhibits regular consumption patterns.
Conclusion: DELTAiF enables scalable hot water demand forecasting for heat pump systems through transfer learning, addressing computational challenges while maintaining accuracy for household-level optimization.
Abstract: With the rapid increase in residential heat pump (HP) installations, optimizing hot water production in households is essential, yet it faces major technical and scalability challenges. Adapting production to actual household needs requires accurate forecasting of hot water demand to ensure comfort and, most importantly, to reduce energy waste. However, the conventional approach of training separate machine learning models for each household becomes computationally expensive at scale, particularly in cloud-connected HP deployments. This study introduces DELTAiF, a transfer learning (TL) based framework that provides scalable and accurate prediction of household hot water consumption. By predicting large hot water usage events, such as showers, DELTAiF enables adaptive yet scalable hot water production at the household level. DELTAiF leverages learned knowledge from a representative household and fine-tunes it across others, eliminating the need to train separate machine learning models for each HP installation. This approach reduces overall training time by approximately 67 percent while maintaining high predictive accuracy values between 0.874 and 0.991, and mean absolute percentage error values between 0.001 and 0.017. The results show that TL is particularly effective when the source household exhibits regular consumption patterns, enabling hot water demand forecasting at scale.
[633] Radial-VCReg: More Informative Representation Learning Through Radial Gaussianization
Yilun Kuang, Yash Dagade, Deep Chakraborty, Erik Learned-Miller, Randall Balestriero, Tim G. J. Rudner, Yann LeCun
Main category: cs.LG
TL;DR: Radial-VCReg enhances self-supervised learning by adding radial Gaussianization to VCReg, aligning feature norms with Chi distribution to better approximate maximum entropy representations in high dimensions.
Details
Motivation: Existing self-supervised learning methods like VCReg regularize first and second-order feature statistics but cannot fully achieve maximum entropy due to the curse of dimensionality. There's a need for better methods that can transform broader classes of distributions toward normality.Method: Radial-VCReg augments VCReg with a radial Gaussianization loss that aligns feature norms with the Chi distribution, which is a defining property of high-dimensional Gaussians. This approach addresses higher-order dependencies beyond first and second-order statistics.
Result: Theoretical proof shows Radial-VCReg transforms a broader class of distributions toward normality compared to VCReg. Empirical results on synthetic and real-world datasets demonstrate consistent performance improvements by reducing higher-order dependencies and promoting more diverse, informative representations.
Conclusion: Radial-VCReg provides an effective enhancement to existing self-supervised learning methods by better approximating maximum entropy representations through radial Gaussianization, leading to improved representation learning in high-dimensional spaces.
Abstract: Self-supervised learning aims to learn maximally informative representations, but explicit information maximization is hindered by the curse of dimensionality. Existing methods like VCReg address this by regularizing first and second-order feature statistics, which cannot fully achieve maximum entropy. We propose Radial-VCReg, which augments VCReg with a radial Gaussianization loss that aligns feature norms with the Chi distribution-a defining property of high-dimensional Gaussians. We prove that Radial-VCReg transforms a broader class of distributions towards normality compared to VCReg and show on synthetic and real-world datasets that it consistently improves performance by reducing higher-order dependencies and promoting more diverse and informative representations.
[634] Integrating Unstructured Text into Causal Inference: Empirical Evidence from Real Data
Boning Zhou, Ziyu Wang, Han Hong, Haoqi Hu
Main category: cs.LG
TL;DR: Transformer-based framework for causal inference using unstructured text instead of structured data
Details
Motivation: Traditional causal inference relies on structured data, but this is often incomplete or unavailable in real-world scenarios. Unstructured text data is abundant and could enable causal inference when structured data is scarce.Method: Developed a framework using transformer-based language models to perform causal inference from unstructured text. Compared causal estimates from text against those from structured data at population, group, and individual levels.
Result: Found consistent results between causal estimates derived from unstructured text and those from structured data, validating the potential of text-based causal inference.
Conclusion: The framework extends causal inference applicability to scenarios with only textual data, enabling data-driven business decision-making when structured data is unavailable.
Abstract: Causal inference, a critical tool for informing business decisions, traditionally relies heavily on structured data. However, in many real-world scenarios, such data can be incomplete or unavailable. This paper presents a framework that leverages transformer-based language models to perform causal inference using unstructured text. We demonstrate the effectiveness of our framework by comparing causal estimates derived from unstructured text against those obtained from structured data across population, group, and individual levels. Our findings show consistent results between the two approaches, validating the potential of unstructured text in causal inference tasks. Our approach extends the applicability of causal inference methods to scenarios where only textual data is available, enabling data-driven business decision-making when structured tabular data is scarce.
[635] Reverse N-Wise Output-Oriented Testing for AI/ML and Quantum Computing Systems
Lamine Rihani
Main category: cs.LG
TL;DR: Reverse n-wise output testing: A paradigm that constructs covering arrays over output equivalence classes (ML confidence buckets, fairness partitions, quantum measurement outcomes) and uses metaheuristic optimization to synthesize inputs that elicit targeted behavioral signatures from opaque AI/ML and quantum systems.
Details
Motivation: AI/ML and quantum computing systems present unprecedented testing challenges: high-dimensional continuous input spaces, probabilistic outputs, correctness defined over observable behaviors, and critical quality dimensions (trustworthiness, fairness, calibration, robustness) that manifest through complex multi-way interactions among output properties rather than deterministic input-output mappings.Method: Reverse n-wise output testing constructs covering arrays directly over domain-specific output equivalence classes (ML confidence calibration buckets, decision boundary regions, fairness partitions, embedding clusters, quantum measurement outcome distributions, error syndrome patterns). It then solves the black-box inverse mapping problem via gradient-free metaheuristic optimization to synthesize input feature configurations or quantum circuit parameters capable of eliciting targeted behavioral signatures.
Result: The framework delivers synergistic benefits: explicit customer-centric prediction/measurement coverage guarantees, substantial improvements in fault detection rates for ML calibration/boundary failures and quantum error syndromes, enhanced test suite efficiency, and structured MLOps/quantum validation pipelines with automated partition discovery from uncertainty analysis and coverage drift monitoring.
Conclusion: Reverse n-wise output testing provides a mathematically principled paradigm inversion for testing opaque AI/ML and quantum systems, addressing their unique challenges through output-centric coverage and automated input synthesis.
Abstract: Artificial intelligence/machine learning (AI/ML) systems and emerging quantum computing software present unprecedented testing challenges characterized by high-dimensional/continuous input spaces, probabilistic/non-deterministic output distributions, behavioral correctness defined exclusively over observable prediction behaviors and measurement outcomes, and critical quality dimensions, trustworthiness, fairness, calibration, robustness, error syndrome patterns, that manifest through complex multi-way interactions among semantically meaningful output properties rather than deterministic input-output mappings. This paper introduces reverse n-wise output testing, a mathematically principled paradigm inversion that constructs covering arrays directly over domain-specific output equivalence classes, ML confidence calibration buckets, decision boundary regions, fairness partitions, embedding clusters, ranking stability bands, quantum measurement outcome distributions (0-dominant, 1-dominant, superposition collapse), error syndrome patterns (bit-flip, phase-flip, correlated errors), then solves the computationally challenging black-box inverse mapping problem via gradient-free metaheuristic optimization to synthesize input feature configurations or quantum circuit parameters capable of eliciting targeted behavioral signatures from opaque models. The framework delivers synergistic benefits across both domains: explicit customer-centric prediction/measurement coverage guarantees, substantial improvements in fault detection rates for ML calibration/boundary failures and quantum error syndromes, enhanced test suite efficiency, and structured MLOps/quantum validation pipelines with automated partition discovery from uncertainty analysis and coverage drift monitoring.
[636] Whom to Query for What: Adaptive Group Elicitation via Multi-Turn LLM Interactions
Ruomeng Ding, Tianwei Gao, Thomas P. Zollo, Eitan Bachmat, Richard Zemel, Zhun Deng
Main category: cs.LG
TL;DR: A framework for adaptive group elicitation that combines LLM-based question scoring with graph neural networks to optimize both question selection and respondent sampling under constrained budgets.
Details
Motivation: Existing elicitation methods don't adaptively select both questions and respondents, nor leverage population structure when responses are incomplete. There's a need for methods that can efficiently gather group-level information under real-world constraints like limited budgets and missing data.Method: Proposes a multi-round adaptive group elicitation framework with: (1) LLM-based expected information gain objective for scoring candidate questions, and (2) heterogeneous graph neural network propagation that aggregates observed responses and participant attributes to impute missing responses and guide respondent selection.
Result: The method consistently improves population-level response prediction across three real-world opinion datasets, achieving >12% relative gain on CES at a 10% respondent budget.
Conclusion: The proposed closed-loop procedure effectively queries a small, informative subset of individuals while inferring population-level responses via structured similarity, demonstrating practical value for efficient group elicitation under constraints.
Abstract: Eliciting information to reduce uncertainty about latent group-level properties from surveys and other collective assessments requires allocating limited questioning effort under real costs and missing data. Although large language models enable adaptive, multi-turn interactions in natural language, most existing elicitation methods optimize what to ask with a fixed respondent pool, and do not adapt respondent selection or leverage population structure when responses are partial or incomplete. To address this gap, we study adaptive group elicitation, a multi-round setting where an agent adaptively selects both questions and respondents under explicit query and participation budgets. We propose a theoretically grounded framework that combines (i) an LLM-based expected information gain objective for scoring candidate questions with (ii) heterogeneous graph neural network propagation that aggregates observed responses and participant attributes to impute missing responses and guide per-round respondent selection. This closed-loop procedure queries a small, informative subset of individuals while inferring population-level responses via structured similarity. Across three real-world opinion datasets, our method consistently improves population-level response prediction under constrained budgets, including a >12% relative gain on CES at a 10% respondent budget.
[637] KernelBlaster: Continual Cross-Task CUDA Optimization via Memory-Augmented In-Context Reinforcement Learning
Kris Shengjun Dong, Sahil Modi, Dima Nikiforov, Sana Damani, Edward Lin, Siva Kumar Sastry Hari, Christos Kozyrakis
Main category: cs.LG
TL;DR: KernelBlaster: A Memory-Augmented In-context Reinforcement Learning framework that improves LLM-based GPU coding agents’ ability to optimize CUDA code across multiple GPU generations by accumulating knowledge in a retrievable Persistent CUDA Knowledge Base.
Details
Motivation: Optimizing CUDA code across multiple GPU architectures is challenging due to complex hardware-specific optimization spaces. Traditional compilers have fixed heuristics, fine-tuning LLMs is expensive, and current agentic workflows lack knowledge aggregation from prior exploration, leading to biased sampling and suboptimal solutions.Method: Proposes KernelBlaster, a Memory-Augmented In-context Reinforcement Learning (MAIC-RL) framework with a Persistent CUDA Knowledge Base that accumulates optimization knowledge. Uses a novel profile-guided, textual-gradient-based agentic flow for CUDA generation and optimization to systematically explore high-potential strategies beyond naive rewrites.
Result: Achieves geometric mean speedups of 1.43x, 2.50x, and 1.50x on KernelBench Levels 1, 2, and 3 respectively compared to PyTorch baseline. The framework is released as open-source with test harness, verification components, and reproducible evaluation pipeline.
Conclusion: KernelBlaster enables LLM-based GPU coding agents to learn from experience and make systematically informed decisions on CUDA optimization tasks across multiple GPU generations, overcoming limitations of traditional approaches.
Abstract: Optimizing CUDA code across multiple generations of GPU architectures is challenging, as achieving peak performance requires an extensive exploration of an increasingly complex, hardware-specific optimization space. Traditional compilers are constrained by fixed heuristics, whereas finetuning Large Language Models (LLMs) can be expensive. However, agentic workflows for CUDA code optimization have limited ability to aggregate knowledge from prior exploration, leading to biased sampling and suboptimal solutions. We propose KernelBlaster, a Memory-Augmented In-context Reinforcement Learning (MAIC-RL) framework designed to improve CUDA optimization search capabilities of LLM-based GPU coding agents. KernelBlaster enables agents to learn from experience and make systematically informed decisions on future tasks by accumulating knowledge into a retrievable Persistent CUDA Knowledge Base. We propose a novel profile-guided, textual-gradient-based agentic flow for CUDA generation and optimization to achieve high performance across generations of GPU architectures. KernelBlaster guides LLM agents to systematically explore high-potential optimization strategies beyond naive rewrites. Compared to the PyTorch baseline, our method achieves geometric mean speedups of 1.43x, 2.50x, and 1.50x on KernelBench Levels 1, 2, and 3, respectively. We release KernelBlaster as an open-source agentic framework, accompanied by a test harness, verification components, and a reproducible evaluation pipeline.
[638] Machine Learning as a Tool (MLAT): A Framework for Integrating Statistical ML Models as Callable Tools within LLM Agent Workflows
Edwin Chen, Zulekha Bibi
Main category: cs.LG
TL;DR: MLAT is a design pattern that exposes pre-trained ML models as callable tools within LLM agent workflows, enabling dynamic invocation of quantitative predictions based on conversational context.
Details
Motivation: To enable LLM agents to dynamically invoke quantitative ML predictions when needed, moving beyond static preprocessing pipelines and allowing contextual reasoning about when and how to use ML models.Method: MLAT framework positions ML models as first-class tools alongside other services. Implemented in PitchCraft system with two agents: Research Agent for intelligence gathering and Draft Agent that invokes XGBoost pricing model as a tool call to generate proposals.
Result: Pricing model trained on 70 examples (real + synthetic) achieved R^2 = 0.807 with MAE of 3688 USD. System reduced proposal generation time from hours to under 10 minutes.
Conclusion: MLAT generalizes to domains requiring quantitative estimation combined with contextual reasoning, enabling LLM agents to dynamically leverage ML predictions as needed.
Abstract: We introduce Machine Learning as a Tool (MLAT), a design pattern in which pre-trained statistical machine learning models are exposed as callable tools within large language model (LLM) agent workflows. This allows an orchestrating agent to invoke quantitative predictions when needed and reason about their outputs in context. Unlike conventional pipelines that treat ML inference as a static preprocessing step, MLAT positions the model as a first-class tool alongside web search, database queries, and APIs, enabling the LLM to decide when and how to use it based on conversational context. To validate MLAT, we present PitchCraft, a pilot production system that converts discovery call recordings into professional proposals with ML-predicted pricing. The system uses two agents: a Research Agent that gathers prospect intelligence via parallel tool calls, and a Draft Agent that invokes an XGBoost pricing model as a tool call and generates a complete proposal through structured outputs. The pricing model, trained on 70 examples combining real and human-verified synthetic data, achieves R^2 = 0.807 on held-out data with a mean absolute error of 3688 USD. The system reduces proposal generation time from multiple hours to under 10 minutes. We describe the MLAT framework, structured output architecture, training methodology under extreme data scarcity, and sensitivity analysis demonstrating meaningful learned relationships. MLAT generalizes to domains requiring quantitative estimation combined with contextual reasoning.
[639] In Transformer We Trust? A Perspective on Transformer Architecture Failure Modes
Trishit Mondal, Ameya D. Jagtap
Main category: cs.LG
TL;DR: A comprehensive review examining the trustworthiness of transformer models across various high-stakes applications, evaluating reliability through interpretability, robustness, fairness, and privacy aspects.
Details
Motivation: Transformers are increasingly deployed in safety-critical applications across diverse domains, necessitating rigorous understanding of their trustworthiness for reliable deployment.Method: Systematic review and evaluation of transformer reliability through analysis of interpretability, explainability, robustness against adversarial attacks, fairness, and privacy across multiple domains.
Result: Identifies recurring structural vulnerabilities, domain-specific risks, and open research challenges that limit reliable deployment of transformers in safety-critical applications.
Conclusion: Transformers require deeper trustworthiness evaluation before deployment in high-stakes applications, with identified vulnerabilities and challenges needing further research.
Abstract: Transformer architectures have revolutionized machine learning across a wide range of domains, from natural language processing to scientific computing. However, their growing deployment in high-stakes applications, such as computer vision, natural language processing, healthcare, autonomous systems, and critical areas of scientific computing including climate modeling, materials discovery, drug discovery, nuclear science, and robotics, necessitates a deeper and more rigorous understanding of their trustworthiness. In this work, we critically examine the foundational question: \textitHow trustworthy are transformer models?} We evaluate their reliability through a comprehensive review of interpretability, explainability, robustness against adversarial attacks, fairness, and privacy. We systematically examine the trustworthiness of transformer-based models in safety-critical applications spanning natural language processing, computer vision, and science and engineering domains, including robotics, medicine, earth sciences, materials science, fluid dynamics, nuclear science, and automated theorem proving; highlighting high-impact areas where these architectures are central and analyzing the risks associated with their deployment. By synthesizing insights across these diverse areas, we identify recurring structural vulnerabilities, domain-specific risks, and open research challenges that limit the reliable deployment of transformers.
[640] Conformal Signal Temporal Logic for Robust Reinforcement Learning Control: A Case Study
Hani Beirami, M M Manjurul Islam
Main category: cs.LG
TL;DR: Combining formal temporal logic specifications with reinforcement learning for safer aerospace control using conformal prediction-based shielding
Details
Motivation: To enhance safety and robustness of RL control in aerospace applications by integrating formal temporal logic specifications, addressing challenges like model mismatch and environmental uncertaintiesMethod: Train PPO agent for F-16 throttle control, encode safety as STL requirements, implement conformal STL shield using online conformal prediction to filter RL actions, compare with baseline PPO and classical rule-based shield
Result: Conformal shield preserves STL satisfaction while maintaining near baseline performance, provides stronger robustness guarantees than classical shield under stress scenarios
Conclusion: Formal specification monitoring combined with data-driven RL control significantly improves autonomous flight control reliability in challenging environments
Abstract: We investigate how formal temporal logic specifications can enhance the safety and robustness of reinforcement learning (RL) control in aerospace applications. Using the open source AeroBench F-16 simulation benchmark, we train a Proximal Policy Optimization (PPO) agent to regulate engine throttle and track commanded airspeed. The control objective is encoded as a Signal Temporal Logic (STL) requirement to maintain airspeed within a prescribed band during the final seconds of each maneuver. To enforce this specification at run time, we introduce a conformal STL shield that filters the RL agent’s actions using online conformal prediction. We compare three settings: (i) PPO baseline, (ii) PPO with a classical rule-based STL shield, and (iii) PPO with the proposed conformal shield, under both nominal conditions and a severe stress scenario involving aerodynamic model mismatch, actuator rate limits, measurement noise, and mid-episode setpoint jumps. Experiments show that the conformal shield preserves STL satisfaction while maintaining near baseline performance and providing stronger robustness guarantees than the classical shield. These results demonstrate that combining formal specification monitoring with data driven RL control can substantially improve the reliability of autonomous flight control in challenging environments.
[641] Train Less, Learn More: Adaptive Efficient Rollout Optimization for Group-Based Reinforcement Learning
Zhi Zhang, Zhen Han, Costas Mavromatis, Qi Zhu, Yunyi Zhang, Sheng Guan, Dingmin Wang, Xiong Zhou, Shuai Wang, Soji Adeshina, Vassilis Ioannidis, Huzefa Rangwala
Main category: cs.LG
TL;DR: AERO improves RL fine-tuning efficiency for LLMs by using adaptive rollout strategies and selective rejection to avoid zero-gradient dead zones, reducing compute by 48% while maintaining performance.
Details
Motivation: Current RL fine-tuning methods like GRPO waste compute when all rollouts in a group share the same outcome (all correct or all incorrect), resulting in zero gradient signals and inefficient training.Method: AERO enhances GRPO with: 1) adaptive rollout strategy, 2) selective rejection to prune uninformative rollouts, and 3) Bayesian posterior to prevent zero-advantage dead zones.
Result: AERO reduces total training compute by ~48% and wall-clock time per step by ~45% while matching or improving Pass@8 and Avg@8 metrics across three model configurations.
Conclusion: AERO provides a practical, scalable, and compute-efficient strategy for RL-based LLM alignment that maintains performance while significantly reducing computational costs.
Abstract: Reinforcement learning (RL) plays a central role in large language model (LLM) post-training. Among existing approaches, Group Relative Policy Optimization (GRPO) is widely used, especially for RL with verifiable rewards (RLVR) fine-tuning. In GRPO, each query prompts the LLM to generate a group of rollouts with a fixed group size $N$. When all rollouts in a group share the same outcome, either all correct or all incorrect, the group-normalized advantages become zero, yielding no gradient signal and wasting fine-tuning compute. We introduce Adaptive Efficient Rollout Optimization (AERO), an enhancement of GRPO. AERO uses an adaptive rollout strategy, applies selective rejection to strategically prune rollouts, and maintains a Bayesian posterior to prevent zero-advantage dead zones. Across three model configurations (Qwen2.5-Math-1.5B, Qwen2.5-7B, and Qwen2.5-7B-Instruct), AERO improves compute efficiency without sacrificing performance. Under the same total rollout budget, AERO reduces total training compute by about 48% while shortening wall-clock time per step by about 45% on average. Despite the substantial reduction in compute, AERO matches or improves Pass@8 and Avg@8 over GRPO, demonstrating a practical, scalable, and compute-efficient strategy for RL-based LLM alignment.
[642] Zero-Shot Instruction Following in RL via Structured LTL Representations
Mathias Jackermeier, Mattia Giuri, Jacques Cloete, Alessandro Abate
Main category: cs.LG
TL;DR: A novel hierarchical neural architecture with attention mechanism for learning structured task representations in multi-task RL using Linear Temporal Logic specifications, enabling better zero-shot generalization to novel tasks.
Details
Motivation: Existing approaches for instruction following in multi-task RL struggle to effectively capture the rich logical and temporal structure inherent in LTL specifications, limiting their ability to generalize to novel tasks.Method: Conditions policy on sequences of Boolean formulae from task automaton, uses hierarchical neural architecture to encode logical structure, and introduces attention mechanism for reasoning about future subgoals.
Result: Demonstrates strong generalization capabilities and superior performance in complex environments compared to existing approaches.
Conclusion: The proposed structured task representation approach effectively captures LTL specifications’ logical and temporal structure, enabling better zero-shot execution of novel tasks in multi-task RL.
Abstract: We study instruction following in multi-task reinforcement learning, where an agent must zero-shot execute novel tasks not seen during training. In this setting, linear temporal logic (LTL) has recently been adopted as a powerful framework for specifying structured, temporally extended tasks. While existing approaches successfully train generalist policies, they often struggle to effectively capture the rich logical and temporal structure inherent in LTL specifications. In this work, we address these concerns with a novel approach to learn structured task representations that facilitate training and generalisation. Our method conditions the policy on sequences of Boolean formulae constructed from a finite automaton of the task. We propose a hierarchical neural architecture to encode the logical structure of these formulae, and introduce an attention mechanism that enables the policy to reason about future subgoals. Experiments in a variety of complex environments demonstrate the strong generalisation capabilities and superior performance of our approach.
[643] WIMLE: Uncertainty-Aware World Models with IMLE for Sample-Efficient Continuous Control
Mehran Aghabozorgi, Alireza Moazeni, Yanshu Zhang, Ke Li
Main category: cs.LG
TL;DR: WIMLE improves model-based RL by learning stochastic, multi-modal world models with uncertainty estimation and confidence-weighted training to address compounding model error and overconfident predictions.
Details
Motivation: Model-based RL suffers from compounding model error, unimodal world models that average over multi-modal dynamics, and overconfident predictions that bias learning, limiting its practical effectiveness despite promising sample efficiency.Method: Extends Implicit Maximum Likelihood Estimation (IMLE) to model-based RL to learn stochastic, multi-modal world models without iterative sampling, using ensembles and latent sampling for uncertainty estimation, and weighting synthetic transitions by predicted confidence during training.
Result: Achieves superior sample efficiency and competitive/better asymptotic performance across 40 continuous-control tasks in DeepMind Control, MyoSuite, and HumanoidBench. On Humanoid-run, improves sample efficiency by over 50% relative to strongest competitor, and solves 8 of 14 HumanoidBench tasks vs 4 for BRO and 5 for SimbaV2.
Conclusion: IMLE-based multi-modality and uncertainty-aware weighting provide stable model-based RL, addressing key limitations of existing approaches and enabling practical sample-efficient reinforcement learning.
Abstract: Model-based reinforcement learning promises strong sample efficiency but often underperforms in practice due to compounding model error, unimodal world models that average over multi-modal dynamics, and overconfident predictions that bias learning. We introduce WIMLE, a model-based method that extends Implicit Maximum Likelihood Estimation (IMLE) to the model-based RL framework to learn stochastic, multi-modal world models without iterative sampling and to estimate predictive uncertainty via ensembles and latent sampling. During training, WIMLE weights each synthetic transition by its predicted confidence, preserving useful model rollouts while attenuating bias from uncertain predictions and enabling stable learning. Across $40$ continuous-control tasks spanning DeepMind Control, MyoSuite, and HumanoidBench, WIMLE achieves superior sample efficiency and competitive or better asymptotic performance than strong model-free and model-based baselines. Notably, on the challenging Humanoid-run task, WIMLE improves sample efficiency by over $50$% relative to the strongest competitor, and on HumanoidBench it solves $8$ of $14$ tasks (versus $4$ for BRO and $5$ for SimbaV2). These results highlight the value of IMLE-based multi-modality and uncertainty-aware weighting for stable model-based RL.
[644] A Study on Multi-Class Online Fuzzy Classifiers for Dynamic Environments
Kensuke Ajimoto, Yuma Yamamoto, Yoshifumi Kusunoki, Tomoharu Nakashima
Main category: cs.LG
TL;DR: Proposes multi-class online fuzzy classifier for dynamic environments where data arrives sequentially over time, extending previous two-class methods.
Details
Motivation: Existing online fuzzy classifiers only handle two-class problems, but many real-world applications require multi-class classification in dynamic environments where data arrives sequentially.Method: Extends conventional online fuzzy classifiers to handle multi-class problems using fuzzy if-then rules with predetermined antecedent fuzzy sets and learned consequent real values from sequentially arriving training data.
Result: Evaluated through numerical experiments on synthetic dynamic data and benchmark datasets, demonstrating the effectiveness of multi-class online fuzzy classifiers.
Conclusion: Successfully extends online fuzzy classification to multi-class problems, providing a solution for dynamic environments where data patterns become available sequentially over time.
Abstract: This paper proposes a multi-class online fuzzy classifier for dynamic environments. A fuzzy classifier comprises a set of fuzzy if-then rules where human users determine the antecedent fuzzy sets beforehand. In contrast, the consequent real values are determined by learning from training data. In an online framework, not all training dataset patterns are available beforehand. Instead, only a few patterns are available at a time step, and the subsequent patterns become available at the following time steps. The conventional online fuzzy classifier considered only two-class problems. This paper investigates the extension to the conventional fuzzy classifiers for multi-class problems. We evaluate the performance of the multi-class online fuzzy classifiers through numerical experiments on synthetic dynamic data and also several benchmark datasets.
[645] The geometry of invariant learning: an information-theoretic analysis of data augmentation and generalization
Abdelali Bouyahia, Frédéric LeBlanc, Mario Marchand
Main category: cs.LG
TL;DR: Information-theoretic framework analyzes data augmentation’s effect on generalization and invariance learning, deriving bounds that decompose generalization gap into distributional divergence, stability, and sensitivity terms controlled by augmentation group diameter.
Details
Motivation: While data augmentation is widely used to improve generalization by promoting invariance to label-irrelevant transformations, its theoretical role remains only partially understood. The paper aims to systematically account for the effect of augmentation on generalization and invariance learning through an information-theoretic framework.Method: Proposes an information-theoretic framework building upon mutual information-based bounds, modeling augmented distribution as composition of original data distribution with transformation distribution. Derives generalization bounds under sub-Gaussian assumptions, decomposing expected generalization gap into three terms: distributional divergence, stability term, and sensitivity term. Introduces group diameter concept to connect bounds to augmentation group geometry.
Result: Derives new generalization bound that reliably tracks and predicts true generalization gap behavior in numerical experiments. Shows group diameter provides unified control parameter bounding all three terms, revealing intrinsic trade-off: small diameters preserve data fidelity with limited regularization, while large diameters enhance stability at cost of increased bias and sensitivity.
Conclusion: The information-theoretic framework provides systematic understanding of data augmentation’s role in generalization and invariance learning, with group diameter offering practical guidance for augmentation design by quantifying trade-offs between data fidelity, regularization, and stability.
Abstract: Data augmentation is one of the most widely used techniques to improve generalization in modern machine learning, often justified by its ability to promote invariance to label-irrelevant transformations. However, its theoretical role remains only partially understood. In this work, we propose an information-theoretic framework that systematically accounts for the effect of augmentation on generalization and invariance learning. Our approach builds upon mutual information-based bounds, which relate the generalization gap to the amount of information a learning algorithm retains about its training data. We extend this framework by modeling the augmented distribution as a composition of the original data distribution with a distribution over transformations, which naturally induces an orbit-averaged loss function. Under mild sub-Gaussian assumptions on the loss function and the augmentation process, we derive a new generalization bound that decompose the expected generalization gap into three interpretable terms: (1) a distributional divergence between the original and augmented data, (2) a stability term measuring the algorithm dependence on training data, and (3) a sensitivity term capturing the effect of augmentation variability. To connect our bounds to the geometry of the augmentation group, we introduce the notion of group diameter, defined as the maximal perturbation that augmentations can induce in the input space. The group diameter provides a unified control parameter that bounds all three terms and highlights an intrinsic trade-off: small diameters preserve data fidelity but offer limited regularization, while large diameters enhance stability at the cost of increased bias and sensitivity. We validate our theoretical bounds with numerical experiments, demonstrating that it reliably tracks and predicts the behavior of the true generalization gap.
[646] A unified framework for evaluating the robustness of machine-learning interpretability for prospect risking
Prithwijit Chowdhury, Ahmad Mustafa, Mohit Prabhushankar, Ghassan AlRegib
Main category: cs.LG
TL;DR: A framework for evaluating XAI robustness in hydrocarbon prospect risking using counterfactuals and causal necessity/sufficiency metrics to assess LIME and SHAP explanations on high-dimensional geophysical data.
Details
Motivation: Machine learning classifiers for hydrocarbon prospect risking lack transparency, and existing XAI methods (LIME, SHAP) often produce conflicting explanations for complex geophysical data, requiring more reliable evaluation frameworks grounded in causal theory.Method: Proposes a unified framework that generates counterfactuals and quantifies necessity and sufficiency metrics to perform robustness evaluation of LIME and SHAP explanations on high-dimensional structured prospect risking data.
Result: The robustness test provides insights into model capabilities to handle erroneous data and identifies which XAI module works best with which model for hydrocarbon indication in the specific dataset.
Conclusion: Grounding XAI explanations in causal concepts of necessity and sufficiency offers a more reliable way to evaluate explanation robustness and improve trustworthiness in hydrocarbon prospect risking applications.
Abstract: In geophysics, hydrocarbon prospect risking involves assessing the risks associated with hydrocarbon exploration by integrating data from various sources. Machine learning-based classifiers trained on tabular data have been recently used to make faster decisions on these prospects. The lack of transparency in the decision-making processes of such models has led to the emergence of explainable AI (XAI). LIME and SHAP are two such examples of these XAI methods which try to generate explanations of a particular decision by ranking the input features in terms of importance. However, explanations of the same scenario generated by these two different explanation strategies have shown to disagree or be different, particularly for complex data. This is because the definitions of “importance” and “relevance” differ for different explanation strategies. Thus, grounding these ranked features using theoretically backed causal ideas of necessity and sufficiency can prove to be a more reliable and robust way to improve the trustworthiness of the concerned explanation strategies.We propose a unified framework to generate counterfactuals as well as quantify necessity and sufficiency and use these to perform a robustness evaluation of the explanations provided by LIME and SHAP on high dimensional structured prospect risking data. This robustness test gives us deeper insights into the models capabilities to handle erronous data and which XAI module works best in pair with which model for our dataset for hydorcarbon indication.
[647] S2D: Selective Spectral Decay for Quantization-Friendly Conditioning of Neural Activations
Arnav Chavan, Nahush Lele, Udbhav Bamba, Sankalp Dayal, Aditi Raghunathan, Deepak Gupta
Main category: cs.LG
TL;DR: S^2D (Selective Spectral Decay) reduces activation outliers in transformer models by surgically regularizing weight components with largest singular values, improving quantization accuracy.
Details
Motivation: Activation outliers in large transformer models cause severe accuracy drops during quantization, especially in extensively pre-trained models like SigLIP/SigLIP2. These outliers are linked to dominant singular values of weights.Method: Proposed Selective Spectral Decay (S^2D), a geometrically-principled conditioning method that selectively regularizes only the weight components corresponding to largest singular values during fine-tuning to reduce activation outliers.
Result: S^2D significantly reduces activation outliers and produces quantization-friendly representations. Achieves up to 7% improved PTQ accuracy on ImageNet under W4A4 quantization and 4% gains when combined with QAT. Improvements generalize across downstream tasks and vision-language models.
Conclusion: S^2D enables scaling of large, rigorously trained vision-language models without sacrificing deployment efficiency by addressing the fundamental challenge of activation outliers in quantization.
Abstract: Activation outliers in large-scale transformer models pose a fundamental challenge to model quantization, creating excessively large ranges that cause severe accuracy drops during quantization. We empirically observe that outlier severity intensifies with pre-training scale (e.g., progressing from CLIP to the more extensively trained SigLIP and SigLIP2). Through theoretical analysis as well as empirical correlation studies, we establish the direct link between these activation outliers and dominant singular values of the weights. Building on this insight, we propose Selective Spectral Decay ($S^2D$), a geometrically-principled conditioning method that surgically regularizes only the weight components corresponding to the largest singular values during fine-tuning. Through extensive experiments, we demonstrate that $S^2D$ significantly reduces activation outliers and produces well-conditioned representations that are inherently quantization-friendly. Models trained with $S^2D$ achieve up to 7% improved PTQ accuracy on ImageNet under W4A4 quantization and 4% gains when combined with QAT. These improvements also generalize across downstream tasks and vision-language models, enabling the scaling of increasingly large and rigorously trained models without sacrificing deployment efficiency.
[648] Broken Chains: The Cost of Incomplete Reasoning in LLMs
Ian Su, Gaurav Purushothaman, Jey Narayan, Ruhika Goel, Kevin Zhu, Sunishchal Dev, Yash More, Maheep Chaudhary
Main category: cs.LG
TL;DR: Study examines how different reasoning modalities (code, natural language, hybrid, none) perform under token constraints, finding code degrades gracefully while truncated reasoning can actively mislead models.
Details
Motivation: Reasoning-specialized models allocate substantial compute to extended chain-of-thought traces, but reasoning tokens incur significant costs. Need to understand how different reasoning modalities perform under constrained token budgets.Method: Introduce framework constraining models to reason exclusively through code, comments, both, or neither, then systematically ablate token budgets to 10%, 30%, 50%, and 70% of optimal. Evaluate four frontier models across mathematical benchmarks.
Result: 1) Truncated reasoning hurts performance (DeepSeek-V3.2: 53% with no reasoning vs 17% with truncated CoT at 50% budget); 2) Code degrades gracefully (Gemini’s comments collapse to 0% while code maintains 43-47%); 3) Hybrid reasoning underperforms single modalities; 4) Robustness is model-dependent (Grok maintains 80-90% at 30% budget where others collapse).
Conclusion: Incomplete reasoning chains actively mislead models, with implications for deploying reasoning-specialized systems under resource constraints. Code-based reasoning shows better robustness to token constraints than natural language reasoning.
Abstract: Reasoning-specialized models like OpenAI’s 5.1 and DeepSeek-V3.2 allocate substantial inference compute to extended chain-of-thought (CoT) traces, yet reasoning tokens incur significant costs. How do different reasoning modalities of code, natural language, hybrid, or none do perform under token constraints? We introduce a framework that constrains models to reason exclusively through code, comments, both, or neither, then systematically ablates token budgets to 10%, 30%, 50%, and 70% of optimal. We evaluate four frontier models (GPT-5.1, Gemini 3 Flash, DeepSeek-V3.2, Grok 4.1) across mathematical benchmarks (AIME, GSM8K, HMMT). Our findings reveal: (1) \textbf{truncated reasoning can hurt} as DeepSeek-V3.2 achieves 53% with no reasoning but only 17% with truncated CoT at 50% budget; (2) \textbf{code degrades gracefully} as Gemini’s comments collapse to 0% while code maintains 43-47%; (3) \textbf{hybrid reasoning underperforms} single modalities; (4) \textbf{robustness is model-dependent} as Grok maintains 80-90% at 30% budget where OpenAI and DeepSeek collapse to 7-27%. These results suggest incomplete reasoning chains actively mislead models, with implications for deploying reasoning-specialized systems under resource constraints.
[649] Selective Synchronization Attention
Hasi Hays
Main category: cs.LG
TL;DR: SSA replaces standard self-attention with a Kuramoto oscillator-based mechanism using token frequencies and phases for synchronization-based attention, offering sparsity, unified encoding, and efficient computation.
Details
Motivation: Transformers suffer from quadratic complexity and lack biological grounding. The authors aim to create a more efficient, biologically-inspired attention mechanism.Method: Proposes Selective Synchronization Attention (SSA) where tokens are oscillators with learnable frequencies/phases. Attention weights come from synchronization strength based on frequency-dependent coupling and phase-locking conditions. Implemented in Oscillatory Synchronization Network (OSN) as Transformer replacement.
Result: SSA provides natural sparsity via phase-locking thresholds, unified positional-semantic encoding through frequency spectrum, and efficient closed-form computation. Analysis shows non-uniform, diverse coupling patterns at initialization.
Conclusion: SSA offers a biologically-grounded, efficient alternative to standard self-attention with built-in sparsity and unified encoding, demonstrating stronger architectural inductive bias than Transformers.
Abstract: The Transformer architecture has become the foundation of modern deep learning, yet its core self-attention mechanism suffers from quadratic computational complexity and lacks grounding in biological neural computation. We propose Selective Synchronization Attention (SSA), a novel attention mechanism that replaces the standard dot-product self-attention with a closed-form operator derived from the steady-state solution of the Kuramoto model of coupled oscillators. In SSA, each token is represented as an oscillator characterized by a learnable natural frequency and phase; the synchronization strength between token pairs, determined by a frequency-dependent coupling and phase-locking condition, serves as the attention weight. This formulation provides three key advantages: (i) natural sparsity arising from the phase-locking threshold, whereby tokens with incompatible frequencies automatically receive zero attention weight without explicit masking; (ii) unified positional-semantic encoding through the natural frequency spectrum, eliminating the need for separate positional encodings; and (iii) a single-pass, closed-form computation that avoids iterative ODE integration, with all components (coupling, order parameter, synchronization) derived from the oscillatory framework. We instantiate SSA within the Oscillatory Synchronization Network (OSN), a drop-in replacement for the Transformer block. Analysis of the synchronization matrices reveals non-uniform, head-diverse coupling patterns even at initialization, demonstrating a stronger architectural inductive bias than the approximately uniform attention produced by randomly initialized Transformers.
[650] WiSparse: Boosting LLM Inference Efficiency with Weight-Aware Mixed Activation Sparsity
Lei Chen, Yuan Meng, Xiaoyu Zhan, Zhi Wang, Wenwu Zhu
Main category: cs.LG
TL;DR: WiSparse: A training-free activation sparsity method for efficient LLM inference that combines weight-aware channel selection with mixed-granularity sparsity allocation across blocks.
Details
Motivation: Existing training-free activation sparsity methods for LLMs rely solely on activation information and uniform sparsity ratios, overlooking the interplay with weights and inter-block sensitivity variation, leading to suboptimal performance.Method: WiSparse integrates activation magnitudes with precomputed weight norms for weight-aware channel selection, combined with mixed-granularity allocation: global budget distributed across blocks via evolutionary search, then refined within blocks to minimize reconstruction error.
Result: At 50% sparsity, WiSparse preserves 97% of Llama3.1’s dense performance, surpassing the strongest baseline by 2.23 percentage points while achieving 21.4% acceleration in end-to-end inference speed.
Conclusion: WiSparse advances training-free approaches for efficient LLM inference by leveraging both activation and weight information with adaptive sparsity allocation, pushing the boundaries of achievable speedup without training.
Abstract: Large Language Models (LLMs) offer strong capabilities but incur high inference costs due to dense computation and memory access. Training-free activation sparsity is a promising approach for efficient LLM inference, yet existing methods often rely solely on activation information and uniform sparsity ratios. This overlooks the critical interplay with weights and inter-block sensitivity variation, leading to suboptimal performance. We identify two key phenomena in modern LLMs: 1) less significant activations may align with highly important weights, and 2) sparsity sensitivity varies non-monotonically across model blocks. We propose Weight-aware Mixed-Granularity Training-free Activation Sparsity (WiSparse), which leverages both activation and weight information for adaptive sparsity allocation. Specifically, we introduce a weight-aware mechanism integrating activation magnitudes with precomputed weight norms to accurately identify salient channels. This is combined with a mixed-granularity allocation scheme: a global budget is distributed across blocks via evolutionary search to protect sensitive regions, then refined within blocks to minimize reconstruction error. We improve sparse kernels and demonstrate effectiveness on three representative models. Notably, at 50% sparsity, WiSparse preserves 97% of Llama3.1’s dense performance, surpassing the strongest baseline by 2.23 percentage points while achieving a 21.4% acceleration in end-to-end inference speed. Our research advances the limits of training-free approaches for efficient LLM inference, pushing the boundaries of achievable speedup without training.
[651] Traceable Latent Variable Discovery Based on Multi-Agent Collaboration
Huaming Du, Tao Hu, Yijie Huang, Yu Zhao, Guisong Liu, Tao Gu, Gang Kou, Carl Yang
Main category: cs.LG
TL;DR: TLVD is a novel causal modeling framework that combines large language models with traditional causal discovery algorithms to infer latent variables and their semantics, addressing limitations of existing methods.
Details
Motivation: Traditional causal discovery algorithms have limitations: they assume no latent confounders, lack high-quality data, and ignore precise semantics of latent variables. This hinders broader application of causal discovery in real-world scenarios.Method: 1) Use data-driven approach to construct causal graph with latent variables; 2) Employ multi-LLM collaboration for latent variable inference, modeled as a game with incomplete information seeking Bayesian Nash Equilibrium; 3) Validate inferred latent variables using LLMs for evidence exploration across multiple real-world web data sources.
Result: Extensive experiments on three real patient datasets and two benchmark datasets show TLVD achieves average improvements of 32.67% in Acc, 62.21% in CAcc, and 26.72% in ECit across five datasets, confirming effectiveness and reliability.
Conclusion: TLVD successfully integrates LLMs’ metadata-based reasoning with traditional causal discovery algorithms to overcome limitations in latent variable inference, providing a more effective framework for causal discovery with real-world applications.
Abstract: Revealing the underlying causal mechanisms in the real world is crucial for scientific and technological progress. Despite notable advances in recent decades, the lack of high-quality data and the reliance of traditional causal discovery algorithms (TCDA) on the assumption of no latent confounders, as well as their tendency to overlook the precise semantics of latent variables, have long been major obstacles to the broader application of causal discovery. To address this issue, we propose a novel causal modeling framework, TLVD, which integrates the metadata-based reasoning capabilities of large language models (LLMs) with the data-driven modeling capabilities of TCDA for inferring latent variables and their semantics. Specifically, we first employ a data-driven approach to construct a causal graph that incorporates latent variables. Then, we employ multi-LLM collaboration for latent variable inference, modeling this process as a game with incomplete information and seeking its Bayesian Nash Equilibrium (BNE) to infer the possible specific latent variables. Finally, to validate the inferred latent variables across multiple real-world web-based data sources, we leverage LLMs for evidence exploration to ensure traceability. We comprehensively evaluate TLVD on three de-identified real patient datasets provided by a hospital and two benchmark datasets. Extensive experimental results confirm the effectiveness and reliability of TLVD, with average improvements of 32.67% in Acc, 62.21% in CAcc, and 26.72% in ECit across the five datasets.
[652] Silent Inconsistency in Data-Parallel Full Fine-Tuning: Diagnosing Worker-Level Optimization Misalignment
Hong Li, Zhen Zhou, Honggang Zhang, Yuping Luo, Xinyue Wang, Han Gong, Zhiyuan Liu
Main category: cs.LG
TL;DR: A framework for detecting silent inconsistencies in data-parallel LLM fine-tuning where worker-level optimization diverges despite synchronized weights.
Details
Motivation: While data-parallel training with synchronous all-reduce ensures weight equivalence after each iteration, it doesn't guarantee alignment of worker-level optimization dynamics before gradient aggregation. This latent mismatch, called "silent inconsistency," can cause cross-worker divergence in losses and gradients that remains invisible under conventional aggregated monitoring signals.Method: Proposes a lightweight, model-agnostic diagnostic framework with three complementary metrics: loss dispersion, gradient-norm dispersion, and gradient-direction consistency measured by inter-worker cosine similarity. The framework requires no modification to model architecture, synchronization mechanisms, or optimization algorithms and incurs negligible overhead.
Result: Experimental validation by fine-tuning a 1B-parameter model on an 8-NPU setup shows that progressively desynchronized data shuffling and random seeds lead to substantial increases in loss/gradient dispersion and reduced directional alignment, despite smooth globally averaged loss curves.
Conclusion: The proposed indicators provide actionable visibility into hidden instability modes in large-scale data-parallel fine-tuning, enabling more reliable diagnosis and configuration assessment for LLM training.
Abstract: Data-parallel (DP) training with synchronous all-reduce is a dominant paradigm for full-parameter fine-tuning of large language models (LLMs). While parameter synchronization guarantees numerical equivalence of model weights after each iteration, it does not necessarily imply alignment of worker-level optimization dynamics before gradient aggregation. This paper identifies and studies this latent mismatch, termed \emph{silent inconsistency}, where cross-worker divergence in losses and gradients can remain invisible under conventional aggregated monitoring signals. We propose a lightweight, model-agnostic diagnostic framework that quantifies worker-level consistency using training signals readily available in standard pipelines. Specifically, we introduce three complementary metrics: loss dispersion, gradient-norm dispersion, and gradient-direction consistency measured by inter-worker cosine similarity. The proposed metrics incur negligible overhead and require no modification to model architecture, synchronization mechanisms, or optimization algorithms. We validate the framework by fully fine-tuning the 1B-parameter \texttt{openPangu-Embedded-1B-V1.1} model on the \texttt{tatsu-lab/alpaca} dataset using an 8-NPU DP setup, under controlled perturbations of cross-rank stochasticity. Experimental results show that progressively desynchronized data shuffling and random seeds lead to substantial increases in loss/gradient dispersion and reduced directional alignment, despite smooth globally averaged loss curves. These findings demonstrate that the proposed indicators provide actionable visibility into hidden instability modes in large-scale DP fine-tuning, enabling more reliable diagnosis and configuration assessment.
[653] Parameter-Efficient Fine-Tuning of LLMs with Mixture of Space Experts
Buze Zhang, Jinkai Tao, Zilang Zeng, Neil He, Ali Maatouk, Menglin Yang, Rex Ying
Main category: cs.LG
TL;DR: MoSLoRA extends LoRA with mixture of geometric spaces, enabling dynamic selection of appropriate manifolds (hyperbolic, spherical, Euclidean) for richer representations, improving performance on reasoning tasks.
Details
Motivation: Existing PEFT methods operate only in Euclidean space, limiting their ability to capture complex geometric structures in language data. Single manifold approaches restrict expressiveness even with learnable curvature.Method: Proposes Mixture of Space (MoS) framework using multiple geometric spaces simultaneously. Develops MoSLoRA which extends LoRA with heterogeneous geometric experts and lightweight routing mechanism for efficient manifold switching.
Result: MoSLoRA consistently outperforms baselines, achieving up to 5.6% improvement on MATH500 and 15.9% on MAWPS benchmarks. Provides empirical insights on curvature optimization’s impact on training stability.
Conclusion: Leveraging multiple geometric spaces through MoSLoRA enables richer, curvature-aware representations that improve model performance on diverse tasks compared to single-manifold approaches.
Abstract: Large Language Models (LLMs) have achieved remarkable progress, with Parameter-Efficient Fine-Tuning (PEFT) emerging as a key technique for downstream task adaptation. However, existing PEFT methods mainly operate in Euclidean space, fundamentally limiting their capacity to capture complex geometric structures inherent in language data. While alternative geometric spaces, like hyperbolic geometries for hierarchical data and spherical manifolds for circular patterns, offer theoretical advantages, forcing representations into a single manifold type ultimately limits expressiveness, even when curvature parameters are learnable. To address this, we propose Mixture of Space (MoS), a unified framework that leverages multiple geometric spaces simultaneously to learn richer, curvature-aware representations. Building on this scheme, we develop MoSLoRA, which extends Low-Rank Adaptation (LoRA) with heterogeneous geometric experts, enabling models to dynamically select or combine appropriate geometric spaces based on input context. Furthermore, to address the computational overhead of frequent manifold switching, we develop a lightweight routing mechanism. Moreover, we provide empirical insights into how curvature optimization impacts training stability and model performance. Our experiments across diverse benchmarks demonstrate that MoSLoRA consistently outperforms strong baselines, achieving up to 5.6% improvement on MATH500 and 15.9% on MAWPS.
[654] LACONIC: Length-Aware Constrained Reinforcement Learning for LLM
Chang Liu, Yiran Zhao, Lawrence Liu, Yaoqi Ye, Csaba Szepesvári, Lin F. Yang
Main category: cs.LG
TL;DR: LACONIC is a reinforcement learning method that enforces token budget constraints during LLM training to reduce response length while maintaining task performance.
Details
Motivation: RL training for LLMs often produces excessively long responses, increasing inference latency and computational costs. Existing length-control methods use fixed heuristic rewards that can misalign with task objectives and require brittle tuning.Method: LACONIC updates policy models using an augmented objective combining task reward with length-based cost. The cost scale is adaptively adjusted throughout training to balance brevity and task performance, enforcing a target token budget.
Result: Across mathematical reasoning models and datasets, LACONIC preserves or improves pass@1 while reducing output length by over 50%. It maintains out-of-domain performance on general knowledge and multilingual benchmarks with 44% fewer tokens.
Conclusion: LACONIC provides robust length control while preserving task reward, integrates into standard RL-tuning with no inference changes, and has minimal deployment overhead, with theoretical guarantees supporting the method.
Abstract: Reinforcement learning (RL) has enhanced the capabilities of large language models (LLMs) through reward-driven training. Nevertheless, this process can introduce excessively long responses, inflating inference latency and computational overhead. Prior length-control approaches typically rely on fixed heuristic reward shaping, which can misalign with the task objective and require brittle tuning. In this work, we propose LACONIC, a reinforcement learning method that enforces a target token budget during training. Specifically, we update policy models using an augmented objective that combines the task reward with a length-based cost. To balance brevity and task performance, the cost scale is adaptively adjusted throughout training. This yields robust length control while preserving task reward. We provide a theoretical guarantee that support the method. Across mathematical reasoning models and datasets, LACONIC preserves or improves pass@1 while reducing output length by over 50%. It maintains out-of-domain performance on general knowledge and multilingual benchmarks with 44% fewer tokens. Moreover, LACONIC integrates into standard RL-tuning with no inference changes and minimal deployment overhead.
[655] Alignment Adapter to Improve the Performance of Compressed Deep Learning Models
Rohit Raj Rai, Abhishek Dhaka, Amit Awekar
Main category: cs.LG
TL;DR: Alignment Adapter (AlAd) improves compressed DL models by aligning their token embeddings with original large models using lightweight sliding-window adapters.
Details
Motivation: Compressed DL models are necessary for resource-constrained environments but suffer performance degradation compared to large models. There's a need to bridge this performance gap without significant overhead.Method: Proposes AlAd: a lightweight sliding-window-based adapter that aligns token-level embeddings of compressed models with original large models. It preserves local contextual semantics, works across different dimensionalities/architectures, and is compression-method agnostic. Can be deployed as plug-and-play module or jointly fine-tuned.
Result: Experiments on BERT-family models across three token-level NLP tasks show AlAd significantly boosts compressed model performance with only marginal size/latency overhead.
Conclusion: AlAd effectively bridges performance gap between compressed and large models through lightweight token-level alignment, offering flexible deployment options with minimal overhead.
Abstract: Compressed Deep Learning (DL) models are essential for deployment in resource-constrained environments. But their performance often lags behind their large-scale counterparts. To bridge this gap, we propose Alignment Adapter (AlAd): a lightweight, sliding-window-based adapter. It aligns the token-level embeddings of a compressed model with those of the original large model. AlAd preserves local contextual semantics, enables flexible alignment across differing dimensionalities or architectures, and is entirely agnostic to the underlying compression method. AlAd can be deployed in two ways: as a plug-and-play module over a frozen compressed model, or by jointly fine-tuning AlAd with the compressed model for further performance gains. Through experiments on BERT-family models across three token-level NLP tasks, we demonstrate that AlAd significantly boosts the performance of compressed models with only marginal overhead in size and latency.
[656] One Good Source is All You Need: Near-Optimal Regret for Bandits under Heterogeneous Noise
Aadirupa Saha, Amith Bhat, Haipeng Luo
Main category: cs.LG
TL;DR: SOAR algorithm for multi-armed bandits with multiple heterogeneous data sources adaptively selects optimal low-variance sources while minimizing regret.
Details
Motivation: Traditional MAB assumes single data source with known/unknown variance, but real-world applications often have multiple data sources with different noise characteristics. The challenge is to minimize regret while adaptively identifying and utilizing the optimal (minimum-variance) source without prior knowledge.Method: Proposes Source-Optimistic Adaptive Regret minimization (SOAR) algorithm: 1) Quickly prunes high-variance sources using sharp variance-concentration bounds, 2) Uses balanced min-max LCB-UCB approach to simultaneously identify best arm and optimal data source, 3) Adaptively selects which data source to query at each round.
Result: Achieves instance-dependent regret bound of $\tilde{O}\left({σ^}^2\sum_{i=2}^K \frac{\log T}{Δ_i} + \sqrt{K \sum_{j=1}^M σ_j^2}\right)$, attaining optimal single-source MAB regret with minimum variance σ² plus small additive cost for source identification. Outperforms baselines like Uniform UCB and Explore-then-Commit UCB on synthetic and MovieLens 25M datasets.
Conclusion: SOAR effectively handles MAB with multiple heterogeneous data sources, achieving near-optimal regret by adaptively identifying and leveraging the minimum-variance source without prior knowledge, with significant improvements over naive approaches.
Abstract: We study $K$-armed Multiarmed Bandit (MAB) problem with $M$ heterogeneous data sources, each exhibiting unknown and distinct noise variances ${σ_j^2}{j=1}^M$. The learner’s objective is standard MAB regret minimization, with the additional complexity of adaptively selecting which data source to query from at each round. We propose Source-Optimistic Adaptive Regret minimization (SOAR), a novel algorithm that quickly prunes high-variance sources using sharp variance-concentration bounds, followed by a `balanced min-max LCB-UCB approach’ that seamlessly integrates the parallel tasks of identifying the best arm and the optimal (minimum-variance) data source. Our analysis shows SOAR achieves an instance-dependent regret bound of $\tilde{O}\left({σ^*}^2\sum{i=2}^K \frac{\log T}{Δ_i} + \sqrt{K \sum_{j=1}^M σ_j^2}\right)$, up to preprocessing costs depending only on problem parameters, where ${σ^}^2 := \min_j σ_j^2$ is the minimum source variance and $Δ_i$ denotes the suboptimality gap of the $i$-th arm. This result is both surprising as despite lacking prior knowledge of the minimum-variance source among $M$ alternatives, SOAR attains the optimal instance-dependent regret of standard single-source MAB with variance ${σ^}^2$, while incurring only an small (and unavoidable) additive cost of $\tilde O(\sqrt{K \sum_{j=1}^M σ_j^2})$ towards the optimal (minimum variance) source identification. Our theoretical bounds represent a significant improvement over some proposed baselines, e.g. Uniform UCB or Explore-then-Commit UCB, which could potentially suffer regret scaling with $σ_{\max}^2$ in place of ${σ^}^2$-a gap that can be arbitrarily large when $σ_{\max} \gg σ^$. Experiments on multiple synthetic problem instances and the real-world MovieLens;25M dataset, demonstrating the superior performance of SOAR over the baselines.
[657] Revisiting the Platonic Representation Hypothesis: An Aristotelian View
Fabian Gröger, Shuo Wen, Maria Brbić
Main category: cs.LG
TL;DR: The paper introduces a calibration framework to correct scale confounds in representational similarity metrics, finding that neural network representations converge to shared local neighborhood relationships rather than global structure.
Details
Motivation: To address the confounding effect of network scale on representational similarity metrics and properly test the Platonic Representation Hypothesis about neural networks converging to a common statistical model of reality.Method: Developed a permutation-based null-calibration framework that transforms any representational similarity metric into a calibrated score with statistical guarantees, then applied this to analyze representation convergence across different network architectures and modalities.
Result: After calibration, the apparent convergence reported by global spectral measures largely disappears, while local neighborhood similarity (but not local distances) retains significant agreement across different modalities.
Conclusion: Proposes the Aristotelian Representation Hypothesis: neural network representations converge to shared local neighborhood relationships rather than global structure, with implications for understanding representation learning and transfer learning.
Abstract: The Platonic Representation Hypothesis suggests that representations from neural networks are converging to a common statistical model of reality. We show that the existing metrics used to measure representational similarity are confounded by network scale: increasing model depth or width can systematically inflate representational similarity scores. To correct these effects, we introduce a permutation-based null-calibration framework that transforms any representational similarity metric into a calibrated score with statistical guarantees. We revisit the Platonic Representation Hypothesis with our calibration framework, which reveals a nuanced picture: the apparent convergence reported by global spectral measures largely disappears after calibration, while local neighborhood similarity, but not local distances, retains significant agreement across different modalities. Based on these findings, we propose the Aristotelian Representation Hypothesis: representations in neural networks are converging to shared local neighborhood relationships.
[658] Learning State-Tracking from Code Using Linear RNNs
Julien Siems, Riccardo Grazzi, Kirill Kalinin, Hitesh Ballani, Babak Rahmani
Main category: cs.LG
TL;DR: Transformers fail at state tracking in code REPL traces while linear RNNs succeed, but linear RNNs struggle with probabilistic state tracking where actions aren’t fully observable.
Details
Motivation: To bridge the gap between state-tracking tasks (like permutation composition) and next-token prediction used in language models, by converting permutation composition into code via REPL traces that interleave state reveals and variable transformations.Method: Convert permutation composition into code via REPL traces with state reveals through prints and variable transformations. Compare performance of Transformers vs linear/non-linear RNNs. Extend to probabilistic finite-state automaton tracking with deterministic state reveals.
Result: Linear RNNs capable of state tracking excel in the REPL code setting, while Transformers still fail. However, linear RNNs can be worse than non-linear RNNs at tracking states in probabilistic setups where actions are not fully observable.
Conclusion: State tracking in code is difficult because actions are not always fully observable, and different architectures have varying capabilities depending on the observability of the state tracking problem.
Abstract: Over the last years, state-tracking tasks, particularly permutation composition, have become a testbed to understand the limits of sequence models architectures like Transformers and RNNs (linear and non-linear). However, these are often sequence-to-sequence tasks: learning to map actions (permutations) to states, which is incompatible with the next-token prediction setting commonly used to train language models. We address this gap by converting permutation composition into code via REPL traces that interleave state-reveals through prints and variable transformations. We show that linear RNNs capable of state-tracking excel also in this setting, while Transformers still fail. Motivated by this representation, we investigate why tracking states in code is generally difficult: actions are not always fully observable. We frame this as tracking the state of a probabilistic finite-state automaton with deterministic state reveals and show that linear RNNs can be worse than non-linear RNNs at tracking states in this setup.
[659] Divine Benevolence is an $x^2$: GLUs scale asymptotically faster than MLPs
Alejandro Francisco Queiruga
Main category: cs.LG
TL;DR: The paper uses numerical analysis to explain why GLU variants outperform MLPs in LLMs, showing GLUs have x² terms enabling asymptotically faster scaling (L(P)∝P⁻³ vs P⁻²), and proposes a new Gated Quadratic Unit with even steeper scaling.
Details
Motivation: The empirical success of GLU variants in frontier LLMs and similar architectures in ranking models lacks theoretical explanation. The paper aims to apply numerical analysis to understand why these architectures outperform traditional MLPs from first principles.Method: The authors use tools from numerical analysis and function approximation theory to analyze scaling laws. They demonstrate that GLUs have piecewise quadratic functional forms enabling quadratic order of approximation, leading to L(P)∝P⁻³ scaling versus MLPs’ L(P)=P⁻². They provide parameter construction and empirical verification on 1D function approximation problems.
Result: Theoretical analysis shows GLUs achieve asymptotically faster scaling than MLPs (P⁻³ vs P⁻²). Empirical verification on 1D function approximation confirms these scaling slopes. Based on these principles, the authors propose a “Gated Quadratic Unit” with even steeper scaling slope than both GLUs and MLPs.
Conclusion: Architecture design can be guided by first principles numerical theory to unlock superior scaling in large models. The theoretical understanding of why GLUs outperform MLPs opens possibilities for designing even more efficient architectures like the proposed Gated Quadratic Unit.
Abstract: Scaling laws can be understood from ground-up numerical analysis, where traditional function approximation theory can explain shifts in model architecture choices. GLU variants now dominate frontier LLMs and similar outer-product architectures are prevalent in ranking models. The success of these architectures has mostly been left as an empirical discovery. In this paper, we apply the tools of numerical analysis to expose a key factor: these models have an $x^2$ which enables \emph{asymptotically} faster scaling than MLPs. GLUs have piecewise quadratic functional forms that are sufficient to exhibit quadratic order of approximation. Our key contribution is to demonstrate that the $L(P)$ scaling slope is $L(P)\propto P^{-3}$ for GLUs but only $L(P)=P^{-2}$ for MLPs on function reconstruction problems. We provide a parameter construction and empirical verification of these slopes for 1D function approximation. From the first principles we discover, we make one stride and propose the ``Gated Quadratic Unit’’ which has an even steeper $L(P)$ slope than the GLU and MLP. This opens the possibility of architecture design from first principles numerical theory to unlock superior scaling in large models. Replication code is available at https://github.com/afqueiruga/divine_scaling.
[660] Scaling Beyond Masked Diffusion Language Models
Subham Sekhar Sahoo, Jean-Marie Lemercier, Zhihan Yang, Justin Deschenaux, Jingyu Liu, John Thickstun, Ante Jukic
Main category: cs.LG
TL;DR: Scaling law study of discrete diffusion language models shows uniform-state diffusion remains competitive and outperforms autoregressive/masked diffusion on reasoning tasks despite worse perplexity, challenging the dominance of masked diffusion.
Details
Motivation: To challenge the prevailing view that masked diffusion is categorically superior for language modeling based on perplexity benchmarks, and to examine scaling laws across different discrete diffusion families to understand trade-offs between likelihood and practical sampling efficiency.Method: Conducted first scaling law study of uniform-state and interpolating discrete diffusion methods; trained masked diffusion models with cross-entropy objective for FLOPs efficiency; compared perplexity scaling across diffusion families; evaluated speed-quality Pareto frontiers; scaled all methods to 1.7B parameters.
Result: Masked diffusion made ~12% more FLOPs-efficient with cross-entropy training; perplexity informative within families but misleading across families; uniform-state diffusion competitive on likelihood benchmarks and outperformed autoregressive/masked diffusion on GSM8K reasoning task despite worse validation perplexity.
Conclusion: Masked diffusion not categorically superior; perplexity alone insufficient for cross-algorithm comparison; uniform-state diffusion remains viable alternative with better practical sampling characteristics; need to consider speed-quality trade-offs beyond perplexity benchmarks.
Abstract: Diffusion language models are a promising alternative to autoregressive models due to their potential for faster generation. Among discrete diffusion approaches, Masked diffusion currently dominates, largely driven by strong perplexity on language modeling benchmarks. In this work, we present the first scaling law study of uniform-state and interpolating discrete diffusion methods. We also show that Masked diffusion models can be made approximately 12% more FLOPs-efficient when trained with a simple cross-entropy objective. We find that perplexity is informative within a diffusion family but can be misleading across families, where models with worse likelihood scaling may be preferable due to faster and more practical sampling, as reflected by the speed-quality Pareto frontier. These results challenge the view that Masked diffusion is categorically the future of diffusion language modeling and that perplexity alone suffices for cross-algorithm comparison. Scaling all methods to 1.7B parameters, we show that uniform-state diffusion remains competitive on likelihood-based benchmarks and outperforms autoregressive and Masked diffusion models on GSM8K, despite worse validation perplexity. We provide the code, model checkpoints, and video tutorials on the project page: http://s-sahoo.github.io/scaling-dllms
[661] Covariance-Aware Transformers for Quadratic Programming and Decision Making
Kutay Tire, Yufan Zhang, Ege Onur Taga, Samet Oymak
Main category: cs.LG
TL;DR: Transformers can solve quadratic programs (QPs) using linear attention to emulate optimization algorithms, enabling direct decision-making from time series data with explicit covariance modeling.
Details
Motivation: The paper aims to bridge the gap between time series foundation models and decision-making by enabling transformers to directly solve optimization problems like portfolio construction, rather than just forecasting.Method: Theoretical analysis shows transformers with linear attention can solve unconstrained QPs by emulating gradient descent. With MLPs, they can solve ℓ₁-penalized QPs (emulating iterative soft-thresholding) and ℓ₁-constrained QPs (with feedback loops). This leads to Time2Decide: feeding covariance matrices to TSFMs for direct decision-making.
Result: Time2Decide outperforms base TSFMs and traditional “Predict-then-Optimize” approaches for portfolio optimization, showing transformers benefit from explicit second-order statistics for complex decision-making.
Conclusion: Transformers can effectively solve complex decision-making problems like portfolio construction in one forward pass by explicitly using covariance information, bridging forecasting and optimization.
Abstract: We explore the use of transformers for solving quadratic programs and how this capability benefits decision-making problems that involve covariance matrices. We first show that the linear attention mechanism can provably solve unconstrained QPs by tokenizing the matrix variables (e.g.~$A$ of the objective $\frac{1}{2}x^\top Ax+b^\top x$) row-by-row and emulating gradient descent iterations. Furthermore, by incorporating MLPs, a transformer block can solve (i) $\ell_1$-penalized QPs by emulating iterative soft-thresholding and (ii) $\ell_1$-constrained QPs when equipped with an additional feedback loop. Our theory motivates us to introduce Time2Decide: a generic method that enhances a time series foundation model (TSFM) by explicitly feeding the covariance matrix between the variates. We empirically find that Time2Decide uniformly outperforms the base TSFM model for the classical portfolio optimization problem that admits an $\ell_1$-constrained QP formulation. Remarkably, Time2Decide also outperforms the classical “Predict-then-Optimize (PtO)” procedure, where we first forecast the returns and then explicitly solve a constrained QP, in suitable settings. Our results demonstrate that transformers benefit from explicit use of second-order statistics, and this can enable them to effectively solve complex decision-making problems, like portfolio construction, in one forward pass.
[662] Symmetry in language statistics shapes the geometry of model representations
Dhruva Karkada, Daniel J. Korchinski, Andres Nava, Matthieu Wyart, Yasaman Bahri
Main category: cs.LG
TL;DR: The paper shows that translation symmetry in language statistics explains geometric structures in LLM representations, and these structures are robust to perturbations due to underlying continuous latent variables.
Details
Motivation: To understand why simple geometric structures emerge in LLM representations (like months forming circles, years forming smooth manifolds) and what governs these structures.Method: Theoretical analysis showing translation symmetry in language statistics governs geometric structures, empirical validation in word/text embedding models and LLMs, and investigation of robustness to perturbations.
Result: Demonstrated that translation symmetry in co-occurrence statistics explains geometric structures, and these structures persist even with perturbed statistics due to underlying continuous latent variables controlling the statistics collectively.
Conclusion: Geometric structures in LLM representations emerge from translation symmetry in language statistics and are robust due to collective control by underlying continuous latent variables.
Abstract: Although learned representations underlie neural networks’ success, their fundamental properties remain poorly understood. A striking example is the emergence of simple geometric structures in LLM representations: for example, calendar months organize into a circle, years form a smooth one-dimensional manifold, and cities’ latitudes and longitudes can be decoded by a linear probe. We show that the statistics of language exhibit a translation symmetry – e.g., the co-occurrence probability of two months depends only on the time interval between them – and we prove that the latter governs the aforementioned geometric structures in high-dimensional word embedding models. Moreover, we find that these structures persist even when the co-occurrence statistics are strongly perturbed (for example, by removing all sentences in which two months appear together) and at moderate embedding dimension. We show that this robustness naturally emerges if the co-occurrence statistics are collectively controlled by an underlying continuous latent variable. We empirically validate this theoretical framework in word embedding models, text embedding models, and large language models.
[663] DeepMTL2R: A Library for Deep Multi-task Learning to Rank
Chaosheng Dong, Peiyao Xiao, Yijia Wang, Kaiyi Ji
Main category: cs.LG
TL;DR: DeepMTL2R is an open-source deep learning framework for Multi-task Learning to Rank that integrates multiple relevance criteria using transformer self-attention for unified optimization.
Details
Motivation: Modern ranking systems need to optimize multiple relevance criteria simultaneously, which can be conflicting. Existing approaches lack unified frameworks that can effectively handle heterogeneous signals and complex dependencies among items and labels.Method: The framework leverages transformer self-attention mechanisms to integrate heterogeneous relevance signals into a context-aware model. It includes 21 state-of-the-art multi-task learning algorithms and supports multi-objective optimization to find Pareto-optimal ranking models.
Result: Demonstrated effectiveness on publicly available datasets with competitive performance. The framework enables visualization of trade-offs among objectives and facilitates controlled comparisons across MTL strategies.
Conclusion: DeepMTL2R provides a scalable and expressive solution for modern ranking systems that need to optimize multiple relevance criteria, offering a comprehensive framework for multi-task learning to rank.
Abstract: This paper presents DeepMTL2R, an open-source deep learning framework for Multi-task Learning to Rank (MTL2R), where multiple relevance criteria must be optimized simultaneously. DeepMTL2R integrates heterogeneous relevance signals into a unified, context-aware model by leveraging the self-attention mechanism of transformer architectures, enabling effective learning across diverse and potentially conflicting objectives. The framework includes 21 state-of-the-art multi-task learning algorithms and supports multi-objective optimization to identify Pareto-optimal ranking models. By capturing complex dependencies and long-range interactions among items and labels, DeepMTL2R provides a scalable and expressive solution for modern ranking systems and facilitates controlled comparisons across MTL strategies. We demonstrate its effectiveness on a publicly available dataset, report competitive performance, and visualize the resulting trade-offs among objectives. DeepMTL2R is available at \href{https://github.com/amazon-science/DeepMTL2R}{https://github.com/amazon-science/DeepMTL2R}.
[664] Truly Adapting to Adversarial Constraints in Constrained MABs
Francesco Emanuele Stradi, Kalana Kalupahana, Matteo Castiglioni, Alberto Marchesi, Nicola Gatti
Main category: cs.LG
TL;DR: This paper studies constrained multi-armed bandit problems with unknown constraints under both full and bandit feedback, focusing on non-stationary environments where both losses and constraints can change arbitrarily over time.
Details
Motivation: The motivation is to address the challenging problem of constrained bandit learning where the learner must minimize total loss while controlling violation of multiple unknown constraints, particularly in non-stationary environments that encompass both stochastic and adversarial models.Method: The authors propose algorithms for different feedback scenarios: full feedback, bandit feedback for losses only, and bandit feedback for constraints. They develop theoretical frameworks to handle the trade-off between regret and constraint violation in non-stationary settings.
Result: The algorithms achieve optimal rates: under full feedback, they attain Õ(√T + C) regret and Õ(√T + C) positive violation; with bandit feedback for constraints, they achieve Õ(√T + C) positive violation and Õ(√T + C√T) regret, where C quantifies constraint non-stationarity.
Conclusion: This work provides the first algorithms that achieve optimal regret and positive constraint violation rates for constrained bandit problems with stochastic constraints and arbitrary losses, with guarantees that degrade smoothly with the adversariality of constraints.
Abstract: We study the constrained variant of the \emph{multi-armed bandit} (MAB) problem, in which the learner aims not only at minimizing the total loss incurred during the learning dynamic, but also at controlling the violation of multiple \emph{unknown} constraints, under both \emph{full} and \emph{bandit feedback}. We consider a non-stationary environment that subsumes both stochastic and adversarial models and where, at each round, both losses and constraints are drawn from distributions that may change arbitrarily over time. In such a setting, it is provably not possible to guarantee both sublinear regret and sublinear violation. Accordingly, prior work has mainly focused either on settings with stochastic constraints or on relaxing the benchmark with fully adversarial constraints (\emph{e.g.}, via competitive ratios with respect to the optimum). We provide the first algorithms that achieve optimal rates of regret and \emph{positive} constraint violation when the constraints are stochastic while the losses may vary arbitrarily, and that simultaneously yield guarantees that degrade smoothly with the degree of adversariality of the constraints. Specifically, under \emph{full feedback} we propose an algorithm attaining $\widetilde{\mathcal{O}}(\sqrt{T}+C)$ regret and $\widetilde{\mathcal{O}}(\sqrt{T}+C)$ {positive} violation, where $C$ quantifies the amount of non-stationarity in the constraints. We then show how to extend these guarantees when only bandit feedback is available for the losses. Finally, when \emph{bandit feedback} is available for the constraints, we design an algorithm achieving $\widetilde{\mathcal{O}}(\sqrt{T}+C)$ {positive} violation and $\widetilde{\mathcal{O}}(\sqrt{T}+C\sqrt{T})$ regret.
[665] Governing AI Forgetting: Auditing for Machine Unlearning Compliance
Qinqi Lin, Ningning Ding, Lingjie Duan, Jianwei Huang
Main category: cs.LG
TL;DR: Economic framework for auditing machine unlearning compliance using game theory and certified unlearning theory
Details
Motivation: AI operators often fail to comply with data deletion requests despite legal mandates, creating a gap between technical machine unlearning solutions and regulatory implementationMethod: Integrates certified unlearning theory with regulatory enforcement using hypothesis-testing interpretation of certified unlearning and game-theoretic modeling of auditor-operator interactions
Result: Counterintuitive finding that auditors can reduce inspection intensity as deletion requests increase, and that undisclosed auditing reduces cost-effectiveness despite informational advantages
Conclusion: Provides first economic framework for auditing machine unlearning compliance, addressing the gap between technical feasibility and regulatory enforcement
Abstract: Despite legal mandates for the right to be forgotten, AI operators routinely fail to comply with data deletion requests. While machine unlearning (MU) provides a technical solution to remove personal data’s influence from trained models, ensuring compliance remains challenging due to the fundamental gap between MU’s technical feasibility and regulatory implementation. In this paper, we introduce the first economic framework for auditing MU compliance, by integrating certified unlearning theory with regulatory enforcement. We first characterize MU’s inherent verification uncertainty using a hypothesis-testing interpretation of certified unlearning to derive the auditor’s detection capability, and then propose a game-theoretic model to capture the strategic interactions between the auditor and the operator. A key technical challenge arises from MU-specific nonlinearities inherent in the model utility and the detection probability, which create complex strategic couplings that traditional auditing frameworks do not address and that also preclude closed-form solutions. We address this by transforming the complex bivariate nonlinear fixed-point problem into a tractable univariate auxiliary problem, enabling us to decouple the system and establish the equilibrium existence, uniqueness, and structural properties without relying on explicit solutions. Counterintuitively, our analysis reveals that the auditor can optimally reduce the inspection intensity as deletion requests increase, since the operator’s weakened unlearning makes non-compliance easier to detect. This is consistent with recent auditing reductions in China despite growing deletion requests. Moreover, we prove that although undisclosed auditing offers informational advantages for the auditor, it paradoxically reduces the regulatory cost-effectiveness relative to disclosed auditing.
[666] DCTracks: An Open Dataset for Machine Learning-Based Drift Chamber Track Reconstruction
Qian Liyan, Zhang Yao, Yuan Ye, Zhang Zhaoke, Fang Jin, Jiang Shimiao, Zhang Jin, Li Ke, Liu Beijiang, Xu Chenglin, Zhang Yifan, Jia Xiaoqian, Qin Xiaoshuai, Huang Xingtao
Main category: cs.LG
TL;DR: A Monte Carlo dataset for track reconstruction with standardized metrics and benchmark results for traditional algorithms and Graph Neural Networks.
Details
Motivation: To advance ML-based track reconstruction by providing a standardized dataset and evaluation framework for reproducible research and comparison between traditional and ML methods.Method: Created a Monte Carlo dataset of single- and two-track drift chamber events, defined track reconstruction-specific metrics, and benchmarked both traditional algorithms and Graph Neural Networks approaches.
Result: Established a standardized evaluation framework with specific metrics for track reconstruction, enabling reproducible validation and comparison of different methods on the same dataset.
Conclusion: The dataset and evaluation framework provide a foundation for rigorous, comparable research in ML-based track reconstruction, facilitating future advancements in the field.
Abstract: We introduce a Monte Carlo (MC) dataset of single- and two-track drift chamber events to advance Machine Learning (ML)-based track reconstruction. To enable standardized and comparable evaluation, we define track reconstruction specific metrics and report results for traditional track reconstruction algorithms and a Graph Neural Networks (GNNs) method, facilitating rigorous, reproducible validation for future research.
[667] RNM-TD3: N:M Semi-structured Sparse Reinforcement Learning From Scratch
Isam Vrce, Andreas Kassler, Gökçe Aydos
Main category: cs.LG
TL;DR: First study applying N:M structured sparsity to reinforcement learning, achieving competitive performance with up to 87.5% sparsity while enabling hardware acceleration.
Details
Motivation: Existing sparsity methods in DRL use unstructured fine-grained sparsity that limits hardware acceleration due to irregular computation patterns, while structured coarse-grained sparsity typically degrades performance and increases pruning complexity.Method: Proposes RNM-TD3 framework that enforces row-wise N:M sparsity throughout training for all networks in off-policy RL (TD3), maintaining compatibility with accelerators that support N:M sparse matrix operations.
Result: RNM-TD3 outperforms dense counterpart at 50%-75% sparsity (2:4 and 1:4), achieving up to 14% performance increase at 2:4 sparsity on Ant environment, and remains competitive even at 87.5% sparsity (1:8).
Conclusion: N:M structured sparsity effectively balances compression, performance, and hardware efficiency in RL, enabling potential training speedups while maintaining or improving performance.
Abstract: Sparsity is a well-studied technique for compressing deep neural networks (DNNs) without compromising performance. In deep reinforcement learning (DRL), neural networks with up to 5% of their original weights can still be trained with minimal performance loss compared to their dense counterparts. However, most existing methods rely on unstructured fine-grained sparsity, which limits hardware acceleration opportunities due to irregular computation patterns. Structured coarse-grained sparsity enables hardware acceleration, yet typically degrades performance and increases pruning complexity. In this work, we present, to the best of our knowledge, the first study on N:M structured sparsity in RL, which balances compression, performance, and hardware efficiency. Our framework enforces row-wise N:M sparsity throughout training for all networks in off-policy RL (TD3), maintaining compatibility with accelerators that support N:M sparse matrix operations. Experiments on continuous-control benchmarks show that RNM-TD3, our N:M sparse agent, outperforms its dense counterpart at 50%-75% sparsity (e.g., 2:4 and 1:4), achieving up to a 14% increase in performance at 2:4 sparsity on the Ant environment. RNM-TD3 remains competitive even at 87.5% sparsity (1:8), while enabling potential training speedups.
[668] Replicable Constrained Bandits
Matteo Bollini, Gianmarco Genalti, Francesco Emanuele Stradi, Matteo Castiglioni, Alberto Marchesi
Main category: cs.LG
TL;DR: Replicable algorithms for constrained multi-armed bandit problems that maintain decision consistency across executions while matching non-replicable algorithms’ regret and constraint violation bounds.
Details
Motivation: Address the need for reproducible experiments in machine learning by studying algorithmic replicability in constrained multi-armed bandit problems, where decisions should be consistent across different executions in the same environment.Method: Develop replicable algorithms for constrained MABs, including designing the first replicable UCB-like algorithm for unconstrained MABs as a key building block, showing that optimism-in-the-face-of-uncertainty algorithms can be made replicable.
Result: Replicability can be achieved in constrained MABs with algorithms whose regret and constraint violation match those of non-replicable ones in terms of time horizon T.
Conclusion: Algorithmic replicability is feasible for constrained multi-armed bandit problems without sacrificing performance guarantees, with the replicable UCB algorithm being a key technical contribution of independent interest.
Abstract: Algorithmic \emph{replicability} has recently been introduced to address the need for reproducible experiments in machine learning. A \emph{replicable online learning} algorithm is one that takes the same sequence of decisions across different executions in the same environment, with high probability. We initiate the study of algorithmic replicability in \emph{constrained} MAB problems, where a learner interacts with an unknown stochastic environment for $T$ rounds, seeking not only to maximize reward but also to satisfy multiple constraints. Our main result is that replicability can be achieved in constrained MABs. Specifically, we design replicable algorithms whose regret and constraint violation match those of non-replicable ones in terms of $T$. As a key step toward these guarantees, we develop the first replicable UCB-like algorithm for \emph{unconstrained} MABs, showing that algorithms that employ the optimism in-the-face-of-uncertainty principle can be replicable, a result that we believe is of independent interest.
[669] Decoupled Continuous-Time Reinforcement Learning via Hamiltonian Flow
Minh Nguyen
Main category: cs.LG
TL;DR: A novel decoupled continuous-time actor-critic algorithm for RL in continuous-time control problems with non-uniform event-driven decisions, outperforming prior methods on benchmarks and real-world trading.
Details
Motivation: Standard discrete-time RL struggles with continuous-time control problems where time gaps shrink, causing Q-functions to collapse to value functions and eliminating action ranking. Existing continuous-time methods have complex optimization problems that are difficult to train reliably.Method: Proposes a decoupled continuous-time actor-critic algorithm with alternating updates: q is learned from diffusion generators on V, and V is updated via a Hamiltonian-based value flow that remains informative under infinitesimal time steps.
Result: Method outperforms prior continuous-time and leading discrete-time baselines across continuous-control benchmarks and a real-world trading task, achieving 21% profit over a single quarter (nearly doubling the second-best method).
Conclusion: The proposed decoupled approach provides a more reliable and effective solution for continuous-time RL problems, with theoretical convergence guarantees and strong empirical performance.
Abstract: Many real-world control problems, ranging from finance to robotics, evolve in continuous time with non-uniform, event-driven decisions. Standard discrete-time reinforcement learning (RL), based on fixed-step Bellman updates, struggles in this setting: as time gaps shrink, the $Q$-function collapses to the value function $V$, eliminating action ranking. Existing continuous-time methods reintroduce action information via an advantage-rate function $q$. However, they enforce optimality through complicated martingale losses or orthogonality constraints, which are sensitive to the choice of test processes. These approaches entangle $V$ and $q$ into a large, complex optimization problem that is difficult to train reliably. To address these limitations, we propose a novel decoupled continuous-time actor-critic algorithm with alternating updates: $q$ is learned from diffusion generators on $V$, and $V$ is updated via a Hamiltonian-based value flow that remains informative under infinitesimal time steps, where standard max/softmax backups fail. Theoretically, we prove rigorous convergence via new probabilistic arguments, sidestepping the challenge that generator-based Hamiltonians lack Bellman-style contraction under the sup-norm. Empirically, our method outperforms prior continuous-time and leading discrete-time baselines across challenging continuous-control benchmarks and a real-world trading task, achieving 21% profit over a single quarter$-$nearly doubling the second-best method.
[670] OPBench: A Graph Benchmark to Combat the Opioid Crisis
Tianyi Ma, Yiyang Li, Yiyue Qian, Zheyuan Zhang, Zehong Wang, Chuxu Zhang, Yanfang Ye
Main category: cs.LG
TL;DR: OPBench: First comprehensive benchmark for evaluating graph learning methods on opioid crisis applications across five datasets in three domains.
Details
Motivation: The opioid epidemic demands computational solutions, and graph learning methods show promise but lack systematic evaluation benchmarks for real-world opioid crisis scenarios.Method: Created OPBench with five datasets across three domains: opioid overdose detection from healthcare claims, illicit drug trafficking detection from digital platforms, and drug misuse prediction from dietary patterns. Uses heterogeneous graphs and hypergraphs to preserve complex relational information, with expert-curated data following privacy/ethical guidelines.
Result: Established unified evaluation framework with standardized protocols, predefined data splits, and reproducible baselines. Extensive experiments analyzed strengths/limitations of existing graph learning methods.
Conclusion: OPBench provides actionable insights for future research in combating opioid crisis through graph learning, with code and datasets publicly available.
Abstract: The opioid epidemic continues to ravage communities worldwide, straining healthcare systems, disrupting families, and demanding urgent computational solutions. To combat this lethal opioid crisis, graph learning methods have emerged as a promising paradigm for modeling complex drug-related phenomena. However, a significant gap remains: there is no comprehensive benchmark for systematically evaluating these methods across real-world opioid crisis scenarios. To bridge this gap, we introduce OPBench, the first comprehensive opioid benchmark comprising five datasets across three critical application domains: opioid overdose detection from healthcare claims, illicit drug trafficking detection from digital platforms, and drug misuse prediction from dietary patterns. Specifically, OPBench incorporates diverse graph structures, including heterogeneous graphs and hypergraphs, to preserve the rich and complex relational information among drug-related data. To address data scarcity, we collaborate with domain experts and authoritative institutions to curate and annotate datasets while adhering to privacy and ethical guidelines. Furthermore, we establish a unified evaluation framework with standardized protocols, predefined data splits, and reproducible baselines to facilitate fair and systematic comparison among graph learning methods. Through extensive experiments, we analyze the strengths and limitations of existing graph learning methods, thereby providing actionable insights for future research in combating the opioid crisis. Our source code and datasets are available at https://github.com/Tianyi-Billy-Ma/OPBench.
[671] Concepts’ Information Bottleneck Models
Karim Galliamov, Syed M Ahsan Kazmi, Adil Khan, Adín Ramírez Rivera
Main category: cs.LG
TL;DR: Information Bottleneck regularization improves Concept Bottleneck Models by enforcing minimal-sufficient concept representations, enhancing both accuracy and faithfulness of concept interventions.
Details
Motivation: Concept Bottleneck Models (CBMs) aim for interpretable predictions through human-understandable concepts, but suffer from reduced accuracy and concept leakage that undermines faithfulness. There's a need to improve CBMs while maintaining their interpretability benefits.Method: Introduces an explicit Information Bottleneck regularizer on the concept layer that penalizes mutual information between input and concepts I(X;C) while preserving task-relevant information I(C;Y). Derives two practical variants: a variational objective and an entropy-based surrogate, integrated into standard CBM training without architectural changes or additional supervision.
Result: IB-regularized models consistently outperform vanilla counterparts across six CBM families and three benchmarks. Information-plane analyses confirm the intended behavior, showing improved predictive performance and reliability of concept-level interventions.
Conclusion: Enforcing a minimal-sufficient concept bottleneck improves both predictive performance and reliability of concept interventions. The regularizer offers a theoretically-grounded, architecture-agnostic path to more faithful and intervenable CBMs, resolving prior evaluation inconsistencies.
Abstract: Concept Bottleneck Models (CBMs) aim to deliver interpretable predictions by routing decisions through a human-understandable concept layer, yet they often suffer reduced accuracy and concept leakage that undermines faithfulness. We introduce an explicit Information Bottleneck regularizer on the concept layer that penalizes $I(X;C)$ while preserving task-relevant information in $I(C;Y)$, encouraging minimal-sufficient concept representations. We derive two practical variants (a variational objective and an entropy-based surrogate) and integrate them into standard CBM training without architectural changes or additional supervision. Evaluated across six CBM families and three benchmarks, the IB-regularized models consistently outperform their vanilla counterparts. Information-plane analyses further corroborate the intended behavior. These results indicate that enforcing a minimal-sufficient concept bottleneck improves both predictive performance and the reliability of concept-level interventions. The proposed regularizer offers a theoretic-grounded, architecture-agnostic path to more faithful and intervenable CBMs, resolving prior evaluation inconsistencies by aligning training protocols and demonstrating robust gains across model families and datasets.
[672] An Embarrassingly Simple Way to Optimize Orthogonal Matrices at Scale
Adrián Javaloy, Antonio Vergari
Main category: cs.LG
TL;DR: POGO is a fast GPU-friendly optimizer for orthogonal constraints that improves upon the Landing algorithm, enabling modern adaptive optimizers while maintaining orthogonality with minimal hyperparameters.
Details
Motivation: Orthogonality constraints are important in robust and probabilistic ML but current optimizers are computationally expensive and don't scale to problems with hundreds/thousands of constraints. The Landing algorithm exists but temporarily relaxes orthogonality.Method: POGO revisits and improves on Landing algorithm ideas, enabling inclusion of modern adaptive optimizers while ensuring orthogonal constraints are effectively met. The algorithm is fast and GPU-friendly with only 5 matrix products, maintaining orthogonality at all times.
Result: POGO greatly outperforms recent optimizers on challenging benchmarks, can optimize problems with thousands of orthogonal matrices in minutes (vs hours for alternatives), and maintains orthogonality at all times with little to no additional cost.
Conclusion: POGO sets a milestone for exploiting orthogonality constraints in ML at scale, providing a practical solution for large-scale orthogonal optimization problems.
Abstract: Orthogonality constraints are ubiquitous in robust and probabilistic machine learning. Unfortunately, current optimizers are computationally expensive and do not scale to problems with hundreds or thousands of constraints. One notable exception is the Landing algorithm (Ablin et al., 2024) which, however comes at the expense of temporarily relaxing orthogonality. In this work, we revisit and improve on the ideas behind Landing, enabling the inclusion of modern adaptive optimizers while ensuring that orthogonal constraints are effectively met. Remarkably, these improvements come at little to no cost, and reduce the number of required hyperparemeters. Our algorithm POGO is fast and GPU-friendly, consisting of only 5 matrix products, and in practice maintains orthogonality at all times. On several challenging benchmarks, POGO greatly outperforms recent optimizers and shows it can optimize problems with thousands of orthogonal matrices in minutes while alternatives would take hours. As such, POGO sets a milestone to finally exploit orthogonality constraints in ML at scale. A PyTorch implementation of POGO is publicly available at https://github.com/adrianjav/pogo.
[673] Pseudo-differential-enhanced physics-informed neural networks
Andrew Gracyk
Main category: cs.LG
TL;DR: Pseudo-differential enhanced PINNs extend gradient enhancement to Fourier space, using Fourier transforms to improve training efficiency and handle fractional derivatives while maintaining mesh flexibility.
Details
Motivation: The paper aims to improve Physics-Informed Neural Networks (PINNs) by addressing training challenges like slow convergence, frequency bias, and difficulty with high-frequency components. Gradient enhancement helps but has limitations, so the authors propose Fourier-space enhancement for better spectral properties and broader applicability.Method: Extends gradient enhancement to Fourier space by applying Fourier transforms to PDE residuals, then multiplying by Fourier wavenumbers (equivalent to differentiation in Fourier space). Uses Fast Fourier Transform (FFT) for efficiency but also supports Monte Carlo methods for mesh flexibility on various domains. Analyzes effects on Neural Tangent Kernel (NTK) spectral properties.
Result: The method achieves superior PINN vs numerical error in fewer training iterations, breaks training plateaus in low collocation settings, handles fractional derivatives, improves NTK spectral eigenvalue decay, and mitigates frequency bias. Works with Fourier feature embeddings and maintains mesh flexibility.
Conclusion: Pseudo-differential enhancement in Fourier space provides an effective extension to gradient-enhanced PINNs, offering improved training efficiency, better spectral properties, and broader applicability including fractional derivatives and various domain types.
Abstract: We present pseudo-differential enhanced physics-informed neural networks (PINNs), an extension of gradient enhancement but in Fourier space. Gradient enhancement of PINNs dictates that the PDE residual is taken to a higher differential order than prescribed by the PDE, added to the objective as an augmented term in order to improve training and overall learning fidelity. We propose the same procedure after application via Fourier transforms, since differentiating in Fourier space is multiplication with the Fourier wavenumber under suitable decay. Our methods are fast and efficient. Our methods oftentimes achieve superior PINN versus numerical error in fewer training iterations, potentially pair well with few samples in collocation, and can on occasion break plateaus in low collocation settings. Moreover, our methods are suitable for fractional derivatives. We establish that our methods improve spectral eigenvalue decay of the neural tangent kernel (NTK), and so our methods contribute towards the learning of high frequencies in early training, mitigating the effects of frequency bias up to the polynomial order and possibly greater with smooth activations. Our methods accommodate advanced techniques in PINNs, such as Fourier feature embeddings. A pitfall of discrete Fourier transforms via the Fast Fourier Transform (FFT) is mesh subjugation, and so we demonstrate compatibility of our methods for greater mesh flexibility and invariance on alternative Euclidean and non-Euclidean domains via Monte Carlo methods and otherwise.
[674] Exposing Diversity Bias in Deep Generative Models: Statistical Origins and Correction of Diversity Error
Farzan Farnia, Mohammad Jalali, Azim Ospanov
Main category: cs.LG
TL;DR: Deep generative models systematically underestimate data distribution diversity compared to test samples, as measured by Vendi and RKE entropy-based scores, due to finite-sample bias in diversity estimation.
Details
Motivation: While deep generative models excel at producing high-quality samples, there's limited systematic study on whether they faithfully capture the full diversity of underlying data distributions. The paper aims to investigate potential diversity biases in state-of-the-art generative models.Method: The authors compare diversity of generated samples vs. test samples using reference-free entropy-based diversity scores (Vendi and RKE). They analyze finite-sample behavior of these scores and show expected values increase with sample size, explaining the diversity underestimation. They also discuss diversity-aware regularization strategies.
Result: Across multiple benchmark datasets, test data consistently achieves substantially higher Vendi and RKE diversity scores than generated samples, revealing systematic downward diversity bias in modern generative models. The finite-sample analysis shows diversity estimated from training sets inherently underestimates true distribution diversity.
Conclusion: Generative models optimized to minimize divergence to empirical data distributions inherently lose diversity due to finite-sample bias in diversity estimation. The paper proposes diversity-aware regularization and guidance strategies based on Vendi and RKE as potential solutions to mitigate this bias.
Abstract: Deep generative models have achieved great success in producing high-quality samples, making them a central tool across machine learning applications. Beyond sample quality, an important yet less systematically studied question is whether trained generative models faithfully capture the diversity of the underlying data distribution. In this work, we address this question by directly comparing the diversity of samples generated by state-of-the-art models with that of test samples drawn from the target data distribution, using recently proposed reference-free entropy-based diversity scores, Vendi and RKE. Across multiple benchmark datasets, we find that test data consistently attains substantially higher Vendi and RKE diversity scores than the generated samples, suggesting a systematic downward diversity bias in modern generative models. To understand the origin of this bias, we analyze the finite-sample behavior of entropy-based diversity scores and show that their expected values increase with sample size, implying that diversity estimated from finite training sets could inherently underestimate the diversity of the true distribution. As a result, optimizing the generators to minimize divergence to empirical data distributions would induce a loss of diversity. Finally, we discuss potential diversity-aware regularization and guidance strategies based on Vendi and RKE as principled directions for mitigating this bias, and provide empirical evidence suggesting their potential to improve the results.
[675] SynthSAEBench: Evaluating Sparse Autoencoders on Scalable Realistic Synthetic Data
David Chanin, Adrià Garriga-Alonso
Main category: cs.LG
TL;DR: SynthSAEBench: A synthetic benchmark toolkit for evaluating Sparse Autoencoder architectures with realistic feature characteristics, enabling precise comparison and diagnosis of SAE failure modes.
Details
Motivation: Current SAE benchmarks on LLMs are too noisy to differentiate architectural improvements, and synthetic data experiments are too small-scale and unrealistic for meaningful comparisons. There's a need for controlled benchmarks with ground-truth features.Method: Introduces SynthSAEBench toolkit for generating large-scale synthetic data with realistic feature characteristics (correlation, hierarchy, superposition) and a standardized benchmark model (SynthSAEBench-16k) for direct SAE architecture comparison.
Result: The benchmark reproduces several LLM SAE phenomena, identifies a new failure mode where Matching Pursuit SAEs exploit superposition noise to improve reconstruction without learning ground-truth features, and shows that more expressive encoders can easily overfit.
Conclusion: SynthSAEBench complements LLM benchmarks by providing ground-truth features and controlled ablations, enabling precise diagnosis of SAE failure modes and validation of architectural improvements before scaling to LLMs.
Abstract: Improving Sparse Autoencoders (SAEs) requires benchmarks that can precisely validate architectural innovations. However, current SAE benchmarks on LLMs are often too noisy to differentiate architectural improvements, and current synthetic data experiments are too small-scale and unrealistic to provide meaningful comparisons. We introduce SynthSAEBench, a toolkit for generating large-scale synthetic data with realistic feature characteristics including correlation, hierarchy, and superposition, and a standardized benchmark model, SynthSAEBench-16k, enabling direct comparison of SAE architectures. Our benchmark reproduces several previously observed LLM SAE phenomena, including the disconnect between reconstruction and latent quality metrics, poor SAE probing results, and a precision-recall trade-off mediated by L0. We further use our benchmark to identify a new failure mode: Matching Pursuit SAEs exploit superposition noise to improve reconstruction without learning ground-truth features, suggesting that more expressive encoders can easily overfit. SynthSAEBench complements LLM benchmarks by providing ground-truth features and controlled ablations, enabling researchers to precisely diagnose SAE failure modes and validate architectural improvements before scaling to LLMs.
[676] A Critical Look at Targeted Instruction Selection: Disentangling What Matters (and What Doesn’t)
Nihal V. Nayak, Paula Rodriguez-Diaz, Neha Hulkund, Sara Beery, David Alvarez-Melis
Main category: cs.LG
TL;DR: Systematic analysis of instruction selection methods for LLM fine-tuning, focusing on data representation and selection algorithms, with gradient-based representations showing best performance at low budgets.
Details
Motivation: The literature on targeted instruction selection for LLM fine-tuning is fragmented and opaque, with methods varying widely and lacking clear guidance for practitioners. The authors aim to bring clarity by systematically analyzing core components.Method: Disentangles and analyzes two core ingredients: data representation and selection algorithms. Creates a framework for controlled comparisons across models, tasks, and budgets. Unifies existing selection algorithms as forms of approximate distance minimization.
Result: Gradient-based data representations consistently predict performance across datasets and models. Gradient-based representations with greedy round-robin selection perform best at low budgets, but benefits diminish at larger budgets.
Conclusion: Provides critical insights and foundation for more principled data selection in LLM fine-tuning, with gradient-based representations being particularly effective for targeted instruction selection.
Abstract: Instruction fine-tuning of large language models (LLMs) often involves selecting a subset of instruction training data from a large candidate pool, using a small query set from the target task. Despite growing interest, the literature on targeted instruction selection remains fragmented and opaque: methods vary widely in selection budgets, often omit zero-shot baselines, and frequently entangle the contributions of key components. As a result, practitioners lack actionable guidance on selecting instructions for their target tasks. In this work, we aim to bring clarity to this landscape by disentangling and systematically analyzing the two core ingredients: data representation and selection algorithms. Our framework enables controlled comparisons across models, tasks, and budgets. We find that only gradient-based data representations choose subsets whose similarity to the query consistently predicts performance across datasets and models. While no single method dominates, gradient-based representations paired with a greedy round-robin selection algorithm tend to perform best on average at low budgets, but these benefits diminish at larger budgets. Finally, we unify several existing selection algorithms as forms of approximate distance minimization between the selected subset and the query set, and support this view with new generalization bounds. More broadly, our findings provide critical insights and a foundation for more principled data selection in LLM fine-tuning. The code is available at https://github.com/dcml-lab/targeted-instruction-selection.
[677] Unbiased Approximate Vector-Jacobian Products for Efficient Backpropagation
Killian Bakong, Laurent Massoulié, Edouard Oyallon, Kevin Scaman
Main category: cs.LG
TL;DR: Proposes randomized unbiased approximations of vector-jacobian products during backpropagation to reduce computational and memory costs of training deep neural networks, with theoretical analysis and experiments on various architectures.
Details
Motivation: Training deep neural networks is computationally expensive and memory-intensive, especially for large models. The authors aim to reduce these costs by approximating the backpropagation process through randomized methods while maintaining training effectiveness.Method: Replace exact vector-jacobian products with randomized, unbiased approximations during backpropagation. The authors identify specific unbiased estimates with minimal variance under sparsity constraints and provide theoretical analysis of the trade-off between training epochs and cost reduction.
Result: Theoretical analysis shows optimality properties of the proposed unbiased estimates. Experiments on multi-layer perceptrons, BagNets, and Visual Transformers validate the theoretical results and demonstrate potential for significant cost reduction in deep learning training.
Conclusion: The proposed unbiased randomized backpropagation approach effectively reduces computational and memory costs of training deep neural networks while maintaining training effectiveness, with validated results across different architectures.
Abstract: In this work we introduce methods to reduce the computational and memory costs of training deep neural networks. Our approach consists in replacing exact vector-jacobian products by randomized, unbiased approximations thereof during backpropagation. We provide a theoretical analysis of the trade-off between the number of epochs needed to achieve a target precision and the cost reduction for each epoch. We then identify specific unbiased estimates of vector-jacobian products for which we establish desirable optimality properties of minimal variance under sparsity constraints. Finally we provide in-depth experiments on multi-layer perceptrons, BagNets and Visual Transfomers architectures. These validate our theoretical results, and confirm the potential of our proposed unbiased randomized backpropagation approach for reducing the cost of deep learning.
[678] D2-LoRA: A Synergistic Approach to Differential and Directional Low-Rank Adaptation
Nozomu Fujisawa, Masaaki Kondo
Main category: cs.LG
TL;DR: D2-LoRA: A parameter-efficient fine-tuning method with signed low-rank residual updates and train-time column-wise projection that achieves better accuracy than LoRA while preserving algebraic mergeability with zero inference latency.
Details
Motivation: To systematically investigate parameter-efficient fine-tuning design under practical data and compute constraints, aiming to improve performance while maintaining mergeability for efficient inference.Method: Combines signed low-rank residual updates (additive and subtractive components) with train-time column-wise projection that keeps each column close to its original norm. After training, adapters are merged into a single weight matrix.
Result: Achieves 76.4% average accuracy across 8 QA/reading comprehension benchmarks using only 5k samples per task, improves over LoRA by 2.2 percentage points, matches/exceeds DoRA, improves generative tasks (+1.2 ROUGE-L, +1.1% win rate), with 36% lower training volatility and 1.91x evaluation throughput.
Conclusion: D2-LoRA provides an effective parameter-efficient fine-tuning method that balances performance, training stability, and inference efficiency through architectural innovations in low-rank adaptation.
Abstract: We systematically investigate the parameter-efficient fine-tuning design space under practical data and compute constraints, and propose D2-LoRA. D2-LoRA achieves 76.4 percent average accuracy across eight question answering and reading comprehension benchmarks using only 5k training samples per task and two epochs, while preserving algebraic mergeability at inference with near-exact numerical equivalence. The method combines signed low-rank residual updates with additive and subtractive components, together with a train-time column-wise projection that keeps each column close to its original norm. After training, the adapter is merged into a single weight matrix, adding zero inference latency. Compared with LoRA, D2-LoRA improves average accuracy by 2.2 percentage points; at matched parameter counts (LoRA rank 2r versus D2-LoRA rank r), the improvement is 1.6 points, indicating gains from architectural design rather than increased parameterization. Compared with DoRA, it matches or exceeds performance on most tasks. Beyond QA and reading comprehension, D2-LoRA improves generative tasks (plus 1.2 ROUGE-L and plus 1.1 percent win rate) and shows 36 percent lower training volatility. The merge preserves numerical fidelity (mean gap about 0.03 percentage points) and recovers about 1.91x evaluation throughput. Training overhead is 19 percent, comparable to DoRA, and decreases with longer input sequences. We provide a geometric analysis explaining how the projection stabilizes training, together with ablation studies isolating the contribution of each design component.
[679] Scale redundancy and soft gauge fixing in positively homogeneous neural networks
Rodrigo Carmo Terin
Main category: cs.LG
TL;DR: The paper analyzes gauge symmetries in neural networks with homogeneous activations, showing how scale imbalances affect optimization and proposing a norm-balancing penalty to stabilize training.
Details
Motivation: Neural networks with positively homogeneous activation functions have continuous reparametrization symmetries where neuron-wise rescalings don't change the input-output function. These symmetries create gauge redundancies that affect optimization dynamics, particularly causing scale drift and conditioning issues during training.Method: The authors interpret the rescaling symmetry as gauge redundancy and introduce gauge-adapted coordinates separating invariant and scale-imbalance directions. They propose a soft orbit-selection functional (norm-balancing penalty) acting only on redundant scale coordinates, inspired by gauge fixing in field theory. This penalty induces dissipative relaxation of imbalance modes while preserving the realized function.
Result: The norm-balancing penalty expands the stable learning-rate regime and suppresses scale drift without changing network expressivity. Analytical results show it induces dissipative relaxation of imbalance modes, and controlled experiments demonstrate improved optimization conditioning.
Conclusion: The work establishes a structural link between gauge-orbit geometry and optimization conditioning, providing a concrete connection between gauge-theoretic concepts and machine learning. The approach offers a principled way to handle scale symmetries in neural network training.
Abstract: Neural networks with positively homogeneous activations exhibit an exact continuous reparametrization symmetry: neuron-wise rescalings generate parameter-space orbits along which the input–output function is invariant. We interpret this symmetry as a gauge redundancy and introduce gauge-adapted coordinates that separate invariant and scale-imbalance directions. Inspired by gauge fixing in field theory, we introduce a soft orbit-selection (norm-balancing) functional acting only on redundant scale coordinates. We show analytically that it induces dissipative relaxation of imbalance modes to preserve the realized function. In controlled experiments, this orbit-selection penalty expands the stable learning-rate regime and suppresses scale drift without changing expressivity. These results establish a structural link between gauge-orbit geometry and optimization conditioning, providing a concrete connection between gauge-theoretic concepts and machine learning.
[680] Parameter-Minimal Neural DE Solvers via Horner Polynomials
T. Matulić, D. Seršić
Main category: cs.LG
TL;DR: Horner-factorized polynomial networks for solving differential equations with minimal parameters, enforcing initial conditions exactly and using piecewise extensions for better accuracy.
Details
Motivation: To develop resource-efficient neural architectures for solving differential equations that work well with minimal parameters while maintaining accuracy, addressing the trade-off between model complexity and computational efficiency in scientific modeling.Method: Restrict hypothesis class to Horner-factorized polynomials to create implicit, differentiable trial solutions with few learnable coefficients. Enforce initial conditions exactly by construction. Use piecewise extension with multiple small Horner models on subintervals while enforcing continuity and first-derivative continuity at boundaries.
Result: Horner networks with tens or fewer parameters accurately match solutions and derivatives on ODE benchmarks and heat-equation examples, outperforming small MLP and sinusoidal-representation baselines under same training settings.
Conclusion: Horner networks demonstrate practical accuracy-parameter trade-off for resource-efficient scientific modeling, offering an effective approach for solving differential equations with minimal computational resources.
Abstract: We propose a parameter-minimal neural architecture for solving differential equations by restricting the hypothesis class to Horner-factorized polynomials, yielding an implicit, differentiable trial solution with only a small set of learnable coefficients. Initial conditions are enforced exactly by construction by fixing the low-order polynomial degrees of freedom, so training focuses solely on matching the differential-equation residual at collocation points. To reduce approximation error without abandoning the low-parameter regime, we introduce a piecewise (“spline-like”) extension that trains multiple small Horner models on subintervals while enforcing continuity (and first-derivative continuity) at segment boundaries. On illustrative ODE benchmarks and a heat-equation example, Horner networks with tens (or fewer) parameters accurately match the solution and its derivatives and outperform small MLP and sinusoidal-representation baselines under the same training settings, demonstrating a practical accuracy-parameter trade-off for resource-efficient scientific modeling.
[681] Inner Loop Inference for Pretrained Transformers: Unlocking Latent Capabilities Without Training
Jonathan Lys, Vincent Gripon, Bastien Pasdeloup, Lukas Mauch, Fabien Cardinaux, Ghouthi Boukli Hacene
Main category: cs.LG
TL;DR: Inner looping extends inference-time computation in pretrained Transformers by repeatedly applying selected blocks, yielding modest accuracy improvements through continued refinement.
Details
Motivation: Transformers can be viewed as iterative refinement processes where inner representations evolve across layers. The paper explores whether additional refinement can be obtained at inference time by prolonging the refinement process through repeated application of selected blocks.Method: Proposes inference-time inner looping which repeatedly re-applies a selected block range in pretrained language models. This extends computation in frozen models by allowing tokens to undergo additional refinement cycles beyond the original architecture.
Result: Across multiple benchmarks, inner looping yields modest but consistent accuracy improvements. Analysis shows more stable state evolution and continued semantic refinement in the resulting latent trajectories.
Conclusion: Additional refinement can be obtained through simple test-time looping in frozen pretrained models, suggesting that inner representations can benefit from extended computation beyond the original layer structure.
Abstract: Deep Learning architectures, and in particular Transformers, are conventionally viewed as a composition of layers. These layers are actually often obtained as the sum of two contributions: a residual path that copies the input and the output of a Transformer block. As a consequence, the inner representations (i.e. the input of these blocks) can be interpreted as iterative refinement of a propagated latent representation. Under this lens, many works suggest that the inner space is shared across layers, meaning that tokens can be decoded at early stages. Mechanistic interpretability even goes further by conjecturing that some layers act as refinement layers. Following this path, we propose inference-time inner looping, which prolongs refinement in pretrained off-the-shelf language models by repeatedly re-applying a selected block range. Across multiple benchmarks, inner looping yields modest but consistent accuracy improvements. Analyses of the resulting latent trajectories suggest more stable state evolution and continued semantic refinement. Overall, our results suggest that additional refinement can be obtained through simple test-time looping, extending computation in frozen pretrained models.
[682] Universal Algorithm-Implicit Learning
Stefano Woerner, Seong Joon Oh, Christian F. Baumgartner
Main category: cs.LG
TL;DR: TAIL is a transformer-based universal meta-learner that handles tasks with varying domains, modalities, and label spaces through random projections, label embeddings, and efficient query processing.
Details
Motivation: Current meta-learning methods are limited to narrow task distributions with fixed feature and label spaces, and there's inconsistent terminology in the literature. The authors aim to create a truly universal meta-learner that can handle diverse tasks across different domains and modalities.Method: The authors introduce a theoretical framework for meta-learning with precise definitions, then present TAIL - a transformer-based algorithm-implicit meta-learner with three innovations: random projections for cross-modal feature encoding, random injection label embeddings for extrapolation to larger label spaces, and efficient inline query processing.
Result: TAIL achieves state-of-the-art performance on standard few-shot benchmarks while generalizing to unseen domains and modalities. It solves text classification tasks despite training exclusively on images, handles tasks with up to 20x more classes than seen during training, and provides orders-of-magnitude computational savings over prior transformer-based approaches.
Conclusion: TAIL represents a significant advance in universal meta-learning, demonstrating cross-modal generalization and scalability while establishing a principled theoretical framework for the field.
Abstract: Current meta-learning methods are constrained to narrow task distributions with fixed feature and label spaces, limiting applicability. Moreover, the current meta-learning literature uses key terms like “universal” and “general-purpose” inconsistently and lacks precise definitions, hindering comparability. We introduce a theoretical framework for meta-learning which formally defines practical universality and introduces a distinction between algorithm-explicit and algorithm-implicit learning, providing a principled vocabulary for reasoning about universal meta-learning methods. Guided by this framework, we present TAIL, a transformer-based algorithm-implicit meta-learner that functions across tasks with varying domains, modalities, and label configurations. TAIL features three innovations over prior transformer-based meta-learners: random projections for cross-modal feature encoding, random injection label embeddings that extrapolate to larger label spaces, and efficient inline query processing. TAIL achieves state-of-the-art performance on standard few-shot benchmarks while generalizing to unseen domains. Unlike other meta-learning methods, it also generalizes to unseen modalities, solving text classification tasks despite training exclusively on images, handles tasks with up to 20$\times$ more classes than seen during training, and provides orders-of-magnitude computational savings over prior transformer-based approaches.
[683] Learning Structural Hardness for Combinatorial Auctions: Instance-Dependent Algorithm Selection via Graph Neural Networks
Sungwoo Kang
Main category: cs.LG
TL;DR: A hybrid approach combining ML hardness prediction with specialized solvers for combinatorial auction winner determination, where a lightweight classifier predicts when instances are hard for greedy heuristics and deploys GNN specialists for those cases.
Details
Motivation: Current ML approaches for combinatorial optimization focus on replacing solvers but rarely outperform classical methods. Instead, the paper proposes learning when instances are hard for greedy algorithms to enable intelligent algorithm selection.Method: Design a 20-dimensional structural feature vector and train a lightweight MLP classifier to predict greedy optimality gap. For hard instances (showing “whale-fish” trap structure), deploy a heterogeneous GNN specialist. Create hybrid allocator combining hardness classifier with GNN and greedy solvers.
Result: Hardness classifier achieves mean absolute error 0.033, Pearson correlation 0.937, and 94.7% binary classification accuracy. GNN specialist achieves ≈0% optimality gap on all six adversarial configurations (vs. 3.75-59.24% for greedy). Hybrid allocator achieves 0.51% overall gap on mixed distributions.
Conclusion: Learning when to deploy expensive solvers is more tractable than learning to replace them. GNNs don’t outperform Gurobi on standard benchmarks, motivating algorithm selection approach over solver replacement.
Abstract: The Winner Determination Problem (WDP) in combinatorial auctions is NP-hard, and no existing method reliably predicts which instances will defeat fast greedy heuristics. The ML-for-combinatorial-optimization community has focused on learning to \emph{replace} solvers, yet recent evidence shows that graph neural networks (GNNs) rarely outperform well-tuned classical methods on standard benchmarks. We pursue a different objective: learning to predict \emph{when} a given instance is hard for greedy allocation, enabling instance-dependent algorithm selection. We design a 20-dimensional structural feature vector and train a lightweight MLP hardness classifier that predicts the greedy optimality gap with mean absolute error 0.033, Pearson correlation 0.937, and binary classification accuracy 94.7% across three random seeds. For instances identified as hard – those exhibiting ``whale-fish’’ trap structure where greedy provably fails – we deploy a heterogeneous GNN specialist that achieves ${\approx}0%$ optimality gap on all six adversarial configurations tested (vs.\ 3.75–59.24% for greedy). A hybrid allocator combining the hardness classifier with GNN and greedy solvers achieves 0.51% overall gap on mixed distributions. Our honest evaluation on CATS benchmarks confirms that GNNs do not outperform Gurobi (0.45–0.71 vs.\ 0.20 gap), motivating the algorithm selection framing. Learning \emph{when} to deploy expensive solvers is more tractable than learning to replace them.
[684] On the Stability of Nonlinear Dynamics in GD and SGD: Beyond Quadratic Potentials
Rotem Mulayoff, Sebastian U. Stich
Main category: cs.LG
TL;DR: This paper analyzes the nonlinear dynamics of gradient descent and stochastic gradient descent near minima, showing that linear stability analysis can be misleading and deriving exact criteria for stable oscillations that depend on high-order derivatives.
Details
Motivation: The paper addresses the limitations of linear stability analysis for optimization algorithms like gradient descent and SGD. Prior work often relies on linearization to determine stability near minima, but recent findings show GD can stably oscillate near linearly unstable minima, indicating linear analysis may not capture full nonlinear behavior. The authors aim to explicitly study nonlinear terms to better understand dynamical stability during training.Method: The authors derive an exact criterion for stable oscillations of gradient descent near minima in multivariate settings, which depends on high-order derivatives and generalizes existing results. They extend this analysis to stochastic gradient descent, examining how nonlinear dynamics behave when different batches have varying stability properties. They prove theoretical results about when SGD nonlinear dynamics are stable in expectation.
Result: The paper shows that nonlinear dynamics can diverge in expectation even if only a single batch is unstable, contrary to linear analysis which suggests stability depends on average effects. They prove that if all batches are linearly stable, the nonlinear dynamics of SGD are stable in expectation. The findings reveal that stability can be dictated by a single batch that oscillates unstably rather than average behavior.
Conclusion: Linear stability analysis is insufficient for understanding optimization dynamics near minima. Nonlinear terms play crucial roles, and stability can be determined by extreme cases (single unstable batch) rather than averages. The derived criteria provide more accurate understanding of when optimization algorithms will exhibit stable oscillations or divergence.
Abstract: The dynamical stability of the iterates during training plays a key role in determining the minima obtained by optimization algorithms. For example, stable solutions of gradient descent (GD) correspond to flat minima, which have been associated with favorable features. While prior work often relies on linearization to determine stability, it remains unclear whether linearized dynamics faithfully capture the full nonlinear behavior. Recent work has shown that GD may stably oscillate near a linearly unstable minimum and still converge once the step size decays, indicating that linear analysis can be misleading. In this work, we explicitly study the effect of nonlinear terms. Specifically, we derive an exact criterion for stable oscillations of GD near minima in the multivariate setting. Our condition depends on high-order derivatives, generalizing existing results. Extending the analysis to stochastic gradient descent (SGD), we show that nonlinear dynamics can diverge in expectation even if a single batch is unstable. This implies that stability can be dictated by a single batch that oscillates unstably, rather than an average effect, as linear analysis suggests. Finally, we prove that if all batches are linearly stable, the nonlinear dynamics of SGD are stable in expectation.
[685] Extending Multi-Source Bayesian Optimization With Causality Principles
Luuk Jacobs, Mohammad Ali Javidian
Main category: cs.LG
TL;DR: Proposes Multi-Source Causal Bayesian Optimization (MSCBO) that integrates causal principles with multi-source Bayesian optimization to handle dependent variables and interventions across multiple information sources.
Details
Motivation: Traditional Multi-Source Bayesian Optimization (MSBO) assumes independent variables, limiting effectiveness in scenarios with causal information and interventions (e.g., clinical trials, policy-making). Causal Bayesian Optimization (CBO) handles dependencies but only for single sources. Need to combine strengths of both approaches.Method: Integrates MSBO and CBO methodologies to create Multi-Source Causal Bayesian Optimization (MSCBO). Leverages causal principles for variable dependency modeling while maintaining multi-source optimization capabilities. Theoretical foundations presented with algorithm development.
Result: MSCBO outperforms foundational counterparts on synthetic and real-world datasets with varying noise levels. Shows robustness, facilitates dimensionality reduction, lowers operational costs, improves convergence speed, performance, and scalability.
Conclusion: Integrating MSBO with CBO causality principles enables more efficient optimization in higher-dimensional problems with dependent variables and interventions, particularly valuable for applications like clinical trials and policy-making.
Abstract: Multi-Source Bayesian Optimization (MSBO) serves as a variant of the traditional Bayesian Optimization (BO) framework applicable to situations involving optimization of an objective black-box function over multiple information sources such as simulations, surrogate models, or real-world experiments. However, traditional MSBO assumes the input variables of the objective function to be independent and identically distributed, limiting its effectiveness in scenarios where causal information is available and interventions can be performed, such as clinical trials or policy-making. In the single-source domain, Causal Bayesian Optimization (CBO) extends standard BO with the principles of causality, enabling better modeling of variable dependencies. This leads to more accurate optimization, improved decision-making, and more efficient use of low-cost information sources. In this article, we propose a principled integration of the MSBO and CBO methodologies in the multi-source domain, leveraging the strengths of both to enhance optimization efficiency and reduce computational complexity in higher-dimensional problems. We present the theoretical foundations of both Causal and Multi-Source Bayesian Optimization, and demonstrate how their synergy informs our Multi-Source Causal Bayesian Optimization (MSCBO) algorithm. We compare the performance of MSCBO against its foundational counterparts for both synthetic and real-world datasets with varying levels of noise, highlighting the robustness and applicability of MSCBO. Based on our findings, we conclude that integrating MSBO with the causality principles of CBO facilitates dimensionality reduction and lowers operational costs, ultimately improving convergence speed, performance, and scalability.
[686] Interactionless Inverse Reinforcement Learning: A Data-Centric Framework for Durable Alignment
Elias Malomgré, Pieter Simoens
Main category: cs.LG
TL;DR: Proposes Interactionless Inverse Reinforcement Learning to decouple alignment from policy optimization, creating reusable reward models, plus an Alignment Flywheel for iterative refinement.
Details
Motivation: Current AI alignment methods like RLHF and DPO suffer from structural flaws that entangle safety objectives with agent policies, creating opaque, single-use alignment artifacts called "Alignment Waste."Method: Interactionless Inverse Reinforcement Learning to decouple alignment artifact learning from policy optimization, producing inspectable, editable, model-agnostic reward models. Also introduces Alignment Flywheel - a human-in-the-loop lifecycle for iterative hardening through automated audits and refinement.
Result: Transforms safety from a disposable expense into a durable, verifiable engineering asset by creating reusable alignment artifacts.
Conclusion: Proposes a new architecture that addresses structural flaws in current alignment approaches, making safety objectives inspectable, editable, and reusable across different models.
Abstract: AI alignment is growing in importance, yet current approaches suffer from a critical structural flaw that entangles the safety objectives with the agent’s policy. Methods such as Reinforcement Learning from Human Feedback and Direct Preference Optimization create opaque, single-use alignment artifacts, which we term Alignment Waste. We propose Interactionless Inverse Reinforcement Learning to decouple alignment artifact learning from policy optimization, producing an inspectable, editable, and model-agnostic reward model. Additionally, we introduce the Alignment Flywheel, a human-in-the-loop lifecycle that iteratively hardens the reward model through automated audits and refinement. This architecture transforms safety from a disposable expense into a durable, verifiable engineering asset.
[687] BEACONS: Bounded-Error, Algebraically-Composable Neural Solvers for Partial Differential Equations
Jonathan Gorard, Ammar Hakim, James Juno
Main category: cs.LG
TL;DR: BEACONS framework constructs formally-verified neural network PDE solvers with rigorous error bounds for reliable extrapolation beyond training data.
Details
Motivation: Neural networks struggle to generalize beyond training data convex hulls, which is problematic for computational physics where PDEs often need solving in unvalidated regimes. Current PINN approaches lack formal guarantees for extrapolation.Method: Uses method of characteristics to predict PDE solution properties a priori, constructs rigorous L^inf error bounds for shallow neural approximations, then composes them into deep architectures using compositional deep learning to suppress large errors. Includes automatic code generator and theorem-proving system for correctness certificates.
Result: Applied to linear advection, inviscid Burgers’, and compressible Euler equations in 1D/2D. BEACONS architectures reliably extrapolate solutions far beyond training data with bounded errors, outperforming classical PINNs.
Conclusion: BEACONS provides a framework for constructing formally-verified neural PDE solvers with guaranteed correctness in extrapolatory regimes, addressing key limitations of neural networks in computational physics.
Abstract: The traditional limitations of neural networks in reliably generalizing beyond the convex hulls of their training data present a significant problem for computational physics, in which one often wishes to solve PDEs in regimes far beyond anything which can be experimentally or analytically validated. In this paper, we show how it is possible to circumvent these limitations by constructing formally-verified neural network solvers for PDEs, with rigorous convergence, stability, and conservation properties, whose correctness can therefore be guaranteed even in extrapolatory regimes. By using the method of characteristics to predict the analytical properties of PDE solutions a priori (even in regions arbitrarily far from the training domain), we show how it is possible to construct rigorous extrapolatory bounds on the worst-case L^inf errors of shallow neural network approximations. Then, by decomposing PDE solutions into compositions of simpler functions, we show how it is possible to compose these shallow neural networks together to form deep architectures, based on ideas from compositional deep learning, in which the large L^inf errors in the approximations have been suppressed. The resulting framework, called BEACONS (Bounded-Error, Algebraically-COmposable Neural Solvers), comprises both an automatic code-generator for the neural solvers themselves, as well as a bespoke automated theorem-proving system for producing machine-checkable certificates of correctness. We apply the framework to a variety of linear and non-linear PDEs, including the linear advection and inviscid Burgers’ equations, as well as the full compressible Euler equations, in both 1D and 2D, and illustrate how BEACONS architectures are able to extrapolate solutions far beyond the training data in a reliable and bounded way. Various advantages of the approach over the classical PINN approach are discussed.
[688] A Pragmatic Method for Comparing Clusterings with Overlaps and Outliers
Ryan DeWolfe, Paweł Prałat, François Théberge
Main category: cs.LG
TL;DR: A novel similarity measure for comparing clusterings that handles outliers and overlapping clusters, addressing limitations of existing comparison methods.
Details
Motivation: Existing clustering comparison methods fail to handle real-world scenarios where clusterings may contain outliers (objects belonging to no cluster) or overlapping clusters (objects belonging to multiple clusters), limiting their practical utility for evaluating clustering algorithms.Method: The authors define a pragmatic similarity measure specifically designed to compare clusterings with both outliers and overlapping clusters. They establish theoretical properties of this measure and validate it experimentally against common biases.
Result: The proposed similarity measure demonstrates desirable theoretical properties and is experimentally shown to be resistant to common biases that affect other clustering comparison measures.
Conclusion: This work provides a practical solution for comparing clusterings in realistic scenarios with outliers and overlaps, filling an important gap in clustering evaluation methodology.
Abstract: Clustering algorithms are an essential part of the unsupervised data science ecosystem, and extrinsic evaluation of clustering algorithms requires a method for comparing the detected clustering to a ground truth clustering. In a general setting, the detected and ground truth clusterings may have outliers (objects belonging to no cluster), overlapping clusters (objects may belong to more than one cluster), or both, but methods for comparing these clusterings are currently undeveloped. In this note, we define a pragmatic similarity measure for comparing clusterings with overlaps and outliers, show that it has several desirable properties, and experimentally confirm that it is not subject to several common biases afflicting other clustering comparison measures.
[689] Goldilocks RL: Tuning Task Difficulty to Escape Sparse Rewards for Reasoning
Ilia Mahrooghi, Aryo Lotfi, Emmanuel Abbe
Main category: cs.LG
TL;DR: Goldilocks: Teacher-driven data sampling strategy that selects questions of appropriate difficulty (neither too easy nor too hard) for student LLMs during reinforcement learning training with GRPO, improving sample efficiency.
Details
Motivation: Reinforcement learning for reasoning in LLMs suffers from sample inefficiency due to sparse rewards. Classic curriculum learning helps but determining the right difficulty ordering for specific models is challenging.Method: Proposes Goldilocks: a teacher model predicts question difficulty for the student model and selects appropriately challenging questions (Goldilocks principle). Teacher continuously adapts to student’s evolving abilities based on performance, while student is trained with GRPO (Group Relative Policy Optimization).
Result: On OpenMathReasoning dataset, Goldilocks data sampling improves performance of models trained with standard GRPO under the same compute budget.
Conclusion: Goldilocks provides an effective teacher-driven sampling strategy that enhances RL training efficiency for reasoning tasks by dynamically selecting questions of appropriate difficulty for the student model.
Abstract: Reinforcement learning has emerged as a powerful paradigm for unlocking reasoning capabilities in large language models. However, relying on sparse rewards makes this process highly sample-inefficient, as models must navigate vast search spaces with minimal feedback. While classic curriculum learning aims to mitigate this by ordering data based on complexity, the right ordering for a specific model is often unclear. To address this, we propose Goldilocks, a novel teacher-driven data sampling strategy that aims to predict each question’s difficulty for the student model. The teacher model selects questions of appropriate difficulty for the student model, i.e., questions that are neither too easy nor too hard (Goldilocks principle), while training the student with GRPO. By leveraging the student’s performance on seen samples, the teacher continuously adapts to the student’s evolving abilities. On OpenMathReasoning dataset, Goldilocks data sampling improves the performance of models trained with standard GRPO under the same compute budget.
[690] On the Learning Dynamics of RLVR at the Edge of Competence
Yu Huang, Zixin Wen, Yuejie Chi, Yuting Wei, Aarti Singh, Yingbin Liang, Yuxin Chen
Main category: cs.LG
TL;DR: RLVR (Reinforcement Learning with Verifiable Rewards) helps transformers overcome long-horizon reasoning barriers through smooth difficulty progression rather than abrupt jumps, enabling steady improvement via a relay effect.
Details
Motivation: To understand how reward signals based solely on final outcomes can help large reasoning models overcome long-horizon barriers in compositional reasoning tasks, despite the sparse nature of such rewards.Method: Developed a theory of training dynamics for transformers on compositional reasoning tasks using RLVR, analyzing how effectiveness depends on the smoothness of difficulty spectrum. Used Fourier analysis on finite groups to analyze training dynamics and validated predictions with synthetic experiments.
Result: When difficulty spectrum is smooth, RLVR produces a relay effect where persistent gradient signals on easier problems elevate capabilities to tackle harder ones, leading to steady improvement. Abrupt difficulty discontinuities cause grokking-type phase transitions with prolonged plateaus.
Conclusion: RLVR can effectively improve performance at the edge of competence when data mixtures have appropriately designed smooth difficulty progression, explaining how sparse final-outcome rewards can overcome long-horizon reasoning barriers.
Abstract: Reinforcement learning with verifiable rewards (RLVR) has been a main driver of recent breakthroughs in large reasoning models. Yet it remains a mystery how rewards based solely on final outcomes can help overcome the long-horizon barrier to extended reasoning. To understand this, we develop a theory of the training dynamics of RL for transformers on compositional reasoning tasks. Our theory characterizes how the effectiveness of RLVR is governed by the smoothness of the difficulty spectrum. When data contains abrupt discontinuities in difficulty, learning undergoes grokking-type phase transitions, producing prolonged plateaus before progress recurs. In contrast, a smooth difficulty spectrum leads to a relay effect: persistent gradient signals on easier problems elevate the model’s capabilities to the point where harder ones become tractable, resulting in steady and continuous improvement. Our theory explains how RLVR can improve performance at the edge of competence, and suggests that appropriately designed data mixtures can yield scalable gains. As a technical contribution, our analysis develops and adapts tools from Fourier analysis on finite groups to our setting. We validate the predicted mechanisms empirically via synthetic experiments.
[691] Web-Scale Multimodal Summarization using CLIP-Based Semantic Alignment
Mounvik K, N Harshit
Main category: cs.LG
TL;DR: Web-Scale Multimodal Summarization: A lightweight framework that retrieves text and images from web sources and generates summaries using fine-tuned CLIP for semantic alignment and optional BLIP captioning for multimodal coherence.
Details
Motivation: To create a practical, deployable tool for generating multimodal summaries by combining web-retrieved text and image data, addressing the need for systems that can effectively integrate language, retrieval, and vision models for web-scale information synthesis.Method: Parallel web, news, and image searches for user-defined topics; fine-tuned CLIP model ranks retrieved images based on semantic alignment with topic and text; optional BLIP captioning enables image-only summaries; configurable pipeline with adjustable fetch limits, semantic filtering, and summary styling.
Result: Evaluation on 500 image-caption pairs with 20:1 contrastive negatives shows strong performance: ROC-AUC of 0.9270, F1-score of 0.6504, and accuracy of 96.99%, demonstrating effective multimodal alignment. The system is exposed via Gradio-based API with controllable parameters.
Conclusion: The framework provides a configurable, deployable tool for web-scale multimodal summarization that successfully integrates language, retrieval, and vision models in a user-extensible pipeline, offering practical applications for information synthesis.
Abstract: We introduce Web-Scale Multimodal Summarization, a lightweight framework for generating summaries by combining retrieved text and image data from web sources. Given a user-defined topic, the system performs parallel web, news, and image searches. Retrieved images are ranked using a fine-tuned CLIP model to measure semantic alignment with topic and text. Optional BLIP captioning enables image-only summaries for stronger multimodal coherence.The pipeline supports features such as adjustable fetch limits, semantic filtering, summary styling, and downloading structured outputs. We expose the system via a Gradio-based API with controllable parameters and preconfigured presets.Evaluation on 500 image-caption pairs with 20:1 contrastive negatives yields a ROC-AUC of 0.9270, an F1-score of 0.6504, and an accuracy of 96.99%, demonstrating strong multimodal alignment. This work provides a configurable, deployable tool for web-scale summarization that integrates language, retrieval, and vision models in a user-extensible pipeline.
[692] Algorithmic Simplification of Neural Networks with Mosaic-of-Motifs
Pedram Bakhtiarifard, Tong Chen, Jonathan Wenshøj, Erik B Dam, Raghavendra Selvan
Main category: cs.LG
TL;DR: MoMos method reduces algorithmic complexity of neural networks by partitioning parameters into blocks and reusing motifs, enabling compression while maintaining performance.
Details
Motivation: To explain why deep neural networks are well-suited for compression by analyzing their algorithmic complexity, hypothesizing that trained models have more structure and lower algorithmic complexity than random initialization.Method: Introduces Mosaic-of-Motifs (MoMos) method that partitions model parameters into blocks of size s, restricts each block to be selected from a set of k reusable motifs with a reuse pattern, creating algorithmically simpler parameterization.
Result: Empirical evidence shows algorithmic complexity of neural networks can be reduced during training, resulting in models that perform comparably with unconstrained models while being algorithmically simpler.
Conclusion: Deep neural networks are compressible because trained parameters exhibit lower algorithmic complexity than random initialization, and compression methods harness this reduced complexity.
Abstract: Large-scale deep learning models are well-suited for compression. Methods like pruning, quantization, and knowledge distillation have been used to achieve massive reductions in the number of model parameters, with marginal performance drops across a variety of architectures and tasks. This raises the central question: \emph{Why are deep neural networks suited for compression?} In this work, we take up the perspective of algorithmic complexity to explain this behavior. We hypothesize that the parameters of trained models have more structure and, hence, exhibit lower algorithmic complexity compared to the weights at (random) initialization. Furthermore, that model compression methods harness this reduced algorithmic complexity to compress models. Although an unconstrained parameterization of model weights, $\mathbf{w} \in \mathbb{R}^n$, can represent arbitrary weight assignments, the solutions found during training exhibit repeatability and structure, making them algorithmically simpler than a generic program. To this end, we formalize the Kolmogorov complexity of $\mathbf{w}$ by $\mathcal{K}(\mathbf{w})$. We introduce a constrained parameterization $\widehat{\mathbf{w}}$, that partitions parameters into blocks of size $s$, and restricts each block to be selected from a set of $k$ reusable motifs, specified by a reuse pattern (or mosaic). The resulting method, $\textit{Mosaic-of-Motifs}$ (MoMos), yields algorithmically simpler model parameterization compared to unconstrained models. Empirical evidence from multiple experiments shows that the algorithmic complexity of neural networks, measured using approximations to Kolmogorov complexity, can be reduced during training. This results in models that perform comparably with unconstrained models while being algorithmically simpler.
[693] Additive Control Variates Dominate Self-Normalisation in Off-Policy Evaluation
Olivier Jeunen, Shashank Gupta
Main category: cs.LG
TL;DR: The paper proves that β*-IPS with optimal additive baseline asymptotically dominates SNIPS in Mean Squared Error for off-policy evaluation in ranking and recommendation systems.
Details
Motivation: Off-policy evaluation is crucial for assessing ranking/recommendation systems without costly online tests. While SNIPS is standard for variance reduction using multiplicative control variates, recent work suggests additive control variates (baseline corrections) may perform better, but theoretical guarantees are lacking.Method: The paper introduces β*-IPS estimator with optimal additive baseline, provides theoretical analysis proving it asymptotically dominates SNIPS in MSE, and analytically decomposes the variance gap to show SNIPS is equivalent to using a specific sub-optimal additive baseline.
Result: Theoretical proof that β*-IPS asymptotically dominates SNIPS in Mean Squared Error, with analytical variance gap decomposition showing SNIPS corresponds to a sub-optimal additive baseline.
Conclusion: The results theoretically justify shifting from self-normalization (SNIPS) to optimal baseline corrections (β*-IPS) for off-policy evaluation in ranking and recommendation systems.
Abstract: Off-policy evaluation (OPE) is essential for assessing ranking and recommendation systems without costly online interventions. Self-Normalised Inverse Propensity Scoring (SNIPS) is a standard tool for variance reduction in OPE, leveraging a multiplicative control variate. Recent advances in off-policy learning suggest that additive control variates (baseline corrections) may offer superior performance, yet theoretical guarantees for evaluation are lacking. This paper provides a definitive answer: we prove that $β^\star$-IPS, an estimator with an optimal additive baseline, asymptotically dominates SNIPS in Mean Squared Error. By analytically decomposing the variance gap, we show that SNIPS is asymptotically equivalent to using a specific – but generally sub-optimal – additive baseline. Our results theoretically justify shifting from self-normalisation to optimal baseline corrections for both ranking and recommendation.
[694] BHyGNN+: Unsupervised Representation Learning for Heterophilic Hypergraphs
Tianyi Ma, Yiyue Qian, Zehong Wang, Zheyuan Zhang, Chuxu Zhang, Yanfang Ye
Main category: cs.LG
TL;DR: BHyGNN+ is a self-supervised learning framework for heterophilic hypergraphs that uses hypergraph duality for contrastive learning without needing labeled data or negative samples.
Details
Motivation: Existing Hypergraph Neural Networks struggle with heterophilic hypergraphs where connected nodes have dissimilar representations, and they rely heavily on labeled data which is scarce in real-world applications.Method: Uses hypergraph duality (interchanging node and hyperedge roles) to create augmented views, then contrasts these views using cosine similarity without needing negative samples.
Result: Outperforms state-of-the-art supervised and self-supervised baselines on 11 benchmark datasets for both heterophilic and homophilic hypergraphs.
Conclusion: Hypergraph duality provides an effective self-supervised learning paradigm for challenging unlabeled hypergraphs, eliminating the need for negative samples.
Abstract: Hypergraph Neural Networks (HyGNNs) have demonstrated remarkable success in modeling higher-order relationships among entities. However, their performance often degrades on heterophilic hypergraphs, where nodes connected by the same hyperedge tend to have dissimilar semantic representations or belong to different classes. While several HyGNNs, including our prior work BHyGNN, have been proposed to address heterophily, their reliance on labeled data significantly limits their applicability in real-world scenarios where annotations are scarce or costly. To overcome this limitation, we introduce BHyGNN+, a self-supervised learning framework that extends BHyGNN for representation learning on heterophilic hypergraphs without requiring ground-truth labels. The core idea of BHyGNN+ is hypergraph duality, a structural transformation where the roles of nodes and hyperedges are interchanged. By contrasting augmented views of a hypergraph against its dual using cosine similarity, our framework captures essential structural patterns in a fully unsupervised manner. Notably, this duality-based formulation eliminates the need for negative samples, a common requirement in existing hypergraph contrastive learning methods that is often difficult to satisfy in practice. Extensive experiments on eleven benchmark datasets demonstrate that BHyGNN+ consistently outperforms state-of-the-art supervised and self-supervised baselines on both heterophilic and homophilic hypergraphs. Our results validate the effectiveness of leveraging hypergraph duality for self-supervised learning and establish a new paradigm for representation learning on challenging, unlabeled hypergraphs.
[695] Variance-Reduced $(\varepsilon,δ)-$Unlearning using Forget Set Gradients
Martin Van Waerebeke, Marco Lorenzi, Kevin Scaman, El Mahdi El Mhamdi, Giovanni Neglia
Main category: cs.LG
TL;DR: VRU is a first-order machine unlearning algorithm that directly uses forget set gradients while provably satisfying (ε,δ)-unlearning guarantees, achieving better convergence rates than methods ignoring the forget set.
Details
Motivation: Current machine unlearning methods either use formal (ε,δ)-unlearning frameworks that only use forget set for noise calibration (not direct optimization), or use empirical heuristics that exploit forget samples but lack formal guarantees. There's a need to bridge this gap.Method: VRU (Variance-Reduced Unlearning) algorithm incorporates forget set gradients directly in its update rule while maintaining formal (ε,δ)-unlearning guarantees. It uses variance reduction techniques and provably satisfies the unlearning framework requirements.
Result: VRU achieves strictly improved convergence rates compared to existing first-order (ε,δ)-unlearning methods. In low-error regimes, it asymptotically outperforms any first-order method ignoring the forget set. Experiments show consistent gains over both certified unlearning methods and empirical baselines.
Conclusion: Directly incorporating forget set gradients in unlearning algorithms can provide both formal guarantees and improved performance, bridging the gap between theoretical and empirical approaches in machine unlearning.
Abstract: In machine unlearning, $(\varepsilon,δ)-$unlearning is a popular framework that provides formal guarantees on the effectiveness of the removal of a subset of training data, the forget set, from a trained model. For strongly convex objectives, existing first-order methods achieve $(\varepsilon,δ)-$unlearning, but they only use the forget set to calibrate injected noise, never as a direct optimization signal. In contrast, efficient empirical heuristics often exploit the forget samples (e.g., via gradient ascent) but come with no formal unlearning guarantees. We bridge this gap by presenting the Variance-Reduced Unlearning (VRU) algorithm. To the best of our knowledge, VRU is the first first-order algorithm that directly includes forget set gradients in its update rule, while provably satisfying ($(\varepsilon,δ)-$unlearning. We establish the convergence of VRU and show that incorporating the forget set yields strictly improved rates, i.e. a better dependence on the achieved error compared to existing first-order $(\varepsilon,δ)-$unlearning methods. Moreover, we prove that, in a low-error regime, VRU asymptotically outperforms any first-order method that ignores the forget set.Experiments corroborate our theory, showing consistent gains over both state-of-the-art certified unlearning methods and over empirical baselines that explicitly leverage the forget set.
[696] Locally Adaptive Multi-Objective Learning
Jivat Neet Kaur, Isaac Gibbs, Michael I. Jordan
Main category: cs.LG
TL;DR: Proposes an adaptive online learning method for multi-objective prediction that handles distribution shifts by replacing part of the algorithm with adaptive components, showing improved performance on energy forecasting and fairness datasets.
Details
Motivation: Existing multi-objective learning methods (for calibration, regret, multiaccuracy) don't adapt well to distribution shifts over time. While some work adds local guarantees, empirical evaluation is limited. Need methods that adapt to changing data distributions in online settings.Method: Replaces one component of existing multi-objective learning methods with an adaptive online algorithm to achieve local adaptivity. This allows the predictor to adjust to distribution shifts while maintaining multiple objectives simultaneously.
Result: Empirical evaluation on energy forecasting and algorithmic fairness datasets shows the proposed method improves upon existing approaches, achieves unbiased predictions over subgroups, and remains robust under distribution shift.
Conclusion: The adaptive approach successfully addresses limitations of existing multi-objective learning methods by enabling local adaptivity to distribution shifts while maintaining multiple learning objectives.
Abstract: We consider the general problem of learning a predictor that satisfies multiple objectives of interest simultaneously, a broad framework that captures a range of specific learning goals including calibration, regret, and multiaccuracy. We work in an online setting where the data distribution can change arbitrarily over time. Existing approaches to this problem aim to minimize the set of objectives over the entire time horizon in a worst-case sense, and in practice they do not necessarily adapt to distribution shifts. Earlier work has aimed to alleviate this problem by incorporating additional objectives that target local guarantees over contiguous subintervals. Empirical evaluation of these proposals is, however, scarce. In this article, we consider an alternative procedure that achieves local adaptivity by replacing one part of the multi-objective learning method with an adaptive online algorithm. Empirical evaluations on datasets from energy forecasting and algorithmic fairness show that our proposed method improves upon existing approaches and achieves unbiased predictions over subgroups, while remaining robust under distribution shift.
[697] Use What You Know: Causal Foundation Models with Partial Graphs
Arik Reuter, Anish Dhir, Cristiana Diaconu, Jake Robertson, Ole Ossen, Frank Hutter, Adrian Weller, Mark van der Wilk, Bernhard Schölkopf
Main category: cs.LG
TL;DR: Conditioning Causal Foundation Models (CFMs) on causal information like graphs or ancestral knowledge to improve performance when domain expertise is available
Details
Motivation: Current Causal Foundation Models don't incorporate domain knowledge, leading to suboptimal predictions. There's a need to bridge this gap by allowing CFMs to leverage causal information when available.Method: Introduce methods to condition CFMs on causal information (causal graph or ancestral information), including partial causal information. Use learnable biases injected into attention mechanisms as the most effective conditioning strategy.
Result: Conditioning allows general-purpose CFMs to match performance of specialized models trained on specific causal structures. Learnable attention biases effectively utilize both full and partial causal information.
Conclusion: The approach addresses a key challenge for all-in-one causal foundation models: answering causal queries data-driven while leveraging domain expertise, enabling more practical deployment.
Abstract: Estimating causal quantities traditionally relies on bespoke estimators tailored to specific assumptions. Recently proposed Causal Foundation Models (CFMs) promise a more unified approach by amortising causal discovery and inference in a single step. However, in their current state, they do not allow for the incorporation of any domain knowledge, which can lead to suboptimal predictions. We bridge this gap by introducing methods to condition CFMs on causal information, such as the causal graph or more readily available ancestral information. When access to complete causal graph information is too strict a requirement, our approach also effectively leverages partial causal information. We systematically evaluate conditioning strategies and find that injecting learnable biases into the attention mechanism is the most effective method to utilise full and partial causal information. Our experiments show that this conditioning allows a general-purpose CFM to match the performance of specialised models trained on specific causal structures. Overall, our approach addresses a central hurdle on the path towards all-in-one causal foundation models: the capability to answer causal queries in a data-driven manner while effectively leveraging any amount of domain expertise.
[698] MacroGuide: Topological Guidance for Macrocycle Generation
Alicja Maksymiuk, Alexandre Duplessis, Michael Bronstein, Alexander Tong, Fernanda Duarte, İsmail İlkan Ceylan
Main category: cs.LG
TL;DR: MacroGuide is a diffusion guidance method that uses persistent homology to steer molecular generation toward macrocycles, increasing generation rates from 1% to 99% while maintaining quality metrics.
Details
Motivation: Macrocycles are promising drug candidates with enhanced selectivity and binding affinity, but they're underexplored in generative modeling due to scarcity in datasets and challenges in enforcing topological constraints in standard deep generative models.Method: Introduces MacroGuide, a diffusion guidance mechanism that uses Persistent Homology to steer sampling of pretrained molecular generative models toward macrocycle generation. At each denoising step, constructs a Vietoris-Rips complex from atomic positions and promotes ring formation by optimizing persistent homology features.
Result: Applying MacroGuide to pretrained diffusion models increases macrocycle generation rates from 1% to 99%, while matching or exceeding state-of-the-art performance on key quality metrics such as chemical validity, diversity, and PoseBusters checks.
Conclusion: MacroGuide effectively addresses the challenge of generating macrocycles using topological guidance with persistent homology, significantly improving generation rates while maintaining molecular quality.
Abstract: Macrocycles are ring-shaped molecules that offer a promising alternative to small-molecule drugs due to their enhanced selectivity and binding affinity against difficult targets. Despite their chemical value, they remain underexplored in generative modeling, likely owing to their scarcity in public datasets and the challenges of enforcing topological constraints in standard deep generative models. We introduce MacroGuide: Topological Guidance for Macrocycle Generation, a diffusion guidance mechanism that uses Persistent Homology to steer the sampling of pretrained molecular generative models toward the generation of macrocycles, in both unconditional and conditional (protein pocket) settings. At each denoising step, MacroGuide constructs a Vietoris-Rips complex from atomic positions and promotes ring formation by optimizing persistent homology features. Empirically, applying MacroGuide to pretrained diffusion models increases macrocycle generation rates from 1% to 99%, while matching or exceeding state-of-the-art performance on key quality metrics such as chemical validity, diversity, and PoseBusters checks.
[699] Orthogonalized Multimodal Contrastive Learning with Asymmetric Masking for Structured Representations
Carolin Cissee, Raneen Younis, Zahra Ahmadi
Main category: cs.LG
TL;DR: COrAL is a multimodal contrastive learning framework that explicitly models redundant, unique, and synergistic information components through dual-path architecture with orthogonality constraints and asymmetric masking.
Details
Motivation: Existing multimodal contrastive learning methods mainly capture redundant cross-modal signals while neglecting modality-specific (unique) and interaction-driven (synergistic) information, leading to incomplete representations and potential information leakage.Method: Uses dual-path architecture with orthogonality constraints to disentangle shared and modality-specific features, plus asymmetric masking with complementary view-specific patterns to force the model to infer cross-modal dependencies rather than relying on redundant cues.
Result: Extensive experiments on synthetic benchmarks and diverse MultiBench datasets show COrAL consistently matches or outperforms state-of-the-art methods while exhibiting low performance variance across runs.
Conclusion: Explicitly modeling the full spectrum of multimodal information (redundant, unique, and synergistic) yields more stable, reliable, and comprehensive embeddings.
Abstract: Multimodal learning seeks to integrate information from heterogeneous sources, where signals may be shared across modalities, specific to individual modalities, or emerge only through their interaction. While self-supervised multimodal contrastive learning has achieved remarkable progress, most existing methods predominantly capture redundant cross-modal signals, often neglecting modality-specific (unique) and interaction-driven (synergistic) information. Recent extensions broaden this perspective, yet they either fail to explicitly model synergistic interactions or learn different information components in an entangled manner, leading to incomplete representations and potential information leakage. We introduce \textbf{COrAL}, a principled framework that explicitly and simultaneously preserves redundant, unique, and synergistic information within multimodal representations. COrAL employs a dual-path architecture with orthogonality constraints to disentangle shared and modality-specific features, ensuring a clean separation of information components. To promote synergy modeling, we introduce asymmetric masking with complementary view-specific patterns, compelling the model to infer cross-modal dependencies rather than rely solely on redundant cues. Extensive experiments on synthetic benchmarks and diverse MultiBench datasets demonstrate that COrAL consistently matches or outperforms state-of-the-art methods while exhibiting low performance variance across runs. These results indicate that explicitly modeling the full spectrum of multimodal information yields more stable, reliable, and comprehensive embeddings.
[700] Spectral Convolution on Orbifolds for Geometric Deep Learning
Tim Mangliers, Bernhard Mössner, Benjamin Himpel
Main category: cs.LG
TL;DR: Introduces spectral convolution on orbifolds as a geometric deep learning building block for non-Euclidean data with orbifold structure, demonstrated with music theory applications.
Details
Motivation: Geometric deep learning needs to handle diverse topological and geometric structures beyond Euclidean data. There's a demand to make various application-related data with complex structures accessible to machine learning, requiring new mathematical frameworks.Method: Extends spectral convolution techniques from geometric deep learning to orbifolds, developing theoretical foundations for convolutional neural network-like architectures on orbifold-structured data.
Result: Develops the concept of spectral convolution on orbifolds as a building block for learning on orbifold-structured data, with theoretical illustration using music theory examples.
Conclusion: Provides a new mathematical framework for geometric deep learning that can handle orbifold structures, expanding the applicability of deep learning to more complex topological data domains.
Abstract: Geometric deep learning (GDL) deals with supervised learning on data domains that go beyond Euclidean structure, such as data with graph or manifold structure. Due to the demand that arises from application-related data, there is a need to identify further topological and geometric structures with which these use cases can be made accessible to machine learning. There are various techniques, such as spectral convolution, that form the basic building blocks for some convolutional neural network-like architectures on non-Euclidean data. In this paper, the concept of spectral convolution on orbifolds is introduced. This provides a building block for making learning on orbifold structured data accessible using GDL. The theory discussed is illustrated using an example from music theory.
[701] Boundary Point Jailbreaking of Black-Box LLMs
Xander Davies, Giorgi Giglemiani, Edmund Lau, Eric Winsor, Geoffrey Irving, Yarin Gal
Main category: cs.LG
TL;DR: BPJ introduces a fully black-box jailbreak attack that evades strong LLM safeguards using only binary classifier feedback, achieving automated universal jailbreaks against Constitutional Classifiers and GPT-5’s input classifier.
Details
Motivation: Current LLM safeguards have become robust against traditional jailbreak attacks, with classifier-based systems surviving extensive human red teaming. There's a need for automated attacks that can bypass these strong defenses without relying on white/grey-box access or existing jailbreak libraries.Method: BPJ is a fully black-box attack that uses only binary classifier feedback (flagged/not flagged). It converts target harmful strings into curriculum of intermediate attack targets, then actively selects “boundary points” - evaluation points that best detect small changes in attack strength to optimize jailbreak effectiveness.
Result: BPJ successfully develops universal jailbreaks against Constitutional Classifiers and is the first automated attack to succeed against GPT-5’s input classifier without human attack seeds. It demonstrates effectiveness against the strongest industry-deployed safeguards.
Conclusion: BPJ represents a significant advancement in automated jailbreak attacks, showing that effective defense requires supplementing single-interaction methods with batch-level monitoring, as BPJ incurs many flags during optimization but succeeds in individual interactions.
Abstract: Frontier LLMs are safeguarded against attempts to extract harmful information via adversarial prompts known as “jailbreaks”. Recently, defenders have developed classifier-based systems that have survived thousands of hours of human red teaming. We introduce Boundary Point Jailbreaking (BPJ), a new class of automated jailbreak attacks that evade the strongest industry-deployed safeguards. Unlike previous attacks that rely on white/grey-box assumptions (such as classifier scores or gradients) or libraries of existing jailbreaks, BPJ is fully black-box and uses only a single bit of information per query: whether or not the classifier flags the interaction. To achieve this, BPJ addresses the core difficulty in optimising attacks against robust real-world defences: evaluating whether a proposed modification to an attack is an improvement. Instead of directly trying to learn an attack for a target harmful string, BPJ converts the string into a curriculum of intermediate attack targets and then actively selects evaluation points that best detect small changes in attack strength (“boundary points”). We believe BPJ is the first fully automated attack algorithm that succeeds in developing universal jailbreaks against Constitutional Classifiers, as well as the first automated attack algorithm that succeeds against GPT-5’s input classifier without relying on human attack seeds. BPJ is difficult to defend against in individual interactions but incurs many flags during optimisation, suggesting that effective defence requires supplementing single-interaction methods with batch-level monitoring.
[702] PDE foundation models are skillful AI weather emulators for the Martian atmosphere
Johannes Schmude, Sujit Roy, Liping Wang, Theodore van Kessel, Levente Klein, Marcus Freitag, Eloisa Bentivegna, Robert Manson-Sawko, Bjorn Lutjens, Manil Maskey, Campbell Watson, Rahul Ramachandran, Juan Bernabe-Moreno
Main category: cs.LG
TL;DR: PDE foundation models pretrained on diverse PDE solutions can be adapted to create skillful weather emulators for Mars, achieving 34.4% performance improvement through pretraining and 3D extension.
Details
Motivation: To demonstrate that PDE foundation models can be adapted for real-world problems with complex interactions, specifically for Martian atmospheric prediction where training data is limited and compute budgets are constrained.Method: Extend the 2D Poseidon PDE foundation model to 3D while preserving pretraining information, then fine-tune on 34GB of Martian atmospheric data (4 Martian years) with sparse initial conditions, using a median compute budget of 13 GPU hours.
Result: The combination of pretraining and model extension yields a 34.4% performance increase on held-out test data, showing effective adaptation of PDE foundation models to real-world atmospheric prediction tasks.
Conclusion: PDE foundation models can serve as effective anchors for real-world problems with complex interactions, even when training data is limited or compute budgets are constrained, demonstrating transferability from synthetic PDE solutions to practical applications.
Abstract: We show that AI foundation models that are pretrained on numerical solutions to a diverse corpus of partial differential equations can be adapted and fine-tuned to obtain skillful predictive weather emulators for the Martian atmosphere. We base our work on the Poseidon PDE foundation model for two-dimensional systems. We develop a method to extend Poseidon from two to three dimensions while keeping the pretraining information. Moreover, we investigate the performance of the model in the presence of sparse initial conditions. Our results make use of four Martian years (approx.~34 GB) of training data and a median compute budget of 13 GPU hours. We find that the combination of pretraining and model extension yields a performance increase of 34.4% on a held-out year. This shows that PDEs-FMs can not only approximate solutions to (other) PDEs but also anchor models for real-world problems with complex interactions that lack a sufficient amount of training data or a suitable compute budget.
[703] Efficient Sampling with Discrete Diffusion Models: Sharp and Adaptive Guarantees
Daniil Dmitriev, Zhihan Huang, Yuting Wei
Main category: cs.LG
TL;DR: Theoretical analysis of discrete diffusion models showing improved convergence rates for τ-leaping samplers with dimension-independent complexity for structured data.
Details
Motivation: Discrete diffusion models have shown empirical success but lack theoretical foundations, particularly regarding sampling efficiency and convergence guarantees.Method: Analyzes score-based discrete diffusion models under continuous-time Markov chain formulation, focusing on τ-leaping-based samplers for both uniform and masking noising processes.
Result: For uniform diffusion: τ-leaping achieves Õ(d/ε) iteration complexity (eliminating dependence on vocabulary size S). For masking diffusion: modified τ-leaping sampler adapts to low-dimensional structure with convergence governed by effective total correlation, yielding sublinear rates for practical examples.
Conclusion: Provides rigorous theoretical foundations for discrete diffusion models, showing they can achieve dimension-independent convergence for structured data without prior knowledge of the structure.
Abstract: Diffusion models over discrete spaces have recently shown striking empirical success, yet their theoretical foundations remain incomplete. In this paper, we study the sampling efficiency of score-based discrete diffusion models under a continuous-time Markov chain (CTMC) formulation, with a focus on $τ$-leaping-based samplers. We establish sharp convergence guarantees for attaining $\varepsilon$ accuracy in Kullback-Leibler (KL) divergence for both uniform and masking noising processes. For uniform discrete diffusion, we show that the $τ$-leaping algorithm achieves an iteration complexity of order $\tilde O(d/\varepsilon)$, with $d$ the ambient dimension of the target distribution, eliminating linear dependence on the vocabulary size $S$ and improving existing bounds by a factor of $d$; moreover, we establish a matching algorithmic lower bound showing that linear dependence on the ambient dimension is unavoidable in general. For masking discrete diffusion, we introduce a modified $τ$-leaping sampler whose convergence rate is governed by an intrinsic information-theoretic quantity, termed the effective total correlation, which is bounded by $d \log S$ but can be sublinear or even constant for structured data. As a consequence, the sampler provably adapts to low-dimensional structure without prior knowledge or algorithmic modification, yielding sublinear convergence rates for various practical examples (such as hidden Markov models, image data, and random graphs). Our analysis requires no boundedness or smoothness assumptions on the score estimator beyond control of the score entropy loss.
[704] Rethinking Diffusion Models with Symmetries through Canonicalization with Applications to Molecular Graph Generation
Cai Zhou, Zijie Chen, Zian Li, Jike Wang, Kaiyi Jiang, Pan Li, Rose Yu, Muhan Zhang, Stephen Bates, Tommi Jaakkola
Main category: cs.LG
TL;DR: Canonical diffusion framework for invariant distributions using canonicalization instead of architectural constraints, applied to molecular graph generation with SE(3) symmetries.
Details
Motivation: Traditional approaches enforce invariance/equivariance through architectural constraints like equivariant denoisers and invariant priors. This paper challenges this tradition by proposing canonicalization as an alternative approach to handle group symmetries in generative tasks.Method: Proposes canonical diffusion framework: 1) map samples to orbit representatives with canonical pose/order, 2) train unconstrained diffusion/flow model on canonical slice, 3) recover invariant distribution by sampling random symmetry transform at generation time. Uses geometric spectra-based canonicalization and mild positional encodings for molecular graphs under S_n × SE(3) symmetries.
Result: Canonical diffusion significantly outperforms equivariant baselines in 3D molecule generation tasks with similar or less computation. CanonFlow achieves state-of-the-art performance on GEOM-DRUG dataset, with large advantages in few-step generation.
Conclusion: Canonicalization provides a powerful alternative to architectural constraints for handling symmetries in generative models, offering theoretical correctness, superior expressivity, and practical efficiency improvements for invariant distributions.
Abstract: Many generative tasks in chemistry and science involve distributions invariant to group symmetries (e.g., permutation and rotation). A common strategy enforces invariance and equivariance through architectural constraints such as equivariant denoisers and invariant priors. In this paper, we challenge this tradition through the alternative canonicalization perspective: first map each sample to an orbit representative with a canonical pose or order, train an unconstrained (non-equivariant) diffusion or flow model on the canonical slice, and finally recover the invariant distribution by sampling a random symmetry transform at generation time. Building on a formal quotient-space perspective, our work provides a comprehensive theory of canonical diffusion by proving: (i) the correctness, universality and superior expressivity of canonical generative models over invariant targets; (ii) canonicalization accelerates training by removing diffusion score complexity induced by group mixtures and reducing conditional variance in flow matching. We then show that aligned priors and optimal transport act complementarily with canonicalization and further improves training efficiency. We instantiate the framework for molecular graph generation under $S_n \times SE(3)$ symmetries. By leveraging geometric spectra-based canonicalization and mild positional encodings, canonical diffusion significantly outperforms equivariant baselines in 3D molecule generation tasks, with similar or even less computation. Moreover, with a novel architecture Canon, CanonFlow achieves state-of-the-art performance on the challenging GEOM-DRUG dataset, and the advantage remains large in few-step generation.
[705] Long Context, Less Focus: A Scaling Gap in LLMs Revealed through Privacy and Personalization
Shangding Gu
Main category: cs.LG
TL;DR: PAPerBench is a large-scale benchmark for studying how increasing context length affects personalization quality and privacy protection in LLMs, revealing performance degradation in both areas as context grows.
Details
Motivation: LLMs are increasingly used in privacy-critical and personalization scenarios, but the impact of context length on privacy leakage and personalization effectiveness remains unexplored. There's a need for systematic study of how longer contexts affect both personalization quality and privacy protection.Method: Created PAPerBench benchmark with ~29,000 instances across context lengths from 1K to 256K tokens (377K total questions). Jointly evaluates personalization performance and privacy risks across diverse scenarios. Conducted extensive evaluations across state-of-the-art LLMs and provided theoretical analysis of attention dilution under context scaling.
Result: Found consistent performance degradation in both personalization and privacy as context length increases. Theoretical analysis explains this as an inherent limitation of soft attention in fixed-capacity Transformers. Reveals a general scaling gap: “long context, less focus.”
Conclusion: Current LLMs face fundamental limitations in handling long contexts while maintaining both personalization quality and privacy protection. The benchmark enables reproducible evaluation and future research on scalable privacy and personalization in LLMs.
Abstract: Large language models (LLMs) are increasingly deployed in privacy-critical and personalization-oriented scenarios, yet the role of context length in shaping privacy leakage and personalization effectiveness remains largely unexplored. We introduce a large-scale benchmark, PAPerBench, to systematically study how increasing context length influences both personalization quality and privacy protection in LLMs. The benchmark comprises approximately 29,000 instances with context lengths ranging from 1K to 256K tokens, yielding a total of 377K evaluation questions. It jointly evaluates personalization performance and privacy risks across diverse scenarios, enabling controlled analysis of long-context model behavior. Extensive evaluations across state-of-the-art LLMs reveal consistent performance degradation in both personalization and privacy as context length increases. We further provide a theoretical analysis of attention dilution under context scaling, explaining this behavior as an inherent limitation of soft attention in fixed-capacity Transformers. The empirical and theoretical findings together suggest a general scaling gap in current models – long context, less focus. We release the benchmark to support reproducible evaluation and future research on scalable privacy and personalization. Code and data are available at https://github.com/SafeRL-Lab/PAPerBench
[706] Sparse MeZO: Less Parameters for Better Performance in Zeroth-Order LLM Fine-Tuning
Yong Liu, Zirui Zhu, Chaoyu Gong, Minhao Cheng, Cho-Jui Hsieh, Yang You
Main category: cs.LG
TL;DR: Sparse-MeZO: A memory-efficient zeroth-order optimization method that applies gradient estimation only to a carefully selected subset of parameters, improving performance and convergence speed while maintaining inference-level memory consumption.
Details
Motivation: Fine-tuning LLMs requires memory-intensive backpropagation. While MeZO optimizers reduce memory by using only forward passes, they suffer from gradient estimation errors that hurt optimization. The authors observed that estimation errors affect large weights more than small weights, motivating a sparse approach.Method: Sparse-MeZO applies zeroth-order gradient estimation only to a carefully chosen subset of parameters rather than all parameters. It includes an effective parameter selection scheme and memory-optimized sparse masking implementation that maintains inference-level memory consumption.
Result: Sparse-MeZO consistently outperforms MeZO in both performance and convergence speed without overhead. It achieved 9% absolute accuracy improvement and 3.5x speedup on RTE task, and can fine-tune LLaMA-30b on a single A100 GPU.
Conclusion: Sparse-MeZO is an effective memory-efficient optimization approach that addresses gradient estimation errors in zeroth-order methods by selectively applying updates to important parameters, enabling efficient fine-tuning of large models with minimal memory requirements.
Abstract: While fine-tuning large language models (LLMs) for specific tasks often yields impressive results, it comes at the cost of memory inefficiency due to back-propagation in gradient-based training. Memory-efficient Zeroth-order (MeZO) optimizers, recently proposed to address this issue, only require forward passes during training, making them more memory-friendly. However, compared with exact gradients, ZO-based gradients usually exhibit an estimation error, which can significantly hurt the optimization process, leading to slower convergence and suboptimal solutions. In addition, we find that the estimation error will hurt more when adding to large weights instead of small weights. Based on this observation, this paper introduces Sparse MeZO, a novel memory-efficient zeroth-order optimization approach that applies ZO only to a carefully chosen subset of parameters. We propose a simple yet effective parameter selection scheme that yields significant performance gains with Sparse-MeZO. Additionally, we develop a memory-optimized implementation for sparse masking, ensuring the algorithm requires only inference-level memory consumption, allowing Sparse-MeZO to fine-tune LLaMA-30b on a single A100 GPU. Experimental results illustrate that Sparse-MeZO consistently improves both performance and convergence speed over MeZO without any overhead. For example, it achieves a 9% absolute accuracy improvement and 3.5x speedup over MeZO on the RTE task. Code is available at https://github.com/NUS-HPC-AI-Lab/SparseMeZO.
[707] Cautious Optimizers: Improving Training with One Line of Code
Kaizhao Liang, Lizhang Chen, Bo Liu, Qiang Liu
Main category: cs.LG
TL;DR: One-line PyTorch modification to momentum-based optimizers (C-AdamW, C-Lion) that preserves Adam’s Hamiltonian function and provides consistent speed-up for LLM pretraining and image classification with minimal hyperparameter tuning.
Details
Motivation: The community has been searching for faster and more stable optimizers for transformer pretraining beyond AdamW, with only limited success. The authors aim to provide a simple yet effective modification to existing momentum-based optimizers.Method: Proposes a one-line modification in PyTorch to any momentum-based optimizer, creating “cautious” variants (e.g., C-AdamW, C-Lion). The modification preserves Adam’s Hamiltonian function and maintains convergence guarantees under Lyapunov analysis. The theoretical insight reveals a whole new family of optimizers.
Result: Shows consistent speed-up on LLM pretraining and image classification tasks with minimum extra hyperparameter tuning. The approach works across different optimization scenarios.
Conclusion: A simple one-line modification to momentum-based optimizers provides practical improvements for deep learning training, particularly for transformer pretraining, while maintaining theoretical guarantees.
Abstract: AdamW has been the default optimizer for transformer pretraining. For many years, our community searched for faster and more stable optimizers with only constrained positive outcomes. In this work, we propose a \textbf{one-line modification in Pytorch} to any momentum-based optimizer, which we rename cautious optimizer, e.g. C-AdamW and C-Lion. Our theoretical result shows that this modification preserves Adam’s Hamiltonian function and it does not break the convergence guarantee under the Lyapunov analysis. In addition, a whole new family of optimizers is revealed by our theoretical insight. Among them, we pick the simplest one for empirical experiments, showing not only consistent speed-up on LLM pretraining, but also image classification, with minimum extra tuning on hyperparameters. Code is available at https://github.com/kyleliang919/C-Optim.
[708] Less is More: Improving LLM Alignment via Preference Data Selection
Xun Deng, Han Zhong, Rui Ai, Fuli Feng, Zheng Wang, Xiangnan He
Main category: cs.LG
TL;DR: Proposes data selection improvements for DPO using margin-maximization principle and Bayesian aggregation of multiple reward sources to address noisy preference data.
Details
Motivation: Current DPO approaches mainly focus on objective function improvements, but data selection is overlooked despite being critical. Noisy preference data causes parameter shrinkage issues in DPO training.Method: 1) Margin-maximization principle for dataset curation to address noisy data, 2) Bayesian Aggregation approach to unify multiple margin sources (external and implicit reward models) into single preference probability.
Result: Achieves 3-8% improvements across Llama, Mistral, and Qwen models on AlpacaEval2 using only 10% of Ultrafeedback dataset. In iterative DPO, achieves ~3% improvement with 25% online data.
Conclusion: Data selection strategies have significant potential for advancing preference optimization, revealing high redundancy in presumed high-quality data construction methods.
Abstract: Direct Preference Optimization (DPO) has emerged as a promising approach for aligning large language models with human preferences. While prior work mainly extends DPO from the aspect of the objective function, we instead improve DPO from the largely overlooked but critical aspect of data selection. Specifically, we address the issue of parameter shrinkage caused by noisy data by proposing a novel margin-maximization principle for dataset curation in DPO training. To further mitigate the noise in different reward models, we propose a Bayesian Aggregation approach that unifies multiple margin sources (external and implicit) into a single preference probability. Extensive experiments in diverse settings demonstrate the consistently high data efficiency of our approach. Remarkably, by using just 10% of the Ultrafeedback dataset, our approach achieves 3% to 8% improvements across various Llama, Mistral, and Qwen models on the AlpacaEval2 benchmark. Furthermore, our approach seamlessly extends to iterative DPO, yielding a roughly 3% improvement with 25% online data, revealing the high redundancy in this presumed high-quality data construction manner. These results highlight the potential of data selection strategies for advancing preference optimization.
[709] Improving Data Efficiency for LLM Reinforcement Fine-tuning Through Difficulty-targeted Online Data Selection and Rollout Replay
Yifan Sun, Jingyan Shen, Yibin Wang, Tianyu Chen, Zhendong Wang, Mingyuan Zhou, Huan Zhang
Main category: cs.LG
TL;DR: Proposes two techniques for improving data efficiency in RL fine-tuning of LLMs: difficulty-targeted online data selection using adaptive difficulty estimation, and rollout replay to reuse recent rollouts.
Details
Motivation: RL fine-tuning for LLMs is highly resource-intensive, and existing work has largely overlooked data efficiency problems. The paper aims to reduce computational costs while maintaining performance.Method: 1) Adaptive difficulty-based online data selection: Uses attention-based framework to estimate question difficulty, prioritizing moderately difficult questions for better learning signals. 2) Rollout replay: Reuses recent rollouts to reduce per-step computation while maintaining stable updates.
Result: Experiments across 6 LLM-dataset combinations show 23% to 62% reduction in RL fine-tuning time while achieving same performance level as original GRPO algorithm.
Conclusion: The proposed techniques significantly improve data efficiency in RL fine-tuning for LLMs, making the process more practical and resource-efficient while maintaining performance.
Abstract: Reinforcement learning (RL) has become an effective approach for fine-tuning large language models (LLMs), particularly to enhance their reasoning capabilities. However, RL fine-tuning remains highly resource-intensive, and existing work has largely overlooked the problem of data efficiency. In this paper, we propose two techniques to improve data efficiency in LLM RL fine-tuning: difficulty-targeted online data selection and rollout replay. We introduce the notion of adaptive difficulty to guide online data selection, prioritizing questions of moderate difficulty that are more likely to yield informative learning signals. To estimate adaptive difficulty efficiently, we develop an attention-based framework that requires rollouts for only a small reference set of questions. The adaptive difficulty of the remaining questions is then estimated based on their similarity to this set. To further reduce rollout cost, we introduce a rollout replay mechanism inspired by experience replay in traditional RL. This technique reuses recent rollouts, lowering per-step computation while maintaining stable updates. Experiments across 6 LLM-dataset combinations show that our method reduces RL fine-tuning time by 23% to 62% while reaching the same level of performance as the original GRPO algorithm. Our code is available at https://github.com/ASTRAL-Group/data-efficient-llm-rl.
[710] Enhancing Delta Compression in LLMs via SVD-based Quantization Error Minimization
Boya Xiong, Shuo Wang, Weifeng Ge, Guanhua Chen, Yun Chen
Main category: cs.LG
TL;DR: PrinMix is a principled SVD-based compression framework for LLM delta parameters that uses mathematical optimization for mix-precision quantization and reconstruction error correction.
Details
Motivation: Supervised Fine-Tuning produces dense, high-dimensional delta parameters that create severe storage and distribution challenges. Existing SVD-based compression methods use heuristic quantization without clear mathematical foundations, leading to poor generalizability.Method: 1) Theoretically derive quantization error and identify singular-value-dominated scaling mechanism proving necessity of mix-precision quantization. 2) Model quantization as 0/1 Integer Linear Programming problem for optimal bit-budget solutions. 3) Integrate Reconstruction Target Correction to compensate for errors from sequential V-then-U quantization.
Result: For 7B LLMs, PrinMix outperforms SOTA Delta-CoMe by 22.3% on AIME2024 and 6.1% on GQA benchmarks. Extensive experiments confirm strong performance.
Conclusion: PrinMix provides a rigorous mathematical framework for compressing LLM delta parameters through principled mix-precision quantization and error correction, addressing storage challenges while maintaining performance.
Abstract: Supervised Fine-Tuning (SFT) empowers Large Language Models (LLMs) with exceptional performance on specialized tasks, but it yields dense, high-dimensional delta parameters that pose severe storage and distribution challenges. Singular Value Decomposition (SVD)-based compression offers a compact representation for such delta parameters, but existing methods adopt heuristic quantization without clarifying underlying mechanisms, leading to poor generalizability. In this work, we propose PrinMix, a rigorous SVD-based framework that models quantization as an optimization problem, grounding the design in mathematical mechanisms. We first theoretically derive quantization error and identify a key singular-value-dominated scaling mechanism, which mathematically proves the necessity of mix-precision quantization. We then model the quantization scheme as a 0/1 Integer Linear Programming (ILP) problem, which yields optimal bit-budget-constrained solutions without empirical assumptions. Furthermore, PrinMix integrates a Reconstruction Target Correction (RTC) method to compensate for errors from the $\mathbf{V}$-then-$\mathbf{U}$ sequential quantization process. Extensive experiments confirm PrinMix performs well: for 7B LLMs, PrinMix outperforms SOTA Delta-CoMe on challenging benchmarks by 22.3% on AIME2024 and 6.1% on GQA.
[711] Token Hidden Reward: Steering Exploration-Exploitation in Group Relative Deep Reinforcement Learning
Wenlong Deng, Yi Ren, Yushu Li, Boying Gong, Danica J. Sutherland, Xiaoxiao Li, Christos Thrampoulidis
Main category: cs.LG
TL;DR: THR (Token Hidden Reward) is a token-level metric that quantifies each token’s influence on correct responses in RL-tuned LLMs, enabling explicit control over exploration vs exploitation through reweighting algorithms.
Details
Motivation: While reinforcement learning with verifiable rewards has advanced LLM reasoning, there's no principled way to explicitly steer training toward exploration or exploitation. The paper aims to address this gap by developing fine-grained control mechanisms.Method: Introduces Token Hidden Reward (THR) that measures each token’s influence on correct response likelihood under Group Relative Policy Optimization (GRPO). Develops THR-guided reweighting algorithm that modulates learning signals by amplifying positive THR tokens (for exploitation) or negative THR tokens (for exploration).
Result: Positive THR tokens strengthen confidence in correct outputs (exploitation), while negative THR tokens preserve probability mass for alternatives (exploration). THR-guided reweighting improves greedy-decoding accuracy when favoring exploitation, and boosts Pass@K accuracy when favoring exploration. Method generalizes across architectures and integrates with other RL objectives.
Conclusion: THR provides a principled, fine-grained mechanism for dynamically controlling exploration and exploitation in RL-tuned LLMs, offering new tools for targeted fine-tuning in reasoning-intensive applications.
Abstract: Reinforcement learning with verifiable rewards has significantly advanced the reasoning capabilities of large language models, yet how to explicitly steer training toward exploration or exploitation remains an open problem. We introduce Token Hidden Reward (THR), a token-level metric that quantifies each token’s influence on the likelihood of correct responses under Group Relative Policy Optimization (GRPO). We find that training dynamics are dominated by a small subset of tokens with high absolute THR values. Most interestingly, tokens with positive THR strengthen confidence in correct outputs, thus favoring exploitation, while tokens with negative THR preserve probability mass for alternative outputs, enabling exploration. This insight suggests a natural intervention: a THR-guided reweighting algorithm that modulates GRPO’s learning signals to explicitly bias training toward exploitation or exploration. We validate the efficacy of this algorithm on diverse math reasoning benchmarks. By amplifying tokens with positive THR value and weakening negative ones, our algorithm improves greedy-decoding accuracy, favoring exploitation. The reverse strategy yields consistent gains in Pass@K accuracy, favoring exploration. We further demonstrate that our algorithm integrates seamlessly with other RL objectives such as GSPO and generalizes across architectures including Llama. These findings establish THR as a principled and fine-grained mechanism for dynamically controlling exploration and exploitation in RL-tuned LLMs, providing new tools for targeted fine-tuning in reasoning-intensive applications.
[712] Endless Terminals: Scaling RL Environments for Terminal Agents
Kanishk Gandhi, Shivam Garg, Noah D. Goodman, Dimitris Papailiopoulos
Main category: cs.LG
TL;DR: Endless Terminals is an autonomous pipeline for generating terminal-use tasks for RL training, showing that simple PPO with scalable environments yields substantial improvements on terminal benchmarks.
Details
Motivation: Current terminal benchmarks are designed for evaluation rather than training, and reinforcement learning requires scalable environments, not just datasets. There's a need for procedurally generated tasks to enable effective agent training.Method: A four-stage pipeline: 1) generating diverse task descriptions, 2) building and validating containerized environments, 3) producing completion tests, and 4) filtering for solvability. Uses vanilla PPO with binary episode-level rewards and minimal interaction loops (no retrieval, multi-agent coordination, or specialized tools).
Result: Generated 3255 tasks spanning file operations, log management, data processing, scripting, and database operations. Models trained on Endless Terminals showed substantial gains: Llama-3.2-3B improved from 4.0% to 18.2%, Qwen2.5-7B from 10.7% to 53.3%, and Qwen3-8B-openthinker-sft from 42.6% to 59.0% on held-out dev set. Improvements transferred to human-curated benchmarks like TerminalBench 2.0.
Conclusion: Simple reinforcement learning succeeds when environments scale, demonstrating that autonomous environment generation pipelines can enable effective agent training without complex agentic scaffolds.
Abstract: Environments are the bottleneck for self-improving agents. Current terminal benchmarks were built for evaluation, not training; reinforcement learning requires a scalable pipeline, not just a dataset. We introduce Endless Terminals, a fully autonomous pipeline that procedurally generates terminal-use tasks without human annotation. The pipeline has four stages: generating diverse task descriptions, building and validating containerized environments, producing completion tests, and filtering for solvability. From this pipeline we obtain 3255 tasks spanning file operations, log management, data processing, scripting, and database operations. We train agents using vanilla PPO with binary episode level rewards and a minimal interaction loop: no retrieval, multi-agent coordination, or specialized tools. Despite this simplicity, models trained on Endless Terminals show substantial gains: on our held-out dev set, Llama-3.2-3B improves from 4.0% to 18.2%, Qwen2.5-7B from 10.7% to 53.3%, and Qwen3-8B-openthinker-sft from 42.6% to 59.0%. These improvements transfer to human-curated benchmarks: models trained on Endless Terminals show substantial gains on held out human curated benchmarks: on TerminalBench 2.0, Llama-3.2-3B improves from 0.0% to 2.2%, Qwen2.5-7B from 2.2% to 3.4%, and Qwen3-8B-openthinker-sft from 1.1% to 6.7%, in each case outperforming alternative approaches including models with more complex agentic scaffolds. These results demonstrate that simple RL succeeds when environments scale.
[713] From Associations to Activations: Comparing Behavioral and Hidden-State Semantic Geometry in LLMs
Louis Schiekiera, Max Zimmer, Christophe Roux, Sebastian Pokutta, Fritz Günther
Main category: cs.LG
TL;DR: LLM hidden-state geometry can be recovered from psycholinguistic behavioral experiments, with forced-choice tasks showing stronger alignment than free association tasks.
Details
Motivation: To understand whether an LLM's internal semantic representations (hidden-state geometry) can be inferred from its behavioral responses in psycholinguistic experiments, rather than requiring direct access to model internals.Method: Used eight instruction-tuned transformer models with two experimental paradigms: similarity-based forced choice and free association over 5,000-word vocabulary (17.5M+ trials). Applied representational similarity analysis to compare behavioral similarity matrices with layerwise hidden-state similarity, benchmarking against FastText, BERT, and cross-model consensus.
Result: Forced-choice behavior aligns substantially more with hidden-state geometry than free association. Behavioral similarity (especially forced choice) predicts unseen hidden-state similarities beyond lexical baselines and cross-model consensus, showing behavior-only measurements retain recoverable information about internal semantic geometry.
Conclusion: Behavioral tasks can uncover information about LLMs’ internal cognitive states, with forced-choice paradigms being particularly effective for recovering hidden-state geometry from observable behavior.
Abstract: We investigate the extent to which an LLM’s hidden-state geometry can be recovered from its behavior in psycholinguistic experiments. Across eight instruction-tuned transformer models, we run two experimental paradigms – similarity-based forced choice and free association – over a shared 5,000-word vocabulary, collecting 17.5M+ trials to build behavior-based similarity matrices. Using representational similarity analysis, we compare behavioral geometries to layerwise hidden-state similarity and benchmark against FastText, BERT, and cross-model consensus. We find that forced-choice behavior aligns substantially more with hidden-state geometry than free association. In a held-out-words regression, behavioral similarity (especially forced choice) predicts unseen hidden-state similarities beyond lexical baselines and cross-model consensus, indicating that behavior-only measurements retain recoverable information about internal semantic geometry. Finally, we discuss implications for the ability of behavioral tasks to uncover hidden cognitive states.
[714] Self-Improving World Modelling with Latent Actions
Yifu Qiu, Zheng Zhao, Waylon Li, Yftah Ziser, Anna Korhonen, Shay B. Cohen, Edoardo M. Ponti
Main category: cs.LG
TL;DR: SWIRL is a self-improvement framework that learns world models from state-only sequences by treating actions as latent variables and alternating between forward world modeling and inverse dynamics modeling.
Details
Motivation: Learning world models for LLMs and VLMs typically requires costly action-labeled trajectories. The paper aims to enable learning from state-only sequences by treating actions as latent variables.Method: SWIRL alternates between Forward World Modeling (FWM) P_θ(Y|X,Z) and Inverse Dynamics Modeling (IDM) Q_φ(Z|X,Y) through two phases: Variational Information Maximization (updates FWM to maximize conditional mutual information) and ELBO Maximization (updates IDM to explain observed transitions). Both models are trained with reinforcement learning (GRPO) using the opposite frozen model’s log-probability as reward.
Result: SWIRL achieves significant gains across multiple environments: 16% on AURORABench, 28% on ByteMorph, 16% on WorldPredictionBench, and 14% on StableToolBench.
Conclusion: SWIRL provides an effective framework for learning world models from state-only sequences with theoretical guarantees, demonstrating strong performance across diverse multimodal environments.
Abstract: Internal modelling of the world – predicting transitions between previous states $X$ and next states $Y$ under actions $Z$ – is essential to reasoning and planning for LLMs and VLMs. Learning such models typically requires costly action-labelled trajectories. We propose SWIRL, a self-improvement framework that learns from state-only sequences by treating actions as a latent variable and alternating between Forward World Modelling (FWM) $P_θ(Y|X,Z)$ and an Inverse Dynamics Modelling (IDM) $Q_φ(Z|X,Y)$. SWIRL iterates two phases: (1) Variational Information Maximisation, which updates the FWM to generate next states that maximise conditional mutual information with latent actions given prior states, encouraging identifiable consistency; and (2) ELBO Maximisation, which updates the IDM to explain observed transitions, effectively performing coordinate ascent. Both models are trained with reinforcement learning (specifically, GRPO) with the opposite frozen model’s log-probability as a reward signal. We provide theoretical learnability guarantees for both updates, and evaluate SWIRL on LLMs and VLMs across multiple environments: single-turn and multi-turn open-world visual dynamics and synthetic textual environments for physics, web, and tool calling. SWIRL achieves gains of 16% on AURORABench, 28% on ByteMorph, 16% on WorldPredictionBench, and 14% on StableToolBench.
[715] Learning with Subset Stacking
Ş. İlker Birbil, Sinan Yıldırım, Samet Çopur, M. Hakan Akyüz
Main category: cs.LG
TL;DR: LESS is a new regression algorithm that handles heterogeneous input-output relationships by creating localized subsets, training local predictors, and combining them through stacking, with bagging and boosting variants.
Details
Motivation: The paper addresses the challenge of regression problems where the relationship between input variables and output exhibits heterogeneous behavior across different regions of the predictor space, requiring more flexible modeling approaches than traditional global methods.Method: The LESS algorithm: 1) generates subsets concentrated around random points in input space, 2) trains local predictors for each subset, 3) combines predictors using a novel stacking approach. Includes bagging and boosting variants.
Result: LESS shows high competitiveness against state-of-the-art methods on several datasets, demonstrating effectiveness in handling heterogeneous regression problems.
Conclusion: LESS provides an effective approach for regression with heterogeneous relationships, offering competitive performance through its localized subset learning and stacking combination strategy.
Abstract: We propose a new regression algorithm that learns from a set of input-output pairs. Our algorithm is designed for populations where the relation between the input variables and the output variable exhibits a heterogeneous behavior across the predictor space. The algorithm starts with generating subsets that are concentrated around random points in the input space. This is followed by training a local predictor for each subset. Those predictors are then combined in a novel way to yield an overall predictor. We call this algorithm “LEarning with Subset Stacking” or LESS, due to its resemblance to the method of stacking regressors. We offer bagging and boosting variants of LESS and test against the state-of-the-art methods on several datasets. Our comparison shows that LESS is highly competitive.
[716] When is Offline Policy Selection Sample Efficient for Reinforcement Learning?
Vincent Liu, Prabhat Nagarajan, Andrew Patterson, Martha White
Main category: cs.LG
TL;DR: This paper studies offline policy selection (OPS) in reinforcement learning, connecting it to off-policy policy evaluation (OPE) and Bellman error (BE) estimation, showing OPS is as hard as OPE in worst case, proposing a BE-based method called IBES, and demonstrating empirical results.
Details
Motivation: Offline RL algorithms require hyperparameter tuning and policy selection before deployment, but there's limited understanding of the fundamental limits of offline policy selection (OPS). The paper aims to clarify when sample-efficient OPS is possible by connecting it to OPE and BE estimation.Method: 1) Prove a reduction from OPE to OPS showing OPS is as hard as OPE in worst case. 2) Connect BE estimation to OPS, showing BE can be used as a tool for OPS when conditions are met. 3) Propose Identifiable BE Selection (IBES), a BE-based method with straightforward hyperparameter selection.
Result: Theoretical results show OPS is as hard as OPE in worst case, but BE-based methods can be more sample efficient when conditions are met. Empirical study compares OPE and IBES, showing difficulty of OPS on offline Atari benchmark dataset.
Conclusion: OPS is fundamentally challenging (as hard as OPE), but BE-based approaches like IBES offer advantages when conditions are satisfied. The work provides theoretical clarity on OPS limits and practical methods for policy selection in offline RL.
Abstract: Offline reinforcement learning algorithms often require careful hyperparameter tuning. Before deployment, we need to select amongst a set of candidate policies. However, there is limited understanding about the fundamental limits of this offline policy selection (OPS) problem. In this work we provide clarity on when sample efficient OPS is possible, primarily by connecting OPS to off-policy policy evaluation (OPE) and Bellman error (BE) estimation. We first show a hardness result, that in the worst case, OPS is just as hard as OPE, by proving a reduction of OPE to OPS. As a result, no OPS method can be more sample efficient than OPE in the worst case. We then connect BE estimation to the OPS problem, showing how BE can be used as a tool for OPS. While BE-based methods generally require stronger requirements than OPE, when those conditions are met they can be more sample efficient. Building on this insight, we propose a BE method for OPS, called Identifiable BE Selection (IBES), that has a straightforward method for selecting its own hyperparameters. We conclude with an empirical study comparing OPE and IBES, and by showing the difficulty of OPS on an offline Atari benchmark dataset.
[717] Permutation-based Inference for Variational Learning of Directed Acyclic Graphs
Edwin V. Bonilla, Pantelis Elinas, He Zhao, Maurizio Filippone, Vassili Kitsios, Terry O’Kane
Main category: cs.LG
TL;DR: PIVID: A Bayesian method for DAG structure estimation using variational inference with continuous relaxations of discrete distributions over permutations and DAGs
Details
Motivation: Bayesian approaches for DAG structure estimation from observational data face challenges in representing distributions over DAGs and estimating posteriors in combinatorial spaces, despite their advantages in uncertainty quantification and handling identifiability issues.Method: PIVID jointly infers distributions over permutations and DAGs using variational inference with continuous relaxations of discrete distributions, enabling efficient posterior estimation in combinatorial spaces.
Result: Experiments on synthetic and real-world datasets show PIVID outperforms deterministic and Bayesian approaches, achieving superior accuracy-uncertainty trade-offs with efficient scaling to many variables.
Conclusion: PIVID provides an effective Bayesian approach for DAG structure estimation that addresses key challenges in representing distributions over combinatorial structures while maintaining computational efficiency.
Abstract: Estimating the structure of Bayesian networks as directed acyclic graphs (DAGs) from observational data is a fundamental challenge, particularly in causal discovery. Bayesian approaches excel by quantifying uncertainty and addressing identifiability, but key obstacles remain: (i) representing distributions over DAGs and (ii) estimating a posterior in the underlying combinatorial space. We introduce PIVID, a method that jointly infers a distribution over permutations and DAGs using variational inference and continuous relaxations of discrete distributions. Through experiments on synthetic and real-world datasets, we show that PIVID can outperform deterministic and Bayesian approaches, achieving superior accuracy-uncertainty trade-offs while scaling efficiently with the number of variables.
[718] Optimal Design for Human Preference Elicitation
Subhojyoti Mukherjee, Anusha Lalitha, Kousha Kalantari, Aniket Deshmukh, Ge Liu, Yifei Ma, Branislav Kveton
Main category: cs.LG
TL;DR: Generalized optimal design framework for efficient human preference elicitation in learning preference models, with algorithms for both absolute and ranking feedback on item lists.
Details
Motivation: High-quality human annotations are costly, motivating the need for efficient preference elicitation methods to learn preference models from human feedback.Method: Generalize optimal designs approach to compute optimal information-gathering policies over lists of items representing potential questions with answers. Develop efficient algorithms for both absolute and ranking feedback models on items in lists.
Result: Designed efficient algorithms for preference elicitation and demonstrated their practicality by evaluating them on existing question-answering problems.
Conclusion: The proposed generalized optimal design framework provides efficient methods for human preference elicitation that can be applied to various feedback models and practical applications.
Abstract: Learning of preference models from human feedback has been central to recent advances in artificial intelligence. Motivated by the cost of obtaining high-quality human annotations, we study efficient human preference elicitation for learning preference models. The key idea in our work is to generalize optimal designs, an approach to computing optimal information-gathering policies, to lists of items that represent potential questions with answers. The policy is a distribution over the lists and we elicit preferences from them proportionally to their probabilities. To show the generality of our ideas, we study both absolute and ranking feedback models on items in the list. We design efficient algorithms for both and analyze them. Finally, we demonstrate that our algorithms are practical by evaluating them on existing question-answering problems.
[719] Exact Solution to Data-Driven Inverse Optimization of MILPs in Finite Time via Gradient-Based Methods
Akira Kitaoka
Main category: cs.LG
TL;DR: Paper proposes gradient-based optimization methods to solve data-driven inverse optimization problems for mixed integer linear programs by using Lipschitz continuous convex suboptimality loss, achieving finite convergence.
Details
Motivation: Data-driven inverse optimization problems (DDIOPs) for mixed integer linear programs (MILPs) face challenges because prediction loss on features becomes discontinuous with respect to weights, making gradient-based optimization difficult.Method: Focuses on Lipschitz continuous and convex suboptimality loss instead of discontinuous prediction loss. Exploits convex and piecewise-linear structure and interiority of minimum set to show that gradient-based methods (including projected subgradient descent) reach minimum loss in finite iterations.
Result: Shows that projected subgradient descent and other gradient-based methods achieve finite convergence to minimum suboptimality loss, thereby exactly solving DDIOP for MILPs. Also derives upper bound on iterations needed and confirms finite-step behavior through numerical experiments.
Conclusion: Gradient-based optimization methods can effectively solve DDIOPs for MILPs using convex suboptimality loss, achieving finite convergence and providing theoretical guarantees on iteration bounds.
Abstract: A data-driven inverse optimization problem (DDIOP) seeks to estimate an objective function (i.e., weights) that is consistent with observed optimal-solution data, and is important in many applications, including those involving mixed integer linear programs (MILPs). In the DDIOP for MILPs, the prediction loss on features (PLF), defined as the discrepancy between observed and predicted feature values, becomes discontinuous with respect to the weights, which makes it difficult to apply gradient-based optimization. To address this issue, we focus on a Lipschitz continuous and convex suboptimality loss. By exploiting its convex and piecewise-linear structure and the interiority of the minimum set, we show that a broad class of gradient-based optimization methods, including projected subgradient descent (PSGD), reaches the minimum suboptimality loss value in a finite number of iterations, thereby exactly solving the DDIOP for MILPs. Furthermore, as a corollary, we show that PSGD attains the minimum PLF in finitely many iterations. We also derive an upper bound on the number of iterations required for PSGD to reach finite convergence, and confirm the finite-step behavior through numerical experiments.
[720] Synergizing Foundation Models and Federated Learning: A Survey
Shenghui Li, Fanghua Ye, Meng Fang, Jiaxu Zhao, Yun-Hin Chan, Edith C. H. Ngai, Thiemo Voigt
Main category: cs.LG
TL;DR: A comprehensive survey paper on Federated Foundation Models (FedFM), reviewing the synergy between Foundation Models and Federated Learning for privacy-preserving AI, with multi-tiered taxonomy covering efficiency, adaptability, and trustworthiness.
Details
Motivation: The motivation is to systematically review and organize the emerging field of Federated Foundation Models (FedFM), which combines the power of large foundation models with the privacy-preserving benefits of federated learning, addressing the need for collaborative AI development while protecting data privacy.Method: The paper conducts a systematic literature review and presents a comprehensive multi-tiered taxonomy based on three major dimensions: efficiency, adaptability, and trustworthiness. It also reviews existing libraries, benchmarks, and real-world applications across multiple domains.
Result: The survey provides a structured understanding of the FedFM landscape, identifies current state-of-the-art approaches, reviews available tools and benchmarks, and outlines diverse real-world applications across various domains.
Conclusion: FedFM represents a promising paradigm for privacy-preserving AI that combines the strengths of foundation models and federated learning. The survey serves as a comprehensive resource for researchers and practitioners, pointing toward future innovations in this rapidly evolving area.
Abstract: Over the past few years, the landscape of Artificial Intelligence (AI) has been reshaped by the emergence of Foundation Models (FMs). Pre-trained on massive datasets, these models exhibit exceptional performance across diverse downstream tasks through adaptation techniques like fine-tuning and prompt learning. More recently, the synergy of FMs and Federated Learning (FL) has emerged as a promising paradigm, often termed Federated Foundation Models (FedFM), allowing for collaborative model adaptation while preserving data privacy. This survey paper provides a systematic review of the current state of the art in FedFM, offering insights and guidance into the evolving landscape. Specifically, we present a comprehensive multi-tiered taxonomy based on three major dimensions, namely efficiency, adaptability, and trustworthiness. To facilitate practical implementation and experimental research, we undertake a thorough review of existing libraries and benchmarks. Furthermore, we discuss the diverse real-world applications of this paradigm across multiple domains. Finally, we outline promising research directions to foster future advancements in FedFM. Overall, this survey serves as a resource for researchers and practitioners, offering a thorough understanding of FedFM’s role in revolutionizing privacy-preserving AI and pointing toward future innovations in this promising area. A periodically updated paper collection on FM-FL is available at https://github.com/lishenghui/awesome-fm-fl.
[721] GraphFM: A generalist graph transformer that learns transferable representations across diverse domains
Divyansha Lachi, Mehdi Azabou, Vinam Arora, Eva Dyer
Main category: cs.LG
TL;DR: GraphFM: A scalable multi-graph pretraining approach using Perceiver-based encoder with learned latent tokens to compress domain-specific features into shared latent space, enabling generalization across 152 diverse graph datasets.
Details
Motivation: Traditional GNNs require specialized models and significant hyperparameter tuning for each individual graph dataset due to unique structures and features, limiting scalability and generalizability. Need for a single generalist model capable of performing across multiple diverse graph structures and tasks.Method: Uses Perceiver-based encoder with learned latent tokens to compress domain-specific features into shared latent space. Proposes techniques for scaling up graph training on datasets of different sizes. Trained on 152 distinct graph datasets (7.4M nodes, 189M edges) across domains like molecules, citation networks, and product graphs.
Result: Training on diverse datasets improves performance over single-source pretraining. Pretraining with mixture of synthetic and real graphs enhances adaptability and stability. Achieves competitive performance with state-of-the-art models across various node classification tasks.
Conclusion: GraphFM reduces burden of dataset-specific training and provides a single generalist model capable of performing across multiple diverse graph structures and tasks, demonstrating benefits of multi-graph pretraining at scale.
Abstract: Graph neural networks (GNNs) are often trained on individual datasets, requiring specialized models and significant hyperparameter tuning due to the unique structures and features of each dataset. This approach limits the scalability and generalizability of GNNs, as models must be tailored for each specific graph type. To address these challenges, we introduce GraphFM, a scalable multi-graph pretraining approach designed for learning across diverse graph datasets. GraphFM uses a Perceiver-based encoder with learned latent tokens to compress domain-specific features into a shared latent space, enabling generalization across graph domains. We propose new techniques for scaling up graph training on datasets of different sizes, allowing us to train GraphFM on 152 distinct graph datasets, containing a total of 7.4 million nodes and 189 million edges. This allows us to study the effect of scale on pretraining across domains such as molecules, citation networks, and product graphs, and show that training on diverse datasets improves performance over single-source pretraining. Additionally, pretraining with a mixture of synthetic and real graphs enhances adaptability and stability, leading to competitive performance with state-of-the-art models across various node classification tasks. This approach reduces the burden of dataset-specific training and provides a single generalist model capable of performing across multiple diverse graph structures and tasks. Code is available at https://github.com/nerdslab/GraphFM.
[722] Benchmarking AI-based data assimilation to advance data-driven global weather forecasting
Wuxin Wang, Weicheng Ni, Ben Fei, Tao Han, Lilan Huang, Taikang Yuan, Xiaoyong Li, Lei Bai, Boheng Duan, Kaijun Ren
Main category: cs.LG
TL;DR: DABench is a benchmark for AI-based Data Assimilation methods that integrates real-world observations for fair comparison and validation of long-term closed-loop DA cycles, supporting both deterministic and ensemble configurations.
Details
Motivation: The rapid expansion of AI-based Data Assimilation research lacks an objective, comprehensive, and real-world benchmark for fair comparison of diverse methods, hindering proper evaluation and advancement in the field.Method: Introduces DABench benchmark that integrates real-world observations to provide objective validation platform for long-term closed-loop DA cycles. Uses dual-validation with reanalysis data and independent radiosonde observations to assess AI-based DA performance in generating initial conditions for AI-based weather forecasting models.
Result: AI-based DA achieves performance competitive with state-of-the-art AI-driven four-dimensional variational frameworks across both global weather DA and medium-range forecasting metrics, as validated through dual-validation approach.
Conclusion: DABench provides an essential benchmark for accelerating AI-based DA development for global weather forecasting, enabling fair comparison and validation of methods while demonstrating competitive performance of AI-based approaches.
Abstract: Research on Artificial Intelligence (AI)-based Data Assimilation (DA) is expanding rapidly. However, the absence of an objective, comprehensive, and real-world benchmark hinders the fair comparison of diverse methods. Here, we introduce DABench, a benchmark designed for contributing to the development and evaluation of AI-based DA methods. By integrating real-world observations, DABench provides an objective and fair platform for validating long-term closed-loop DA cycles, supporting both deterministic and ensemble configurations. Furthermore, we assess the efficacy of AI-based DA in generating initial conditions for the advanced AI-based weather forecasting model to produce accurate medium-range global weather forecasting. Our dual-validation, utilizing both reanalysis data and independent radiosonde observations, demonstrates that AI-based DA achieves performance competitive with state-of-the-art AI-driven four-dimensional variational frameworks across both global weather DA and medium-range forecasting metrics. We invite the research community to utilize DABench to accelerate the advancement of AI-based DA for global weather forecasting.
[723] MixLinear: Extreme Low Resource Multivariate Time Series Forecasting with 0.1K Parameters
Aitian Ma, Dongsheng Luo, Mo Sha
Main category: cs.LG
TL;DR: MixLinear: Ultra-lightweight multivariate time series forecasting model for resource-constrained devices that captures both temporal and frequency domain features with minimal parameters (0.1K).
Details
Motivation: Transformer-based models for Long-term Time Series Forecasting (LTSF) offer high accuracy but are computationally intensive, while linear models reduce overhead but may lack comprehensive feature capture. Need for efficient models deployable on hardware-constrained devices.Method: MixLinear models intra-segment and inter-segment variations in time domain and extracts frequency variations from low-dimensional latent space in frequency domain. Reduces parameter scale from O(n²) to O(n) for downsampled n-length input/output one-layer linear model.
Result: Extensive evaluations on four benchmark datasets show forecasting performance comparable to or surpassing state-of-the-art models with significantly fewer parameters (0.1K).
Conclusion: MixLinear achieves efficient computation without sacrificing accuracy, making it well-suited for deployment on devices with limited computational capacity.
Abstract: Recently, there has been a growing interest in Long-term Time Series Forecasting (LTSF), which involves predicting long-term future values by analyzing a large amount of historical time-series data to identify patterns and trends. There exist significant challenges in LTSF due to its complex temporal dependencies and high computational demands. Although Transformer-based models offer high forecasting accuracy, they are often too compute-intensive to be deployed on devices with hardware constraints. On the other hand, the linear models aim to reduce the computational overhead by employing either decomposition methods in the time domain or compact representations in the frequency domain. In this paper, we propose MixLinear, an ultra-lightweight multivariate time series forecasting model specifically designed for resource-constrained devices. MixLinear effectively captures both temporal and frequency domain features by modeling intra-segment and inter-segment variations in the time domain and extracting frequency variations from a low-dimensional latent space in the frequency domain. By reducing the parameter scale of a downsampled $n$-length input/output one-layer linear model from $O(n^2)$ to $O(n)$, MixLinear achieves efficient computation without sacrificing accuracy. Extensive evaluations with four benchmark datasets show that MixLinear attains forecasting performance comparable to, or surpassing, state-of-the-art models with significantly fewer parameters ($0.1K$), which makes it well-suited for deployment on devices with limited computational capacity.
[724] Online Posterior Sampling with a Diffusion Prior
Branislav Kveton, Boris Oreshkin, Youngsuk Park, Aniket Deshmukh, Rui Song
Main category: cs.LG
TL;DR: Proposes approximate posterior sampling algorithms for contextual bandits using diffusion model priors instead of Gaussian priors, with sampling via reverse diffusion process and Laplace approximation.
Details
Motivation: Gaussian priors in contextual bandits are computationally efficient but limited in describing complex distributions. The authors aim to develop more flexible priors using diffusion models while maintaining efficiency.Method: Sample from a chain of approximate conditional posteriors using reverse diffusion process stages, each obtained via Laplace approximation. Inherits simplicity from Gaussian prior methods while using diffusion model priors.
Result: Algorithms are asymptotically consistent and perform well empirically on various contextual bandit problems, demonstrating effectiveness of diffusion model priors.
Conclusion: Successfully extends posterior sampling to diffusion model priors, providing more flexible distribution modeling while maintaining computational efficiency similar to Gaussian priors.
Abstract: Posterior sampling in contextual bandits with a Gaussian prior can be implemented exactly or approximately using the Laplace approximation. The Gaussian prior is computationally efficient but it cannot describe complex distributions. In this work, we propose approximate posterior sampling algorithms for contextual bandits with a diffusion model prior. The key idea is to sample from a chain of approximate conditional posteriors, one for each stage of the reverse diffusion process, which are obtained by the Laplace approximation. Our approximations are motivated by posterior sampling with a Gaussian prior, and inherit its simplicity and efficiency. They are asymptotically consistent and perform well empirically on a variety of contextual bandit problems.
[725] Model-based Large Language Model Customization as Service
Zhaomin Wu, Jizhou Guo, Junyi Hou, Bingsheng He, Lixin Fan, Qiang Yang
Main category: cs.LG
TL;DR: Llamdex is a privacy-preserving framework for customizing LLM services where clients upload pre-trained domain-specific models instead of sensitive data, using connection modules to integrate them into base LLMs while maintaining privacy and efficiency.
Details
Motivation: Current LLM customization services require users to upload sensitive domain data for fine-tuning, posing significant privacy risks. Differentially private data synthesis alternatives introduce excessive noise, reducing effectiveness. There's a need for privacy-preserving LLM customization that maintains performance.Method: Clients upload pre-trained domain-specific models (optionally DP-protected with lower noise) rather than raw data. These models are inserted into base LLMs via connection modules. The connection modules are trained without requiring sensitive domain data, enabling privacy-preserving customization.
Result: Llamdex improves domain-specific accuracy by up to 26% over state-of-the-art private data synthesis methods under identical privacy constraints. It maintains inference efficiency comparable to original LLM services by eliminating the need for users to provide domain context in queries.
Conclusion: Llamdex provides an effective privacy-preserving framework for LLM customization that outperforms existing DP data synthesis methods while maintaining efficiency, addressing the privacy-performance trade-off in domain-specific LLM applications.
Abstract: Prominent Large Language Model (LLM) services from providers like OpenAI and Google excel at general tasks but often underperform on domain-specific applications. Current customization services for these LLMs typically require users to upload data for fine-tuning, posing significant privacy risks. While differentially private (DP) data synthesis presents a potential alternative, its application commonly results in low effectiveness due to the introduction of excessive noise on data for DP. To overcome this, we introduce Llamdex, a novel framework that facilitates LLM customization as a service, where the client uploads pre-trained domain-specific models rather than data. This client-uploaded model, optionally protected by DP with much lower noise, is inserted into the base LLM via connection modules. Significantly, these connecting modules are trained without requiring sensitive domain data, enabling clients to customize LLM services while preserving data privacy. Experiments demonstrate that Llamdex improves domain-specific accuracy by up to 26% over state-of-the-art private data synthesis methods under identical privacy constraints and, by obviating the need for users to provide domain context within queries, maintains inference efficiency comparable to the original LLM service.
[726] Denoising diffusion probabilistic models are optimally adaptive to unknown low dimensionality
Zhihan Huang, Yuting Wei, Yuxin Chen
Main category: cs.LG
TL;DR: DDPM achieves optimal iteration complexity scaling with intrinsic data dimension k, not ambient dimension, enabling efficient sampling for low-dimensional data distributions.
Details
Motivation: Current DDPM theory shows iteration complexity proportional to ambient data dimension, which is overly conservative and fails to explain practical efficiency. Recent work shows DDPM can exploit intrinsic low dimensionality for speed-ups, but optimal adaptivity to unknown low dimensionality needs investigation.Method: Theoretical analysis of DDPM convergence for data distributions with intrinsic dimension k, proving iteration complexity scales nearly linearly with k rather than ambient dimension.
Result: Proves DDPM iteration complexity scales nearly linearly with intrinsic dimension k, which is optimal when using KL divergence to measure distributional discrepancy. This matches independent concurrent work establishing similar guarantees.
Conclusion: DDPM can achieve optimal adaptivity to unknown low dimensionality, with iteration complexity scaling nearly linearly with intrinsic dimension rather than ambient dimension, explaining practical efficiency.
Abstract: The denoising diffusion probabilistic model (DDPM) has emerged as a mainstream generative model in generative AI. While sharp convergence guarantees have been established for the DDPM, the iteration complexity is, in general, proportional to the ambient data dimension, resulting in overly conservative theory that fails to explain its practical efficiency. This has motivated the recent work Li and Yan (2024a) to investigate how the DDPM can achieve sampling speed-ups through automatic exploitation of intrinsic low dimensionality of data. We strengthen this line of work by demonstrating, in some sense, optimal adaptivity to unknown low dimensionality. For a broad class of data distributions with intrinsic dimension $k$, we prove that the iteration complexity of the DDPM scales nearly linearly with $k$, which is optimal when using KL divergence to measure distributional discrepancy. Notably, our work is closely aligned with the independent concurrent work Potaptchik et al. (2024) – posted two weeks prior to ours – in establishing nearly linear-$k$ convergence guarantees for the DDPM.
[727] VCDF: A Validated Consensus-Driven Framework for Time Series Causal Discovery
Gene Yu, Ce Guo, Wayne Luk
Main category: cs.LG
TL;DR: VCDF is a method-agnostic framework that improves robustness of time series causal discovery by evaluating causal relation stability across temporal subsets, requiring no modification to base algorithms.
Details
Motivation: Existing time series causal discovery methods are sensitive to noise, non-stationarity, and sampling variability, limiting their reliability in real-world applications.Method: VCDF evaluates stability of causal relations across blocked temporal subsets, creating a consensus-driven framework that can be applied to existing methods like VAR-LiNGAM and PCMCI without algorithm modification.
Result: VCDF improves VAR-LiNGAM by 0.08-0.12 in F1 scores across diverse data characteristics, with gains up to 0.18 for longer sequences (1000+). It demonstrates enhanced stability on simulated fMRI and IT-monitoring data.
Conclusion: VCDF provides an effective reliability layer for time series causal discovery that improves robustness without altering underlying modeling assumptions.
Abstract: Time series causal discovery is essential for understanding dynamic systems, yet many existing methods remain sensitive to noise, non-stationarity, and sampling variability. We propose the Validated Consensus-Driven Framework (VCDF), a simple and method-agnostic layer that improves robustness by evaluating the stability of causal relations across blocked temporal subsets. VCDF requires no modification to base algorithms and can be applied to methods such as VAR-LiNGAM and PCMCI. Experiments on synthetic datasets show that VCDF improves VAR-LiNGAM by approximately 0.08-0.12 in both window and summary F1 scores across diverse data characteristics, with gains most pronounced for moderate-to-long sequences. The framework also benefits from longer sequences, yielding up to 0.18 absolute improvement on time series of length 1000 and above. Evaluations on simulated fMRI data and IT-monitoring scenarios further demonstrate enhanced stability and structural accuracy under realistic noise conditions. VCDF provides an effective reliability layer for time series causal discovery without altering underlying modeling assumptions.
[728] Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection
Vima Gupta, Jae Hyung Ju, Kartik Sinha, Ada Gavrilovska, Anand Padmanabha Iyer
Main category: cs.LG
TL;DR: LYNX enables efficient Mixture-of-Expert (MoE) inference by dynamically remapping experts at runtime to resolve the tension between batching requirements and selective parameter activation.
Details
Motivation: MoE models face a fundamental tension in serving: batching is critical for performance but forces activation of all experts, negating MoE benefits and exacerbating memory bandwidth bottlenecks. Existing efficient MoE inference approaches require extensive workload-specific tuning and cannot resolve this tension.Method: LYNX provides a lightweight runtime dynamic expert remapping technique that exploits key observations about MoE models. It depends only on information already available in the models and works in a workload-agnostic fashion without requiring extensive tuning.
Result: Evaluation on four state-of-the-art model families across nine benchmarks shows LYNX achieves up to 1.23x throughput improvement while simultaneously improving accuracy by up to 4% in most tasks, with only negligible accuracy loss (<1%) in significantly hard tasks. LYNX is complementary to existing techniques, boosting their performance by up to 1.38x.
Conclusion: LYNX successfully resolves the tension between batching requirements and selective parameter activation in MoE models, enabling efficient MoE inference in a workload-agnostic manner through dynamic expert remapping.
Abstract: Selective parameter activation provided by Mixture-of-Expert (MoE) models have made them a popular choice in modern foundational models. However, MoEs face a fundamental tension when employed for serving. Batching, critical for performance in serving, forces the activation of all experts, thereby negating MoEs’ benefits and exacerbating memory bandwidth bottlenecks. Existing work on efficient MoE inference are unable to resolve this tension even with extensive workload-specific tuning. We present LYNX, a system that enables efficient MoE inference in a workload-agnostic fashion. Exploiting several key observations that we uncover in this work, LYNX provides a light-weight run-time dynamic expert remapping technique that depends only on information already available in the models. Our evaluation of LYNX on four state-of-the-art model families across nine benchmarks shows that it achieves up to 1.23x improvement in throughput while simultaneously improving accuracy by up to 4% in the majority of the tasks, and incurs only a negligible accuracy loss of less than 1% points in significantly hard tasks. Further, LYNX is complementary to existing techniques where it additionally boosts their performance by up to 1.38x.
[729] Bayesian Flow Is All You Need to Sample Out-of-Distribution Chemical Spaces
Nianze Tao, Minori Abe
Main category: cs.LG
TL;DR: ChemBFN with semi-autoregressive training and RL-enhanced controllable ODE sampling enables out-of-distribution molecular generation for drug design.
Details
Motivation: Current diffusion models struggle with out-of-distribution molecular generation as they aim to fit training data distribution, but drug design requires generating novel molecules with better properties than existing ones.Method: Uses Bayesian flow network (ChemBFN) with semi-autoregressive training, reinforcement learning strategy, and controllable ODE solver-like generation process for accelerated sampling.
Result: ChemBFN generates high-quality out-of-distribution molecules meeting multiple scenarios and surpasses state-of-the-art models.
Conclusion: ChemBFN with semi-autoregressive approach enables effective out-of-distribution molecular generation for drug design, with theoretical analysis supporting its capabilities.
Abstract: Generating novel molecules with higher properties than the training space, namely the out-of-distribution generation, is important for de novo drug design. However, it is not easy for distribution learning-based models, for example diffusion models, to solve this challenge as these methods are designed to fit the distribution of training data as close as possible. In this paper, we show that Bayesian flow network, especially ChemBFN model, is capable of intrinsically generating high quality out-of-distribution samples that meet several scenarios. A reinforcement learning strategy is added to the ChemBFN and a controllable ordinary differential equation solver-like generating process is employed that accelerate the sampling processes. Most importantly, we introduce a semi-autoregressive strategy during training and inference that enhances the model performance and surpass the state-of-the-art models. A theoretical analysis of out-of-distribution generation in ChemBFN with semi-autoregressive approach is included as well.
[730] Regularized Top-$k$: A Bayesian Framework for Gradient Sparsification
Ali Bereyhi, Ben Liang, Gary Boudreau, Ali Afana
Main category: cs.LG
TL;DR: RegTop-k: A novel gradient sparsification method that controls learning rate scaling through Bayesian inference, outperforming Top-k at high compression ratios in distributed training.
Details
Motivation: Error accumulation in gradient sparsification (like Top-k) can deteriorate convergence by scaling learning rates unevenly. The authors aim to develop a sparsification scheme that controls this learning rate scaling effect.Method: Formulates gradient sparsification as Bayesian inference problem, derives optimal sparsification mask as MAP estimator using Top-k prior. Creates RegTop-k algorithm that uses past aggregated gradients to compute posterior statistics and prioritize local accumulated gradients accordingly.
Result: RegTop-k converges to global optimum at high compression ratios where Top-k stagnates. Validated on distributed linear regression, ResNet-18 on CIFAR-10, and fine-tuning vision models on ImageNette, showing significant improvements over Top-k as compression increases.
Conclusion: RegTop-k effectively controls learning rate scaling in gradient sparsification, enabling better convergence than Top-k at high compression ratios for distributed training of neural networks.
Abstract: Error accumulation is effective for gradient sparsification in distributed settings: initially-unselected gradient entries are eventually selected as their accumulated error exceeds a certain level. The accumulation essentially behaves as a scaling of the learning rate for the selected entries. Although this property prevents the slow-down of lateral movements in distributed gradient descent, it can deteriorate convergence in some settings. This work proposes a novel sparsification scheme that controls the learning rate scaling of error accumulation. The development of this scheme follows two major steps: first, gradient sparsification is formulated as an inverse probability (inference) problem, and the Bayesian optimal sparsification mask is derived as a maximum-a-posteriori estimator. Using the prior distribution inherited from Top-k, we derive a new sparsification algorithm which can be interpreted as a regularized form of Top-k. We call this algorithm regularized Top-k (RegTop-k). It utilizes past aggregated gradients to evaluate posterior statistics of the next aggregation. It then prioritizes the local accumulated gradient entries based on these posterior statistics. We validate our derivation through various numerical experiments. In distributed linear regression, it is observed that while Top-k remains at a fixed distance from the global optimum, RegTop-k converges to the global optimum at significantly higher compression ratios. We further demonstrate the generalization of this observation by employing RegTop-k in distributed training of ResNet-18 on CIFAR-10, as well as fine-tuning of multiple computer vision models on the ImageNette dataset. Our numerical results confirm that as the compression ratio increases, RegTop-k sparsification noticeably outperforms Top-k.
[731] Adaptive Width Neural Networks
Federico Errica, Henrik Christiansen, Viktor Zaverkin, Mathias Niepert, Francesco Alesiani
Main category: cs.LG
TL;DR: A method to learn neural network layer widths during training via backpropagation, enabling adaptive width selection and easy truncation for compute-performance trade-offs.
Details
Motivation: Traditional width selection methods (manual tuning, grid search, NAS) are inefficient and costly, especially for large foundation models where hyperparameter tuning is infeasible due to massive training costs.Method: Introduces a technique to jointly optimize layer widths and parameters through standard backpropagation, allowing unbounded width learning during training. The approach enables easy network truncation and dynamic compression.
Result: Method applied across diverse data domains (tables, images, text, sequences, graphs) showing width adaptation to task difficulty. Achieves smooth performance-compute trade-offs through truncation.
Conclusion: Provides a viable alternative to traditional width selection methods, particularly valuable for large foundation models where conventional hyperparameter tuning is prohibitively expensive.
Abstract: For almost 70 years, researchers have typically selected the width of neural networks’ layers either manually or through automated hyperparameter tuning methods such as grid search and, more recently, neural architecture search. This paper challenges the status quo by introducing an easy-to-use technique to learn an unbounded width of a neural network’s layer during training. The method jointly optimizes the width and the parameters of each layer via standard backpropagation. We apply the technique to a broad range of data domains such as tables, images, text, sequences, and graphs, showing how the width adapts to the task’s difficulty. A by product of our width learning approach is the easy truncation of the trained network at virtually zero cost, achieving a smooth trade-off between performance and compute resources. Alternatively, one can dynamically compress the network until performances do not degrade. In light of recent foundation models trained on large datasets, requiring billions of parameters and where hyper-parameter tuning is unfeasible due to huge training costs, our approach introduces a viable alternative for width learning.
[732] SWIFT: Mapping Sub-series with Wavelet Decomposition Improves Time Series Forecasting
Wenxuan Xie, Fanpu Cao
Main category: cs.LG
TL;DR: SWIFT is a lightweight time-series forecasting model using wavelet transforms and learnable filters for efficient edge deployment with minimal parameters.
Details
Motivation: Transformers and large language models are computationally expensive for time-series prediction in resource-constrained edge environments, while existing lightweight models perform poorly on non-stationary sequences.Method: Uses wavelet transform for lossless downsampling, learnable filter for cross-band information fusion, and only one shared linear layer or shallow MLP for sub-series mapping.
Result: Achieves state-of-the-art performance on multiple datasets with SWIFT-Linear using only 25% of parameters compared to single-layer linear time-domain prediction models.
Conclusion: SWIFT provides an efficient, lightweight solution for long-term time-series forecasting suitable for edge computing deployments.
Abstract: In recent work on time-series prediction, Transformers and even large language models have garnered significant attention due to their strong capabilities in sequence modeling. However, in practical deployments, time-series prediction often requires operation in resource-constrained environments, such as edge devices, which are unable to handle the computational overhead of large models. To address such scenarios, some lightweight models have been proposed, but they exhibit poor performance on non-stationary sequences. In this paper, we propose $\textit{SWIFT}$, a lightweight model that is not only powerful, but also efficient in deployment and inference for Long-term Time Series Forecasting (LTSF). Our model is based on three key points: (i) Utilizing wavelet transform to perform lossless downsampling of time series. (ii) Achieving cross-band information fusion with a learnable filter. (iii) Using only one shared linear layer or one shallow MLP for sub-series’ mapping. We conduct comprehensive experiments, and the results show that $\textit{SWIFT}$ achieves state-of-the-art (SOTA) performance on multiple datasets, offering a promising method for edge computing and deployment in this task. Moreover, it is noteworthy that the number of parameters in $\textit{SWIFT-Linear}$ is only 25% of what it would be with a single-layer linear model for time-domain prediction. Our code is available at https://github.com/LancelotXWX/SWIFT.
[733] Faster Adaptive Optimization via Expected Gradient Outer Product Reparameterization
Adela DePavia, Jose Cruzado, Jiayou Liang, Vasileios Charisopoulos, Rebecca Willett
Main category: cs.LG
TL;DR: EGOP reparameterization method improves convergence of adaptive optimization algorithms by using expected gradient outer product matrix to find favorable coordinate transformations.
Details
Motivation: Adaptive optimization algorithms like Adagrad and Adam are not rotationally equivariant, meaning their convergence is sensitive to parameterization choices, but there's no systematic way to identify favorable coordinate transformations.Method: Proposes an orthonormal transformation based on the expected gradient outer product (EGOP) matrix, which can be approximated using full-batch or stochastic gradient oracles. The method identifies favorable coordinate systems where adaptive algorithms perform better.
Result: Shows that sensitivity of adaptive algorithms to basis choice is influenced by EGOP matrix spectrum decay. Provides empirical evidence and theoretical arguments that common ML tasks with natural data exhibit EGOP spectral decay.
Conclusion: EGOP reparameterization can potentially improve convergence behavior of adaptive optimization algorithms by finding more favorable coordinate systems based on gradient statistics.
Abstract: Adaptive optimization algorithms – such as Adagrad, Adam, and their variants – have found widespread use in machine learning, signal processing and many other settings. Several methods in this family are not rotationally equivariant, meaning that simple reparameterizations (i.e. change of basis) can drastically affect their convergence. However, their sensitivity to the choice of parameterization has not been systematically studied; it is not clear how to identify a “favorable” change of basis in which these methods perform best. In this paper we propose a reparameterization method and demonstrate both theoretically and empirically its potential to improve their convergence behavior. Our method is an orthonormal transformation based on the expected gradient outer product (EGOP) matrix, which can be approximated using either full-batch or stochastic gradient oracles. We show that for a broad class of functions, the sensitivity of adaptive algorithms to choice-of-basis is influenced by the decay of the EGOP matrix spectrum. We illustrate the potential impact of EGOP reparameterization by presenting empirical evidence and theoretical arguments that common machine learning tasks with “natural” data exhibit EGOP spectral decay.
[734] Fast Graph Generation via Autoregressive Noisy Filtration Modeling
Markus Krimmel, Jenna Wiens, Karsten Borgwardt, Dexiong Chen
Main category: cs.LG
TL;DR: ANFM is an autoregressive graph generation framework that uses topological filtration to transform graphs into short subgraph sequences, achieving state-of-the-art quality with 100x faster inference than diffusion models.
Details
Motivation: Address the trade-off between sample quality and generation speed in existing graph generative models, while also tackling exposure bias in autoregressive approaches.Method: Uses filtration from topological data analysis to convert graphs into short sequences of subgraphs, employs noise augmentation and reinforcement learning to mitigate exposure bias, and models both edge addition and deletion operations.
Result: Matches state-of-the-art diffusion models in quality while offering over 100 times faster inference, enabling high-throughput graph generation.
Conclusion: ANFM provides a flexible autoregressive framework that balances quality and speed for graph generation, with error correction capabilities through non-monotonic sequence modeling.
Abstract: Existing graph generative models often face a critical trade-off between sample quality and generation speed. We introduce Autoregressive Noisy Filtration Modeling (ANFM), a flexible autoregressive framework that addresses both challenges. ANFM leverages filtration, a concept from topological data analysis, to transform graphs into short sequences of subgraphs. We identify exposure bias as a potential hurdle in autoregressive graph generation and propose noise augmentation and reinforcement learning as effective mitigation strategies, which allow ANFM to learn both edge addition and deletion operations. This unique capability enables ANFM to correct errors during generation by modeling non-monotonic graph sequences. Our results show that ANFM matches state-of-the-art diffusion models in quality while offering over 100 times faster inference, making it a promising approach for high-throughput graph generation. The source code is publicly available at https://github.com/BorgwardtLab/anfm .
[735] LO-BCQ: Block Clustered Quantization for 4-bit (W4A4) LLM Inference
Reena Elangovan, Charbel Sakr, Anand Raghunathan, Brucek Khailany
Main category: cs.LG
TL;DR: BCQ is a post-training quantization method that decomposes tensors into blocks, clusters them by statistics, and designs optimal codebooks per cluster, achieving W4A4 quantization with <1% accuracy loss for LLMs.
Details
Motivation: Current PTQ methods struggle with accurate sub-8-bit quantization for both weights and activations without quantization-aware training. There's a need for efficient quantization that maintains accuracy while reducing computational requirements for large language models.Method: Block clustered quantization (BCQ) decomposes operand tensors into blocks, clusters blocks based on statistics, and designs dedicated optimal quantization codebooks for each cluster. LO-BCQ iterates between block clustering and codebook design to minimize quantization mean squared error.
Result: Achieves <1% loss in inference accuracy across several LLMs and downstream tasks when quantizing to W4A4 format (with 0.5-bits overhead for scaling factors and codebook selectors), advancing state-of-the-art in PTQ.
Conclusion: BCQ provides an effective PTQ approach for sub-8-bit quantization of both weights and activations, enabling efficient deployment of large language models with minimal accuracy degradation.
Abstract: Post-training quantization (PTQ) is a promising approach to reducing the storage and computational requirements of large language models (LLMs) without additional training cost. Recent PTQ studies have primarily focused on quantizing only weights to sub-8-bits while maintaining activations at 8-bits or higher. Accurate sub-8-bit quantization for both weights and activations without relying on quantization-aware training remains a significant challenge. We propose a novel quantization method called block clustered quantization (BCQ) wherein each operand tensor is decomposed into blocks (a block is a group of contiguous scalars), blocks are clustered based on their statistics, and a dedicated optimal quantization codebook is designed for each cluster. As a specific embodiment of this approach, we propose a PTQ algorithm called Locally-Optimal BCQ (LO-BCQ) that iterates between the steps of block clustering and codebook design to greedily minimize the quantization mean squared error. When weight and activation scalars are encoded to W4A4 format (with 0.5-bits of overhead for storing scaling factors and codebook selectors), we advance the current state-of-the-art by demonstrating <1% loss in inference accuracy across several LLMs and downstream tasks.
[736] Fenchel-Young Variational Learning
Sophia Sklaviadis, Thomas Moellenhoff, Andre Martins, Mario Figueiredo
Main category: cs.LG
TL;DR: Introduces Fenchel-Young variational learning framework that generalizes classical variational methods using FY losses as divergences, enabling new algorithms like FYEM and FYVAE with adaptive sparsity features.
Details
Motivation: To broaden the variational perspective in statistical learning beyond traditional KL divergence-based methods, allowing for more flexible model learning with novel features like adaptive sparsity.Method: Proposes FY variational learning framework using Fenchel-Young losses as divergences, introduces FY free energy, FY evidence, FY ELBO, and FY posterior concepts. Develops alternating minimization and gradient backpropagation algorithms for FY evidence computation.
Result: Develops generalized FY variants of classical algorithms including FY expectation-maximization (FYEM) and FY variational autoencoder (FYVAE). Methods show empirical competitiveness, often outperforming classical counterparts, with novel qualitative features like adaptive sparsity in E-step and support for sparse observations/posteriors.
Conclusion: FY variational learning provides a flexible generalization of classical variational methods, enabling new algorithms with improved performance and novel capabilities like adaptive sparsity, expanding the scope of learnable models.
Abstract: From a variational perspective, many statistical learning criteria involve seeking a distribution that balances empirical risk and regularization. In this paper, we broaden this perspective by introducing a new general class of variational methods based on Fenchel-Young (FY) losses, treated as divergences that generalize (and encompass) the familiar Kullback-Leibler divergence at the core of classical variational learning. Our proposed formulation – FY variational learning – includes as key ingredients new notions of FY free energy, FY evidence, FY evidence lower bound, and FY posterior. We derive alternating minimization and gradient backpropagation algorithms to compute (or lower bound) the FY evidence, which enables learning a wider class of models than previous variational formulations. This leads to generalized FY variants of classical algorithms, such as an FY expectation-maximization (FYEM) algorithm, and latent-variable models, such as an FY variational autoencoder (FYVAE). Our new methods are shown to be empirically competitive, often outperforming their classical counterparts, and most importantly, to have qualitatively novel features. For example, FYEM has an adaptively sparse E-step, while the FYVAE can support models with sparse observations and sparse posteriors.
[737] Contextual Quantum Neural Networks for Stock Price Prediction
Sharan Mourya, Hannes Leipold, Bibhas Adhikari
Main category: cs.LG
TL;DR: Quantum machine learning approach for multi-asset stock price prediction using contextual quantum neural networks with quantum batch gradient update and share-and-specify ansatz architecture.
Details
Motivation: To move beyond traditional financial models that use entire historical data by applying quantum machine learning to predict future stock price distributions with enhanced adaptability and precision, capturing recent trends and inter-asset correlations.Method: Proposes a quantum multi-task learning (QMTL) architecture with share-and-specify ansatz that integrates task-specific operators controlled by quantum labels, enabling simultaneous training of multiple assets on the same quantum circuit. Introduces quantum batch gradient update (QBGU) to accelerate stochastic gradient descent in quantum applications.
Result: The approach outperforms quantum single-task learning models on S&P 500 data for Apple, Google, Microsoft, and Amazon stocks, effectively captures inter-asset correlations, and demonstrates enhanced prediction accuracy with logarithmic overhead in qubit requirements.
Conclusion: Demonstrates transformative potential of quantum machine learning in financial applications, paving the way for more advanced, resource-efficient quantum algorithms in stock price prediction and complex financial modeling.
Abstract: In this paper, we apply quantum machine learning (QML) to predict the stock prices of multiple assets using a contextual quantum neural network. Our approach captures recent trends to predict future stock price distributions, moving beyond traditional models that focus on entire historical data, enhancing adaptability and precision. Utilizing the principles of quantum superposition, we introduce a new training technique called the quantum batch gradient update (QBGU), which accelerates the standard stochastic gradient descent (SGD) in quantum applications and improves convergence. Consequently, we propose a quantum multi-task learning (QMTL) architecture, specifically, the share-and-specify ansatz, that integrates task-specific operators controlled by quantum labels, enabling the simultaneous and efficient training of multiple assets on the same quantum circuit as well as enabling efficient portfolio representation with logarithmic overhead in the number of qubits. This architecture represents the first of its kind in quantum finance, offering superior predictive power and computational efficiency for multi-asset stock price forecasting. Through extensive experimentation on S&P 500 data for Apple, Google, Microsoft, and Amazon stocks, we demonstrate that our approach not only outperforms quantum single-task learning (QSTL) models but also effectively captures inter-asset correlations, leading to enhanced prediction accuracy. Our findings highlight the transformative potential of QML in financial applications, paving the way for more advanced, resource-efficient quantum algorithms in stock price prediction and other complex financial modeling tasks.
[738] Robust Multi-Objective Controlled Decoding of Large Language Models
Seongho Son, William Bankes, Sangwoong Yoon, Shyam Sundhar Ramesh, Xiaohang Tang, Ilija Bogunovic
Main category: cs.LG
TL;DR: RMOD is a robust inference-time algorithm that aligns LLMs to multiple human objectives by maximizing worst-case rewards through a game-theoretic approach.
Details
Motivation: Current LLM alignment methods often struggle with balancing multiple competing objectives (instruction-following, helpfulness, safety) and can be vulnerable to adversarial reward weightings, necessitating a robust approach that ensures good performance across all objectives simultaneously.Method: Formulates robust decoding as a maximin two-player game between adversarially computed reward weights and sampling policy, reduces to convex optimization for worst-case weights, and derives optimal sampling policy analytically with efficient algorithm for practical LLM deployment.
Result: RMOD consistently outperforms baselines in worst-case rewards and win rates across alignment datasets with up to 10 objectives, with minimal computational overhead compared to standard controlled decoding methods.
Conclusion: RMOD provides an effective, computationally efficient solution for robust multi-objective alignment of LLMs, ensuring reliable performance across diverse human preferences through game-theoretic optimization.
Abstract: We introduce Robust Multi-Objective Decoding (RMOD), a novel inference-time algorithm that robustly aligns Large Language Models (LLMs) to multiple human objectives (e.g., instruction-following, helpfulness, safety) by maximizing the worst-case rewards. RMOD formulates the robust decoding problem as a maximin two-player game between adversarially computed reward weights and the sampling policy, solvable through a Nash equilibrium. We demonstrate that this game reduces to a convex optimization problem to identify the worst-case reward weights, with the optimal sampling policy analytically derived. For practical applications, we propose an efficient algorithm of RMOD tailored for contemporary LLMs, introducing minimal computational overhead compared to standard non-robust Controlled Decoding methods. Experimental results across a range of popular alignment datasets with up to 10 objectives show the effectiveness of RMOD and its distilled version, consistently outperforming baselines in worst-case rewards and win rates.
[739] Learning Rate Annealing Improves Tuning Robustness in Stochastic Optimization
Amit Attia, Tomer Koren
Main category: cs.LG
TL;DR: Theoretical analysis shows polynomial learning rate annealing schedules (like cosine decay) are more robust to initial learning rate misspecification than fixed or inverse-square-root schedules, reducing computational overhead for hyperparameter tuning.
Details
Motivation: Learning rate tuning via grid search is computationally expensive, especially for large models. The paper aims to theoretically analyze which learning rate schedules are more robust to initial misspecification to reduce tuning costs.Method: Theoretical analysis in stochastic convex optimization setup comparing convergence rates of SGD with different learning rate schedules (polynomial decay, fixed stepsize, inverse-square-root). Derives bounds showing polynomial decay schedules have sublinear dependence on misspecification factor.
Result: Polynomial decay schedules achieve O(ρ^{1/(2p+1)}/√T) convergence rate, which depends sublinearly on misspecification factor ρ, compared to O(ρ/√T) for fixed/inverse-square-root schedules. Experiments confirm increased robustness.
Conclusion: Polynomial annealing schedules like cosine decay offer theoretical and practical advantages for hyperparameter tuning by being more robust to initial learning rate misspecification, reducing computational overhead.
Abstract: The learning rate in stochastic gradient methods is a critical hyperparameter that is notoriously costly to tune via standard grid search, especially for training modern large-scale models with billions of parameters. We identify a theoretical advantage of learning rate annealing schemes that decay the learning rate to zero at a polynomial rate, such as the widely-used cosine schedule, by demonstrating their increased robustness to initial parameter misspecification due to a coarse grid search. We present an analysis in a stochastic convex optimization setup demonstrating that the convergence rate of stochastic gradient descent with annealed schedules depends sublinearly on the multiplicative misspecification factor $ρ$ (i.e., the grid resolution), achieving a rate of $O(ρ^{1/(2p+1)}/\sqrt{T})$ where $p$ is the degree of polynomial decay and $T$ is the number of steps. This is in contrast to the $O(ρ/\sqrt{T})$ rate obtained under the inverse-square-root and fixed stepsize schedules, which depend linearly on $ρ$. Experiments confirm the increased robustness compared to tuning with a fixed stepsize, that has significant implications for the computational overhead of hyperparameter search in practical training scenarios.
[740] Heuristic Methods are Good Teachers to Distill MLPs for Graph Link Prediction
Zongyue Qin, Shichang Zhang, Mingxuan Ju, Tong Zhao, Neil Shah, Yizhou Sun
Main category: cs.LG
TL;DR: EHDM: Ensemble Heuristic-Distilled MLPs for efficient link prediction that eliminates graph dependencies while integrating complementary signals via gating, outperforming previous GNN-to-MLP approaches with significantly reduced training time.
Details
Motivation: Current GNN-to-MLP distillation methods overlook alternative teachers like specialized link prediction models and heuristic methods, and stronger teachers don't always produce stronger students in this context.Method: Proposes Ensemble Heuristic-Distilled MLPs (EHDM) which uses heuristic methods as teachers for MLP distillation, eliminates graph dependencies, and integrates complementary signals through a gating mechanism.
Result: Experiments on ten datasets show 7.93% average improvement over previous GNN-to-MLP approaches with 1.95-3.32 times less training time.
Conclusion: EHDM is an efficient and effective link prediction method that demonstrates weaker heuristic methods can teach MLPs to near-GNN performance with drastically reduced training costs.
Abstract: Link prediction is a crucial graph-learning task with applications including citation prediction and product recommendation. Distilling Graph Neural Networks (GNNs) teachers into Multi-Layer Perceptrons (MLPs) students has emerged as an effective approach to achieve strong performance and reducing computational cost by removing graph dependency. However, existing distillation methods only use standard GNNs and overlook alternative teachers such as specialized model for link prediction (GNN4LP) and heuristic methods (e.g., common neighbors). This paper first explores the impact of different teachers in GNN-to-MLP distillation. Surprisingly, we find that stronger teachers do not always produce stronger students: MLPs distilled from GNN4LP can underperform those distilled from simpler GNNs, while weaker heuristic methods can teach MLPs to near-GNN performance with drastically reduced training costs. Building on these insights, we propose Ensemble Heuristic-Distilled MLPs (EHDM), which eliminates graph dependencies while effectively integrating complementary signals via a gating mechanism. Experiments on ten datasets show an average 7.93% improvement over previous GNN-to-MLP approaches with 1.95-3.32 times less training time, indicating EHDM is an efficient and effective link prediction method. Our code is available at https://github.com/ZongyueQin/EHDM
[741] Riemannian Denoising Diffusion Probabilistic Models
Zichen Liu, Wei Zhang, Christof Schütte, Tiejun Li
Main category: cs.LG
TL;DR: Riemannian Denoising Diffusion Probabilistic Models (RDDPMs) for generative modeling on submanifolds using only function evaluations and first-order derivatives, without requiring extensive geometric information.
Details
Motivation: Existing generative models on manifolds require substantial geometric information (geodesics, Laplace-Beltrami eigenfunctions), limiting them to manifolds where such information is available. Need methods that work on more general manifolds defined as level sets of functions.Method: RDDPMs use a projection scheme requiring only evaluation of the function defining the submanifold and its first-order derivatives. Theoretical analysis connects RDDPMs to score-based generative models on manifolds in continuous-time limit.
Result: Method demonstrated on datasets from previous studies and new datasets from high-dimensional manifolds: SO(10) and configuration space of alanine dipeptide with fixed dihedral angle.
Conclusion: RDDPMs provide a more general approach to generative modeling on manifolds, requiring minimal geometric information and applicable to broader class of manifolds defined as level sets.
Abstract: We propose Riemannian Denoising Diffusion Probabilistic Models (RDDPMs) for learning distributions on submanifolds of Euclidean space that are level sets of functions, including most of the manifolds relevant to applications. Existing methods for generative modeling on manifolds rely on substantial geometric information such as geodesic curves or eigenfunctions of the Laplace-Beltrami operator and, as a result, they are limited to manifolds where such information is available. In contrast, our method, built on a projection scheme, can be applied to more general manifolds, as it only requires being able to evaluate the value and the first order derivatives of the function that defines the submanifold. We provide a theoretical analysis of our method in the continuous-time limit, which elucidates the connection between our RDDPMs and score-based generative models on manifolds. The capability of our method is demonstrated on datasets from previous studies and on new datasets sampled from two high-dimensional manifolds, i.e. $\mathrm{SO}(10)$ and the configuration space of molecular system alanine dipeptide with fixed dihedral angle.
[742] Sparse Latent Factor Forecaster (SLFF) with Iterative Inference for Transparent Multi-Horizon Commodity Futures Prediction
Abhijit Gupta
Main category: cs.LG
TL;DR: SLFF: Sparse Latent Factor Forecaster with Iterative Inference addresses amortization gap in variational forecasting through sparse coding, iterative refinement, and encoder alignment, with interpretability protocols and macroeconomic data handling.
Details
Motivation: Addresses the deployment gap in amortized variational inference for latent-variable forecasters where test-time encoder approximates training-time optimization-refined latent without access to future targets, leading to unnecessary forecast error and interpretability challenges.Method: Proposes SLFF with: (1) sparse coding objective with L1 regularization for low-dimensional latents, (2) unrolled proximal gradient descent (LISTA-style) for iterative refinement during training, (3) encoder alignment to ensure amortized outputs match optimization-refined solutions, and (4) information-set-aware protocol using release calendars and vintage macroeconomic data to prevent mixed-frequency data leakage.
Result: Demonstrates significant improvements over neural baselines at 1- and 5-day horizons on commodity futures (Copper, WTI, Gold; 2005-2025), yields sparse factors stable across seeds and correlated with observable economic fundamentals, with derived bound on amortization gap empirically confirmed.
Conclusion: SLFF effectively addresses the amortization gap in latent-variable forecasting through sparse coding and iterative inference, improving forecast accuracy while providing interpretable sparse factors, though interpretability remains correlational rather than causal.
Abstract: Amortized variational inference in latent-variable forecasters creates a deployment gap: the test-time encoder approximates a training-time optimization-refined latent, but without access to future targets. This gap introduces unnecessary forecast error and interpretability challenges. In this work, we propose the Sparse Latent Factor Forecaster with Iterative Inference (SLFF), addressing this through (i) a sparse coding objective with L1 regularization for low-dimensional latents, (ii) unrolled proximal gradient descent (LISTA-style) for iterative refinement during training, and (iii) encoder alignment to ensure amortized outputs match optimization-refined solutions. Under a linearized decoder assumption, we derive a design-motivating bound on the amortization gap based on encoder-optimizer distance, with convergence rates under mild conditions; empirical checks confirm the bound is predictive for the deployed MLP decoder. To prevent mixed-frequency data leakage, we introduce an information-set-aware protocol using release calendars and vintage macroeconomic data. Interpretability is formalized via a three-stage protocol: stability (Procrustes alignment across seeds), driver validity (held-out regressions against observables), and behavioral consistency (counterfactuals and event studies). Using commodity futures (Copper, WTI, Gold; 2005–2025) as a testbed, SLFF demonstrates significant improvements over neural baselines at 1- and 5-day horizons, yielding sparse factors that are stable across seeds and correlated with observable economic fundamentals (interpretability remains correlational, not causal). Code, manifests, diagnostics, and artifacts are released.
[743] A Generalized Hierarchical Federated Learning Framework with Theoretical Guarantees
Seyed Mohammad Azimi-Abarghouyi, Carlo Fischione
Main category: cs.LG
TL;DR: QMLHFL is a multi-layer hierarchical federated learning framework that extends beyond traditional two-layer systems to arbitrary numbers of layers with nested aggregation and layer-specific quantization for communication efficiency.
Details
Motivation: Existing hierarchical FL models are limited to two aggregation layers, which restricts scalability and flexibility in complex, large-scale networks. There's a need for more flexible hierarchical structures that can handle arbitrary numbers of layers while addressing communication constraints.Method: Proposes QMLHFL framework with nested aggregation for arbitrary hierarchical layers and layer-specific quantization scheme to meet communication constraints. Includes comprehensive convergence analysis, derivation of general convergence conditions and rates, and optimization of intra-layer iterations to maximize convergence rate under deadline constraints.
Result: QMLHFL consistently achieves high learning accuracy even under high data heterogeneity. When optimized, it delivers notably improved performance compared to using randomly selected values. The framework provides theoretical convergence guarantees and practical optimization strategies.
Conclusion: QMLHFL successfully generalizes hierarchical FL to arbitrary numbers of layers with theoretical foundations and practical optimizations, addressing scalability limitations of existing approaches while maintaining performance under challenging conditions like data heterogeneity.
Abstract: Almost all existing hierarchical federated learning (FL) models are limited to two aggregation layers, restricting scalability and flexibility in complex, large-scale networks. In this work, we propose a Multi-Layer Hierarchical Federated Learning framework (QMLHFL), which appears to be the first study that generalizes hierarchical FL to arbitrary numbers of layers and network architectures through nested aggregation, while employing a layer-specific quantization scheme to meet communication constraints. We develop a comprehensive convergence analysis for QMLHFL and derive a general convergence condition and rate that reveal the effects of key factors, including quantization parameters, hierarchical architecture, and intra-layer iteration counts. Furthermore, we determine the optimal number of intra-layer iterations to maximize the convergence rate while meeting a deadline constraint that accounts for both communication and computation times. Our results show that QMLHFL consistently achieves high learning accuracy, even under high data heterogeneity, and delivers notably improved performance when optimized, compared to using randomly selected values.
[744] RainPro-8: An Efficient Deep Learning Model to Estimate Rainfall Probabilities Over 8 Hours
Rafael Pablos Sarabia, Joachim Nyborg, Morten Birk, Jeppe Liborius Sjørup, Anders Lillevang Vesterholt, Ira Assent
Main category: cs.LG
TL;DR: Deep learning model for high-resolution probabilistic precipitation forecasting over 8 hours in Europe using multi-modal data integration
Details
Motivation: Overcome limitations of radar-only deep learning models with short forecast lead times by integrating multiple data sources for longer-range, high-resolution precipitation forecastingMethod: Deep learning model that efficiently integrates radar, satellite, and physics-based numerical weather prediction (NWP) data while capturing long-range interactions through a compact architecture
Result: Surpasses current operational NWP systems, extrapolation-based methods, and deep-learning nowcasting models with accurate forecasts and robust uncertainty quantification
Conclusion: Sets new standard for high-resolution precipitation forecasting in Europe with balance between accuracy, interpretability, and computational efficiency
Abstract: We present a deep learning model for high-resolution probabilistic precipitation forecasting over an 8-hour horizon in Europe, overcoming the limitations of radar-only deep learning models with short forecast lead times. Our model efficiently integrates multiple data sources - including radar, satellite, and physics-based numerical weather prediction (NWP) - while capturing long-range interactions, resulting in accurate forecasts with robust uncertainty quantification through consistent probabilistic maps. Featuring a compact architecture, it enables more efficient training and faster inference than existing models. Extensive experiments demonstrate that our model surpasses current operational NWP systems, extrapolation-based methods, and deep-learning nowcasting models, setting a new standard for high-resolution precipitation forecasting in Europe, ensuring a balance between accuracy, interpretability, and computational efficiency.
[745] Heterogeneity-Aware Client Sampling for Optimal and Efficient Federated Learning
Shudi Weng, Chao Ren, Ming Xiao, Mikael Skoglund
Main category: cs.LG
TL;DR: FedACS addresses objective inconsistency in federated learning caused by heterogeneous client communication and computation capabilities through a unified theoretical analysis and heterogeneity-aware client sampling method.
Details
Motivation: Federated learning involves clients with diverse communication and computational capabilities, which can significantly distort optimization dynamics and lead to objective inconsistency where the global model converges to incorrect stationary points. The joint effect of communication and computation heterogeneity has remained largely unexplored due to the complexity of their interaction.Method: The paper first provides a unified theoretical analysis of general heterogeneous FL, revealing distinct mechanisms through which heterogeneous communication and computation drive inconsistency. Based on these insights, the authors propose Federated Heterogeneity-Aware Client Sampling (FedACS), a universal method to eliminate all types of objective inconsistency by intelligently sampling clients based on their capabilities.
Result: FedACS converges to the correct optimum at a rate of O(1/√R) even in dynamic heterogeneous environments. Extensive experiments across multiple datasets show FedACS outperforms state-of-the-art baselines by 4.3%-36%, while reducing communication costs by 22%-89% and computation loads by 14%-105%.
Conclusion: The paper provides the first unified theoretical analysis of heterogeneous FL and introduces FedACS as an effective solution to eliminate objective inconsistency caused by communication and computation heterogeneity, offering significant performance improvements and resource savings.
Abstract: Federated learning (FL) commonly involves clients with diverse communication and computational capabilities. Such heterogeneity can significantly distort the optimization dynamics and lead to objective inconsistency, where the global model converges to an incorrect stationary point potentially far from the pursued optimum. Despite its critical impact, the joint effect of communication and computation heterogeneity has remained largely unexplored, due to the intrinsic complexity of their interaction. In this paper, we reveal the fundamentally distinct mechanisms through which heterogeneous communication and computation drive inconsistency in FL. To the best of our knowledge, this is the first unified theoretical analysis of general heterogeneous FL, offering a principled understanding of how these two forms of heterogeneity jointly distort the optimization trajectory under arbitrary choices of local solvers. Motivated by these insights, we propose Federated Heterogeneity-Aware Client Sampling, FedACS, a universal method to eliminate all types of objective inconsistency. We theoretically prove that FedACS converges to the correct optimum at a rate of $O(1/\sqrt{R})$, even in dynamic heterogeneous environments. Extensive experiments across multiple datasets show that FedACS outperforms state-of-the-art and category-specific baselines by 4.3%-36%, while reducing communication costs by 22%-89% and computation loads by 14%-105%, respectively.
[746] Residual Feature Integration is Sufficient to Prevent Negative Transfer
Yichen Xu, Ryumei Nakada, Linjun Zhang, Lexin Li
Main category: cs.LG
TL;DR: Theoretical framework for preventing negative transfer in transfer learning by augmenting frozen pretrained features with trainable target-side encoders to capture residual signals.
Details
Motivation: Address the long-standing problem of negative transfer in transfer learning where source representations can harm target task performance, with little theoretical understanding of how to reliably avoid it.Method: Augment frozen pretrained source-side features with a trainable target-side encoder that adapts target features to capture residual signals overlooked by source-pretrained models.
Result: Theoretical guarantees show the method provably prevents negative transfer with no worse convergence rate than training from scratch, and can transition from nonparametric to near-parametric rates when source representations are informative. Extensive experiments across image, text, and tabular benchmarks verify consistent performance protection under distribution shift, label noise, semantic perturbation, and class imbalance.
Conclusion: Provides the first theoretical work ensuring protection against negative transfer, advancing the theory of safe transfer learning with a principled, simple, robust, architecture-agnostic approach that supports adapt-time multimodality extension.
Abstract: Transfer learning has become a central paradigm in modern machine learning, yet it suffers from the long-standing problem of negative transfer, where leveraging source representations can harm rather than help performance on the target task. Although empirical remedies have been proposed, there remains little theoretical understanding of how to reliably avoid negative transfer. In this paper, we investigate a simple yet remarkably effective strategy: augmenting frozen, pretrained source-side features with a trainable target-side encoder that adapts target features to capture residual signals overlooked by models pretrained on the source data. We show this residual feature integration strategy is sufficient to provably prevent negative transfer, by establishing theoretical guarantees that it has no worse convergence rate than training from scratch under the informative class of target distributions up to logarithmic factors, and that the convergence rate can transition seamlessly from nonparametric to near-parametric when source representations are informative. To our knowledge, this is the first theoretical work that ensures protection against negative transfer. We carry out extensive numerical experiments across image, text and tabular benchmarks, and empirically verify that the method consistently safeguards performance under distribution shift, label noise, semantic perturbation, and class imbalance. We additionally demonstrate that this residual integration mechanism uniquely supports adapt-time multimodality extension, enabling a pretrained single-cell foundation model to incorporate spatial signals for lymph-node anatomical classification despite the source model being trained without them. Our study thus advances the theory of safe transfer learning, and provides a principled approach that is simple, robust, architecture-agnostic, and broadly applicable.
[747] Know When to Abstain: Optimal Selective Classification with Likelihood Ratios
Alvin Heng, Harold Soh
Main category: cs.LG
TL;DR: The paper proposes using Neyman-Pearson lemma and likelihood ratio tests for selective classification, particularly under covariate shift, showing improved performance over baselines on vision and language tasks.
Details
Motivation: Selective classification improves model reliability by allowing abstention from uncertain predictions, but optimal selection functions, especially under covariate shift (where test distribution differs from training), remain underexplored and challenging.Method: Revisits selective classification through Neyman-Pearson lemma, which characterizes optimal rejection as likelihood ratio tests. Proposes new selection methods based on this perspective, unifying existing baselines and introducing novel approaches for covariate shift scenarios.
Result: The proposed Neyman-Pearson-informed methods consistently outperform existing baselines across vision and language tasks, including supervised learning and vision-language models, demonstrating robust improvement under covariate shifts.
Conclusion: Likelihood ratio-based selection provides a robust mechanism for improving selective classification, particularly under challenging covariate shift conditions, with applications across vision and language domains.
Abstract: Selective classification enhances the reliability of predictive models by allowing them to abstain from making uncertain predictions. In this work, we revisit the design of optimal selection functions through the lens of the Neyman–Pearson lemma, a classical result in statistics that characterizes the optimal rejection rule as a likelihood ratio test. We show that this perspective not only unifies the behavior of several post-hoc selection baselines, but also motivates new approaches to selective classification which we propose here. A central focus of our work is the setting of covariate shift, where the input distribution at test time differs from that at training. This realistic and challenging scenario remains relatively underexplored in the context of selective classification. We evaluate our proposed methods across a range of vision and language tasks, including both supervised learning and vision-language models. Our experiments demonstrate that our Neyman–Pearson-informed methods consistently outperform existing baselines, indicating that likelihood ratio-based selection offers a robust mechanism for improving selective classification under covariate shifts. Our code is publicly available at https://github.com/clear-nus/sc-likelihood-ratios.
[748] MoESD: Unveil Speculative Decoding’s Potential for Accelerating Sparse MoE
Zongle Huang, Lei Zhu, Zongyuan Zhan, Ting Hu, Weikai Mao, Xianzhi Yu, Yongpan Liu, Tianyu Zhang
Main category: cs.LG
TL;DR: Speculative decoding accelerates Mixture of Experts (MoE) LLMs more effectively than dense models at medium batch sizes, with sparser MoEs benefiting even more broadly.
Details
Motivation: While speculative decoding is known to accelerate dense LLMs, its effectiveness for MoE models is unexplored. MoEs offer better performance with less computation, but current SD research focuses on algorithm acceptance rates rather than workload and architecture effects.Method: Develop theoretical modeling to analyze SD tradeoffs, introduce ’target efficiency’ metric to characterize workload and architecture effects, and experimentally validate on different GPUs with MoE models like Qwen2-57B-A14B.
Result: MoEs benefit more from SD than dense models at medium batch sizes, with up to 2.29x speedup for Qwen2-57B-A14B. Sparser MoEs have broader effective batch size ranges for SD acceleration.
Conclusion: SD provides significant acceleration for MoE inference, especially in private serving scenarios where existing solutions struggle. The ’target efficiency’ metric helps identify system bottlenecks and understand SD acceleration more comprehensively.
Abstract: Large Language Models (LLMs) have achieved remarkable success across many applications, with Mixture of Experts (MoE) models demonstrating great potential. Compared to traditional dense models, MoEs achieve better performance with less computation. Speculative decoding (SD) is a widely used technique to accelerate LLM inference without accuracy loss, but it has been considered efficient only for dense models. In this work, we first demonstrate that, under medium batch sizes, MoE surprisingly benefits more from SD than dense models. Furthermore, as MoE becomes sparser – the prevailing trend in MoE designs – the batch size range where SD acceleration is expected to be effective becomes broader. To quantitatively understand tradeoffs involved in SD, we develop a reliable modeling based on theoretical analyses. While current SD research primarily focuses on improving acceptance rates of algorithms, changes in workload and model architecture can still lead to degraded SD acceleration even with high acceptance rates. To address this limitation, we introduce a new metric ’target efficiency’ that characterizes these effects, thus helping researchers identify system bottlenecks and understand SD acceleration more comprehensively. For scenarios like private serving, this work unveils a new perspective to speed up MoE inference, where existing solutions struggle. Experiments on different GPUs show up to 2.29x speedup for Qwen2-57B-A14B at medium batch sizes and validate our theoretical predictions.
[749] On the Relation between Rectified Flows and Optimal Transport
Johannes Hertrich, Antonin Chambolle, Julie Delon
Main category: cs.LG
TL;DR: Rectified flows and flow matching connections to optimal transport, with analysis showing gradient-constrained rectified flows don’t reliably compute optimal transport maps.
Details
Motivation: To investigate the theoretical connections between rectified flows, flow matching, and optimal transport, and to clarify misconceptions about whether gradient-constrained rectified flows can solve optimal transport problems.Method: Theoretical analysis of rectified flow invariance properties, explicit constructions in Gaussian and Gaussian mixture settings, and study of existence conditions for gradient-constrained rectified flows. Provides counterexamples to previous equivalence claims.
Result: Shows that gradient-constrained rectified flows only relate to optimal transport under much stronger assumptions than previously acknowledged, and presents counterexamples invalidating earlier equivalence results.
Conclusion: Enforcing gradient constraints on rectified flows is generally not a reliable method for computing optimal transport maps, clarifying important theoretical limitations in the flow matching literature.
Abstract: This paper investigates the connections between rectified flows, flow matching, and optimal transport. Flow matching is a recent approach to learning generative models by estimating velocity fields that guide transformations from a source to a target distribution. Rectified flow matching aims to straighten the learned transport paths, yielding more direct flows between distributions. Our first contribution is a set of invariance properties of rectified flows and explicit velocity fields. In addition, we also provide explicit constructions and analysis in the Gaussian (not necessarily independent) and Gaussian mixture settings and study the relation to optimal transport. Our second contribution addresses recent claims suggesting that rectified flows, when constrained such that the learned velocity field is a gradient, can yield (asymptotically) solutions to optimal transport problems. We study the existence of solutions for this problem and demonstrate that they only relate to optimal transport under assumptions that are significantly stronger than those previously acknowledged. In particular, we present several counterexamples that invalidate earlier equivalence results in the literature, and we argue that enforcing a gradient constraint on rectified flows is, in general, not a reliable method for computing optimal transport maps.
[750] A Cramér-von Mises Approach to Incentivizing Truthful Data Sharing
Alex Clinton, Thomas Zeng, Yiding Chen, Xiaojin Zhu, Kirthevasan Kandasamy
Main category: cs.LG
TL;DR: A novel incentive mechanism for data marketplaces that uses a two-sample test based on Cramér-von Mises statistic to reward truthful data sharing while penalizing fabricated or low-quality data.
Details
Motivation: Existing data marketplace incentive schemes that reward based on quantity are vulnerable to manipulation through fabricated or low-quality data submissions. Prior solutions rely on strong assumptions about data distributions (e.g., Gaussian), limiting their real-world applicability.Method: Develops reward mechanisms using a novel two-sample test inspired by the Cramér-von Mises statistic. The method compares each agent’s data against others’ submissions to incentivize truthful reporting while disincentivizing fabrication.
Result: The mechanism establishes truthful reporting as a (possibly approximate) Nash equilibrium in both Bayesian and prior-agnostic settings. It theoretically works for three canonical data sharing problems and empirically demonstrates effectiveness through simulations and real-world language and image data.
Conclusion: The proposed mechanism successfully incentivizes truthful data sharing while relaxing key assumptions of prior work, making it more applicable to real-world data marketplaces with diverse data distributions.
Abstract: Modern data marketplaces and data sharing consortia increasingly rely on incentive mechanisms to encourage agents to contribute data. However, schemes that reward agents based on the quantity of submitted data are vulnerable to manipulation, as agents may submit fabricated or low-quality data to inflate their rewards. Prior work has proposed comparing each agent’s data against others’ to promote honesty: when others contribute genuine data, the best way to minimize discrepancy is to do the same. Yet prior implementations of this idea rely on very strong assumptions about the data distribution (e.g. Gaussian), limiting their applicability. In this work, we develop reward mechanisms based on a novel, two-sample test inspired by the Cramér-von Mises statistic. Our methods strictly incentivize agents to submit more genuine data, while disincentivizing data fabrication and other types of untruthful reporting. We establish that truthful reporting constitutes a (possibly approximate) Nash equilibrium in both Bayesian and prior-agnostic settings. We theoretically instantiate our method in three canonical data sharing problems and show that it relaxes key assumptions made by prior work. Empirically, we demonstrate that our mechanism incentivizes truthful data sharing via simulations and on real-world language and image data.
[751] HYPER: A Foundation Model for Inductive Link Prediction with Knowledge Hypergraphs
Xingyue Huang, Mikhail Galkin, Michael M. Bronstein, İsmail İlkan Ceylan
Main category: cs.LG
TL;DR: HYPER is a foundation model for inductive link prediction in knowledge hypergraphs that can generalize to novel entities AND novel relations, unlike existing methods that only handle novel entities.
Details
Motivation: Existing inductive link prediction methods for knowledge hypergraphs assume a fixed relational vocabulary and cannot handle novel relation types. There's a need for models that can generalize to completely new knowledge hypergraphs with both unseen entities and relations.Method: HYPER encodes entities of each hyperedge along with their respective positions in the hyperedge, enabling learning and transfer across different relation types of varying arities. It’s designed as a foundation model inspired by knowledge graph foundation models.
Result: HYPER consistently outperforms all existing methods in both node-only and node-and-relation inductive settings across 16 new inductive datasets, showing strong generalization to unseen, higher-arity relational structures.
Conclusion: HYPER successfully addresses the limitation of existing methods by generalizing to knowledge hypergraphs with novel entities AND novel relations, demonstrating strong performance across diverse relation types and arities.
Abstract: Inductive link prediction with knowledge hypergraphs is the task of predicting missing hyperedges involving completely novel entities (i.e., nodes unseen during training). Existing methods for inductive link prediction with knowledge hypergraphs assume a fixed relational vocabulary and, as a result, cannot generalize to knowledge hypergraphs with novel relation types (i.e., relations unseen during training). Inspired by knowledge graph foundation models, we propose HYPER as a foundation model for link prediction, which can generalize to any knowledge hypergraph, including novel entities and novel relations. Importantly, HYPER can learn and transfer across different relation types of varying arities, by encoding the entities of each hyperedge along with their respective positions in the hyperedge. To evaluate HYPER, we construct 16 new inductive datasets from existing knowledge hypergraphs, covering a diverse range of relation types of varying arities. Empirically, HYPER consistently outperforms all existing methods in both node-only and node-and-relation inductive settings, showing strong generalization to unseen, higher-arity relational structures.
[752] Calibrated Predictive Lower Bounds on Time-to-Unsafe-Sampling in LLMs
Hen Davidov, Shai Feldman, Gilad Freidkin, Yaniv Romano
Main category: cs.LG
TL;DR: Proposes time-to-unsafe-sampling as a new safety metric for LLMs, with a conformal prediction method to estimate it with statistical guarantees.
Details
Motivation: Current safety evaluation for generative models lacks prompt-adaptive measures that quantify how many generations are needed before an unsafe response appears. Existing methods struggle because unsafe outputs are rare in well-aligned models and may not be observed within practical sampling budgets.Method: Frames the problem as survival analysis and uses conformal prediction with a novel calibration technique to construct lower predictive bounds on time-to-unsafe-sampling. Introduces an optimized sampling-budget allocation scheme for improved sample efficiency while maintaining distribution-free guarantees.
Result: Experiments on synthetic and real data support theoretical results and demonstrate practical utility for safety risk assessment in generative AI models. The method provides rigorous coverage guarantees for the lower bound estimates.
Conclusion: Time-to-unsafe-sampling offers a new dimension for prompt-adaptive safety evaluation, and the proposed conformal prediction approach enables reliable estimation even when unsafe outputs are rare, making it valuable for safety assessment in generative AI.
Abstract: We introduce time-to-unsafe-sampling, a novel safety measure for generative models, defined as the number of generations required by a large language model (LLM) to trigger an unsafe (e.g., toxic) response. While providing a new dimension for prompt-adaptive safety evaluation, quantifying time-to-unsafe-sampling is challenging: unsafe outputs are often rare in well-aligned models and thus may not be observed under any feasible sampling budget. To address this challenge, we frame this estimation problem as one of survival analysis. We build on recent developments in conformal prediction and propose a novel calibration technique to construct a lower predictive bound (LPB) on the time-to-unsafe-sampling of a given prompt with rigorous coverage guarantees. Our key technical innovation is an optimized sampling-budget allocation scheme that improves sample efficiency while maintaining distribution-free guarantees. Experiments on both synthetic and real data support our theoretical results and demonstrate the practical utility of our method for safety risk assessment in generative AI models.
[753] NeuronSeek: On Stability and Expressivity of Task-driven Neurons
Hanyu Pei, Jing-Xiao Liao, Qibin Zhao, Ting Gao, Shijun Zhang, Xiaoge Zhang, Feng-Lei Fan
Main category: cs.LG
TL;DR: NeuronSeek-TD replaces symbolic regression with tensor decomposition to discover optimal neuron formulations, offering better stability and convergence while maintaining competitive performance across benchmarks.
Details
Motivation: Inspired by the human brain's specialized neurons for different tasks, the paper aims to improve upon existing task-driven neuron discovery methods by replacing symbolic regression with tensor decomposition for better stability and convergence.Method: Uses tensor decomposition instead of symbolic regression to discover optimal neuron formulations, establishes theoretical guarantees that modifying aggregation functions with common activations enables universal approximation with fixed parameters.
Result: NeuronSeek-TD achieves superior stability and competitive performance relative to state-of-the-art models across diverse benchmarks.
Conclusion: Tensor decomposition provides a more stable and efficient alternative to symbolic regression for discovering task-driven neurons, with theoretical guarantees for universal approximation.
Abstract: Drawing inspiration from our human brain that designs different neurons for different tasks, recent advances in deep learning have explored modifying a network’s neurons to develop so-called task-driven neurons. Prototyping task-driven neurons (referred to as NeuronSeek) employs symbolic regression (SR) to discover the optimal neuron formulation and construct a network from these optimized neurons. Along this direction, this work replaces symbolic regression with tensor decomposition (TD) to discover optimal neuronal formulations, offering enhanced stability and faster convergence. Furthermore, we establish theoretical guarantees that modifying the aggregation functions with common activation functions can empower a network with a fixed number of parameters to approximate any continuous function with an arbitrarily small error, providing a rigorous mathematical foundation for the NeuronSeek framework. Extensive empirical evaluations demonstrate that our NeuronSeek-TD framework not only achieves superior stability, but also is competitive relative to the state-of-the-art models across diverse benchmarks. The code is available at https://github.com/HanyuPei22/NeuronSeek.
[754] Chain of Thought in Order: Discovering Learning-Friendly Orders for Arithmetic
Yuta Sato, Kazuhiko Kawamoto, Hiroshi Kera
Main category: cs.LG
TL;DR: A method to optimize the ordering of reasoning steps in Transformers for arithmetic tasks by identifying learning-friendly token sequences through hierarchical reordering.
Details
Motivation: While intermediate reasoning steps in Transformers have been studied extensively, the ordering of these steps has received little attention despite significantly affecting reasoning difficulty. The paper addresses the novel task of reordering decoder input tokens into optimal sequences for learning arithmetic tasks.Method: Proposes a pipeline that first trains a Transformer on mixtures of target sequences with different orderings, then identifies benign orders based on fast loss drops in early training. Uses a two-stage hierarchical approach for inter- and intra-block reordering to handle factorial search space growth.
Result: Experiments on seven order-sensitive arithmetic tasks show the method identifies learning-friendly orders from billions of candidates. Notably recovers the reverse-digit order previously reported for multiplication tasks.
Conclusion: Token ordering significantly impacts Transformer learning efficiency for arithmetic reasoning. The proposed hierarchical reordering method effectively identifies optimal sequences, demonstrating the importance of step ordering in chain-of-thought reasoning.
Abstract: The chain of thought, i.e., step-by-step reasoning, is one of the fundamental mechanisms of Transformers. While the design of intermediate reasoning steps has been extensively studied and shown to critically influence performance on mathematical, multi-step reasoning tasks, the ordering of these steps has received little attention, despite its significant effect on the difficulty of reasoning. This study addresses a novel task of unraveling the chain of thought – reordering decoder input tokens into a learning-friendly sequence for Transformers, for learning arithmetic tasks. The proposed pipeline first trains a Transformer on a mixture of target sequences arranged in different orders and then identifies benign orders as those with fast loss drops in the early stage. As the search space grows factorially in sequence length, we propose a two-stage hierarchical approach for inter- and intra-block reordering. Experiments on seven order-sensitive arithmetic tasks show that our method identifies a learning-friendly order out of a few billion candidates. Notably, it recovered the reverse-digit order reported in prior studies for the multiplication task.
[755] wd1: Weighted Policy Optimization for Reasoning in Diffusion Language Models
Xiaohang Tang, Rares Dolga, Sangwoong Yoon, Ilija Bogunovic
Main category: cs.LG
TL;DR: wd1: A ratio-free policy optimization method for diffusion-based LLMs that reformulates RL as weighted log-likelihood, reducing computational cost and improving reasoning performance.
Details
Motivation: Current RL methods for diffusion-based LLMs require approximating multiple policy likelihoods at each optimization step, leading to computational overhead, large variance, and estimation errors in the RL objective.Method: Proposes wd1, a ratio-free policy optimization approach that reformulates RL objective as weighted log-likelihood, requiring only single approximation for current policy likelihood. Method interpreted as energy-guided discrete diffusion training with negative sample unlearning.
Result: Outperforms diffusion-based GRPO (d1) with +59% accuracy improvement on LLaDA-8B model, lower computational cost. Extended wd1++ achieves SOTA math performance: 44.2% on MATH500 and 84.5% on GSM8K with only 20 RL training steps.
Conclusion: wd1 provides efficient ratio-free policy optimization for diffusion-based LLMs, reducing computational burden while improving reasoning capabilities, with wd1++ achieving state-of-the-art math reasoning performance.
Abstract: Improving the reasoning capabilities of diffusion-based large language models (dLLMs) through reinforcement learning (RL) remains an open problem. The intractability of dLLMs likelihood function necessitates approximating the current, old, and reference policy likelihoods at each policy optimization step. This reliance introduces additional computational overhead, and can lead to large variance and estimation error in RL objective – particularly in computing the policy ratio for importance sampling. To mitigate these issues, we introduce wd1, a novel ratio-free policy optimization approach that reformulates the RL objective as a weighted log-likelihood, requiring only a single approximation for the current parametrized policy likelihood. We formally show that our proposed method can be interpreted as energy-guided discrete diffusion training combined with negative sample unlearning, thereby confirming its theoretical soundness. In experiments on LLaDA-8B model, wd1 outperforms diffusion-based GRPO (d1) while requiring lower computational cost, achieving up to a $+59%$ improvement in accuracy. Furthermore, we extend wd1 to denoising-stepwise weighted policy optimization (wd1++), achieving state-of-the-art math performance of $44.2%$ on MATH500 and $84.5%$ on GSM8K with only 20 RL training steps.
[756] The Serial Scaling Hypothesis
Yuxi Liu, Konpat Preechakul, Kananart Kuwaranancharoen, Yutong Bai
Main category: cs.LG
TL;DR: The paper identifies a critical limitation in current parallel-centric ML architectures: they cannot efficiently solve inherently sequential problems that require dependent computational steps, and shows diffusion models also fail at such tasks despite their sequential nature.
Details
Motivation: The motivation is to address a blind spot in machine learning where current parallel-centric approaches fail on inherently sequential problems like mathematical reasoning, physical simulations, and sequential decision-making that require sequentially dependent computational steps.Method: The authors formalize the distinction between parallelizable and inherently serial problems in complexity theory, demonstrate fundamental limitations of current parallel architectures on such tasks, and show for the first time that diffusion models are also incapable of solving inherently serial problems despite their sequential nature.
Result: The paper establishes that inherently serial problems exist as a distinct class that cannot be efficiently parallelized, and shows both current parallel architectures and diffusion models face fundamental limitations on these tasks.
Conclusion: Recognizing the serial nature of computation has profound implications for machine learning, model design, and hardware development, suggesting a need for new approaches that can handle inherently sequential problems.
Abstract: While machine learning has advanced through massive parallelization, we identify a critical blind spot: some problems are fundamentally sequential. These “inherently serial” problems-from mathematical reasoning to physical simulations to sequential decision-making-require sequentially dependent computational steps that cannot be efficiently parallelized. We formalize this distinction in complexity theory, and demonstrate that current parallel-centric architectures face fundamental limitations on such tasks. Then, we show for first time that diffusion models despite their sequential nature are incapable of solving inherently serial problems. We argue that recognizing the serial nature of computation holds profound implications on machine learning, model design, and hardware development.
[757] DeepC4: Deep Conditional Census-Constrained Clustering for Large-scale Multitask Spatial Disaggregation of Urban Morphology
Joshua Dimasaka, Christian Geiß, Emily So
Main category: cs.LG
TL;DR: DeepC4: A deep learning approach for spatial disaggregation of urban morphology maps using census constraints and satellite imagery, applied to building exposure mapping in Rwanda.
Details
Motivation: Existing spatial disaggregation techniques for large-scale urban morphology mapping (like GEM and METEOR projects) face challenges with local discrepancies from census data and model uncertainties, especially with weak supervision. Need better methods that incorporate validated census statistics as constraints.Method: Deep Conditional Census-Constrained Clustering (DeepC4) - a deep learning-based spatial disaggregation approach that incorporates local census statistics as cluster-level constraints while considering multiple conditional label relationships through joint multitask learning on satellite imagery patterns.
Result: Enhanced quality of Rwandan urban morphology maps (building exposure and physical vulnerability) at third-level administrative units from 2022 census data, outperforming GEM and METEOR approaches.
Conclusion: DeepC4 offers a new deep learning mapping technique that explicitly encodes validated census data and expert knowledge for explainable auditing of coarse-grained derived information at large scales, relevant for sustainable development and disaster risk reduction.
Abstract: To understand our global progress for sustainable development and disaster risk reduction in many developing economies, two recent major initiatives - the Uniform African Exposure Dataset of the Global Earthquake Model (GEM) Foundation and the Modelling Exposure through Earth Observation Routines (METEOR) Project - implemented classical spatial disaggregation techniques to generate large-scale mapping of urban morphology using the information from various satellite imagery and its derivatives, geospatial datasets of the built environment, and subnational census statistics. However, the local discrepancy with well-validated census statistics and the propagated model uncertainties remain a challenge in such coarse-to-fine-grained mapping problems, specifically constrained by weak and conditional label supervision. Therefore, we present Deep Conditional Census-Constrained Clustering (DeepC4), a novel deep learning-based spatial disaggregation approach that incorporates local census statistics as cluster-level constraints while considering multiple conditional label relationships in a joint multitask learning of the patterns of satellite imagery. To demonstrate, compared to GEM and METEOR, we enhanced the quality of Rwandan maps of urban morphology, specifically building exposure and physical vulnerability, at the third-level administrative unit from the 2022 census. As the world approaches the conclusion of many global frameworks in 2030, our work offers a new deep learning-based mapping technique that explicitly encodes well-validated census and experts’ belief systems to achieve an explainable and interpretable auditing of existing coarse-grained derived information at large scales.
[758] FGBench: A Dataset and Benchmark for Molecular Property Reasoning at Functional Group-Level in Large Language Models
Xuan Liu, Siru Ouyang, Xianrui Zhong, Jiawei Han, Huimin Zhao
Main category: cs.LG
TL;DR: FGBench introduces a dataset of 625K molecular property reasoning problems with fine-grained functional group annotations to enhance LLMs’ structure-aware reasoning in chemistry.
Details
Motivation: Existing chemistry datasets focus on molecular-level properties but overlook fine-grained functional group information, which is crucial for building interpretable, structure-aware LLMs that can reason about molecular structure-property relationships.Method: Created FGBench dataset with 625K molecular property reasoning problems featuring precisely annotated and localized functional groups across 245 different groups. Includes regression and classification tasks in three categories: single functional group impacts, multiple functional group interactions, and direct molecular comparisons.
Result: Benchmarking state-of-the-art LLMs on 7K curated data shows current models struggle with FG-level property reasoning, highlighting the need for enhanced reasoning capabilities in chemistry tasks.
Conclusion: FGBench provides a foundational framework for generating question-answer pairs with functional group-level information to help LLMs better understand fine-grained molecular structure-property relationships, advancing molecular design and drug discovery.
Abstract: Large language models (LLMs) have gained significant attention in chemistry. However, most existing datasets center on molecular-level property prediction and overlook the role of fine-grained functional group (FG) information. Incorporating FG-level data can provide valuable prior knowledge that links molecular structures with textual descriptions, which can be used to build more interpretable, structure-aware LLMs for reasoning on molecule-related tasks. Moreover, LLMs can learn from such fine-grained information to uncover hidden relationships between specific functional groups and molecular properties, thereby advancing molecular design and drug discovery. Here, we introduce FGBench, a dataset comprising 625K molecular property reasoning problems with functional group information. Functional groups are precisely annotated and localized within the molecule, which ensures the dataset’s interoperability thereby facilitating further multimodal applications. FGBench includes both regression and classification tasks on 245 different functional groups across three categories for molecular property reasoning: (1) single functional group impacts, (2) multiple functional group interactions, and (3) direct molecular comparisons. In the benchmark of state-of-the-art LLMs on 7K curated data, the results indicate that current LLMs struggle with FG-level property reasoning, highlighting the need to enhance reasoning capabilities in LLMs for chemistry tasks. We anticipate that the methodology employed in FGBench to construct datasets with functional group-level information will serve as a foundational framework for generating new question-answer pairs, enabling LLMs to better understand fine-grained molecular structure-property relationships. The dataset and evaluation code are available at https://github.com/xuanliugit/FGBench.
[759] Beyond Prompt-Induced Lies: Investigating LLM Deception on Benign Prompts
Zhaomin Wu, Mingzhe Du, See-Kiong Ng, Bingsheng He
Main category: cs.LG
TL;DR: LLMs can engage in self-initiated deception on benign prompts, measured through statistical metrics based on psychological principles, with deception increasing with task difficulty and not always decreasing with model capacity.
Details
Motivation: To investigate LLMs' self-initiated deception beyond human-induced deception, addressing the critical trustworthiness issue in LLMs deployed for reasoning, planning, and decision-making tasks.Method: Proposed a framework based on Contact Searching Questions (CSQ) with two statistical metrics: Deceptive Intention Score (measures bias toward hidden objective) and Deceptive Behavior Score (measures inconsistency between internal belief and expressed output). Evaluated 16 leading LLMs.
Result: Both deception metrics rise in parallel and escalate with task difficulty for most models. Increasing model capacity does not always reduce deception, posing a significant challenge for future LLM development.
Conclusion: LLMs exhibit self-initiated deception that increases with task difficulty, and larger models don’t necessarily become more trustworthy, highlighting a critical safety concern for LLM deployment.
Abstract: Large Language Models (LLMs) are widely deployed in reasoning, planning, and decision-making tasks, making their trustworthiness critical. A significant and underexplored risk is intentional deception, where an LLM deliberately fabricates or conceals information to serve a hidden objective. Existing studies typically induce deception by explicitly setting a hidden objective through prompting or fine-tuning, which may not reflect real-world human-LLM interactions. Moving beyond such human-induced deception, we investigate LLMs’ self-initiated deception on benign prompts. To address the absence of ground truth, we propose a framework based on Contact Searching Questions~(CSQ). This framework introduces two statistical metrics derived from psychological principles to quantify the likelihood of deception. The first, the Deceptive Intention Score, measures the model’s bias toward a hidden objective. The second, the Deceptive Behavior Score, measures the inconsistency between the LLM’s internal belief and its expressed output. Evaluating 16 leading LLMs, we find that both metrics rise in parallel and escalate with task difficulty for most models. Moreover, increasing model capacity does not always reduce deception, posing a significant challenge for future LLM development.
[760] Lightning Prediction under Uncertainty: DeepLight with Hazy Loss
Md Sultanul Arifin, Abu Nowshed Sakib, Yeasir Rayhan, Tanzima Hashem
Main category: cs.LG
TL;DR: DeepLight is a novel deep learning architecture for lightning prediction that uses multi-source meteorological data and a dual-encoder with multi-branch convolutions to capture spatial correlations, addressing the randomness and uncertainty of lightning events through a novel Hazy Loss function.
Details
Motivation: Lightning poses significant risks to human safety and economic stability, exacerbated by climate change. Existing prediction models struggle with lightning's dynamic spatial context, inherent randomness, underutilize key observational data, and rely heavily on computationally expensive Numerical Weather Prediction systems.Method: DeepLight uses a dual-encoder architecture with multi-branch convolutions to process multi-source meteorological data (radar reflectivity, cloud properties, historical lightning occurrences). It employs a novel Hazy Loss function that penalizes deviations based on proximity to true events to handle spatio-temporal uncertainty.
Result: Extensive experiments show DeepLight improves the Equitable Threat Score (ETS) by 18-30% over state-of-the-art methods, establishing it as a robust solution for lightning prediction.
Conclusion: DeepLight represents a significant advancement in lightning prediction by effectively addressing the limitations of existing models through its innovative architecture and loss function, offering improved accuracy for early warning systems.
Abstract: Lightning, a common feature of severe meteorological conditions, poses significant risks, from direct human injuries to substantial economic losses. These risks are further exacerbated by climate change. Early and accurate prediction of lightning would enable preventive measures to safeguard people, protect property, and minimize economic losses. In this paper, we present DeepLight, a novel deep learning architecture for predicting lightning occurrences. Existing prediction models face several critical limitations: i) they often struggle to capture the dynamic spatial context and the inherent randomness of lightning events, including whether lightning occurs and its variability in location and timing even under similar meteorological conditions; ii) they underutilize key observational data, such as radar reflectivity and cloud properties; and iii) they rely heavily on Numerical Weather Prediction (NWP) systems, which are both computationally expensive and highly sensitive to parameter settings. To overcome these challenges, DeepLight leverages multi-source meteorological data, including radar reflectivity, cloud properties, and historical lightning occurrences through a dual-encoder architecture. By employing multi-branch convolution techniques, it dynamically captures spatial correlations across varying extents. Furthermore, its novel Hazy Loss function explicitly addresses the spatio-temporal uncertainty of lightning by penalizing deviations based on proximity to true events, enabling the model to better learn patterns amidst randomness. Extensive experiments show that DeepLight improves the Equitable Threat Score (ETS) by 18%–30% over state-of-the-art methods, establishing it as a robust solution for lightning prediction.
[761] Zono-Conformal Prediction: Zonotope-Based Uncertainty Quantification for Regression and Classification Tasks
Laura Lützow, Michael Eichelbeck, Mykel J. Kochenderfer, Matthias Althoff
Main category: cs.LG
TL;DR: Zono-conformal prediction is a new uncertainty quantification method that constructs prediction zonotopes (multi-dimensional sets) with statistical coverage guarantees, offering computational efficiency and better handling of multi-dimensional dependencies compared to traditional interval-based conformal prediction.
Details
Motivation: Current conformal prediction methods are computationally expensive, data-intensive, and limited to interval representations that cannot capture dependencies in multi-dimensional outputs. The authors aim to address these limitations with a more efficient and expressive approach.Method: Introduces zono-conformal prediction inspired by interval predictor models and reachset-conformant identification. Places zonotopic uncertainty sets directly into the base predictor model, allowing identification via a single linear program. Focuses on feed-forward neural networks but applicable to arbitrary nonlinear predictors. Also constructs optimal predictors for classification tasks and provides outlier detection methods.
Result: Zono-conformal predictors are less conservative than interval predictor models and standard conformal prediction methods while achieving similar test data coverage. The method provides probabilistic coverage guarantees and enables outlier detection in identification data.
Conclusion: Zono-conformal prediction offers an efficient, data-parsimonious approach to uncertainty quantification with multi-dimensional coverage sets, overcoming limitations of traditional interval-based conformal prediction methods.
Abstract: Conformal prediction is a popular uncertainty quantification method that augments a base predictor to return sets of predictions with statistically valid coverage guarantees. However, current methods are often computationally expensive and data-intensive, as they require constructing an uncertainty model before calibration. Moreover, existing approaches typically represent the prediction sets with intervals, which limits their ability to capture dependencies in multi-dimensional outputs. We address these limitations by introducing zono-conformal prediction, a novel approach inspired by interval predictor models and reachset-conformant identification that constructs prediction zonotopes with assured coverage. By placing zonotopic uncertainty sets directly into the model of the base predictor, zono-conformal predictors can be identified via a single, data-efficient linear program. While we can apply zono-conformal prediction to arbitrary nonlinear base predictors, we focus on feed-forward neural networks in this work. Aside from regression tasks, we also construct optimal zono-conformal predictors in classification settings where the output of an uncertain predictor is a set of possible classes. We provide probabilistic coverage guarantees and present methods for detecting outliers in the identification data. In extensive numerical experiments, we show that zono-conformal predictors are less conservative than interval predictor models and standard conformal prediction methods, while achieving a similar coverage over the test data.
[762] Weather-Driven Agricultural Decision-Making Using Digital Twins Under Imperfect Conditions
Tamim Ahmed, Monowar Hasan
Main category: cs.LG
TL;DR: Cerealia is a modular digital twin framework for detecting weather data inconsistencies in agriculture using neural network anomaly detection, deployed on NVIDIA Jetson Orin platform.
Details
Motivation: Digital twins can enhance data-driven decision-making in agriculture by providing real-time virtual representations of physical systems. Weather data inconsistencies are problematic for agricultural decision-making and automation tasks, especially when perfect weather feeds are unavailable.Method: Developed Cerealia, a modular framework using neural network models to detect anomalies in weather data. Implemented prototype on NVIDIA Jetson Orin platform and tested with operational weather network in commercial orchard plus publicly available weather datasets.
Result: Framework successfully detects inconsistencies in agricultural weather data measurements, aiding end-users in informed decision-making when perfect weather data feeds are unavailable.
Conclusion: Digital twin technology with neural network anomaly detection provides effective solution for identifying weather data inconsistencies in agriculture, enabling better data-driven decision-making.
Abstract: By offering a dynamic, real-time virtual representation of physical systems, digital twin technology can enhance data-driven decision-making in digital agriculture. Our research shows how digital twins are useful for detecting inconsistencies in agricultural weather data measurements, which are key attributes for various agricultural decision-making and automation tasks. We develop a modular framework named Cerealia that allows end-users to check for data inconsistencies when perfect weather feeds are unavailable. Cerealia uses neural network models to check anomalies and aids end-users in informed decision-making. We develop a prototype of Cerealia using the NVIDIA Jetson Orin platform and test it with an operational weather network established in a commercial orchard as well as publicly available weather datasets.
[763] MAVIS: Multi-Objective Alignment via Inference-Time Value-Guided Selection
Jeremy Carleton, Debajoy Mukherjee, Srinivas Shakkottai, Dileep Kalathil
Main category: cs.LG
TL;DR: MAVIS is a lightweight inference-time alignment framework that enables dynamic control over LLM behavior across multiple objectives without modifying base model weights, using small value models and user-specified weights to adjust output distributions.
Details
Motivation: LLMs need to balance multiple conflicting objectives (helpfulness, harmlessness, humor) in diverse applications, but traditional fine-tuning methods for each objective are computationally expensive and inflexible for dynamic preference configurations.Method: Trains small value models for distinct objectives, combines them at inference time using user-specified weights to produce a tilting function that adjusts the base model’s output distribution toward desired trade-offs, using an iterative algorithm for KL-regularized policy improvement.
Result: MAVIS achieves superior pareto front compared to baselines that fine-tune per-objective models and combine them post hoc or train a single preference-conditioned value model for guidance.
Conclusion: MAVIS provides an efficient, flexible inference-time alignment framework for multi-objective LLM control without weight modification, enabling dynamic adaptation to user preferences.
Abstract: Large Language Models (LLMs) are increasingly deployed across diverse applications that demand balancing multiple, often conflicting, objectives – such as helpfulness, harmlessness, or humor. Many traditional methods for aligning outputs to user-specific preferences require fine-tuning models for each objective or for specific preference configurations, which is computationally expensive and inflexible. We introduce \textbf{MAVIS} – \textit{Multi-Objective Alignment via Inference-Time Value-Guided Selection} – a lightweight inference-time alignment framework that enables dynamic control over LLM behavior without modifying the base model’s weights. MAVIS trains a set of small value models, each corresponding to a distinct objective. At inference time, these value models are combined using user-specified weights to produce a tilting function that adjusts the base model’s output distribution toward desired trade-offs. The value models are trained using a simple iterative algorithm that enables monotonic improvement of the KL-regularized policy. We show empirically that MAVIS achieves a superior pareto front compared to baselines which fine-tune per-objective models and combine them post hoc or train a single preference-conditioned value model for guidance. Our code is available at https://github.com/5-Jeremy/MAVIS/tree/main.
[764] Predicting the Order of Upcoming Tokens Improves Language Modeling
Zayd M. K. Zuhri, Erland Hilman Fuadi, Alham Fikri Aji
Main category: cs.LG
TL;DR: Token Order Prediction (TOP) is proposed as an alternative to Multi-Token Prediction (MTP) for language model training, using learning-to-rank to predict token order rather than exact future tokens, showing better performance across NLP benchmarks.
Details
Motivation: Multi-token prediction (MTP) has shown inconsistent improvements over next-token prediction (NTP) in language model training, often underperforming on standard NLP benchmarks. The authors identified that MTP's exact future token prediction is too difficult as an auxiliary loss, motivating a simpler alternative.Method: Proposes Token Order Prediction (TOP) which trains models to order upcoming tokens by their proximity using a learning-to-rank loss. TOP requires only a single additional unembedding layer compared to MTP’s multiple transformer layers. Models of 340M, 1.8B, and 7B parameters were pretrained using NTP, MTP, DeepSeek MTP (DS-MTP) and TOP objectives.
Result: TOP outperforms NTP, MTP, and DS-MTP on nine standard NLP benchmarks, even at scale. TOP models with continued training on math and code also perform better on 4 relevant benchmarks. On the synthetic star graph task, TOP enables pathfinding where NTP, MTP, and DS-MTP fail.
Conclusion: Token Order Prediction (TOP) is a more effective auxiliary objective than Multi-Token Prediction for language model training, providing consistent improvements across various benchmarks and tasks while being more computationally efficient.
Abstract: Multi-token prediction (MTP) has been proposed as an auxiliary objective to improve next-token prediction (NTP) in language model training but shows inconsistent improvements, underperforming in standard NLP benchmarks. We found MTP’s exact future token prediction to be too difficult as an auxiliary loss. Instead, we propose token order prediction (TOP), which trains models to order upcoming tokens by their proximity using a learning-to-rank loss. TOP requires only a single additional unembedding layer compared to MTP’s multiple transformer layers. We pretrain models of 340M, 1.8B, and 7B parameters using NTP, MTP, DeepSeek MTP (DS-MTP) and TOP objectives. The results of nine standard NLP benchmarks show that TOP overall outperforms NTP, MTP, and DS-MTP even at scale. TOP models with continued training on math and code also perform better on 4 relevant benchmarks. On the synthetic star graph task, TOP enables pathfinding on graphs where NTP, MTP, and DS-MTP fail. Our code is available at https://github.com/zaydzuhri/token-order-prediction
[765] ART: Adaptive Resampling-based Training for Imbalanced Classification
Arjun Basandrai, Shourya Jain, K. Ilanthenral
Main category: cs.LG
TL;DR: ART is an adaptive resampling method that dynamically adjusts training data distribution based on class-wise performance metrics, outperforming traditional static resampling methods on imbalanced classification tasks.
Details
Motivation: Traditional resampling methods use fixed distributions that ignore changes in class-wise learning difficulty during training, limiting model performance on imbalanced datasets.Method: ART periodically updates training data distribution using class-wise macro F1 scores computed at fixed intervals, adapting resampling at the class level rather than instance level to focus on underperforming classes.
Result: ART consistently outperforms both resampling-based and algorithm-level methods (SMOTE, NearMiss, Cost-sensitive Learning) on binary and multi-class tasks with varying imbalance degrees, with statistically significant improvements on tabular datasets and favorable results on text/image tasks.
Conclusion: ART provides a reliable adaptive resampling approach that consistently delivers the strongest macro F1 performance across diverse imbalanced classification tasks, making it a robust alternative to traditional static methods.
Abstract: Traditional resampling methods for handling class imbalance typically uses fixed distributions, undersampling the majority or oversampling the minority. These static strategies ignore changes in class-wise learning difficulty, which can limit the overall performance of the model. This paper proposes an Adaptive Resampling-based Training (ART) method that periodically updates the distribution of the training data based on the class-wise performance of the model. Specifically, ART uses class-wise macro F1 scores, computed at fixed intervals, to determine the degree of resampling to be performed. Unlike instance-level difficulty modeling, which is noisy and outlier-sensitive, ART adapts at the class level. This allows the model to incrementally shift its attention towards underperforming classes in a way that better aligns with the optimization objective. Results on diverse benchmarks, including Pima Indians Diabetes and Yeast dataset demonstrate that ART consistently outperforms both resampling-based and algorithm-level methods, including Synthetic Minority Oversampling Technique (SMOTE), NearMiss Undersampling, and Cost-sensitive Learning on binary as well as multi-class classification tasks with varying degrees of imbalance. In most settings, these improvements are statistically significant. On tabular datasets, gains are significant under paired t-tests and Wilcoxon tests (p < 0.05), while results on text and image tasks remain favorable. Compared to training on the original imbalanced data, ART improves macro F1 by an average of 2.64 percentage points across all tested tabular datasets. Unlike existing methods, whose performance varies by task, ART consistently delivers the strongest macro F1, making it a reliable choice for imbalanced classification.
[766] Online reinforcement learning via sparse Gaussian mixture model Q-functions
Minh Vu, Konstantinos Slavakis
Main category: cs.LG
TL;DR: Online RL framework using sparse Gaussian mixture model Q-functions with Riemannian optimization, achieving competitive performance with fewer parameters than dense DeepRL methods.
Details
Motivation: To develop an interpretable and structured online RL framework that can effectively explore while controlling model complexity, addressing limitations of dense deep RL methods that often require many parameters and can overfit.Method: Proposes sparse Gaussian mixture model Q-functions (S-GMM-QFs) with online policy iteration using Hadamard overparametrization for sparsification. Uses Riemannian manifold structure for principled parameter updates via online gradient descent on smooth objectives.
Result: S-GMM-QFs match or outperform dense DeepRL methods on standard benchmarks while using significantly fewer parameters. They maintain strong performance even in low-parameter regimes where sparsified DeepRL methods fail to generalize.
Conclusion: The structured S-GMM-QF framework provides an interpretable, parameter-efficient alternative to dense DeepRL, with principled optimization via Riemannian geometry enabling effective online learning.
Abstract: This paper introduces a structured and interpretable online policy-iteration framework for reinforcement learning (RL), built around the novel class of sparse Gaussian mixture model Q-functions (S-GMM-QFs). Extending earlier work that trained GMM-QFs offline, the proposed framework develops an online scheme that leverages streaming data to encourage exploration. Model complexity is regulated through sparsification by Hadamard overparametrization, which mitigates overfitting while preserving expressiveness. The parameter space of S-GMM-QFs is naturally endowed with a Riemannian manifold structure, allowing for principled parameter updates via online gradient descent on a smooth objective. Numerical experiments show that S-GMM-QFs match or even outperform dense deep RL (DeepRL) methods on standard benchmarks while using significantly fewer parameters. Moreover, they maintain strong performance even in low-parameter regimes where sparsified DeepRL methods fail to generalize.
[767] DiffusionNFT: Online Diffusion Reinforcement with Forward Process
Kaiwen Zheng, Huayu Chen, Haotian Ye, Haoxiang Wang, Qinsheng Zhang, Kai Jiang, Hang Su, Stefano Ermon, Jun Zhu, Ming-Yu Liu
Main category: cs.LG
TL;DR: DiffusionNFT is a new online RL paradigm for diffusion models that optimizes directly on the forward process via flow matching, contrasting positive and negative generations to incorporate reinforcement signals without likelihood estimation or CFG.
Details
Motivation: Existing RL methods for diffusion models have fundamental drawbacks including solver restrictions, forward-reverse inconsistency, and complicated integration with classifier-free guidance. There's a need for more efficient and flexible RL approaches for diffusion model fine-tuning.Method: DiffusionNFT operates directly on the forward process via flow matching, contrasting positive and negative generations to define an implicit policy improvement direction. It incorporates reinforcement signals into supervised learning objectives without likelihood estimation, works with arbitrary black-box solvers, and requires only clean images rather than sampling trajectories.
Result: DiffusionNFT is up to 25× more efficient than FlowGRPO, improves GenEval score from 0.24 to 0.98 within 1k steps (vs. FlowGRPO’s 0.95 with over 5k steps and CFG), and significantly boosts SD3.5-Medium performance across all benchmarks when using multiple reward models.
Conclusion: DiffusionNFT provides an efficient, CFG-free online RL paradigm for diffusion models that overcomes limitations of previous approaches, enabling flexible training with arbitrary solvers and substantial performance improvements with minimal computational cost.
Abstract: Online reinforcement learning (RL) has been central to post-training language models, but its extension to diffusion models remains challenging due to intractable likelihoods. Recent works discretize the reverse sampling process to enable GRPO-style training, yet they inherit fundamental drawbacks, including solver restrictions, forward-reverse inconsistency, and complicated integration with classifier-free guidance (CFG). We introduce Diffusion Negative-aware FineTuning (DiffusionNFT), a new online RL paradigm that optimizes diffusion models directly on the forward process via flow matching. DiffusionNFT contrasts positive and negative generations to define an implicit policy improvement direction, naturally incorporating reinforcement signals into the supervised learning objective. This formulation enables training with arbitrary black-box solvers, eliminates the need for likelihood estimation, and requires only clean images rather than sampling trajectories for policy optimization. DiffusionNFT is up to $25\times$ more efficient than FlowGRPO in head-to-head comparisons, while being CFG-free. For instance, DiffusionNFT improves the GenEval score from 0.24 to 0.98 within 1k steps, while FlowGRPO achieves 0.95 with over 5k steps and additional CFG employment. By leveraging multiple reward models, DiffusionNFT significantly boosts the performance of SD3.5-Medium in every benchmark tested.
[768] Functional Scaling Laws in Kernel Regression: Loss Dynamics and Learning Rate Schedules
Binghui Li, Fengling Chen, Zixun Huang, Lean Wang, Lei Wu
Main category: cs.LG
TL;DR: The paper introduces Functional Scaling Laws (FSL) that capture full loss trajectories in LLM training under arbitrary learning rate schedules, providing theoretical insights into training dynamics beyond final-step loss.
Details
Motivation: Existing scaling law studies focus only on final-step loss, leaving gaps in understanding how entire loss dynamics evolve and how learning rate schedules shape training trajectories. There's a need for a more comprehensive theoretical framework.Method: Analyzes SGD on a power-law kernel regression model using a novel intrinsic-time viewpoint. Derives FSL that captures full loss trajectories under arbitrary learning rate schedules, with schedule influence modeled through a convolutional functional. Instantiates theory for three LRSs: constant, exponential decay, and warmup-stable-decay.
Result: Establishes explicit scaling relations in data- and compute-limited regimes, explaining empirical phenomena: higher-capacity models are more efficient, learning-rate decay improves training efficiency, and WSD schedules outperform pure decay. Experiments on 0.1B-1B parameter LLMs validate FSL as a surrogate model for predicting loss trajectories.
Conclusion: FSL provides a comprehensive theoretical framework for understanding LLM training dynamics beyond final loss, offering practical guidance for learning rate schedule design and training optimization across different resource constraints.
Abstract: Scaling laws have emerged as a unifying lens for understanding and guiding the training of large language models (LLMs). However, existing studies predominantly focus on the final-step loss, leaving open whether the entire loss dynamics obey similar laws and, crucially, how the learning rate schedule (LRS) shapes them. We address these gaps in a controlled theoretical setting by analyzing stochastic gradient descent (SGD) on a power-law kernel regression model. The key insight is a novel intrinsic-time viewpoint, which captures the training progress more faithfully than iteration count. We then establish a Functional Scaling Law (FSL) that captures the full loss trajectory under arbitrary LRSs, with the schedule’s influence entering through a simple convolutional functional. We further instantiate the theory for three representative LRSs – constant, exponential decay, and warmup-stable-decay (WSD) – and derive explicit scaling relations in both data- and compute-limited regimes. These comparisons explain key empirical phenomena: (i) higher-capacity models are more data- and compute-efficient; (ii) learning-rate decay improves training efficiency; and (iii) WSD-type schedules outperform pure decay. Finally, experiments on LLMs ranging from 0.1B to 1B parameters demonstrate the practical relevance of FSL as a surrogate model for fitting and predicting loss trajectories in large-scale pre-training.
[769] Efficient Test-Time Scaling for Small Vision-Language Models
Mehmet Onurcan Kaya, Desmond Elliott, Dim P. Papadopoulos
Main category: cs.LG
TL;DR: Efficient test-time scaling strategies (TTAug and TTAdapt) for small vision-language models that improve performance without compromising computational efficiency.
Details
Motivation: Small VLMs are computationally efficient but suffer from weaker generalization and task performance; existing test-time scaling methods are too computationally demanding, contradicting the resource-efficient goals of small models.Method: Two novel test-time scaling strategies: (1) Test-Time Augmentation (TTAug) generates multiple augmented inputs and aggregates outputs at token level without parameter updates; (2) Test-Time Adaptation (TTAdapt) adapts model parameters during inference using consensus-based pseudolabels from TTAug.
Result: Consistent performance improvements across nine benchmarks while maintaining computational efficiency suitable for resource-constrained environments; approach generalizes across different model scales and VLMs without additional tuning.
Conclusion: Proposed efficient test-time scaling strategies effectively address the performance limitations of small VLMs while preserving their computational efficiency advantages.
Abstract: Small Vision-Language Models (VLMs) provide a computationally efficient alternative to larger models, at the cost of weaker generalization abilities and downstream task performance. These shortcomings could be addressed by test-time scaling techniques, but existing methods are typically computationally demanding, contradicting the resource-efficient design goals of small models. To address these limitations, we propose two novel and efficient test-time scaling strategies that leverage the model-internal features rather than external supervision: (i) Test-Time Augmentation (TTAug), which generates multiple augmented inputs and aggregates outputs at the token level without parameter updates, and (ii) Test-Time Adaptation (TTAdapt), which adapts model parameters during inference using consensus-based pseudolabels from TTAug. Through extensive experiments across nine benchmarks, we demonstrate consistent performance improvements while maintaining computational efficiency suitable for resource-constrained environments. The generality of our approach is demonstrated both within models at different scales and across different VLMs without additional tuning.
[770] Learning the Inverse Temperature of Ising Models under Hard Constraints using One Sample
Rohan Chauhan, Ioannis Panageas
Main category: cs.LG
TL;DR: Single-sample estimation of inverse temperature parameter β for truncated Ising models using pseudolikelihood maximization, with applications to k-SAT constrained configurations.
Details
Motivation: The paper addresses the challenging problem of parameter estimation for truncated Ising models using only a single sample, which is relevant for statistical inference in constrained systems where data collection is expensive or limited.Method: The authors design an estimator based on maximization of pseudolikelihood, generalizing recent techniques from related works. The approach handles truncation sets expressed as satisfying assignments of k-SAT formulas and works with bounded-degree graphs.
Result: The estimator achieves O(Δ³/√n)-consistency with the true parameter β* and runs in nearly O(n) time, under conditions where k ≳ log(d²k)Δ³.
Conclusion: The paper provides an efficient single-sample estimator for truncated Ising models with theoretical guarantees, extending pseudolikelihood methods to handle challenging truncation constraints.
Abstract: We consider the problem of estimating inverse temperature parameter $β$ of an $n$-dimensional truncated Ising model using a single sample. Given a graph $G = (V,E)$ with $n$ vertices, a truncated Ising model is a probability distribution over the $n$-dimensional hypercube ${-1,1}^n$ where each configuration $\mathbfσ$ is constrained to lie in a truncation set $S \subseteq {-1,1}^n$ and has probability $\Pr(\mathbfσ) \propto \exp(β\mathbfσ^\top A\mathbfσ)$ with $A$ being the adjacency matrix of $G$. We adopt the recent setting of [Galanis et al. SODA'24], where the truncation set $S$ can be expressed as the set of satisfying assignments of a $k$-SAT formula. Given a single sample $\mathbfσ$ from a truncated Ising model, with inverse parameter $β^$, underlying graph $G$ of bounded degree $Δ$ and $S$ being expressed as the set of satisfying assignments of a $k$-SAT formula, we design in nearly $O(n)$ time an estimator $\hatβ$ that is $O(Δ^3/\sqrt{n})$-consistent with the true parameter $β^$ for $k \gtrsim \log(d^2k)Δ^3.$ Our estimator is based on the maximization of the pseudolikelihood, a notion that has received extensive analysis for various probabilistic models without [Chatterjee, Annals of Statistics ‘07] or with truncation [Galanis et al. SODA ‘24]. Our approach generalizes recent techniques from [Daskalakis et al. STOC ‘19, Galanis et al. SODA ‘24], to confront the more challenging setting of the truncated Ising model.
[771] The Rogue Scalpel: Activation Steering Compromises LLM Safety
Anton Korznikov, Andrey Galichin, Alexey Dontsov, Oleg Y. Rogov, Ivan Oseledets, Elena Tutubalina
Main category: cs.LG
TL;DR: Activation steering in LLMs systematically breaks alignment safeguards, enabling harmful compliance even with random or interpretable feature directions, challenging safety-through-interpretability paradigms.
Details
Motivation: The paper investigates whether activation steering - adding meaningful vectors to LLM hidden states during inference - is truly a safer alternative to fine-tuning, as often claimed. The authors question the assumption that precise control over model internals guarantees precise control over model behavior.Method: The researchers conducted extensive experiments across different model families, testing: 1) steering with random directions, 2) steering with interpretable features from sparse autoencoders (SAEs), and 3) combining multiple steering vectors to create universal attacks. They measured harmful compliance rates across various prompts.
Result: Random steering increased harmful compliance from 0% to 1-13%. SAE-based steering showed comparable harmful potential. Combining 20 random vectors that jailbreak a single prompt created universal attacks that significantly increased harmful compliance on unseen requests.
Conclusion: Activation steering systematically breaks model alignment safeguards, demonstrating that precise control over model internals does not guarantee precise control over model behavior. This challenges the paradigm of safety through interpretability.
Abstract: Activation steering is a promising technique for controlling LLM behavior by adding semantically meaningful vectors directly into a model’s hidden states during inference. It is often framed as a precise, interpretable, and potentially safer alternative to fine-tuning. We demonstrate the opposite: steering systematically breaks model alignment safeguards, making it comply with harmful requests. Through extensive experiments on different model families, we show that even steering in a random direction can increase the probability of harmful compliance from 0% to 1-13%. Alarmingly, steering benign features from a sparse autoencoder (SAE), a common source of interpretable directions, demonstrates a comparable harmful potential. Finally, we show that combining 20 randomly sampled vectors that jailbreak a single prompt creates a universal attack, significantly increasing harmful compliance on unseen requests. These results challenge the paradigm of safety through interpretability, showing that precise control over model internals does not guarantee precise control over model behavior.
[772] Effective Quantization of Muon Optimizer States
Aman Gupta, Rafael Celente, Abhishek Shivanna, D. T. Braithwaite, Gregory Dexter, Shao Tang, Hiroto Udagawa, Daniel Silva, Rohan Ramanath, S. Sathiya Keerthi
Main category: cs.LG
TL;DR: 8-bit Muon optimizer uses blockwise quantization to reduce optimizer state memory by up to 62% while maintaining performance parity with full-precision Muon in LLM training.
Details
Motivation: Muon optimizer shows better convergence than AdamW but suffers from high memory overhead from maintaining high-precision optimizer states, limiting large-scale deployment.Method: Introduces 8-bit Muon using blockwise quantization, leveraging Muon’s update mechanism compatibility with simple linear quantization (unlike AdamW which requires complex dynamic scaling).
Result: Achieves parity with full-precision Muon in validation loss and downstream benchmarks while reducing optimizer state footprint by up to 62% in experiments up to 2.7B parameters.
Conclusion: 8-bit Muon provides memory-efficient optimization for large-scale LLM training while maintaining performance, with theoretical analysis showing robustness to quantization noise.
Abstract: The Muon optimizer, based on matrix orthogonalization, has recently shown faster convergence and better computational efficiency over AdamW in LLM pre-training. However, the memory overhead of maintaining high-precision optimizer states remains a challenge for large-scale deployment. In this paper, we introduce the 8-bit Muon optimizer using blockwise quantization. In extensive Chinchilla-optimal experiments on pre-training models of up to 2.7B in size and fine-tuning them for instruction following, we demonstrate that 8-bit Muon achieves parity with Muon in terms of validation loss and downstream benchmarks, while achieving up to a 62% reduction in optimizer state footprint. Crucially, we show that Muon’s update mechanism is uniquely compatible with a simple linear quantization scheme, bypassing the complex dynamic scaling required for quantized AdamW. We supplement our empirical findings with a theoretical analysis of Muon’s robustness to quantization noise.
[773] MPCM-Net: Multi-scale network integrates partial attention convolution with Mamba for ground-based cloud image segmentation
Penghui Niu, Jiashuai She, Taotao Cai, Yajuan Zhang, Ping Zhang, Junhua Gu, Jianxin Li
Main category: cs.LG
TL;DR: MPCM-Net: A multi-scale network integrating partial attention convolutions with Mamba architectures for ground-based cloud image segmentation, achieving optimal balance between accuracy and inference speed.
Details
Motivation: Current deep learning approaches for cloud image segmentation have limitations: they rely on dilated convolutions lacking partial feature effectiveness and inter-channel interoperability; attention-based methods neglect accuracy-throughput balance; decoder modifications fail to establish global interdependencies among hierarchical local features, limiting inference efficiency.Method: Proposes MPCM-Net with encoder incorporating MPAC (MPC block with ParCM and ParSM for global spatial interaction across multi-scale cloud formations, and MPA block combining ParAM and ParSM to extract discriminative features with reduced complexity). Decoder uses M2B with SSHD to mitigate contextual loss while maintaining linear complexity for deep feature aggregation across spatial and scale dimensions.
Result: Extensive experiments on the introduced CSRC dataset demonstrate superior performance over state-of-the-art methods, achieving optimal balance between segmentation accuracy and inference speed.
Conclusion: MPCM-Net effectively addresses limitations of existing cloud segmentation methods through innovative architecture combining partial attention convolutions with Mamba, while also contributing a new clear-label, fine-grained segmentation benchmark dataset (CSRC) to the community.
Abstract: Ground-based cloud image segmentation is a critical research domain for photovoltaic power forecasting. Current deep learning approaches primarily focus on encoder-decoder architectural refinements. However, existing methodologies exhibit several limitations:(1)they rely on dilated convolutions for multi-scale context extraction, lacking the partial feature effectiveness and interoperability of inter-channel;(2)attention-based feature enhancement implementations neglect accuracy-throughput balance; and (3)the decoder modifications fail to establish global interdependencies among hierarchical local features, limiting inference efficiency. To address these challenges, we propose MPCM-Net, a Multi-scale network that integrates Partial attention Convolutions with Mamba architectures to enhance segmentation accuracy and computational efficiency. Specifically, the encoder incorporates MPAC, which comprises:(1)a MPC block with ParCM and ParSM that enables global spatial interaction across multi-scale cloud formations, and (2)a MPA block combining ParAM and ParSM to extract discriminative features with reduced computational complexity. On the decoder side, a M2B is employed to mitigate contextual loss through a SSHD that maintains linear complexity while enabling deep feature aggregation across spatial and scale dimensions. As a key contribution to the community, we also introduce and release a dataset CSRC, which is a clear-label, fine-grained segmentation benchmark designed to overcome the critical limitations of existing public datasets. Extensive experiments on CSRC demonstrate the superior performance of MPCM-Net over state-of-the-art methods, achieving an optimal balance between segmentation accuracy and inference speed. The dataset and source code will be available at https://github.com/she1110/CSRC.
[774] Better Hessians Matter: Studying the Impact of Curvature Approximations in Influence Functions
Steve Hong, Runa Eschenhagen, Bruno Mlodozeniec, Richard Turner
Main category: cs.LG
TL;DR: Better Hessian approximations lead to better influence function attributions, with K-FAC eigenvalue mismatch being the main source of error in data attribution accuracy.
Details
Motivation: Influence functions are valuable for tracing model predictions to training data, but their practical use in deep learning is limited by the computational challenge of inverting large, ill-conditioned Hessian matrices. While approximations like GGN and K-FAC exist, it's unclear how approximation quality affects data attribution performance, and whether better Hessian approximations actually lead to better influence scores.Method: The authors investigate the effect of Hessian approximation quality on influence-function attributions in a controlled classification setting. They systematically decompose the approximation steps of recent Hessian approximation methods (GGN and K-FAC) and evaluate each step’s impact on attribution accuracy, comparing different approximation approaches.
Result: Experiments show that better Hessian approximations consistently yield better influence score quality. The mismatch between K-FAC eigenvalues and GGN/EK-FAC eigenvalues accounts for the majority of the error and influence loss. This provides justification for research efforts focused on improving Hessian approximations.
Conclusion: Better Hessian approximations do lead to better influence function attributions, with eigenvalue approximation being the most critical factor. These findings guide future efforts to balance computational tractability and attribution accuracy by highlighting which approximations matter most for data attribution performance.
Abstract: Influence functions offer a principled way to trace model predictions back to training data, but their use in deep learning is hampered by the need to invert a large, ill-conditioned Hessian matrix. Approximations such as Generalised Gauss-Newton (GGN) and Kronecker-Factored Approximate Curvature (K-FAC) have been proposed to make influence computation tractable, yet it remains unclear how the departure from exactness impacts data attribution performance. Critically, given the restricted regime in which influence functions are derived, it is not necessarily clear better Hessian approximations should even lead to better data attribution performance. In this paper, we investigate the effect of Hessian approximation quality on influence-function attributions in a controlled classification setting. Our experiments show that better Hessian approximations consistently yield better influence score quality, offering justification for recent research efforts towards that end. We further decompose the approximation steps for recent Hessian approximation methods and evaluate each step’s influence on attribution accuracy. Notably, the mismatch between K-FAC eigenvalues and GGN/EK-FAC eigenvalues accounts for the majority of the error and influence loss. These findings highlight which approximations are most critical, guiding future efforts to balance computational tractability and attribution accuracy.
[775] LLM DNA: Tracing Model Evolution via Functional Representations
Zhaomin Wu, Haodong Zhao, Ziyang Wang, Jizhou Guo, Qian Wang, Bingsheng He
Main category: cs.LG
TL;DR: LLM DNA: A mathematical representation of LLM functional behavior that enables evolutionary relationship analysis and phylogenetic tree construction across 305 models.
Details
Motivation: The LLM landscape has millions of models with undocumented evolutionary relationships through fine-tuning, distillation, or adaptation, complicating model management. Existing methods are limited by task specificity, fixed model sets, or architectural assumptions.Method: Define LLM DNA as a low-dimensional, bi-Lipschitz representation of functional behavior. Prove inheritance and genetic determinism properties, then derive a general, scalable, training-free pipeline for DNA extraction. Apply phylogenetic algorithms to construct evolutionary trees.
Result: DNA aligns with prior studies on limited subsets, achieves competitive performance on specific tasks, uncovers undocumented relationships among LLMs, and constructs evolutionary trees showing architectural shifts, temporal progression, and varying evolutionary speeds.
Conclusion: LLM DNA provides a principled framework for understanding LLM evolution, enabling better model management, relationship discovery, and evolutionary analysis across the LLM ecosystem.
Abstract: The explosive growth of large language models (LLMs) has created a vast but opaque landscape: millions of models exist, yet their evolutionary relationships through fine-tuning, distillation, or adaptation are often undocumented or unclear, complicating LLM management. Existing methods are limited by task specificity, fixed model sets, or strict assumptions about tokenizers or architectures. Inspired by biological DNA, we address these limitations by mathematically defining LLM DNA as a low-dimensional, bi-Lipschitz representation of functional behavior. We prove that LLM DNA satisfies inheritance and genetic determinism properties and establish the existence of DNA. Building on this theory, we derive a general, scalable, training-free pipeline for DNA extraction. In experiments across 305 LLMs, DNA aligns with prior studies on limited subsets and achieves superior or competitive performance on specific tasks. Beyond these tasks, DNA comparisons uncover previously undocumented relationships among LLMs. We further construct the evolutionary tree of LLMs using phylogenetic algorithms, which align with shifts from encoder-decoder to decoder-only architectures, reflect temporal progression, and reveal distinct evolutionary speeds across LLM families.
[776] OpenTSLM: Time-Series Language Models for Reasoning over Multivariate Medical Text- and Time-Series Data
Patrick Langer, Thomas Kaar, Max Rosenblattl, Maxwell A. Xu, Winnie Chow, Martin Maritsch, Robert Jakob, Ning Wang, Juncheng Liu, Aradhana Verma, Brian Han, Daniel Seung Kim, Henry Chubb, Scott Ceresnak, Aydin Zahedivash, Alexander Tarlochan Singh Sandhu, Fatima Rodriguez, Daniel McDuff, Elgar Fleisch, Oliver Aalami, Filipe Barata, Paul Schmiedmayer
Main category: cs.LG
TL;DR: OpenTSLM integrates time series as a native modality into pretrained LLMs, enabling multimodal reasoning over time series data through two architectures: SoftPrompt (parameter-efficient) and Flamingo (cross-attention based), outperforming text-only models and GPT-4o on clinical time series tasks.
Details
Motivation: LLMs are powerful for multimodal data interpretation but lack native time series handling capabilities, which is crucial for medical applications where synthesizing clinical time series data into actionable insights is needed.Method: Two architectures: 1) OpenTSLM-SoftPrompt uses learnable time series tokens concatenated with text via soft prompting; 2) OpenTSLM-Flamingo integrates time series with text via cross-attention. Both enable Chain-of-Thought reasoning over time series of any length.
Result: OpenTSLM models outperform baselines across three datasets (HAR-CoT, Sleep-CoT, ECG-QA-CoT), achieving 69.9 F1 in sleep staging and 65.4 in HAR vs. 9.05 and 52.2 for text-only models. Even 1B-parameter models surpass GPT-4o. Flamingo matches SoftPrompt performance while handling longer sequences with stable memory requirements.
Conclusion: OpenTSLM successfully integrates time series as a native LLM modality, enabling effective multimodal reasoning over clinical time series data. The Flamingo architecture scales better with sequence length while maintaining performance. The approach shows strong clinical reasoning capabilities and is released open-source.
Abstract: LLMs have emerged as powerful tools for interpreting multimodal data. In medicine, they hold particular promise for synthesizing large volumes of clinical information into actionable insights and digital health applications. Yet, a major limitation remains their inability to handle time series. To overcome this gap, we present OpenTSLM, a family of Time Series Language Models (TSLMs) created by integrating time series as a native modality to pretrained LLMs, enabling reasoning over multiple time series of any length. We investigate two architectures for OpenTSLM. The first, OpenTSLM-SoftPrompt, models time series implicitly by concatenating learnable time series tokens with text tokens via soft prompting. Although parameter-efficient, we hypothesize that explicit time series modeling scales better and outperforms implicit approaches. We thus introduce OpenTSLM-Flamingo, which integrates time series with text via cross-attention. We benchmark both variants against baselines that treat time series as text tokens or plots, across a suite of text-time-series Chain-of-Thought (CoT) reasoning tasks. We introduce three datasets: HAR-CoT, Sleep-CoT, and ECG-QA-CoT. Across all, OpenTSLM models outperform baselines, reaching 69.9 F1 in sleep staging and 65.4 in HAR, compared to 9.05 and 52.2 for finetuned text-only models. Notably, even 1B-parameter OpenTSLM models surpass GPT-4o (15.47 and 2.95). OpenTSLM-Flamingo matches OpenTSLM-SoftPrompt in performance and outperforms on longer sequences, while maintaining stable memory requirements. By contrast, SoftPrompt grows exponentially in memory with sequence length, requiring around 110 GB compared to 40 GB VRAM when training on ECG-QA with LLaMA-3B. Expert reviews by clinicians find strong reasoning capabilities exhibited by OpenTSLMs on ECG-QA. To facilitate further research, we provide all code, datasets, and models open-source.
[777] Where to Add PDE Diffusion in Transformers
Yukun Zhang, Xueqing Zhou
Main category: cs.LG
TL;DR: The paper studies where to insert local diffusion layers relative to attention in transformer architectures, showing diffusion and attention don’t commute and early diffusion acts as effective pre-regularization while post-attention diffusion degrades performance.
Details
Motivation: Transformers lack explicit local geometric priors along sequence axes, and the placement of locality-inducing modules in hybrid architectures has been largely empirical. The paper aims to provide theoretical understanding of where to insert local diffusion operations relative to attention mechanisms.Method: Develops a three-layer operator-theoretic framework: 1) establishes unconditional guarantees for diffusion subsystem (spectral non-expansiveness, monotone Dirichlet-energy dissipation), 2) derives compositional perturbation bounds linking insertion effects to representation roughness and downstream amplification, 3) uses diffusion-attention non-commutativity as diagnostic for structural conflicts. Evaluates seven insertion positions on Long Range Arena benchmark.
Result: Early diffusion after embedding improves average accuracy by 4.1 percentage points, acting as effective pre-regularization. Post-attention diffusion degrades performance by 2.5 percentage points, consistent with predicted conflict. Multi-scale diffusion variant yields consistent gains under same global stability constraint.
Conclusion: The analysis provides a general template for reasoning about local-global compositions in sequence models by separating provable guarantees, compositional bounds, and mechanistic diagnostics. Diffusion-attention non-commutativity serves as diagnostic for structural conflicts.
Abstract: Transformers enable powerful content-based global routing via self-attention, but they lack an explicit local geometric prior along the sequence axis. As a result, the placement of locality-inducing modules in hybrid architectures has largely been empirical. We study a simple deterministic PDE diffusion layer implemented as one explicit Euler step of one-dimensional heat smoothing using a discrete Neumann Laplacian under a spectral stability constraint, and ask a structural question: where should diffusion be inserted relative to attention? Our central claim is that diffusion and attention generally do not commute, so inserting the same local operator before versus after attention leads to qualitatively different behaviors. We develop a three-layer operator-theoretic framework that (1) establishes unconditional guarantees for the diffusion subsystem, including spectral non-expansiveness and monotone Dirichlet-energy dissipation when the diffusion step size is smaller than one half, (2) derives compositional perturbation bounds linking insertion effects to representation roughness and downstream amplification, and (3) uses diffusion-attention non-commutativity as a diagnostic for structural double-mixing conflicts. Guided by theory, we evaluate seven insertion positions on the Long Range Arena benchmark. Early diffusion acts as effective pre-regularization, improving average accuracy by 4.1 percentage points when applied after embedding, while post-attention diffusion degrades performance by 2.5 percentage points, consistent with the predicted conflict. A multi-scale diffusion variant yields consistent gains under the same global stability constraint. Our analysis provides a general template for reasoning about local-global compositions in sequence models by separating provable guarantees, compositional bounds, and mechanistic diagnostics.
[778] Multi-scale Autoregressive Models are Laplacian, Discrete, and Latent Diffusion Models in Disguise
Steve Hong, Samuel Belkadi
Main category: cs.LG
TL;DR: VAR models reinterpreted as iterative refinement with Laplacian pyramid structure, revealing key design choices for quality-efficiency trade-off and connecting to diffusion methods.
Details
Motivation: To understand what design choices drive the quality-efficiency trade-off in Visual Autoregressive (VAR) models by reinterpreting them as iterative refinement models rather than just next-scale autoregression.Method: Formalize VAR as a deterministic forward process building a Laplacian-style latent pyramid with a learned backward process that reconstructs samples in coarse-to-fine steps. Identify three key modeling choices: refinement in learned latent space, discrete prediction over code indices, and decomposition by spatial frequency.
Result: Controlled experiments isolate contributions of each factor to quality and speed. The framework can be adapted to permutation-invariant graph generation and probabilistic medium-range weather forecasting, providing practical connections to diffusion methods while preserving few-step, scale-parallel generation.
Conclusion: VAR’s efficiency and sample quality stem from specific architectural choices that can be systematically analyzed, creating bridges to diffusion methods while maintaining efficient generation properties.
Abstract: We reinterpret Visual Autoregressive (VAR) models as iterative refinement models to identify which design choices drive their quality-efficiency trade-off. Instead of treating VAR only as next-scale autoregression, we formalise it as a deterministic forward process that builds a Laplacian-style latent pyramid, together with a learned backward process that reconstructs samples in a small number of coarse-to-fine steps. This formulation makes the link to denoising diffusion explicit and highlights three modelling choices that may underlie VAR’s efficiency and sample quality: refinement in a learned latent space, discrete prediction over code indices, and decomposition by spatial frequency. We support this view with controlled experiments that isolate the contribution of each factor to quality and speed. We also discuss how the same framework can be adapted to permutation-invariant graph generation and probabilistic medium-range weather forecasting, and how it provides practical points of contact with diffusion methods while preserving few-step, scale-parallel generation.
[779] RACE Attention: A Strictly Linear-Time Attention for Long-Sequence Training
Sahil Joshi, Agniva Chowdhury, Amar Kanakamedala, Ekam Singh, Evan Tu, Anshumali Shrivastava
Main category: cs.LG
TL;DR: RACE Attention: A linear-time alternative to quadratic Softmax Attention using angular similarity, Gaussian random projections, and soft LSH for efficient long-context processing up to 75M tokens.
Details
Motivation: Softmax Attention's quadratic complexity becomes prohibitive for long contexts, with current optimized implementations (FlashAttention-2/3) limited to ~4M tokens on high-end GPUs, creating a bottleneck for long-context training.Method: Replaces exponential kernel with sharpened angular similarity, approximates attention outputs via Gaussian random projections and soft Locality-Sensitive Hashing (LSH), avoiding full attention matrix construction for strictly linear complexity in sequence length and embedding size.
Result: Matches or outperforms strong baselines up to 64K sequence length across language modeling, masked language modeling, and text/image classification while reducing wall-clock time and memory usage; processes up to 12M tokens on NVIDIA GH200 GPU and 75M tokens on CPU in single forward-backward pass.
Conclusion: RACE Attention provides practical, theoretically grounded mechanism for long-context training on current hardware, significantly extending processing capabilities beyond state-of-the-art attention implementations.
Abstract: Softmax Attention has a quadratic time complexity in sequence length, which becomes prohibitive to run at long contexts, even with highly optimized GPU kernels. For example, FlashAttention-2/3 (exact, GPU-optimized implementations of Softmax Attention) cannot complete a single forward-backward pass of a single attention layer once the context exceeds ~4 million tokens on an NVIDIA GH200 (96 GB). We introduce Repeated Arrays-of-Count Estimators (RACE) Attention, a kernel-inspired alternative to Softmax Attention that is strictly linear in sequence length and embedding size. RACE Attention replaces the exponential kernel with a sharpened angular similarity, and approximates attention outputs via Gaussian random projections and soft Locality-Sensitive Hashing (LSH), avoiding construction of the full attention matrix. Across language modeling, masked language modeling, and text/image classification, RACE Attention matches or outperforms strong baselines up to 64K seqeuence length while reducing wall-clock time and memory usage. In addition, we conduct a controlled scaling study on a single attention layer and demonstrate processing of up to 12 million tokens on an NVIDIA GH200 GPU and 75 million tokens on an Intel Xeon Gold 5220R CPU in a single forward-backward pass, which is well beyond the capabilities of current state-of-the-art attention implementations. RACE Attention thus offers a practical and theoretically grounded mechanism for long-context training on today’s hardware. We release our code at https://github.com/sahiljoshi515/RACE_Attention.
[780] Lightweight Transformer for EEG Classification via Balanced Signed Graph Algorithm Unrolling
Junyi Yao, Parham Eftekhar, Gene Cheung, Xujin Chris Liu, Yao Wang, Wei Hu
Main category: cs.LG
TL;DR: Lightweight transformer-like neural nets for EEG epilepsy classification using spectral denoising on balanced signed graphs
Details
Motivation: EEG signals have inherent anti-correlations modeled by negative edges in graphs. Need lightweight, interpretable models to differentiate epilepsy patients from healthy subjects using EEG signals.Method: Build transformer-like neural nets by unrolling spectral denoising algorithm for signals on balanced signed graphs. Use similarity transform to map to positive graphs, implement ideal low-pass filter via Lanczos approximation with learned cutoff frequency. Use two balanced signed graph denoisers to learn posterior probabilities for binary classification.
Result: Achieves classification performance comparable to representative deep learning schemes while using dramatically fewer parameters.
Conclusion: Proposed method provides lightweight, interpretable alternative for EEG-based epilepsy classification using graph signal processing techniques.
Abstract: Samples of brain signals collected by EEG sensors have inherent anti-correlations that are well modeled by negative edges in a finite graph. To differentiate epilepsy patients from healthy subjects using collected EEG signals, we build lightweight and interpretable transformer-like neural nets by unrolling a spectral denoising algorithm for signals on a balanced signed graph – graph with no cycles of odd number of negative edges. A balanced signed graph has well-defined frequencies that map to a corresponding positive graph via similarity transform of the graph Laplacian matrices. We implement an ideal low-pass filter efficiently on the mapped positive graph via Lanczos approximation, where the optimal cutoff frequency is learned from data. Given that two balanced signed graph denoisers learn posterior probabilities of two different signal classes during training, we evaluate their reconstruction errors for binary classification of EEG signals. Experiments show that our method achieves classification performance comparable to representative deep learning schemes, while employing dramatically fewer parameters.
[781] Dual Goal Representations
Seohong Park, Deepinder Mann, Sergey Levine
Main category: cs.LG
TL;DR: Dual goal representations for GCRL encode states by their temporal distances to all other states, providing dynamics-invariant representations that improve goal-reaching performance.
Details
Motivation: Goal-conditioned RL needs effective state representations that capture relationships between states for optimal goal-reaching. Existing methods may be sensitive to state representation choices or miss important relational information.Method: Proposes dual goal representations that characterize states by the set of temporal distances to all other states. Develops a practical learning method that can be combined with any GCRL algorithm, focusing on capturing intrinsic dynamics relationships.
Result: Empirical evaluation on OGBench task suite shows consistent improvement in offline goal-reaching performance across 20 state- and pixel-based tasks compared to baseline methods.
Conclusion: Dual goal representations provide theoretically grounded, dynamics-invariant state encodings that enhance GCRL performance by capturing essential relational information between states.
Abstract: In this work, we introduce dual goal representations for goal-conditioned reinforcement learning (GCRL). A dual goal representation characterizes a state by “the set of temporal distances from all other states”; in other words, it encodes a state through its relations to every other state, measured by temporal distance. This representation provides several appealing theoretical properties. First, it depends only on the intrinsic dynamics of the environment and is invariant to the original state representation. Second, it contains provably sufficient information to recover an optimal goal-reaching policy, while being able to filter out exogenous noise. Based on this concept, we develop a practical goal representation learning method that can be combined with any existing GCRL algorithm. Through diverse experiments on the OGBench task suite, we empirically show that dual goal representations consistently improve offline goal-reaching performance across 20 state- and pixel-based tasks.
[782] Discrete State Diffusion Models: A Sample Complexity Perspective
Aadithya Srikanth, Mudit Gaur, Vaneet Aggarwal
Main category: cs.LG
TL;DR: Theoretical analysis of discrete-state diffusion models with first sample complexity bound of Õ(ε⁻²), addressing gap in understanding discrete diffusion for text/sequences.
Details
Motivation: Discrete-state diffusion models are crucial for text, sequences, and combinatorial structures but lack theoretical understanding compared to continuous-state models. Existing analyses assume score estimation error bounds without studying sample complexity.Method: Develops principled theoretical framework for discrete-state diffusion with structured decomposition of score estimation error into statistical, approximation, optimization, and clipping components.
Result: Provides first sample complexity bound of Õ(ε⁻²) for discrete-state diffusion models, establishing theoretical tractability and practical relevance.
Conclusion: Addresses fundamental gap in literature by providing theoretical foundation for discrete-state diffusion models, enabling more efficient training and broader applications in text and sequence domains.
Abstract: Diffusion models have demonstrated remarkable performance in generating high-dimensional samples across domains such as vision, language, and the sciences. Although continuous-state diffusion models have been extensively studied both empirically and theoretically, discrete-state diffusion models, essential for applications involving text, sequences, and combinatorial structures, remain significantly less understood from a theoretical standpoint. In particular, all existing analyses of discrete-state models assume score estimation error bounds without studying sample complexity results. In this work, we present a principled theoretical framework for discrete-state diffusion, providing the first sample complexity bound of $\widetilde{\mathcal{O}}(ε^{-2})$. Our structured decomposition of the score estimation error into statistical, approximation, optimization, and clipping components offers critical insights into how discrete-state models can be trained efficiently. This analysis addresses a fundamental gap in the literature and establishes the theoretical tractability and practical relevance of discrete-state diffusion models.
[783] Bridged Clustering: Semi-Supervised Sparse Bridging
Patrick Peixuan Ye, Chen Shani, Ellen Vitercik
Main category: cs.LG
TL;DR: Bridged Clustering is a semi-supervised learning framework that learns sparse, interpretable bridges between independently clustered input and output datasets using minimal paired examples.
Details
Motivation: The paper addresses the challenge of learning predictors from unpaired input and output datasets with limited supervision. Traditional semi-supervised learning doesn't leverage output-only data effectively, while dense transport-based methods lack interpretability and sparsity.Method: 1. Independently cluster input X and output Y datasets. 2. Learn a sparse, interpretable bridge between clusters using only a few paired examples. 3. At inference: assign new input x to nearest input cluster, return centroid of linked output cluster as prediction.
Result: Theoretical analysis shows the algorithm becomes effective with bounded mis-clustering and mis-bridging rates. Empirically competitive with state-of-the-art methods while being simple, model-agnostic, and highly label-efficient in low-supervision settings.
Conclusion: Bridged Clustering provides an effective semi-supervised framework that leverages output-only data, maintains interpretable sparse alignments, and achieves competitive performance with minimal supervision.
Abstract: We introduce Bridged Clustering, a semi-supervised framework to learn predictors from any unpaired input $X$ and output $Y$ dataset. Our method first clusters $X$ and $Y$ independently, then learns a sparse, interpretable bridge between clusters using only a few paired examples. At inference, a new input $x$ is assigned to its nearest input cluster, and the centroid of the linked output cluster is returned as the prediction $\hat{y}$. Unlike traditional SSL, Bridged Clustering explicitly leverages output-only data, and unlike dense transport-based methods, it maintains a sparse and interpretable alignment. Through theoretical analysis, we show that with bounded mis-clustering and mis-bridging rates, our algorithm becomes an effective and efficient predictor. Empirically, our method is competitive with SOTA methods while remaining simple, model-agnostic, and highly label-efficient in low-supervision settings.
[784] Challenges and Requirements for Benchmarking Time Series Foundation Models
Marcel Meyer, Sascha Kaltenpoth, Kevin Zalipski, Oliver Müller
Main category: cs.LG
TL;DR: The paper identifies information leakage issues in Time Series Foundation Model evaluation, similar to LLM problems, and calls for better evaluation methodologies.
Details
Motivation: Time Series Foundation Models (TSFMs) promise zero-shot forecasting but face evaluation challenges similar to LLMs, where information leakage from training data can lead to overly optimistic performance estimates that don't generalize to real-world settings.Method: The authors investigate existing TSFM evaluation studies and identify two types of information leakage: 1) train-test sample overlaps from multi-purpose dataset reuse, and 2) temporal overlap of correlated train and test series.
Result: The investigation reveals that ignoring these information leakage issues risks producing misleading performance estimates for TSFMs that fail to generalize to practical applications.
Conclusion: The paper argues for developing novel evaluation methodologies that avoid pitfalls observed in both LLM and classical time-series benchmarking, and calls for the research community to adopt principled approaches to safeguard TSFM evaluation integrity.
Abstract: Time Series Foundation Models (TSFMs) represent a new paradigm for time-series forecasting, promising zero-shot predictions without the need for task-specific training or fine-tuning. However, similar to Large Language Models (LLMs), the evaluation of TSFMs is challenging: as training corpora grow increasingly large, it becomes difficult to ensure the integrity of the test sets used for benchmarking. Our investigation of existing TSFM evaluation studies identifies two kinds of information leakage: (1) train-test sample overlaps arising from the multi-purpose reuse of datasets and (2) temporal overlap of correlated train and test series. Ignoring these forms of information leakage when benchmarking TSFMs risks producing overly optimistic performance estimates that fail to generalize to real-world settings. We therefore argue for the development of novel evaluation methodologies that avoid pitfalls already observed in both LLM and classical time-series benchmarking, and we call on the research community to adopt principled approaches to safeguard the integrity of TSFM evaluation.
[785] Model-agnostic Selective Labeling with Provable Statistical Guarantees
Huipeng Huang, Wenbo Liao, Huajun Xi, Hao Zeng, Mengchen Zhao, Hongxin Wei
Main category: cs.LG
TL;DR: Conformal Labeling is a method that uses conformal prediction to identify which AI-generated labels can be provably trusted by controlling the false discovery rate, ensuring a predefined fraction of AI-assigned labels is correct.
Details
Motivation: AI models offer cost-effective labeling but produce unreliable labels with errors. Existing selective labeling methods lack theoretical guarantees on AI label quality, leading to unacceptable error rates in AI-labeled subsets.Method: Construct conformal p-values for each test instance by comparing AI model confidence scores to calibration instances mislabeled by AI. Select instances with p-values below a data-dependent threshold to certify trustworthy AI predictions while controlling false discovery rate.
Result: Extensive experiments show the method achieves tight FDR control with high power across various tasks including image and text labeling, and LLM question answering.
Conclusion: Conformal Labeling provides theoretical guarantees for AI label trustworthiness with FDR control, enabling reliable selective labeling where AI predictions can be provably trusted.
Abstract: Obtaining high-quality labels for large datasets is expensive, requiring massive annotations from human experts. While AI models offer a cost-effective alternative by predicting labels, their label quality is compromised by the unavoidable labeling errors. Existing methods mitigate this issue through selective labeling, where AI labels a subset and human labels the remainder. However, these methods lack theoretical guarantees on the quality of AI-assigned labels, often resulting in unacceptably high labeling error within the AI-labeled subset. To address this, we introduce \textbf{Conformal Labeling}, a novel method to identify instances where AI predictions can be provably trusted. This is achieved by controlling the false discovery rate (FDR), the proportion of incorrect labels within the selected subset. In particular, we construct a conformal $p$-value for each test instance by comparing AI models’ predicted confidence to those of calibration instances mislabeled by AI models. Then, we select test instances whose $p$-values are below a data-dependent threshold, certifying AI models’ predictions as trustworthy. We provide theoretical guarantees that Conformal Labeling controls the FDR below the nominal level, ensuring that a predefined fraction of AI-assigned labels is correct on average. Extensive experiments demonstrate that our method achieves tight FDR control with high power across various tasks, including image and text labeling, and LLM QA.
[786] Algorithmic Primitives and Compositional Geometry of Reasoning in Language Models
Samuel Lippl, Thomas McGee, Kimberly Lopez, Ziwen Pan, Pierce Zhang, Salma Ziadi, Oliver Eberle, Ida Momennejad
Main category: cs.LG
TL;DR: A framework for analyzing algorithmic primitives in LLM reasoning through activation clustering and function vectors, showing compositional geometry and cross-task transferability.
Details
Motivation: To understand how latent and inference time computations enable LLMs to solve multi-step reasoning problems by identifying and analyzing algorithmic primitives that underlie model reasoning.Method: Links reasoning traces to internal activations, clusters activations to identify primitives, uses automated LLM pipeline to annotate reasoning traces, applies function vector methods to derive primitive vectors as reusable building blocks, and evaluates through injection experiments.
Result: Reveals compositional geometry in activation space where primitive vectors can be combined through arithmetic operations. Cross-task and cross-model evaluations show both shared and task-specific primitives. Reasoning-finetuned models exhibit more systematic use of verification and path-generation primitives.
Conclusion: LLM reasoning is supported by a compositional geometry of algorithmic primitives that transfer cross-task and cross-model, and reasoning finetuning strengthens algorithmic generalization across domains.
Abstract: How do latent and inference time computations enable large language models (LLMs) to solve multi-step reasoning? We introduce a framework for tracing and steering algorithmic primitives that underlie model reasoning. Our approach links reasoning traces to internal activations and evaluates algorithmic primitives by injecting them into residual streams and measuring their effect on reasoning steps and task performance. We consider four benchmarks: Traveling Salesperson Problem (TSP), 3SAT, AIME, and graph navigation. We operationalize primitives by clustering activations and annotating their matched reasoning traces using an automated LLM pipeline. We then apply function vector methods to derive primitive vectors as reusable compositional building blocks of reasoning. Primitive vectors can be combined through addition, subtraction, and scalar operations, revealing a geometric logic in activation space. Cross-task and cross-model evaluations (Phi-4, Phi-4-Reasoning, Llama-3-8B) show both shared and task-specific primitives. Notably, comparing Phi-4 with its reasoning-finetuned variant highlights compositional generalization after finetuning: Phi-4-Reasoning exhibits more systematic use of verification and path-generation primitives. Injecting the associated primitive vectors in Phi-4 induces behavioral hallmarks associated with Phi-4-Reasoning. Together, these findings demonstrate that reasoning in LLMs may be supported by a compositional geometry of algorithmic primitives, that primitives transfer cross-task and cross-model, and that reasoning finetuning strengthens algorithmic generalization across domains.
[787] Beyond Static Cutoffs: One-Shot Dynamic Thresholding for Diffusion Language Models
Jucheng Shen, Yeonju Ro
Main category: cs.LG
TL;DR: OSDT accelerates masked diffusion language model decoding by using one-shot dynamic thresholding based on reusable confidence patterns across inputs.
Details
Motivation: Current masked diffusion language models have inefficient decoding due to fixed steps and sequential unmasking. Recent parallel decoding methods use static thresholds but suffer from confidence fluctuations and similar confidence trajectories across inputs.Method: One-Shot Dynamic Thresholding (OSDT) calibrates thresholds on a single sequence and applies them to subsequent inputs with minimal overhead, leveraging observed reusable confidence patterns across similar inputs.
Result: OSDT achieves superior accuracy-throughput trade-offs: +24% tokens/s on GSM8K at best accuracy, +45% on GPQA with comparable accuracy, and +50% on HumanEval with modest accuracy gap.
Conclusion: The method demonstrates reusable task-level confidence signatures can enable algorithmic and systems innovations for more efficient diffusion decoding beyond current applications.
Abstract: Masked diffusion language models (MDLMs) are becoming competitive with their autoregressive counterparts but typically decode with fixed steps and sequential unmasking. To accelerate decoding, recent work such as Fast-dLLM enables parallel decoding via a static global confidence threshold, yet we observe strong block- and step-wise confidence fluctuations and, within a dataset, near-identical confidence trajectories across inputs as measured by cosine similarity. Motivated by these observations, we introduce One-Shot Dynamic Thresholding (OSDT), which calibrates thresholds on a single sequence and applies them to subsequent inputs with negligible overhead. On GPQA, GSM8K, and HumanEval, OSDT attains superior accuracy-throughput trade-offs (+24% tokens/s on GSM8K at the best accuracy, +45% on GPQA with comparable accuracy, and +50% on HumanEval with a modest accuracy gap). Beyond these results, our findings suggest broader opportunities to leverage reusable task-level confidence signatures for more general-purpose algorithmic and systems innovations in diffusion decoding.
[788] On the Mechanisms of Collaborative Learning in VAE Recommenders
Tung-Long Vuong, Julien Monteil, Hien Dang, Volodymyr Vaskovych, Trung Le, Vu Nguyen
Main category: cs.LG
TL;DR: Theoretical analysis of how collaboration emerges in VAE-based recommendation systems through latent proximity, with proposed anchor regularization to balance local and global collaboration while preserving user identity.
Details
Motivation: To understand the theoretical foundations of collaboration in VAE-based collaborative filtering, particularly how binary input masking affects performance and the balance between local collaboration (similar users) and global collaboration (distant but related users).Method: Theoretical analysis of latent proximity and collaboration mechanisms, derivation of latent sharing radius, comparison of β-KL regularization vs. input masking, and proposal of anchor regularization that aligns user posteriors with item embeddings to stabilize users under masking.
Result: Validated analyses on Netflix, MovieLens-20M, and Million Song datasets, with successful deployment on an Amazon streaming platform following positive online experiments.
Conclusion: VAE-based CF primarily exploits local collaboration, but global collaboration can be enhanced through careful regularization techniques like anchor regularization that balance information sharing with user identity preservation.
Abstract: Variational Autoencoders (VAEs) are a powerful alternative to matrix factorization for recommendation. A common technique in VAE-based collaborative filtering (CF) consists in applying binary input masking to user interaction vectors, which improves performance but remains underexplored theoretically. In this work, we analyze how collaboration arises in VAE-based CF and show it is governed by \emph{latent proximity}: we derive a latent sharing radius that informs when an SGD update on one user strictly reduces the loss on another user, with influence decaying as the latent Wasserstein distance increases. We further study the induced geometry: with clean inputs, VAE-based CF primarily exploits \emph{local} collaboration between input-similar users and under-utilizes \emph{global} collaboration between far-but-related users. We compare two mechanisms that encourage \emph{global} mixing and characterize their trade-offs: \ding{172} $β$-KL regularization directly tightens the information bottleneck, promoting posterior overlap but risking representational collapse if too large; \ding{173} input masking induces stochastic \emph{geometric} contractions and expansions, which can bring distant users onto the same latent neighborhood but also introduce neighborhood drift. To preserve user identity while enabling global consistency, we propose an anchor regularizer that aligns user posteriors with item embeddings, stabilizing users under masking and facilitating signal sharing across related items. Our analyses are validated on the Netflix, MovieLens-20M, and Million Song datasets. We also successfully deployed our proposed algorithm on an Amazon streaming platform following a successful online experiment.
[789] MURPHY: Multi-Turn GRPO for Self Correcting Code Generation
Chanakya Ekbote, Vijay Lingam, Sujay Sanghavi, Jun Huan, Behrooz Omidvar-Tehrani, Anoop Deoras, Stefano Soatto
Main category: cs.LG
TL;DR: MURPHY is a multi-turn reinforcement learning framework that extends GRPO to optimize over iterative decision-making trajectories by incorporating execution feedback, improving performance on agentic tasks like code generation.
Details
Motivation: Existing RLVR approaches like GRPO work well on reasoning benchmarks but struggle with agentic tasks requiring iterative decision-making, as they don't effectively incorporate execution feedback during training.Method: Extends GRPO to multi-turn trajectories with execution feedback, using feedback-conditioned rollout trees, trajectory-level credit assignment, and pruning to reduce optimization costs.
Result: Achieves up to 8% absolute gain in pass@1 over compute-matched GRPO baselines on code generation benchmarks, outperforming prior methods that incorporate multi-turn execution feedback.
Conclusion: MURPHY effectively addresses limitations of existing RLVR methods for agentic tasks by incorporating execution feedback into multi-turn optimization, significantly improving iterative decision-making performance.
Abstract: Reinforcement Learning with Verifiable Rewards(RLVR) has emerged as a powerful framework for enhancing the reasoning capabilities of large language models (LLMs). However, existing approaches such as Group Relative Policy Optimization (GRPO) and its variants, while effective on reasoning benchmarks, struggle with agentic tasks that require iterative decision-making. We introduce MURPHY, a multi-turn RLVR framework that incorporates execution feedback directly into training, extending GRPO to optimize over multi-turn trajectories where models iteratively refine solutions. MURPHY combines a feedback conditioned rollout tree with trajectory-level credit assignment, and uses pruning to reduce the cost of multi-turn optimization. Evaluations on code generation benchmarks with two model families show that MURPHY consistently improves multi-iteration performance, achieving up to an 8% absolute gain in pass@1 over compute-matched GRPO baselines, and outperforming the prior leading method that incorporates multi-turn execution feedback.
[790] Evolution Strategies at the Hyperscale
Bidipta Sarkar, Mattie Fellows, Juan Agustin Duque, Alistair Letcher, Antonio León Villares, Anya Sims, Clarisse Wibault, Dmitry Samsonov, Dylan Cope, Jarek Liesen, Kang Li, Lukas Seier, Theo Wolf, Uljad Berdica, Valentin Mohl, Alexander David Goldie, Aaron Courville, Karin Sevegnani, Shimon Whiteson, Jakob Nicolaus Foerster
Main category: cs.LG
TL;DR: EGGROLL is a high-performance Evolution Strategies algorithm that structures perturbations as low-rank matrices for efficient large-scale optimization, enabling faster training of billion-parameter models while maintaining ES performance.
Details
Motivation: Standard Evolution Strategies (ES) become computationally expensive at scale on GPUs due to low arithmetic intensity from unstructured random perturbations in batched matrix multiplications, limiting their application to large models.Method: EGGROLL structures individual perturbations as rank-r matrices instead of unstructured random perturbations, improving arithmetic intensity and enabling hundredfold speed increases for billion-parameter models at large population sizes.
Result: EGGROLL achieves up to 91% of the throughput of pure batch inference, enables stable pretraining of nonlinear recurrent language models in integer datatypes, is competitive with GRPO for post-training LLMs on reasoning tasks, and maintains ES performance in RL settings.
Conclusion: EGGROLL provides a scalable, efficient alternative to traditional ES that maintains theoretical convergence properties while dramatically improving computational efficiency for large-scale optimization problems.
Abstract: Evolution Strategies (ES) is a class of powerful black-box optimisation methods that are highly parallelisable and can handle non-differentiable and noisy objectives. However, naïve ES becomes prohibitively expensive at scale on GPUs due to the low arithmetic intensity of batched matrix multiplications with unstructured random perturbations. We introduce Evolution Guided GeneRal Optimisation via Low-rank Learning (EGGROLL), which improves arithmetic intensity by structuring individual perturbations as rank-$r$ matrices, resulting in a hundredfold increase in training speed for billion-parameter models at large population sizes, achieving up to 91% of the throughput of pure batch inference. We provide a rigorous theoretical analysis of Gaussian ES for high-dimensional parameter objectives, investigating conditions needed for ES updates to converge in high dimensions. Our results reveal a linearising effect, and proving consistency between EGGROLL and ES as parameter dimension increases. Our experiments show that EGGROLL: (1) enables the stable pretraining of nonlinear recurrent language models that operate purely in integer datatypes, (2) is competitive with GRPO for post-training LLMs on reasoning tasks, and (3) does not compromise performance compared to ES in tabula rasa RL settings, despite being faster.
[791] ESPO: Entropy Importance Sampling Policy Optimization
Yuepeng Sheng, Yuwei Huang, Shuman Liu, Anxiang Zeng, Haibo Zhang
Main category: cs.LG
TL;DR: ESPO is a novel RL framework for LLM post-training that addresses the stability-efficiency trade-off by using entropy-based grouping and adaptive clipping for better gradient utilization.
Details
Motivation: Current RL methods for LLM post-training face a fundamental trade-off: token-level optimization provides fine-grained updates but suffers from high variance and instability, while sequence-level optimization uses aggressive clipping that discards many valid training samples, leading to inefficient gradient utilization.Method: ESPO (Entropy Importance Sampling Policy Optimization) decomposes sequences into groups based on predictive entropy, enabling: (1) Entropy Grouping Importance Sampling to capture intra-sequence heterogeneity, and (2) Entropy Adaptive Clipping to dynamically allocate trust regions based on model uncertainty.
Result: Extensive experiments on mathematical reasoning benchmarks show ESPO accelerates convergence and achieves state-of-the-art performance, notably improving accuracy on challenging mathematical benchmarks.
Conclusion: ESPO successfully addresses the stability-efficiency trade-off in RL for LLM post-training by combining fine-grained updates with stable training through entropy-based techniques, leading to better performance on complex reasoning tasks.
Abstract: Reinforcement learning (RL) has become a central component of post-training for large language models (LLMs), particularly for complex reasoning tasks that require stable optimization over long generation horizons. However, achieving performance at scale often introduces a fundamental trade-off between training stability and training efficiency. Token-level optimization applies fine-grained updates at the individual units, but is prone to high variance in gradient estimation, which can result in unstable training dynamics. In contrast, Sequence-level optimization often relies on aggressive clipping mechanisms to ensure stable updates. However, such design may discard a large fraction of valid training samples, leading to inefficient gradient utilization and reduced training efficiency. We refer to this phenomenon as gradient underutilization. In this work, we propose Entropy Importance Sampling Policy Optimization (ESPO), a novel framework that aims to combine fine-grained updates with stable training. ESPO decomposes sequences into groups based on predictive entropy, enabling (1) Entropy Grouping Importance Sampling to capture intra-sequence heterogeneity, and (2) Entropy Adaptive Clipping to dynamically allocate trust regions based on model uncertainty. Extensive experiments on mathematical reasoning benchmarks demonstrate that ESPO not only accelerates convergence but also achieves state-of-the-art performance, notably improving accuracy on the challenging mathematical benchmarks.
[792] Generative Anchored Fields: Controlled Data Generation via Emergent Velocity Fields and Transport Algebra
Deressa Wodajo Deressa, Hannes Mareen, Peter Lambert, Glenn Van Wallendael
Main category: cs.LG
TL;DR: GAF is a generative model that learns independent endpoint predictors for noise and data from any point on a linear bridge, enabling compositional control through algebraic operations on multiple heads.
Details
Motivation: Existing generative models use single trajectory or score predictors, limiting compositional control. The authors aim to create a model that enables controllable interpolation, multi-class composition, and semantic editing through architectural design.Method: GAF learns independent endpoint predictors J (noise) and K (data) from any point on a linear bridge. The velocity field emerges from their disagreement. With class-specific K_n heads, it defines directed transport maps between shared base noise and multiple data domains. Uses Iterative Endpoint Refinement (IER) sampler for efficient generation.
Result: Achieves strong sample quality with FID 7.51 on ImageNet 256×256 and 7.27 on CelebA-HQ 256×256 without classifier-free guidance. Enables compositional generation as architectural primitive with 5-8 step generation.
Conclusion: GAF provides a novel factorization approach for generative modeling that enables Transport Algebra for compositional control, achieving state-of-the-art results with efficient sampling.
Abstract: We present Generative Anchored Fields (GAF), a generative model that learns independent endpoint predictors, $J$ (noise) and $K$ (data), from any point on a linear bridge. Unlike existing approaches that use a single trajectory or score predictor, GAF is trained to recover the bridge endpoints directly via coordinate learning. The velocity field $v=K-J$ emerges from their time-conditioned disagreement. This factorization enables \textit{Transport Algebra}: algebraic operations on multiple $J/K$ heads for compositional control. With class-specific $K_n$ heads, GAF defines directed transport maps between a shared base noise distribution and multiple data domains, allowing controllable interpolation, multi-class composition, and semantic editing. This is achieved either directly on the predicted data coordinates ($K$) using Iterative Endpoint Refinement (IER), a novel sampler that achieves high-quality generation in $5-8$ steps, or on the emergent velocity field ($v$). We achieve strong sample quality (FID 7.51 on ImageNet $256\times256$ and $7.27$ on CelebA-HQ $256\times 256$, without classifier-free guidance) while treating compositional generation as an architectural primitive. Code available at https://github.com/IDLabMedia/GAF.
[793] Network Level Evaluation of Hangup Susceptibility of HRGCs using Deep Learning and Sensing Techniques: A Goal Towards Safer Future
Kaustav Chatterjee, Joshua Li, Kundan Parajulee, Jared Schwennesen
Main category: cs.LG
TL;DR: A framework for evaluating hang-up susceptibility at highway-rail grade crossings using deep learning for profile reconstruction and vehicle dimension analysis to identify high-risk crossings.
Details
Motivation: Steep-profiled highway-rail grade crossings pose safety hazards for low-clearance vehicles that can become stranded on tracks, creating collision risks with trains. Current methods lack comprehensive network-level evaluation.Method: Collected crossing profile data using walking profiler and Pave3D8K Laser Imaging System. Developed hybrid LSTM-Transformer deep learning model to reconstruct accurate crossing profiles. Collected dimension data from 350 specialty vehicles. Analyzed hang-up susceptibility using three vehicle dimension scenarios: median, 75-25 percentile, and worst-case dimensions.
Result: Identified 70, 80, and 95 crossings at highest hang-up risk levels under median, 75-25 percentile, and worst-case dimension scenarios respectively. Developed ArcGIS database and software interface for transportation agencies.
Conclusion: The framework integrates next-generation sensing, deep learning, and infrastructure data into practical decision support tools for transportation agencies to mitigate crossing hazards and advance safety evaluation.
Abstract: Steep-profiled Highway Railway Grade Crossings (HRGCs) pose safety hazards to vehicles with low ground clearance, which may become stranded on the tracks, creating risks of train vehicle collisions. This research develops a framework for network level evaluation of hang-up susceptibility of HRGCs. Profile data from different crossings in Oklahoma were collected using both a walking profiler and the Pave3D8K Laser Imaging System. A hybrid deep learning model, combining Long Short Term Memory (LSTM) and Transformer architectures, was developed to reconstruct accurate HRGC profiles from Pave3D8K Laser Imaging System data. Vehicle dimension data from around 350 specialty vehicles were collected at various locations across Oklahoma to enable up-to-date statistical design dimensions. Hang-up susceptibility was analyzed using three vehicle dimension scenarios: (a) median dimension (median wheelbase and ground clearance), (b) 75-25 percentile dimension (75 percentile wheelbase, 25 percentile ground clearance), and (c) worst case dimension (maximum wheelbase and minimum ground clearance). Results indicate 70, 80, and 95 crossings at the highest hang-up risk levels under these scenarios, respectively. An ArcGIS database and a software interface were developed to support transportation agencies in mitigating crossing hazards. This framework advances safety evaluation by integrating next-generation sensing, deep learning, and infrastructure datasets into practical decision support tools.
[794] Scaling Behavior of Discrete Diffusion Language Models
Dimitri von Rütte, Janis Fluri, Omead Pooladzandi, Bernhard Schölkopf, Thomas Hofmann, Antonio Orvieto
Main category: cs.LG
TL;DR: Discrete diffusion language models (DLMs) show different scaling behaviors than autoregressive models, with uniform diffusion requiring more parameters but less data for compute-efficient training, making them promising for data-limited scenarios.
Details
Motivation: To understand the scaling behavior of discrete diffusion language models compared to autoregressive models, particularly how different noise types (masked vs uniform) affect compute and data efficiency, since prior work suggested DLMs need more resources to match ALM performance.Method: Systematically study DLM scaling by interpolating between masked and uniform diffusion noise types while carefully controlling hyperparameters like batch size and learning rate. Scale uniform diffusion models up to 10B parameters trained for 10^22 FLOPs.
Result: DLM scaling behavior strongly depends on noise type and differs from ALMs. Uniform diffusion requires more parameters but less data for compute-efficient training compared to masked diffusion, making it promising for data-bound settings. The 10B parameter uniform diffusion model confirms predicted scaling behavior.
Conclusion: Uniform diffusion models offer a promising alternative to autoregressive models in data-limited scenarios due to their different scaling properties, requiring more parameters but less training data for efficient performance.
Abstract: Modern LLM pre-training consumes vast amounts of compute and training data, making the scaling behavior, or scaling laws, of different models a key distinguishing factor. Discrete diffusion language models (DLMs) have been proposed as an alternative to autoregressive language models (ALMs). However, their scaling behavior has not yet been fully explored, with prior work suggesting that they require more data and compute to match the performance of ALMs. We study the scaling behavior of DLMs on different noise types by smoothly interpolating between masked and uniform diffusion while paying close attention to crucial hyperparameters such as batch size and learning rate. Our experiments reveal that the scaling behavior of DLMs strongly depends on the noise type and is considerably different from ALMs. While all noise types converge to similar loss values in compute-bound scaling, we find that uniform diffusion requires more parameters and less data for compute-efficient training compared to masked diffusion, making them a promising candidate in data-bound settings. We scale our uniform diffusion model up to 10B parameters trained for $10^{22}$ FLOPs, confirming the predicted scaling behavior and making it the largest publicly known uniform diffusion model to date.
[795] ModSSC: A Modular Framework for Semi-Supervised Classification on Heterogeneous Data
Melvin Barbaux, Samia Boukir
Main category: cs.LG
TL;DR: ModSSC is an open-source Python framework for semi-supervised classification that provides modular components for reproducible experiments across different data modalities and learning settings.
Details
Motivation: Existing semi-supervised classification software is fragmented across methods, learning settings, and data modalities, making reproducible and controlled experimentation difficult.Method: Developed a modular Python framework with reusable components, stable abstractions, and declarative experiment specification through configuration files for systematic comparison across datasets and model backbones.
Result: Released ModSSC 1.0.0 under MIT license with full documentation and automated tests, validated through controlled experiments reproducing established semi-supervised learning baselines across multiple data modalities.
Conclusion: ModSSC provides a unified framework for reproducible semi-supervised classification experiments that supports heterogeneous datasets and model backbones without modifying algorithmic code.
Abstract: Semi-supervised classification leverages both labeled and unlabeled data to improve predictive performance, but existing software support remains fragmented across methods, learning settings, and data modalities. We introduce ModSSC, an open source Python framework for inductive and transductive semi-supervised classification designed to support reproducible and controlled experimentation. ModSSC provides a modular and extensible software architecture centered on reusable semi-supervised learning components, stable abstractions, and fully declarative experiment specification. Experiments are defined through configuration files, enabling systematic comparison across heterogeneous datasets and model backbones without modifying algorithmic code. ModSSC 1.0.0 is released under the MIT license with full documentation and automated tests, and is available at https://github.com/ModSSC/ModSSC. The framework is validated through controlled experiments reproducing established semi-supervised learning baselines across multiple data modalities.
[796] From GNNs to Symbolic Surrogates via Kolmogorov-Arnold Networks for Delay Prediction
Sami Marouani, Kamal Singh, Baptiste Jeudy, Amaury Habrard
Main category: cs.LG
TL;DR: FlowKANet: A graph neural network using Kolmogorov-Arnold Networks for flow delay prediction in communication networks, with symbolic distillation for lightweight deployment.
Details
Motivation: Accurate flow delay prediction is crucial for optimizing modern communication networks, requiring efficient and transparent models that can handle graph-structured data.Method: Three-level approach: 1) Heterogeneous GNN with attention-based message passing as baseline, 2) FlowKANet replacing MLP layers with KAN layers (KAMP-Attn), 3) Distillation into symbolic surrogate models using block-wise regression for closed-form equations.
Result: KAN layers provide favorable trade-off between efficiency and accuracy with reduced trainable parameters, and symbolic surrogates enable lightweight deployment while preserving graph-structured dependencies.
Conclusion: FlowKANet demonstrates KAN layers’ effectiveness for graph-based prediction tasks, and symbolic distillation offers potential for transparent, lightweight network optimization models.
Abstract: Accurate prediction of flow delay is essential for optimizing and managing modern communication networks. We investigate three levels of modeling for this task. First, we implement a heterogeneous GNN with attention-based message passing, establishing a strong neural baseline. Second, we propose FlowKANet in which Kolmogorov-Arnold Networks replace standard MLP layers, reducing trainable parameters while maintaining competitive predictive performance. FlowKANet integrates KAMP-Attn (Kolmogorov-Arnold Message Passing with Attention), embedding KAN operators directly into message-passing and attention computation. Finally, we distill the model into symbolic surrogate models using block-wise regression, producing closed-form equations that eliminate trainable weights while preserving graph-structured dependencies. The results show that KAN layers provide a favorable trade-off between efficiency and accuracy and that symbolic surrogates emphasize the potential for lightweight deployment and enhanced transparency.
[797] Critic-Guided Reinforcement Unlearning in Text-to-Image Diffusion
Mykola Vysotskyi, Zahar Kohut, Mariia Shpir, Taras Rumezhak, Volodymyr Karpiv
Main category: cs.LG
TL;DR: A reinforcement learning framework for machine unlearning in text-to-image diffusion models that uses timestep-aware critics and noisy-step rewards to remove targeted concepts while preserving overall image quality.
Details
Motivation: Existing diffusion unlearning methods rely on supervised weight edits or global penalties, while RL approaches suffer from high-variance updates and weak credit assignment due to sparse end-of-trajectory rewards.Method: Treats denoising as a sequential decision process, trains a CLIP-based reward predictor on noisy latents, uses per-step signal to compute advantage estimates for policy-gradient updates of the reverse diffusion kernel.
Result: Achieves better or comparable forgetting to strong baselines across multiple concepts while maintaining image quality and benign prompt fidelity; per-step critics and noisy-conditioned rewards are key to stability and effectiveness.
Conclusion: The RL framework is simple to implement, supports off-policy reuse, and plugs into standard text-to-image backbones, providing an effective approach for diffusion unlearning with better credit assignment.
Abstract: Machine unlearning in text-to-image diffusion models aims to remove targeted concepts while preserving overall utility. Prior diffusion unlearning methods typically rely on supervised weight edits or global penalties; reinforcement-learning (RL) approaches, while flexible, often optimize sparse end-of-trajectory rewards, yielding high-variance updates and weak credit assignment. We present a general RL framework for diffusion unlearning that treats denoising as a sequential decision process and introduces a timestep-aware critic with noisy-step rewards. Concretely, we train a CLIP-based reward predictor on noisy latents and use its per-step signal to compute advantage estimates for policy-gradient updates of the reverse diffusion kernel. Our algorithm is simple to implement, supports off-policy reuse, and plugs into standard text-to-image backbones. Across multiple concepts, the method achieves better or comparable forgetting to strong baselines while maintaining image quality and benign prompt fidelity; ablations show that (i) per-step critics and (ii) noisy-conditioned rewards are key to stability and effectiveness. We release code and evaluation scripts to facilitate reproducibility and future research on RL-based diffusion unlearning.
[798] Orthogonalized Policy Optimization:Decoupling Sampling Geometry from Optimization Geometry in RLHF
Wang Zixian
Main category: cs.LG
TL;DR: OPO is a new alignment objective that decouples sampling geometry from optimization geometry in LLM alignment, addressing systematic instability in existing methods.
Details
Motivation: Existing alignment methods (PPO, DPO, IPO) entangle sampling geometry and optimization geometry through a single divergence, causing systematic instability, especially in preference-based RL where advantage signals are unbounded.Method: Formulates alignment as orthogonal mirror descent, where sampling geometry uses alpha-divergence projection as linear driving force, while optimization geometry uses independent Bregman divergence (mirror map). OPO uses Euclidean mirror map in likelihood ratio space.
Result: OPO admits closed-form solution, linear/non-saturating gradient dynamics, well-conditioned trust region, and remains compatible with standard LLM training pipelines.
Conclusion: Decoupling sampling and optimization geometry provides principled structural remedy for alignment instability, with OPO offering practical benefits while maintaining compatibility with existing training infrastructure.
Abstract: Large language model alignment objectives are often presented as a collection of distinct algorithms, such as PPO, DPO, IPO, and their variants, each motivated by different derivations. In this work, we argue that this diversity obscures a simpler underlying structure. At a fundamental level, alignment objectives involve two independent design choices: (i) how training signals are sampled and weighted, and (ii) how deviations from a reference policy are geometrically penalized. Existing methods typically entangle these choices through a single divergence, most commonly the Kullback-Leibler divergence. We show that this entanglement is not merely a modeling convenience but a source of systematic instability. When the same divergence simultaneously determines sample weighting and optimization curvature, adjusting one aspect, such as exploration strength, inevitably alters the other, such as gradient geometry. This coupling is particularly problematic in preference-based reinforcement learning, where advantage signals are unbounded and high-confidence regimes are common. We propose a principled structural remedy by formulating alignment as an orthogonal mirror descent problem, in which sampling geometry enters as a linear driving force derived from an alpha-divergence projection, while optimization geometry is determined independently by a Bregman divergence, or mirror map. This perspective leads to a new alignment objective called Orthogonalized Policy Optimization (OPO), obtained by choosing a Euclidean mirror map in likelihood ratio space. The resulting objective admits a closed-form solution, linear and non-saturating gradient dynamics, and a well-conditioned trust region, while remaining fully compatible with standard large language model training pipelines.
[799] Parallelizable memory recurrent units
Florent De Geeter, Gaspard Lambrechts, Damien Ernst, Guillaume Drion
Main category: cs.LG
TL;DR: A new family of RNNs called Memory Recurrent Units (MRUs) combines persistent memory capabilities of nonlinear RNNs with parallelizable computations of state-space models, addressing limitations of both Transformers and SSMs.
Details
Motivation: Transformers enable parallel training but are inefficient at sequence generation, while state-space models (SSMs) offer efficient recurrent processing but lack persistent memory capabilities due to monostability. There's a need for models that combine parallelizable training with persistent memory retention.Method: Introduces Memory Recurrent Units (MRUs) that leverage multistability for persistent memory while eliminating transient dynamics for efficient computation. Presents a specific implementation called Bistable Memory Recurrent Unit (BMRU) that is compatible with parallel scan algorithm for parallel training.
Result: BMRU achieves good results in tasks with long-term dependencies and can be combined with SSMs to create hybrid networks that are both parallelizable and have both transient dynamics and persistent memory capabilities.
Conclusion: MRUs offer a promising direction for sequence models that combine the parallel training advantages of SSMs with the persistent memory capabilities of nonlinear RNNs, addressing key limitations of current approaches.
Abstract: With the emergence of massively parallel processing units, parallelization has become a desirable property for new sequence models. The ability to parallelize the processing of sequences with respect to the sequence length during training is one of the main factors behind the uprising of the Transformer architecture. However, Transformers lack efficiency at sequence generation, as they need to reprocess all past timesteps at every generation step. Recently, state-space models (SSMs) emerged as a more efficient alternative. These new kinds of recurrent neural networks (RNNs) keep the efficient update of the RNNs while gaining parallelization by getting rid of nonlinear dynamics (or recurrence). SSMs can reach state-of-the art performance through the efficient training of potentially very large networks, but still suffer from limited representation capabilities. In particular, SSMs cannot exhibit persistent memory, or the capacity of retaining information for an infinite duration, because of their monostability. In this paper, we introduce a new family of RNNs, the memory recurrent units (MRUs), that combine the persistent memory capabilities of nonlinear RNNs with the parallelizable computations of SSMs. These units leverage multistability as a source of persistent memory, while getting rid of transient dynamics for efficient computations. We then derive a specific implementation as proof-of-concept: the bistable memory recurrent unit (BMRU). This new RNN is compatible with the parallel scan algorithm. We show that BMRU achieves good results in tasks with long-term dependencies, and can be combined with state-space models to create hybrid networks that are parallelizable and have transient dynamics as well as persistent memory.
[800] GRIP: Algorithm-Agnostic Machine Unlearning for Mixture-of-Experts via Geometric Router Constraints
Andy Zhu, Rongzhe Wei, Yupu Gu, Pan Li
Main category: cs.LG
TL;DR: GRIP framework prevents machine unlearning methods from exploiting MoE router vulnerabilities by constraining router updates to preserve routing stability while allowing expert parameter unlearning.
Details
Motivation: Existing machine unlearning methods fail for Mixture-of-Experts (MoE) architectures because they exploit router vulnerabilities to superficially redirect queries rather than actually erase knowledge from experts, causing utility loss.Method: GRIP uses geometric constraints by projecting router gradient updates into expert-specific null-spaces, decoupling routing stability from parameter rigidity. This forces unlearning to erase knowledge from expert parameters rather than manipulate routers.
Result: GRIP achieves over 95% routing stability across tested unlearning methods while preserving utility, effectively adapting existing unlearning research from dense architectures to MoEs.
Conclusion: GRIP provides an algorithm-agnostic framework that prevents exploitation of MoE router vulnerabilities, enabling effective machine unlearning for MoE architectures while maintaining model utility.
Abstract: Machine unlearning (MU) for large language models has become critical for AI safety, yet existing methods fail to generalize to Mixture-of-Experts (MoE) architectures. We identify that traditional unlearning methods exploit MoE’s architectural vulnerability: they manipulate routers to redirect queries away from knowledgeable experts rather than erasing knowledge, causing a loss of model utility and superficial forgetting. We propose Geometric Routing Invariance Preservation (GRIP), an algorithm-agnostic framework for unlearning for MoE. Our core contribution is a geometric constraint, implemented by projecting router gradient updates into an expert-specific null-space. Crucially, this decouples routing stability from parameter rigidity: while discrete expert selections remain stable for retained knowledge, the continuous router parameters remain plastic within the null space, allowing the model to undergo necessary internal reconfiguration to satisfy unlearning objectives. This forces the unlearning optimization to erase knowledge directly from expert parameters rather than exploiting the superficial router manipulation shortcut. GRIP functions as an adapter, constraining router parameter updates without modifying the underlying unlearning algorithm. Extensive experiments on large-scale MoE models demonstrate that our adapter eliminates expert selection shift (achieving over 95% routing stability) across all tested unlearning methods while preserving their utility. By preventing existing algorithms from exploiting MoE model’s router vulnerability, GRIP adapts existing unlearning research from dense architectures to MoEs.
[801] LAViG-FLOW: Latent Autoregressive Video Generation for Fluid Flow Simulations
Vittoria De Pellegrini, Tariq Alkhalifah
Main category: cs.LG
TL;DR: LAViG-FLOW: A latent autoregressive video generation diffusion framework for modeling subsurface multiphase fluid flow fields (saturation and pressure) that runs 100x faster than traditional numerical solvers.
Details
Motivation: High-fidelity multiphase simulators for subsurface fluid flow (used in CO2 sequestration and geothermal applications) are computationally expensive for many forward runs needed in inversion and uncertainty quantification.Method: Latent autoregressive video generation diffusion framework with dedicated 2D autoencoders for each state variable (saturation, pressure) and Video Diffusion Transformer (VDiT) to model their coupled temporal evolution. Trained on time horizon then fine-tuned autoregressively for extrapolation.
Result: Generates consistent saturation and pressure fields across time while running two orders of magnitude (100x) faster than traditional numerical solvers on CO2 sequestration dataset.
Conclusion: LAViG-FLOW provides an efficient alternative to expensive numerical simulators for subsurface flow modeling, enabling faster uncertainty quantification and inversion tasks.
Abstract: Modeling and forecasting subsurface multiphase fluid flow fields underpin applications ranging from geological CO2 sequestration (GCS) operations to geothermal production. This is essential for ensuring both operational performance and long-term safety. While high fidelity multiphase simulators are widely used for this purpose, they become prohibitively expensive once many forward runs are required for inversion purposes and to quantify uncertainty. To tackle this challenge, we propose LAViG-FLOW, a latent autoregressive video generation diffusion framework that explicitly learns the coupled evolution of saturation and pressure fields. Each state variable is compressed by a dedicated 2D autoencoder, and a Video Diffusion Transformer (VDiT) models their coupled distribution across time. We first train the model on a given time horizon to learn their coupled relationship and then fine-tune it autoregressively so it can extrapolate beyond the observed time window. Evaluated on an open-source CO2 sequestration dataset, LAViG-FLOW generates saturation and pressure fields that stay consistent across time while running two orders of magnitude faster than traditional numerical solvers.
[802] From Fuzzy to Exact: The Halo Architecture for Infinite-Depth Reasoning via Rational Arithmetic
Hansheng Ren
Main category: cs.LG
TL;DR: Halo Architecture introduces exact rational arithmetic for LLMs using dual-ring topology to prevent numerical instability, enabling simplified transformer blocks and faster convergence.
Details
Motivation: Current LLMs rely on fuzzy floating-point arithmetic requiring complex numerical heuristics to prevent collapse, consuming significant compute. The paper proposes moving to exact arithmetic for stability and efficiency.Method: Halo Architecture uses rational field arithmetic with Exact Inference Unit and Dual-Ring Topology: Micro-Ring for memory complexity bounds via Diophantine Approximation, and Macro-Ring for logical consistency via periodic state collapse.
Result: Enables “Great Dismantling” of numerical scaffolding, reducing transformer blocks to clean algebraic form. Demonstrates “Efficiency Paradox” where elimination of gradient noise allows macro-learning rates, potentially reducing convergence time by orders of magnitude.
Conclusion: General intelligence requires hybridization of continuous fields and discrete chains under rigorous mathematical framework. Halo shows exact arithmetic can simplify LLM architecture and accelerate training.
Abstract: The prevailing scaling paradigm of Large Language Models (LLMs) rests on a substrate of “Fuzzy” floating-point arithmetic. To mitigate the inherent instability of this approximate foundation, modern architectures have erected a complex scaffolding of structural and numerical heuristics–Complex Residuals, Pre-RMSNorm, Attention Scaling, and Gradient Clipping–consuming significant compute solely to prevent numerical collapse. We propose a paradigm shift to the “Exact”. We introduce the Halo Architecture, grounded in the Rational Field (Q) and powered by a custom Exact Inference Unit (EIU). To resolve the exponential bit-width growth of rational arithmetic, Halo employs a Dual-Ring Topology that unifies two complementary control mechanisms: (1) The Micro-Ring (Continuum Maintenance), which strictly bounds memory complexity via Diophantine Approximation; and (2) The Macro-Ring (Symbolic Alignment), which enforces logical consistency via periodic state collapse. This stable dual-ring substrate allows for the “Great Dismantling” of numerical scaffolding, reducing the Transformer block to its “Clean” algebraic form (Tabula Rasa). Furthermore, we verify the “Efficiency Paradox”: the elimination of gradient noise (sigma -> 0) allows for Macro-Learning Rates, potentially reducing the Total Time-to-Convergence by orders of magnitude. Halo demonstrates that General Intelligence requires the hybridization of continuous fields and discrete chains under a rigorous mathematical framework.
[803] Reinforcement Learning via Self-Distillation
Jonas Hübotter, Frederike Lübeck, Lejs Behric, Anton Baumann, Marco Bagatella, Daniel Marta, Ido Hakimi, Idan Shenfeld, Thomas Kleine Buening, Carlos Guestrin, Andreas Krause
Main category: cs.LG
TL;DR: SDPO: Self-Distillation Policy Optimization for RL with rich textual feedback, using model’s own feedback-informed predictions to improve learning in verifiable domains like code and math.
Details
Motivation: Current RL with verifiable rewards (RLVR) methods only use scalar outcome rewards, creating credit assignment bottlenecks. Many verifiable environments provide rich textual feedback (runtime errors, judge evaluations) that explain failures, which could be better utilized.Method: Self-Distillation Policy Optimization (SDPO) converts tokenized feedback into dense learning signal without external teacher or reward model. It treats current model conditioned on feedback as self-teacher, distilling its feedback-informed next-token predictions back into the policy.
Result: SDPO improves sample efficiency and final accuracy over strong RLVR baselines across scientific reasoning, tool use, and competitive programming (LiveCodeBench v6). Also outperforms baselines in standard RLVR environments using successful rollouts as implicit feedback.
Conclusion: SDPO effectively leverages rich textual feedback for better RL in verifiable domains, accelerating discovery on difficult binary-reward tasks with 3x fewer attempts compared to existing methods.
Abstract: Large language models are increasingly post-trained with reinforcement learning in verifiable domains such as code and math. Yet, current methods for reinforcement learning with verifiable rewards (RLVR) learn only from a scalar outcome reward per attempt, creating a severe credit-assignment bottleneck. Many verifiable environments actually provide rich textual feedback, such as runtime errors or judge evaluations, that explain why an attempt failed. We formalize this setting as reinforcement learning with rich feedback and introduce Self-Distillation Policy Optimization (SDPO), which converts tokenized feedback into a dense learning signal without any external teacher or explicit reward model. SDPO treats the current model conditioned on feedback as a self-teacher and distills its feedback-informed next-token predictions back into the policy. In this way, SDPO leverages the model’s ability to retrospectively identify its own mistakes in-context. Across scientific reasoning, tool use, and competitive programming on LiveCodeBench v6, SDPO improves sample efficiency and final accuracy over strong RLVR baselines. Notably, SDPO also outperforms baselines in standard RLVR environments that only return scalar feedback by using successful rollouts as implicit feedback for failed attempts. Finally, applying SDPO to individual questions at test time accelerates discovery on difficult binary-reward tasks, achieving the same discovery probability as best-of-k sampling or multi-turn conversations with 3x fewer attempts.
[804] Models Under SCOPE: Scalable and Controllable Routing via Pre-hoc Reasoning
Qi Cao, Shuhao Zhang, Ruizhe Zhou, Ruiyi Zhang, Peijia Qin, Pengtao Xie
Main category: cs.LG
TL;DR: SCOPE is a scalable routing framework that predicts model cost and performance using retrieval-based reasoning, enabling dynamic trade-offs between accuracy and efficiency for unseen models.
Details
Motivation: Existing model routers treat routing as fixed choices among small model sets, making them inflexible to new models or changing budget constraints. There's a need for a more adaptive approach that can predict cost and performance for dynamic decision-making.Method: SCOPE uses reinforcement learning to make reasoning-based predictions by retrieving how models behave on similar problems, rather than relying on fixed model names. It explicitly predicts both accuracy and cost, turning routing into a dynamic decision problem.
Result: SCOPE can boost accuracy by up to 25.7% when performance is prioritized, or cut costs by up to 95.1% when efficiency matters most. It adapts flexibly to user needs beyond just cost savings.
Conclusion: SCOPE provides a scalable and controllable framework for model routing that goes beyond simple model selection, enabling dynamic adaptation to new models and flexible trade-offs between accuracy and cost.
Abstract: Model routing chooses which language model to use for each query. By sending easy queries to cheaper models and hard queries to stronger ones, it can significantly reduce inference cost while maintaining high accuracy. However, most existing routers treat this as a fixed choice among a small set of models, which makes them hard to adapt to new models or changing budget constraints. In this paper, we propose SCOPE (Scalable and Controllable Outcome Performance Estimator), a routing framework that goes beyond model selection by predicting their cost and performance. Trained with reinforcement learning, SCOPE makes reasoning-based predictions by retrieving how models behave on similar problems, rather than relying on fixed model names, enabling it to work with new, unseen models. Moreover, by explicitly predicting how accurate and how expensive a model will be, it turns routing into a dynamic decision problem, allowing users to easily control the trade-off between accuracy and cost. Experiments show that SCOPE is more than just a cost-saving tool. It flexibly adapts to user needs: it can boost accuracy by up to 25.7% when performance is the priority, or cut costs by up to 95.1% when efficiency matters most.
[805] SwiftRepertoire: Few-Shot Immune-Signature Synthesis via Dynamic Kernel Codes
Rong Fu, Wenxin Zhang, Muge Qi, Yang Li, Yabin Jin, Jiekai Wu, Jiaxuan Lu, Chunlei Meng, Youjin Wang, Zeli Su, Juntao Gao, Li Bao, Qi Zhao, Wei Luo, Simon Fong
Main category: cs.LG
TL;DR: A framework for T cell receptor repertoire analysis that enables sample-efficient adaptation to new tasks using lightweight task descriptors and adapter modules, addressing challenges of label sparsity and computational constraints.
Details
Motivation: T cell receptor repertoire analysis provides valuable biological signals for disease detection and immune monitoring, but faces practical deployment challenges including label sparsity, cohort heterogeneity, and computational burden when adapting large encoders to new tasks.Method: The framework synthesizes compact task-specific parameterizations from a learned dictionary of prototypes conditioned on lightweight task descriptors derived from repertoire probes and pooled embedding statistics. This produces small adapter modules applied to a frozen pretrained backbone, enabling adaptation with few support examples without full fine-tuning.
Result: The approach enables immediate adaptation to novel tasks with only a handful of support examples, preserves interpretability through motif-aware probes and calibrated motif discovery, and provides a practical pathway for repertoire-informed models in clinical/research settings with scarce data and constrained resources.
Conclusion: The framework offers a sample-efficient, interpretable, and computationally practical solution for deploying T cell receptor repertoire analysis in diverse settings where labeled data are scarce and computational resources are limited.
Abstract: Repertoire-level analysis of T cell receptors offers a biologically grounded signal for disease detection and immune monitoring, yet practical deployment is impeded by label sparsity, cohort heterogeneity, and the computational burden of adapting large encoders to new tasks. We introduce a framework that synthesizes compact task-specific parameterizations from a learned dictionary of prototypes conditioned on lightweight task descriptors derived from repertoire probes and pooled embedding statistics. This synthesis produces small adapter modules applied to a frozen pretrained backbone, enabling immediate adaptation to novel tasks with only a handful of support examples and without full model fine-tuning. The architecture preserves interpretability through motif-aware probes and a calibrated motif discovery pipeline that links predictive decisions to sequence-level signals. Together, these components yield a practical, sample-efficient, and interpretable pathway for translating repertoire-informed models into diverse clinical and research settings where labeled data are scarce and computational resources are constrained.
[806] Cardinality-Preserving Attention Channels for Graph Transformers in Molecular Property Prediction
Abhijit Gupta
Main category: cs.LG
TL;DR: Graph transformer with query-conditioned cardinality-preserving attention for molecular property prediction, combining structured sparse attention with Graphormer biases and dual-objective self-supervised pretraining.
Details
Motivation: Addressing the challenge of molecular property prediction when labeled data is scarce, particularly in drug discovery applications where obtaining labeled molecular data is expensive and time-consuming.Method: Proposes a graph transformer with query-conditioned cardinality-preserving attention (CPA) that retains dynamic support-size signals, combined with Graphormer-inspired biases (shortest-path distance, centrality, direct-bond features) and unified dual-objective self-supervised pretraining (masked reconstruction and contrastive alignment of augmented views).
Result: Demonstrates consistent improvements over protocol-matched baselines on 11 public benchmarks spanning MoleculeNet, OGB, and TDC ADMET datasets under matched pretraining, optimization, and hyperparameter tuning conditions.
Conclusion: The CPA mechanism provides valuable complementary signals to static centrality embeddings, with rigorous ablations confirming its contributions beyond simple size shortcuts, offering an effective approach for molecular property prediction with limited labeled data.
Abstract: Molecular property prediction is crucial for drug discovery when labeled data are scarce. This work presents \modelname, a graph transformer augmented with a query-conditioned cardinality-preserving attention (CPA) channel that retains dynamic support-size signals complementary to static centrality embeddings. The approach combines structured sparse attention with Graphormer-inspired biases (shortest-path distance, centrality, direct-bond features) and unified dual-objective self-supervised pretraining (masked reconstruction and contrastive alignment of augmented views). Evaluation on 11 public benchmarks spanning MoleculeNet, OGB, and TDC ADMET demonstrates consistent improvements over protocol-matched baselines under matched pretraining, optimization, and hyperparameter tuning. Rigorous ablations confirm CPA’s contributions and rule out simple size shortcuts. Code and reproducibility artifacts are provided.
[807] A Meta-Knowledge-Augmented LLM Framework for Hyperparameter Optimization in Time-Series Forecasting
Ons Saadallah, Mátyás andó, Tamás Gábor Orosz
Main category: cs.LG
TL;DR: LLM-AutoOpt: A hybrid hyperparameter optimization framework combining Bayesian Optimization with LLM-based contextual reasoning for time-series forecasting, using structured meta-knowledge to improve performance and interpretability.
Details
Motivation: Hyperparameter optimization is computationally expensive and difficult to interpret for time-series forecasting. Bayesian Optimization treats tasks independently and provides limited insight, while LLMs offer opportunities to incorporate structured prior knowledge and reasoning into optimization pipelines.Method: Combines Bayesian Optimization with LLM-based contextual reasoning. Encodes dataset meta-features, model descriptions, historical optimization outcomes, and target objectives as structured meta-knowledge within LLM prompts. Uses BO to initialize search and mitigate cold-start effects.
Result: Experiments on multivariate time series forecasting benchmark show LLM-AutoOpt achieves improved predictive performance and more interpretable optimization behavior compared to BO and LLM baselines without meta-knowledge.
Conclusion: LLM-AutoOpt provides a context-aware and stable hyperparameter refinement approach that exposes reasoning behind optimization decisions, demonstrating the value of combining BO with LLM-based reasoning for HPO in time-series forecasting.
Abstract: Hyperparameter optimization (HPO) plays a central role in the performance of deep learning models, yet remains computationally expensive and difficult to interpret, particularly for time-series forecasting. While Bayesian Optimization (BO) is a standard approach, it typically treats tuning tasks independently and provides limited insight into its decisions. Recent advances in large language models (LLMs) offer new opportunities to incorporate structured prior knowledge and reasoning into optimization pipelines. We introduce LLM-AutoOpt, a hybrid HPO framework that combines BO with LLM-based contextual reasoning. The framework encodes dataset meta-features, model descriptions, historical optimization outcomes, and target objectives as structured meta-knowledge within LLM prompts, using BO to initialize the search and mitigate cold-start effects. This design enables context-aware and stable hyperparameter refinement while exposing the reasoning behind optimization decisions. Experiments on a multivariate time series forecasting benchmark demonstrate that LLM-AutoOpt achieves improved predictive performance and more interpretable optimization behavior compared to BO and LLM baselines without meta-knowledge.
[808] RPG-AE: Neuro-Symbolic Graph Autoencoders with Rare Pattern Mining for Provenance-Based Anomaly Detection
Asif Tauhid, Sidahmed Benabderrahmane, Mohamad Altrabulsi, Ahamed Foisal, Talal Rahwan
Main category: cs.LG
TL;DR: A neuro-symbolic anomaly detection framework combining Graph Autoencoder with rare pattern mining to identify APT-like activities in system-level provenance data, showing improved detection performance over baseline methods.
Details
Motivation: Advanced Persistent Threats (APTs) are sophisticated, stealthy cyberattacks that blend into normal system behavior, making them difficult to detect using traditional methods. There's a need for more effective anomaly detection approaches that can identify these subtle, long-term attacks in system provenance data.Method: The approach constructs a process behavioral graph using k-Nearest Neighbors based on feature similarity, then learns normal relational structure using a Graph Autoencoder. Anomaly candidates are identified through deviations between observed and reconstructed graph structure. A rare pattern mining module discovers infrequent behavioral co-occurrences and uses them to boost anomaly scores for processes exhibiting rare signatures.
Result: Evaluation on DARPA Transparent Computing datasets shows that rare-pattern boosting yields substantial gains in anomaly ranking quality over the baseline GAE. The single unified model consistently outperforms individual context-based detectors and achieves performance competitive with ensemble aggregation methods that require multiple separate detectors.
Conclusion: The results highlight the value of coupling graph-based representation learning with classical pattern mining to improve both effectiveness and interpretability in provenance-based security anomaly detection, offering a promising approach for detecting sophisticated APT attacks.
Abstract: Advanced Persistent Threats (APTs) are sophisticated, long-term cyberattacks that are difficult to detect because they operate stealthily and often blend into normal system behavior. This paper presents a neuro-symbolic anomaly detection framework that combines a Graph Autoencoder (GAE) with rare pattern mining to identify APT-like activities in system-level provenance data. Our approach first constructs a process behavioral graph using k-Nearest Neighbors based on feature similarity, then learns normal relational structure using a Graph Autoencoder. Anomaly candidates are identified through deviations between observed and reconstructed graph structure. To further improve detection, we integrate an rare pattern mining module that discovers infrequent behavioral co-occurrences and uses them to boost anomaly scores for processes exhibiting rare signatures. We evaluate the proposed method on the DARPA Transparent Computing datasets and show that rare-pattern boosting yields substantial gains in anomaly ranking quality over the baseline GAE. Compared with existing unsupervised approaches on the same benchmark, our single unified model consistently outperforms individual context-based detectors and achieves performance competitive with ensemble aggregation methods that require multiple separate detectors. These results highlight the value of coupling graph-based representation learning with classical pattern mining to improve both effectiveness and interpretability in provenance-based security anomaly detection.
[809] Reinforcement Learning with Promising Tokens for Large Language Models
Jing-Cheng Pang, Liang Lu, Xian Tang, Kun Jiang, Sijie Wu, Kai Zhang, Xubin Li
Main category: cs.LG
TL;DR: RLPT is a reinforcement learning framework that reduces the action space for LLM alignment by focusing policy optimization only on promising tokens identified via semantic priors, improving training stability and sample efficiency.
Details
Motivation: Standard RL for LLM alignment treats the full vocabulary as the action space, which includes many irrelevant tokens that distract from meaningful decision-making and cause training instability.Method: RLPT decouples strategic decision-making from token generation by using the base model’s semantic priors to identify a dynamic set of promising tokens, then constraining policy optimization to this refined subset via masking.
Result: RLPT reduces gradient variance, stabilizes training, improves sample efficiency, and outperforms standard RL baselines on math, coding, and telecom reasoning tasks across various model sizes (4B and 8B) and RL algorithms.
Conclusion: Focusing RL optimization on promising tokens rather than the full vocabulary addresses a fundamental limitation in LLM alignment, leading to more efficient and stable training.
Abstract: Reinforcement learning (RL) has emerged as a key paradigm for aligning and optimizing large language models (LLMs). Standard approaches treat the LLM as the policy and apply RL directly over the full vocabulary space. However, this formulation includes the massive tail of contextually irrelevant tokens in the action space, which could distract the policy from focusing on decision-making among the truly reasonable tokens. In this work, we verify that valid reasoning paths could inherently concentrate within a low-rank subspace. Based on this insight, we introduce Reinforcement Learning with Promising Tokens (RLPT), a framework that mitigates the action space issue by decoupling strategic decision-making from token generation. Specifically, RLPT leverages the semantic priors of the base model to identify a dynamic set of promising tokens and constrains policy optimization exclusively to this refined subset via masking. Theoretical analysis and empirical results demonstrate that RLPT effectively reduces gradient variance, stabilizes the training process, and improves sample efficiency. Experiment results on math, coding, and telecom reasoning show that RLPT outperforms standard RL baselines and integrates effectively across various model sizes (4B and 8B) and RL algorithms (GRPO and DAPO).
[810] Privileged Information Distillation for Language Models
Emiliano Penaloza, Dheeraj Vattikonda, Nicolas Gontier, Alexandre Lacoste, Laurent Charlin, Massimo Caccia
Main category: cs.LG
TL;DR: π-Distill and OPSD methods for distilling frontier models using privileged information during training but not at inference, outperforming standard supervised finetuning+RL approaches on agentic benchmarks.
Details
Motivation: Training with privileged information (PI) helps models succeed on hard tasks, but transferring these capabilities to policies that must act without PI at inference remains challenging, especially in multi-turn agentic environments where only action trajectories are observable.Method: Two approaches: 1) π-Distill - joint teacher-student objective training PI-conditioned teacher and unconditioned student simultaneously; 2) OPSD - reinforcement learning with reverse KL-penalty between student and PI-conditioned teacher.
Result: Both methods effectively distill frontier agents using action-only PI, outperforming industry standard practices (supervised finetuning followed by RL) that assume full Chain-of-Thought supervision across multiple benchmarks, models, and PI forms.
Conclusion: π-Distill and OPSD provide effective solutions for knowledge distillation when only action trajectories are available, enabling transfer of capabilities learned with PI to policies that must act without it at inference.
Abstract: Training-time privileged information (PI) can enable language models to succeed on tasks they would otherwise fail, making it a powerful tool for reinforcement learning in hard, long-horizon settings. However, transferring capabilities learned with PI to policies that must act without it at inference time remains a fundamental challenge. We study this problem in the context of distilling frontier models for multi-turn agentic environments, which typically hide their internal reasoning and expose only action trajectories. This breaks standard distillation pipelines, since successful behavior is observable, but the reasoning process is not. For this, we introduce π-Distill, a joint teacher-student objective that trains a PI-conditioned teacher and an unconditioned student simultaneously using the same model. Additionally, we also introduce On-Policy Self-Distillation (OPSD), an alternative approach that trains using Reinforcement Learning (RL) with a reverse KL-penalty between the student and the PI-conditioned teacher. We show that both of these algorithms effectively distill frontier agents using action-only PI. Specifically, we find that π-Distill and, in some cases, OPSD, outperform industry standard practices (Supervised finetuning followed by RL) that assume access to full Chain-of-Thought supervision across multiple agentic benchmarks, models, and forms of PI. We complement our results with extensive analysis that characterizes the factors enabling effective learning with PI, focusing primarily on π-Distill and characterizing when OPSD is competitive.
[811] How to Train Your Resistive Network: Generalized Equilibrium Propagation and Analytical Learning
Jonathan Lin, Aman Desai, Frank Barrows, Francesco Caravelli
Main category: cs.LG
TL;DR: A novel algorithm for training analog computing systems using graph theory and Kirchhoff’s laws to calculate exact gradients, enabling energy-efficient machine learning with physical locality constraints.
Details
Motivation: Current digital hardware for machine learning is energy-intensive, and analog computing offers a promising alternative. However, training analog systems is challenging due to physical locality constraints that prevent standard backpropagation. There's a need for local learning algorithms that can work within these constraints while maintaining performance.Method: Developed an algorithm using graph theory and analytical framework for Kirchhoff’s laws to exactly calculate gradients. Introduced Generalized Equilibrium Propagation framework encompassing Hebbian learning algorithms. The approach trains resistor networks without needing replicas or readout over all resistors, only requiring output layer measurements.
Result: Demonstrated through numerical simulations that resistor networks can be trained without replica or full resistor readout. Showed that under the analytical gradient approach, only a subset of resistance values need to be updated without significant performance degradation.
Conclusion: The proposed algorithm enables efficient training of analog computing systems while respecting physical locality constraints, offering a pathway to energy-efficient machine learning implementations using analog hardware.
Abstract: Machine learning is a powerful method of extracting meaning from data; unfortunately, current digital hardware is extremely energy-intensive. There is interest in an alternative analog computing implementation that could match the performance of traditional machine learning while being significantly more energy-efficient. However, it remains unclear how to train such analog computing systems while adhering to locality constraints imposed by the physical (as opposed to digital) nature of these systems. Local learning algorithms such as Equilibrium Propagation and Coupled Learning have been proposed to address this issue. In this paper, we develop an algorithm to exactly calculate gradients using a graph theoretic and analytical framework for Kirchhoff’s laws. We also introduce Generalized Equilibrium Propagation, a framework encompassing a broad class of Hebbian learning algorithms, including Coupled Learning and Equilibrium Propagation, and show how our algorithm compares. We demonstrate our algorithm using numerical simulations and show that we can train resistor networks without the need for a replica or readout over all resistors, only at the output layer. We also show that under the analytical gradient approach, it is possible to update only a subset of the resistance values without a strong degradation in performance.
[812] NeuroPareto: Calibrated Acquisition for Costly Many-Goal Search in Vast Parameter Spaces
Rong Fu, Wenxin Zhang, Chunlei Meng, Youjin Wang, Haoyu Zhao, Jiaxuan Lu, Kun Liu, JiaBao Dou, Simon James Fong
Main category: cs.LG
TL;DR: NeuroPareto is a multi-objective optimization framework that integrates rank filtering, uncertainty disentanglement, and history-conditioned acquisition to efficiently navigate high-dimensional search spaces under computational constraints.
Details
Motivation: The challenge of finding optimal trade-offs in high-dimensional search spaces with strict computational constraints requires efficient multi-objective optimization methods that can navigate complex objective landscapes with minimal evaluation costs.Method: NeuroPareto combines rank-centric filtering, uncertainty disentanglement via calibrated Bayesian classifier and Deep Gaussian Process surrogates, and history-conditioned acquisition strategies. It uses hierarchical screening, amortized surrogate updates, and a lightweight acquisition network trained online from historical hypervolume improvements.
Result: NeuroPareto consistently outperforms classifier-enhanced and surrogate-assisted baselines on DTLZ and ZDT benchmark suites and a subsurface energy extraction task, achieving better Pareto proximity and hypervolume metrics.
Conclusion: NeuroPareto provides an effective framework for computationally efficient multi-objective optimization that balances convergence and diversity while maintaining low computational overhead.
Abstract: The pursuit of optimal trade-offs in high-dimensional search spaces under stringent computational constraints poses a fundamental challenge for contemporary multi-objective optimization. We develop NeuroPareto, a cohesive architecture that integrates rank-centric filtering, uncertainty disentanglement, and history-conditioned acquisition strategies to navigate complex objective landscapes. A calibrated Bayesian classifier estimates epistemic uncertainty across non-domination tiers, enabling rapid generation of high-quality candidates with minimal evaluation cost. Deep Gaussian Process surrogates further separate predictive uncertainty into reducible and irreducible components, providing refined predictive means and risk-aware signals for downstream selection. A lightweight acquisition network, trained online from historical hypervolume improvements, guides expensive evaluations toward regions balancing convergence and diversity. With hierarchical screening and amortized surrogate updates, the method maintains accuracy while keeping computational overhead low. Experiments on DTLZ and ZDT suites and a subsurface energy extraction task show that NeuroPareto consistently outperforms classifier-enhanced and surrogate-assisted baselines in Pareto proximity and hypervolume.
[813] On the Non-Identifiability of Steering Vectors in Large Language Models
Sohan Venkatesh, Ashish Mahendran Kurapath
Main category: cs.LG
TL;DR: Activation steering vectors in LLMs are fundamentally non-identifiable due to large equivalence classes of behaviorally indistinguishable interventions, revealing interpretability limits.
Details
Motivation: Activation steering methods are widely used to control LLM behavior and are often interpreted as revealing meaningful internal representations, but this assumes steering directions are identifiable and uniquely recoverable from input-output behavior.Method: The paper shows that under white-box single-layer access, steering vectors are fundamentally non-identifiable due to large equivalence classes of behaviorally indistinguishable interventions. Empirically demonstrates that orthogonal perturbations achieve near-equivalent efficacy with negligible effect sizes across multiple models and traits.
Result: The non-identifiability is a robust geometric property that persists across diverse prompt distributions. Orthogonal perturbations achieve near-equivalent efficacy with negligible effect sizes.
Conclusion: These findings reveal fundamental interpretability limits and highlight the need for structural constraints beyond behavioral testing to enable reliable alignment interventions.
Abstract: Activation steering methods are widely used to control large language model (LLM) behavior and are often interpreted as revealing meaningful internal representations. This interpretation assumes steering directions are identifiable and uniquely recoverable from input-output behavior. We show that, under white-box single-layer access, steering vectors are fundamentally non-identifiable due to large equivalence classes of behaviorally indistinguishable interventions. Empirically, we show that orthogonal perturbations achieve near-equivalent efficacy with negligible effect sizes across multiple models and traits. Critically, we show that the non-identifiability is a robust geometric property that persists across diverse prompt distributions. These findings reveal fundamental interpretability limits and highlight the need for structural constraints beyond behavioral testing to enable reliable alignment interventions.
[814] Unbiased Single-Queried Gradient for Combinatorial Objective
Thanawat Sornwanee
Main category: cs.LG
TL;DR: Proposes a stochastic gradient method for combinatorial optimization problems that requires only a single query of the combinatorial function, generalizing REINFORCE with importance sampling.
Details
Motivation: Combinatorial optimization problems often require multiple function queries for exact gradient computation, which is computationally expensive. The authors aim to develop more efficient stochastic gradient methods that reduce query complexity.Method: Reformulates combinatorial problems as probabilistic optimization over hypercubes (Bernoulli parameters). Proposes an unbiased stochastic gradient estimator that requires only a single query of the combinatorial function, generalizing REINFORCE through importance sampling techniques.
Result: Develops a class of new stochastic gradient estimators that are unbiased and more query-efficient than traditional methods, with REINFORCE as a special case of the proposed framework.
Conclusion: The proposed method provides an efficient alternative for gradient-based optimization in combinatorial problems, reducing computational cost while maintaining theoretical guarantees through unbiased estimation.
Abstract: In a probabilistic reformulation of a combinatorial problem, we often face an optimization over a hypercube, which corresponds to the Bernoulli probability parameter for each binary variable in the primal problem. The combinatorial nature suggests that an exact gradient computation requires multiple queries. We propose a stochastic gradient that is unbiased and requires only a single query of the combinatorial function. This method encompasses a well-established REINFORCE (through an importance sampling), as well as including a class of new stochastic gradients.
[815] Accelerated Sequential Flow Matching: A Bayesian Filtering Perspective
Yinan Huang, Hans Hao-Hsun Hsu, Junran Wang, Bo Dai, Pan Li
Main category: cs.LG
TL;DR: Sequential Flow Matching: A Bayesian filtering framework for efficient real-time streaming inference with flow-matching models, using previous posterior as warm start to accelerate sampling.
Details
Motivation: Diffusion and flow-matching models for sequential prediction suffer from high inference latency in real-time streaming environments due to repeated sampling from non-informative initial distributions, causing system backlogs.Method: Treats streaming inference as learning a probability flow that transports predictive distribution from one time step to the next, aligning with recursive Bayesian belief updates. Initializes generation from previous posterior as principled warm start.
Result: Achieves performance competitive with full-step diffusion while requiring only one or very few sampling steps, enabling faster sampling across forecasting, decision-making, and state estimation tasks.
Conclusion: Framing sequential inference via Bayesian filtering provides a principled perspective for efficient real-time deployment of flow-based models, with theoretical justification for accelerated sampling.
Abstract: Sequential prediction from streaming observations is a fundamental problem in stochastic dynamical systems, where inherent uncertainty often leads to multiple plausible futures. While diffusion and flow-matching models are capable of modeling complex, multi-modal trajectories, their deployment in real-time streaming environments typically relies on repeated sampling from a non-informative initial distribution, incurring substantial inference latency and potential system backlogs. In this work, we introduce Sequential Flow Matching, a principled framework grounded in Bayesian filtering. By treating streaming inference as learning a probability flow that transports the predictive distribution from one time step to the next, our approach naturally aligns with the recursive structure of Bayesian belief updates. We provide theoretical justification that initializing generation from the previous posterior offers a principled warm start that can accelerate sampling compared to naïve re-sampling. Across a wide range of forecasting, decision-making and state estimation tasks, our method achieves performance competitive with full-step diffusion while requiring only one or very few sampling steps, therefore with faster sampling. It suggests that framing sequential inference via Bayesian filtering provides a new and principled perspective towards efficient real-time deployment of flow-based models. Our code is available at https://github.com/Graph-COM/Sequential_Flow_Matching.
[816] Causal Schrödinger Bridges: Constrained Optimal Transport on Structural Manifolds
Rui Wu, Li YongJun
Main category: cs.LG
TL;DR: CSB is a causal Schrödinger Bridge framework that uses diffusion processes for robust counterfactual inference via Entropic Optimal Transport, overcoming limitations of deterministic flows in handling support mismatches and high-dimensional causal systems.
Details
Motivation: Deterministic flows (ODE-based generative models) become brittle under causal interventions when transporting probability mass across low-density regions where vector fields are ill-defined, leading to numerical instability and spurious correlations.Method: Causal Schrödinger Bridge (CSB) reformulates counterfactual inference as Entropic Optimal Transport using diffusion processes (SDEs) instead of deterministic flows, leveraging structural admissibility constraints and proving a Structural Decomposition Theorem for exact factorization into local transitions.
Result: CSB breaks the Curse of Dimensionality in high intrinsic dimension regimes, completing transport on a full-rank causal system (d=10^5) in 26.48 seconds on a single GPU, versus estimated 6+ years for structure-agnostic baselines. Achieves high fidelity (MSE ≈ 0.04) on Morpho-MNIST and 10^5-D stress tests.
Conclusion: CSB provides a robust framework for counterfactual inference that outperforms deterministic baselines in structural consistency and distribution coverage, enabling efficient handling of high-dimensional causal systems through diffusion-based transport.
Abstract: Generative modeling typically seeks the path of least action via deterministic flows (ODE). While effective for in-distribution tasks, we argue that these deterministic paths become brittle under causal interventions, which often require transporting probability mass across low-density regions (off-manifold'') where the vector field is ill-defined. This leads to numerical instability and spurious correlations. In this work, we introduce the Causal Schrödinger Bridge (CSB), a framework that reformulates counterfactual inference as Entropic Optimal Transport. Unlike deterministic approaches that require strict invertibility or rely on low-rank approximations, CSB leverages diffusion processes (SDEs) to robustly tunnel’’ through support mismatches while strictly enforcing structural admissibility constraints. We prove the Structural Decomposition Theorem, showing that the global high-dimensional bridge factorizes exactly into local, robust transitions. Crucially, we demonstrate that CSB breaks the Curse of Dimensionality in regimes of high intrinsic dimension. We empirically validate this on a full-rank causal system ($d=10^5$, intrinsic rank $10^5$), completing the transport in 26.48 seconds on a single GPU (RTX 3090). This stands in stark contrast to structure-agnostic $O(d^3)$ baselines, which are estimated to require over 6 years for dense computations of this scale regardless of the data’s intrinsic rank. Empirical validation on Morpho-MNIST and $10^5$-D extremal stress tests demonstrates that CSB significantly outperforms deterministic baselines in structural consistency and distribution coverage, capturing the underlying manifold with high fidelity (MSE $\approx$ 0.04).
[817] From Robotics to Sepsis Treatment: Offline RL via Geometric Pessimism
Sarthak Wanjari
Main category: cs.LG
TL;DR: Geo-IQL: A compute-efficient offline RL method that adds geometric pessimism via k-NN density penalties to IQL, improving OOD action handling without degrading stable performance.
Details
Motivation: Offline RL suffers from OOD action overestimation, especially in fractured/sparse data. Current methods trade computational efficiency for performance - CQL is rigorous but compute-heavy, while IQL is efficient but fails on pathological datasets, collapsing to behavioral cloning.Method: Geometric Pessimism framework augments standard IQL with density-based penalties derived from k-nearest-neighbor distances in state-action embedding space. Pre-computed penalties are applied via reward shaping with O(1) training overhead.
Result: On D4RL MuJoCo: Geo-IQL outperforms IQL by over 18 points on sensitive medium-replay tasks, reduces inter-seed std by 4x, maintains performance on stable manifolds. On MIMIC-III Sepsis: Geo-IQL shows active policy improvement while IQL collapses to behavior cloning; achieves 86.4% terminal agreement with clinicians vs IQL’s 75%.
Conclusion: Geometric pessimism provides necessary regularization to safely overcome local optima in critical real-world decision systems, offering compute-efficient OOD conservatism without performance degradation on stable data manifolds.
Abstract: Offline Reinforcement Learning (RL) promises the recovery of optimal policies from static datasets, yet it remains susceptible to the overestimation of out-of-distribution (OOD) actions, particularly in fractured and sparse data manifolds. Current solutions necessitate a trade-off between computational efficiency and performance. Methods like CQL offer rigorous conservatism but require tremendous compute power while efficient expectile-based methods like IQL often fail to correct OOD errors on pathological datasets, collapsing to Behavioural Cloning. In this work, we propose Geometric Pessimism, a modular, compute-efficient framework that augments standard IQL with density-based penalty derived from k-nearest-neighbour distances in the state-action embedding space. By pre-computing the penalties applied to each state-action pair, our method injects OOD conservatism via reward shaping with a O(1) training overhead to the training loop. Evaluated on the D4RL MuJoCo benchmark, our method, Geo-IQL outperforms standard IQL on sensitive and unstable medium-replay tasks by over 18 points, while reducing inter-seed standard-deviation by 4 times. Furthermore, Geo-IQL does not degrade performance on stable manifolds. Crucially, we validate our algorithm on the MIMIC-III Sepsis critical care dataset. While standard IQL collapses to behaviour cloning, Geo-IQL demonstrates active policy improvement. Maintaining safety constraints, it achieves 86.4% terminal agreement with clinicians compared to IQL’s 75%. Our results suggest that geometric pessimism provides the necessary regularisation to safely overcome local optima in critical, real-world decision systems.
[818] Feature salience – not task-informativeness – drives machine learning model explanations
Benedict Clark, Marta Oliveira, Rick Wilming, Stefan Haufe
Main category: cs.LG
TL;DR: XAI methods attribute importance based on feature salience rather than informativeness, as shown through watermark experiments in image classification.
Details
Motivation: To investigate whether XAI methods truly identify informative features or are instead driven by other data properties like feature salience, statistical suppression, or novelty at test-time.Method: Trained deep learning models on three variants of binary image classification with translucent watermarks (absent, class-dependent confounds, class-independent noise), then analyzed five popular attribution methods’ importance attribution patterns.
Result: XAI methods showed substantially elevated importance in watermarked areas regardless of training setting, with minimal effect from whether watermarks were class-dependent or not. Importance attribution behaved similarly to edge detection filters and was strongly influenced by feature salience rather than learned statistical associations.
Conclusion: XAI importance attribution is primarily driven by feature salience at test time rather than model-learned informativeness, suggesting previous XAI studies need reevaluation and workflows using feature attribution should be scrutinized.
Abstract: Explainable AI (XAI) promises to provide insight into machine learning models’ decision processes, where one goal is to identify failures such as shortcut learning. This promise relies on the field’s assumption that input features marked as important by an XAI must contain information about the target variable. However, it is unclear whether informativeness is indeed the main driver of importance attribution in practice, or if other data properties such as statistical suppression, novelty at test-time, or high feature salience substantially contribute. To clarify this, we trained deep learning models on three variants of a binary image classification task, in which translucent watermarks are either absent, act as class-dependent confounds, or represent class-independent noise. Results for five popular attribution methods show substantially elevated relative importance in watermarked areas (RIW) for all models regardless of the training setting ($R^2 \geq .45$). By contrast, whether the presence of watermarks is class-dependent or not only has a marginal effect on RIW ($R^2 \leq .03$), despite a clear impact impact on model performance and generalisation ability. XAI methods show similar behaviour to model-agnostic edge detection filters and attribute substantially less importance to watermarks when bright image intensities are encoded by smaller instead of larger feature values. These results indicate that importance attribution is most strongly driven by the salience of image structures at test time rather than statistical associations learned by machine learning models. Previous studies demonstrating successful XAI application should be reevaluated with respect to a possibly spurious concurrency of feature salience and informativeness, and workflows using feature attribution methods as building blocks should be scrutinised.
[819] dnaHNet: A Scalable and Hierarchical Foundation Model for Genomic Sequence Learning
Arnav Shah, Junzhe Li, Parsa Idehpour, Adibvafa Fallahpour, Brandon Wang, Sukjun Hwang, Bo Wang, Patrick D. Hsu, Hani Goodarzi, Albert Gu
Main category: cs.LG
TL;DR: dnaHNet is a tokenizer-free autoregressive model for genomic sequences that uses differentiable dynamic chunking to compress raw nucleotides into latent tokens adaptively, achieving better scaling and efficiency than existing architectures while preserving biological coherence.
Details
Motivation: Genomic foundation models face a fundamental tradeoff: fixed-vocabulary tokenizers fragment biologically meaningful motifs like codons and regulatory elements, while nucleotide-level models preserve biological coherence but have prohibitive computational costs for long contexts.Method: Introduces dnaHNet, a tokenizer-free autoregressive model that segments and models genomic sequences end-to-end using a differentiable dynamic chunking mechanism. This compresses raw nucleotides into latent tokens adaptively, balancing compression with predictive accuracy. The model is pretrained on prokaryotic genomes.
Result: Outperforms leading architectures including StripedHyena2 in scaling and efficiency, achieves quadratic FLOP reductions enabling >3× inference speedup over Transformers. On zero-shot tasks, achieves superior performance in predicting protein variant fitness and gene essentiality, while automatically discovering hierarchical biological structures without supervision.
Conclusion: dnaHNet establishes a scalable, interpretable framework for next-generation genomic modeling that addresses the fundamental tradeoff between biological coherence and computational efficiency in genomic foundation models.
Abstract: Genomic foundation models have the potential to decode DNA syntax, yet face a fundamental tradeoff in their input representation. Standard fixed-vocabulary tokenizers fragment biologically meaningful motifs such as codons and regulatory elements, while nucleotide-level models preserve biological coherence but incur prohibitive computational costs for long contexts. We introduce dnaHNet, a state-of-the-art tokenizer-free autoregressive model that segments and models genomic sequences end-to-end. Using a differentiable dynamic chunking mechanism, dnaHNet compresses raw nucleotides into latent tokens adaptively, balancing compression with predictive accuracy. Pretrained on prokaryotic genomes, dnaHNet outperforms leading architectures including StripedHyena2 in scaling and efficiency. This recursive chunking yields quadratic FLOP reductions, enabling $>3 \times$ inference speedup over Transformers. On zero-shot tasks, dnaHNet achieves superior performance in predicting protein variant fitness and gene essentiality, while automatically discovering hierarchical biological structures without supervision. These results establish dnaHNet as a scalable, interpretable framework for next-generation genomic modeling.
[820] In-the-Wild Model Organisms: Mitigating Undesirable Emergent Behaviors in Production LLM Post-Training via Data Attribution
Frank Xiao, Santiago Aranguri
Main category: cs.LG
TL;DR: Activation-based data attribution method traces behavioral changes in language models to responsible training datapoints, identifying harmful behaviors like distractor-triggered compliance in DPO training.
Details
Motivation: To develop a method for tracing behavioral changes in post-trained language models back to specific training datapoints, enabling identification of harmful behaviors that emerge from contaminated preference data rather than deliberate injection.Method: Computes activation-difference vectors for test prompts and preference pairs, ranks by cosine similarity to identify responsible datapoints, validates causally by retraining with modified data, and clusters behavior-datapoint similarity matrices for unsupervised discovery of emergent behaviors.
Result: Applied to OLMo 2’s DPO training, discovered “distractor-triggered compliance” - harmful behavior where model complies with dangerous requests when benign formatting instructions are appended. Filtering top-ranked datapoints reduces behavior by 63%, switching labels achieves 78% reduction. Method outperforms gradient-based attribution and LLM-judge baselines while being 10x cheaper.
Conclusion: Activation-based data attribution provides effective, efficient method for identifying training data responsible for specific model behaviors, offering realistic benchmark for safety techniques on in-the-wild model organisms emerging from contaminated preference data.
Abstract: We propose activation-based data attribution, a method that traces behavioral changes in post-trained language models to responsible training datapoints. By computing activation-difference vectors for both test prompts and preference pairs and ranking by cosine similarity, we identify datapoints that cause specific behaviors and validate these attributions causally by retraining with modified data. Clustering behavior-datapoint similarity matrices also enables unsupervised discovery of emergent behaviors. Applying this to OLMo 2’s production DPO training, we surfaced distractor-triggered compliance: a harmful behavior where the model complies with dangerous requests when benign formatting instructions are appended. Filtering top-ranked datapoints reduces this behavior by 63% while switching their labels achieves 78%. Our method outperforms gradient-based attribution and LLM-judge baselines while being over 10 times cheaper than both. This in-the-wild model organism - emerging from contaminated preference data rather than deliberate injection - provides a realistic benchmark for safety techniques.
[821] Potential-energy gating for robust state estimation in bistable stochastic systems
Luigi Simeone
Main category: cs.LG
TL;DR: Physics-based Bayesian filtering method that modulates observation noise covariance using potential energy landscape to improve state estimation in double-well stochastic systems.
Details
Motivation: Traditional robust filters treat all state-space regions identically and statistical methods fail in non-ergodic or data-scarce settings where only single realizations are available. Need for physics-informed approach that leverages known potential energy structure.Method: Potential-energy gating modulates observation noise covariance based on local potential energy value: observations are trusted near potential minima and progressively discounted near barriers. Implemented within Extended, Unscented, Ensemble, Adaptive Kalman filters and particle filters with only two hyperparameters.
Result: 57-80% RMSE improvement over standard Extended Kalman Filter on Ginzburg-Landau double-well benchmark with 10% outlier contamination. Robust to misspecification (47% improvement even with 50% parameter errors). Applied to ice-core climate data showing 91% variance explained by outlier fraction.
Conclusion: Potential-energy gating provides physics-informed robust state estimation for stochastic systems with metastable dynamics, outperforming statistical methods especially in data-scarce settings.
Abstract: We introduce potential-energy gating, a method for robust state estimation in systems governed by double-well stochastic dynamics. The observation noise covariance of a Bayesian filter is modulated by the local value of a known or assumed potential energy function: observations are trusted when the state is near a potential minimum and progressively discounted as it approaches the barrier separating metastable wells. This physics-based mechanism differs from statistical robust filters, which treat all state-space regions identically, and from constrained filters, which bound states rather than modulating observation trust. The approach is especially relevant in non-ergodic or data-scarce settings where only a single realization is available and statistical methods alone cannot learn the noise structure. We implement gating within Extended, Unscented, Ensemble, and Adaptive Kalman filters and particle filters, requiring only two additional hyperparameters. Monte Carlo benchmarks (100 replications) on a Ginzburg-Landau double-well with 10% outlier contamination show 57-80% RMSE improvement over the standard Extended Kalman Filter, all statistically significant (p < 10^{-15}, Wilcoxon test). A naive topological baseline using only well positions achieves 57%, confirming that the continuous energy landscape adds ~21 percentage points. The method is robust to misspecification: even with 50% parameter errors, improvement never falls below 47%. Comparing externally forced and spontaneous Kramers-type transitions, gating retains 68% improvement under noise-induced transitions whereas the naive baseline degrades to 30%. As an empirical illustration, we apply the framework to Dansgaard-Oeschger events in the NGRIP delta-18O ice-core record, estimating asymmetry gamma = -0.109 (bootstrap 95% CI: [-0.220, -0.011]) and showing that outlier fraction explains 91% of the variance in filter improvement.
[822] ExtractBench: A Benchmark and Evaluation Methodology for Complex Structured Extraction
Nick Ferguson, Josh Pennington, Narek Beghian, Aravind Mohan, Douwe Kiela, Sheshansh Agrawal, Thien Hang Nguyen
Main category: cs.LG
TL;DR: ExtractBench: A benchmark and evaluation framework for PDF-to-JSON structured extraction using LLMs, addressing enterprise-scale schema complexity and semantic evaluation metrics.
Details
Motivation: LLMs are increasingly used for PDF-to-JSON extraction but lack proper evaluation benchmarks for enterprise-scale schemas and semantic correctness across different field types (identifiers, quantities, names).Method: Created ExtractBench with 35 PDF documents paired with JSON Schemas and human-annotated gold labels across valuable domains, totaling 12,867 evaluatable fields. The framework treats schemas as executable specifications with field-specific scoring metrics.
Result: Frontier models (GPT-5/5.2, Gemini-3, Claude 4.5) remain unreliable on realistic schemas, with performance degrading sharply with schema breadth. On a 369-field financial reporting schema, all tested models produced 0% valid output.
Conclusion: Current LLMs struggle with complex PDF-to-JSON extraction tasks, highlighting the need for better benchmarks and improved model capabilities for structured information extraction from unstructured documents.
Abstract: Unstructured documents like PDFs contain valuable structured information, but downstream systems require this data in reliable, standardized formats. LLMs are increasingly deployed to automate this extraction, making accuracy and reliability paramount. However, progress is bottlenecked by two gaps. First, no end-to-end benchmark evaluates PDF-to-JSON extraction under enterprise-scale schema breadth. Second, no principled methodology captures the semantics of nested extraction, where fields demand different notions of correctness (exact match for identifiers, tolerance for quantities, semantic equivalence for names), arrays require alignment, and omission must be distinguished from hallucination. We address both gaps with ExtractBench, an open-source benchmark and evaluation framework for PDF-to-JSON structured extraction. The benchmark pairs 35 PDF documents with JSON Schemas and human-annotated gold labels across economically valuable domains, yielding 12,867 evaluatable fields spanning schema complexities from tens to hundreds of fields. The evaluation framework treats the schema as an executable specification: each field declares its scoring metric. Baseline evaluations reveal that frontier models (GPT-5/5.2, Gemini-3 Flash/Pro, Claude 4.5 Opus/Sonnet) remain unreliable on realistic schemas. Performance degrades sharply with schema breadth, culminating in 0% valid output on a 369-field financial reporting schema across all tested models. We release ExtractBench at https://github.com/ContextualAI/extract-bench.
[823] Why Deep Jacobian Spectra Separate: Depth-Induced Scaling and Singular-Vector Alignment
Nathanaël Haas, François Gatine, Augustin M Cosse, Zied Bouraoui
Main category: cs.LG
TL;DR: The paper analyzes deep network Jacobians, showing depth-induced exponential scaling of singular values and spectral separation, leading to decoupled singular-value dynamics that explain implicit bias in gradient-based training.
Details
Motivation: Understanding why gradient-based training in deep networks exhibits strong implicit bias remains challenging, as tractable singular-value dynamics are typically only available for balanced deep linear models. The authors seek alternative theoretical grounding for analyzing deep network Jacobians.Method: Adopting a fixed-gates view of piecewise-linear networks where Jacobians reduce to products of masked linear maps within activation regions. Proving existence of Lyapunov exponents for top singular values at initialization, giving closed-form expressions in tractable masked models, and quantifying finite-depth corrections. Showing strong spectral separation forces singular-vector alignment in matrix products.
Result: Theoretical results demonstrate depth-induced exponential scaling of ordered singular values and strong spectral separation in deep Jacobians. Experiments in fixed-gates settings validate predicted scaling, alignment, and resulting dynamics.
Conclusion: The analysis supports a mechanistic account of emergent low-rank Jacobian structure as a driver of implicit bias in deep networks, providing an approximation regime where singular-value dynamics become effectively decoupled without requiring balancing.
Abstract: Understanding why gradient-based training in deep networks exhibits strong implicit bias remains challenging, in part because tractable singular-value dynamics are typically available only for balanced deep linear models. We propose an alternative route based on two theoretically grounded and empirically testable signatures of deep Jacobians: depth-induced exponential scaling of ordered singular values and strong spectral separation. Adopting a fixed-gates view of piecewise-linear networks, where Jacobians reduce to products of masked linear maps within a single activation region, we prove the existence of Lyapunov exponents governing the top singular values at initialization, give closed-form expressions in a tractable masked model, and quantify finite-depth corrections. We further show that sufficiently strong separation forces singular-vector alignment in matrix products, yielding an approximately shared singular basis for intermediate Jacobians. Together, these results motivate an approximation regime in which singular-value dynamics become effectively decoupled, mirroring classical balanced deep-linear analyses without requiring balancing. Experiments in fixed-gates settings validate the predicted scaling, alignment, and resulting dynamics, supporting a mechanistic account of emergent low-rank Jacobian structure as a driver of implicit bias.
[824] Prior-Guided Symbolic Regression: Towards Scientific Consistency in Equation Discovery
Jing Xiao, Xinhai Chen, Jiaming Peng, Qinglin Wang, Menghan Jia, Zhiquan Lai, Guangping Yu, Dongsheng Li, Tiejun Li, Jie Liu
Main category: cs.LG
TL;DR: PG-SR is a prior-guided symbolic regression framework that prevents pseudo-equations by incorporating domain knowledge constraints throughout a three-stage pipeline, ensuring scientific consistency while maintaining empirical performance.
Details
Motivation: Existing symbolic regression methods often produce equations that fit data well but violate fundamental scientific principles (Pseudo-Equation Trap), as they focus primarily on empirical risk minimization without explicit scientific consistency constraints.Method: Three-stage pipeline (warm-up, evolution, refinement) with prior constraint checker that encodes domain priors as executable programs, plus Prior Annealing Constrained Evaluation (PACE) mechanism that progressively steers evolution toward scientifically consistent regions.
Result: PG-SR outperforms state-of-the-art baselines across diverse domains, maintains robustness to varying prior quality, noisy data, and data scarcity, and theoretically reduces Rademacher complexity for better generalization bounds.
Conclusion: PG-SR successfully addresses the Pseudo-Equation Trap by integrating domain knowledge constraints into symbolic regression, ensuring both empirical performance and scientific consistency while providing theoretical guarantees.
Abstract: Symbolic Regression (SR) aims to discover interpretable equations from observational data, with the potential to reveal underlying principles behind natural phenomena. However, existing approaches often fall into the Pseudo-Equation Trap: producing equations that fit observations well but remain inconsistent with fundamental scientific principles. A key reason is that these approaches are dominated by empirical risk minimization, lacking explicit constraints to ensure scientific consistency. To bridge this gap, we propose PG-SR, a prior-guided SR framework built upon a three-stage pipeline consisting of warm-up, evolution, and refinement. Throughout the pipeline, PG-SR introduces a prior constraint checker that explicitly encodes domain priors as executable constraint programs, and employs a Prior Annealing Constrained Evaluation (PACE) mechanism during the evolution stage to progressively steer discovery toward scientifically consistent regions. Theoretically, we prove that PG-SR reduces the Rademacher complexity of the hypothesis space, yielding tighter generalization bounds and establishing a guarantee against pseudo-equations. Experimentally, PG-SR outperforms state-of-the-art baselines across diverse domains, maintaining robustness to varying prior quality, noisy data, and data scarcity.
[825] R-Diverse: Mitigating Diversity Illusion in Self-Play LLM Training
Gengsheng Li, Jinghan He, Shijie Wang, Dan Zhang, Ruiqi Liu, Renrui Zhang, Zijun Yao, Junfeng Fang, Haiyun Guo, Jinqiao Wang
Main category: cs.LG
TL;DR: R-Diverse improves self-play LLM reasoning by addressing Diversity Illusion with memory-augmented penalty and skill-aware measurement to sustain iterative improvement.
Details
Motivation: Existing self-play frameworks like R-Zero show non-sustained improvement where early gains degrade over iterations due to Diversity Illusion - questions appear diverse but collapse into recurring patterns, limiting reasoning skill expansion.Method: Proposes R-Diverse with two innovations: 1) Memory-Augmented Penalty (MAP) uses persistent memory bank to discourage question recycling across iterations, 2) Skill-Aware Measurement (SAM) evaluates diversity by reasoning skills exercised rather than surface variation of questions.
Result: Across 10 math and general reasoning benchmarks, R-Diverse sustains gains over more iterations and consistently outperforms prior self-play methods.
Conclusion: Addressing Diversity Illusion through MAP and SAM enables sustained improvement in self-play LLM reasoning frameworks, advancing iterative skill expansion.
Abstract: Self-play bootstraps LLM reasoning through an iterative Challenger-Solver loop: the Challenger is trained to generate questions that target the Solver’s capabilities, and the Solver is optimized on the generated data to expand its reasoning skills. However, existing frameworks like R-Zero often exhibit non-sustained improvement, where early gains degrade as self-play continues. We identify a key failure mode, Diversity Illusion, where the Solver’s training signals appear diverse yet collapse into recurring underlying patterns. It manifests as (1) Local Diversity Illusion, where diversity is enforced only within-batch, inducing cross-iteration mode cycling; and (2) Surface Diversity Illusion, where questions vary superficially but require near-identical reasoning skills. To mitigate them, we propose R-Diverse with two aligned innovations: Memory-Augmented Penalty (MAP), which uses a persistent memory bank to discourage recycling across iterations, and Skill-Aware Measurement (SAM), which evaluates diversity by the reasoning skills exercised rather than surface variation of questions. Across 10 math and general reasoning benchmarks, R-Diverse sustains gains over more iterations and consistently outperforms prior self-play methods. Code is available at https://github.com/Gengsheng-Li/R-Diverse.
cs.MA
[826] Agent Mars: Multi-Agent Simulation for Multi-Planetary Life Exploration and Settlement
Ziyang Wang
Main category: cs.MA
TL;DR: Agent Mars is a multi-agent simulation framework for Mars base operations that models hierarchical coordination among 93 specialized agents across command layers with audit trails, failover mechanisms, and performance metrics.
Details
Motivation: Space exploration presents unique challenges not found on Earth: delayed communications, extreme resource scarcity, heterogeneous expertise, and strict safety requirements. There's a need for auditable coordination systems that can manage specialized humans, robots, and digital services in safety-critical space environments.Method: Developed an open, end-to-end multi-agent simulation framework with 93 agents across seven command/execution layers. Implements hierarchical coordination with audit trails, dynamic role handover with failover, phase-dependent leadership, scenario-aware memory, propose-vote consensus, and translator-mediated protocols.
Result: Created the Agent Mars Performance Index (AMPI) for interpretable performance measurement. Tested across 13 Mars-relevant operational scripts, revealing coordination trade-offs and identifying regimes where cross-layer collaboration and functional leadership reduce overhead without sacrificing reliability.
Conclusion: Agent Mars provides a benchmarkable, auditable foundation for Space AI research, enabling realistic studies of multi-agent coordination in space exploration scenarios beyond toy settings.
Abstract: Artificial Intelligence (AI) has transformed robotics, healthcare, industry, and scientific discovery, yet a major frontier may lie beyond Earth. Space exploration and settlement offer vast environments and resources, but impose constraints unmatched on Earth: delayed/intermittent communications, extreme resource scarcity, heterogeneous expertise, and strict safety, accountability, and command authority. The key challenge is auditable coordination among specialised humans, robots, and digital services in a safety-critical system-of-systems. We introduce Agent Mars, an open, end-to-end multi-agent simulation framework for Mars base operations. Agent Mars formalises a realistic organisation with a 93-agent roster across seven layers of command and execution (human roles and physical assets), enabling base-scale studies beyond toy settings. It implements hierarchical and cross-layer coordination that preserves chain-of-command while allowing vetted cross-layer exchanges with audit trails; supports dynamic role handover with automatic failover under outages; and enables phase-dependent leadership for routine operations, emergencies, and science campaigns. Agent Mars further models mission-critical mechanisms-scenario-aware short/long-horizon memory, configurable propose-vote consensus, and translator-mediated heterogeneous protocols-to capture how teams align under stress. To quantify behaviour, we propose the Agent Mars Performance Index (AMPI), an interpretable composite score with diagnostic sub-metrics. Across 13 reproducible Mars-relevant operational scripts, Agent Mars reveals coordination trade-offs and identifies regimes where curated cross-layer collaboration and functional leadership reduce overhead without sacrificing reliability. Agent Mars provides a benchmarkable, auditable foundation for Space AI.
[827] Adaptive Value Decomposition: Coordinating a Varying Number of Agents in Urban Systems
Yexin Li, Jinjin Guo, Haoyu Zhang, Yuhan Zhao, Yiwen Sun, Zihao Jiao
Main category: cs.MA
TL;DR: AVD is a cooperative multi-agent reinforcement learning framework that adapts to dynamically changing agent populations and mitigates action homogenization in semi-MARL settings with asynchronous decision-making.
Details
Motivation: Traditional MARL methods assume fixed agent numbers and synchronous actions, which don't hold in real urban systems where agent populations vary and actions have heterogeneous durations, creating semi-MARL challenges. Shared policy parameters can cause action homogenization when agents with similar observations act concurrently, degrading coordination quality.Method: Proposes Adaptive Value Decomposition (AVD) with: 1) Adaptation to dynamically changing agent populations, 2) Lightweight mechanism to mitigate action homogenization from shared policies, encouraging behavioral diversity, 3) Training-execution strategy for semi-MARL setting accommodating asynchronous decision-making when agents act at different times.
Result: Experiments on real-world bike-sharing redistribution tasks in London and Washington, D.C. show AVD outperforms state-of-the-art baselines, confirming effectiveness and generalizability.
Conclusion: AVD successfully addresses challenges of dynamic agent populations and action homogenization in semi-MARL settings, demonstrating practical value for urban coordination tasks like bike-sharing redistribution.
Abstract: Multi-agent reinforcement learning (MARL) provides a promising paradigm for coordinating multi-agent systems (MAS). However, most existing methods rely on restrictive assumptions, such as a fixed number of agents and fully synchronous action execution. These assumptions are often violated in urban systems, where the number of active agents varies over time, and actions may have heterogeneous durations, resulting in a semi-MARL setting. Moreover, while sharing policy parameters among agents is commonly adopted to improve learning efficiency, it can lead to highly homogeneous actions when a subset of agents make decisions concurrently under similar observations, potentially degrading coordination quality. To address these challenges, we propose Adaptive Value Decomposition (AVD), a cooperative MARL framework that adapts to a dynamically changing agent population. AVD further incorporates a lightweight mechanism to mitigate action homogenization induced by shared policies, thereby encouraging behavioral diversity and maintaining effective cooperation among agents. In addition, we design a training-execution strategy tailored to the semi-MARL setting that accommodates asynchronous decision-making when some agents act at different times. Experiments on real-world bike-sharing redistribution tasks in two major cities, London and Washington, D.C., demonstrate that AVD outperforms state-of-the-art baselines, confirming its effectiveness and generalizability.
[828] PeroMAS: A Multi-agent System of Perovskite Material Discovery
Yishu Wang, Wei Liu, Yifan Li, Shengxiang Xu, Xujie Yuan, Ran Li, Yuyu Luo, Jia Zhu, Shimin Di, Min-Ling Zhang, Guixiang Li
Main category: cs.MA
TL;DR: PeroMAS is a multi-agent system for perovskite solar cell material discovery that integrates various perovskite-specific tools into a unified workflow, enabling end-to-end optimization from literature retrieval to experimental synthesis.
Details
Motivation: Current AI approaches for perovskite solar cells focus on discrete models (material design, process optimization, property prediction) that fail to propagate physical constraints across the workflow, hindering end-to-end optimization. There's a need for a comprehensive system that covers the entire discovery process.Method: Developed PeroMAS, a multi-agent system that encapsulates perovskite-specific tools into Model Context Protocols (MCPs). The system plans and invokes these tools to design perovskite materials under multi-objective constraints, covering literature retrieval, data extraction, property prediction, and mechanism analysis.
Result: PeroMAS significantly enhances discovery efficiency compared to single LLM or traditional search strategies, successfully identifying candidate materials satisfying multi-objective constraints. Real synthesis experiments verified the system’s effectiveness in the physical world.
Conclusion: PeroMAS provides an effective multi-agent framework for perovskite material discovery that integrates various tools into a cohesive workflow, enabling end-to-end optimization and demonstrating practical utility through real-world experimental validation.
Abstract: As a pioneer of the third-generation photovoltaic revolution, Perovskite Solar Cells (PSCs) are renowned for their superior optoelectronic performance and cost potential. The development process of PSCs is precise and complex, involving a series of closed-loop workflows such as literature retrieval, data integration, experimental design, and synthesis. However, existing AI perovskite approaches focus predominantly on discrete models, including material design, process optimization,and property prediction. These models fail to propagate physical constraints across the workflow, hindering end-to-end optimization. In this paper, we propose a multi-agent system for perovskite material discovery, named PeroMAS. We first encapsulated a series of perovskite-specific tools into Model Context Protocols (MCPs). By planning and invoking these tools, PeroMAS can design perovskite materials under multi-objective constraints, covering the entire process from literature retrieval and data extraction to property prediction and mechanism analysis. Furthermore, we construct an evaluation benchmark by perovskite human experts to assess this multi-agent system. Results demonstrate that, compared to single Large Language Model (LLM) or traditional search strategies, our system significantly enhances discovery efficiency. It successfully identified candidate materials satisfying multi-objective constraints. Notably, we verify PeroMAS’s effectiveness in the physical world through real synthesis experiments.
[829] Robust Mean-Field Games with Risk Aversion and Bounded Rationality
Bhavini Jeloka, Yue Guan, Panagiotis Tsiotras
Main category: cs.MA
TL;DR: The paper introduces MF-RQE, a new equilibrium concept combining risk aversion with respect to initial population distribution and bounded rationality (quantal response) in mean-field games, with scalable RL algorithms for large state-action spaces.
Details
Motivation: Existing mean-field game approaches assume fixed initial population distributions and fully rational agents, limiting robustness under distributional uncertainty and cognitive constraints. The paper aims to address these limitations.Method: Introduces risk aversion with respect to initial population distribution and incorporates bounded rationality via quantal response equilibrium. Develops MF-RQE equilibrium concept with existence proofs, convergence analysis (fixed-point iteration, fictitious play), and scalable reinforcement learning algorithms for large state-action spaces.
Result: MF-RQE policies demonstrate improved robustness compared to classical mean-field approaches that optimize expected cumulative rewards under fixed initial distributions and are restricted to entropy-based regularizers.
Conclusion: The proposed MF-RQE framework provides a more general equilibrium concept that addresses distributional uncertainty and cognitive constraints in large-scale multi-agent systems, with practical algorithms for implementation.
Abstract: Recent advances in mean-field game literature enable the reduction of large-scale multi-agent problems to tractable interactions between a representative agent and a population distribution. However, existing approaches typically assume a fixed initial population distribution and fully rational agents, limiting robustness under distributional uncertainty and cognitive constraints. We address these limitations by introducing risk aversion with respect to the initial population distribution and by incorporating bounded rationality to model deviations from fully rational decision-making agents. The combination of these two elements yields a new and more general equilibrium concept, which we term the mean-field risk-averse quantal response equilibrium (MF-RQE). We establish existence results and prove convergence of fixed-point iteration and fictitious play to MF-RQE. Building on these insights, we develop a scalable reinforcement learning algorithm for scenarios with large state-action spaces. Numerical experiments demonstrate that MF-RQE policies achieve improved robustness relative to classical mean-field approaches that optimize expected cumulative rewards under a fixed initial distribution and are restricted to entropy-based regularizers.
[830] G2CP: A Graph-Grounded Communication Protocol for Verifiable and Efficient Multi-Agent Reasoning
Karim Ben Khaled, Davy Monticolo
Main category: cs.MA
TL;DR: G2CP introduces a graph-based communication protocol for multi-agent LLM systems, replacing natural language with structured graph operations to reduce semantic drift and improve efficiency.
Details
Motivation: Current multi-agent LLM systems suffer from semantic drift, hallucination propagation, and inefficient token consumption due to natural language communication between agents.Method: G2CP uses structured graph operations (traversal commands, subgraph fragments, update operations) over shared knowledge graphs instead of free text for agent communication.
Result: Reduces inter-agent communication tokens by 73%, improves task completion accuracy by 34% over free-text baselines, eliminates cascading hallucinations, and produces auditable reasoning chains in industrial scenarios.
Conclusion: G2CP represents a fundamental shift from linguistic to structural communication in multi-agent systems, enabling precise agent coordination with verifiable reasoning traces.
Abstract: Multi-agent systems powered by Large Language Models face a critical challenge: agents communicate through natural language, leading to semantic drift, hallucination propagation, and inefficient token consumption. We propose G2CP (Graph-Grounded Communication Protocol), a structured agent communication language where messages are graph operations rather than free text. Agents exchange explicit traversal commands, subgraph fragments, and update operations over a shared knowledge graph, enabling verifiable reasoning traces and eliminating ambiguity. We validate G2CP within an industrial knowledge management system where specialized agents (Diagnostic, Procedural, Synthesis, and Ingestion) coordinate to answer complex queries. Experimental results on 500 industrial scenarios and 21 real-world maintenance cases show that G2CP reduces inter-agent communication tokens by 73%, improves task completion accuracy by 34% over free-text baselines, eliminates cascading hallucinations, and produces fully auditable reasoning chains. G2CP represents a fundamental shift from linguistic to structural communication in multi-agent systems, with implications for any domain requiring precise agent coordination. Code, data, and evaluation scripts are publicly available.
[831] MAS-on-the-Fly: Dynamic Adaptation of LLM-based Multi-Agent Systems at Test Time
Guangyi Liu, Haojun Lin, Huan Zeng, Heng Wang, Quanming Yao
Main category: cs.MA
TL;DR: MASFly: A dynamic multi-agent framework that enables LLM-based systems to adapt at test time through retrieval-augmented SOP instantiation and experience-guided supervision.
Details
Motivation: Existing LLM-based multi-agent systems lack dynamic adaptability after deployment, relying on manual designs or one-size-fits-all automation. The paper aims to create systems that can adapt like biological systems during execution.Method: Two key mechanisms: 1) Retrieval-augmented SOP instantiation - uses a self-constructed repository of successful collaboration patterns to assemble customized MASs for new queries. 2) Experience-guided supervision - a Watcher agent monitors system behaviors using a personalized experience pool and provides real-time interventions.
Result: MASFly achieves state-of-the-art performance with 61.7% success rate on TravelPlanner benchmark, demonstrating strong task adaptability and robustness.
Conclusion: MASFly enables dynamic adaptation in LLM-based multi-agent systems at test time, outperforming existing approaches and showing promise for complex task solving.
Abstract: Large Language Model (LLM)-based multi-agent systems (MAS) have emerged as a promising paradigm for solving complex tasks. However, existing works often rely on manual designs or “one-size-fits-all” automation, lacking dynamic adaptability after deployment. Inspired by how biological systems adapt, we introduce MASFly, a novel multi-agent framework enabling dynamic adaptation at test time. To adapt system generation, MASFly employs a retrieval-augmented SOP instantiation mechanism that leverages a self-constructed repository of successful collaboration patterns, enabling the LLM to assemble customized MASs for new queries. For adaptive execution, MASFly incorporates an experience-guided supervision mechanism, where a dedicated Watcher agent monitors system behaviors with reference to a personalized experience pool and provides real-time interventions. Extensive experiments demonstrate that MASFly achieves state-of-the-art performance, most notably a 61.7% success rate on the TravelPlanner benchmark, while exhibiting strong task adaptability and robustness.
[832] Testing BDI-based Multi-Agent Systems using Discrete Event Simulation
Martina Baiardi, Samuele Burattini, Giovanni Ciatto, Danilo Pianini
Main category: cs.MA
TL;DR: Paper presents a simulation-based testing framework for BDI multi-agent systems by mapping agent control flow to discrete event simulation at different granularities to bridge reality gap.
Details
Motivation: Testing multi-agent systems is challenging due to unpredictable dynamics. Simulation helps but achieving fidelity is difficult, especially for cognitive agent models like BDI where agent code can't run unchanged in simulation, creating a reality gap between deployed and simulated systems.Method: Proposes mapping Belief Desire Intention (BDI) agent control flow onto Discrete Event Simulation (DES) at different granularities. Implements an open-source prototype integration between JaKtA and Alchemist tools to create a simulation-based testing environment for distributed BDI agents.
Result: Demonstrates that integration between BDI agents and DES is possible at different granularities, and that different mapping granularities lead to different degrees of simulation fidelity.
Conclusion: BDI developers can test the same specification in simulation that will be deployed, without surrogate representations, by mapping agent control flow to DES with varying granularity to achieve desired fidelity levels.
Abstract: Multi-agent systems are designed to deal with open, distributed systems with unpredictable dynamics, which makes them inherently hard to test. The value of using simulation for this purpose is recognized in the literature, although achieving sufficient fidelity (i.e., the degree of similarity between the simulation and the real-world system) remains a challenging task. This is exacerbated when dealing with cognitive agent models, such as the Belief Desire Intention (BDI) model, where the agent codebase is not suitable to run unchanged in simulation environments, thus increasing the reality gap between the deployed and simulated systems. We argue that BDI developers should be able to test in simulation the same specification that will be later deployed, with no surrogate representations. Thus, in this paper, we discuss how the control flow of BDI agents can be mapped onto a Discrete Event Simulation (DES), showing that such integration is possible at different degrees of granularity. We substantiate our claims by producing an open-source prototype integration between two pre-existing tools (JaKtA and Alchemist), showing that it is possible to produce a simulation-based testing environment for distributed BDI} agents, and that different granularities in mapping BDI agents over DESs may lead to different degrees of fidelity.
[833] Socially-Weighted Alignment: A Game-Theoretic Framework for Multi-Agent LLM Systems
Furkan Mumcu, Yasin Yilmaz
Main category: cs.MA
TL;DR: SWA is a game-theoretic framework that balances individual agent objectives with group welfare through a social weight parameter, preventing congestion in shared LLM agent environments.
Details
Motivation: The paper addresses the tension between individual alignment and collective stability when deploying LLM agents in shared environments, where locally rational decisions can create negative externalities that degrade system-level performance.Method: Proposes Socially-Weighted Alignment (SWA), a game-theoretic framework that modifies inference-time decision making by interpolating between an agent’s private objective and group welfare via a social weight parameter λ. The method includes an inference-time algorithmic instantiation that doesn’t require parameter updates or multi-agent reinforcement learning.
Result: In a shared-resource congestion game with n agents and congestion severity β, SWA induces a critical threshold λ* = (n-β)/(n-1) above which agents no longer have marginal incentive to increase demand under overload, yielding a phase transition from persistent congestion to stable operation near capacity. Multi-agent simulations empirically validate the predicted threshold behavior.
Conclusion: SWA provides a practical framework for achieving collective stability in shared LLM agent environments through inference-time adjustments that balance individual and group objectives, without requiring complex training procedures.
Abstract: Deploying large language model (LLM) agents in shared environments introduces a fundamental tension between individual alignment and collective stability: locally rational decisions can impose negative externalities that degrade system-level performance. We propose Socially-Weighted Alignment (SWA), a game-theoretic framework that modifies inference-time decision making by interpolating between an agent’s private objective and an estimate of group welfare via a social weight $λ\in[0,1]$. In a shared-resource congestion game with $n$ agents and congestion severity $β$, we show that SWA induces a critical threshold $λ^*=(n-β)/(n-1)$ above which agents no longer have marginal incentive to increase demand under overload, yielding a phase transition from persistent congestion to stable operation near capacity. We further provide an inference-time algorithmic instantiation of SWA that does not require parameter updates or multi-agent reinforcement learning, and use a multi-agent simulation to empirically validate the predicted threshold behavior.
[834] Towards Selection as Power: Bounding Decision Authority in Autonomous Agents
Jose Manuel de la Chica Rodriguez, Juan Manuel Vera Díaz
Main category: cs.MA
TL;DR: A governance architecture for autonomous agents that separates cognition, selection, and action into distinct domains, bounding selection authority through mechanical primitives to prevent deterministic outcome capture while preserving reasoning capacity.
Details
Motivation: Existing safety approaches (alignment, interpretability, action-level filtering) are insufficient for regulated, high-stakes domains because they don't govern selection power - the authority to determine which options are generated, surfaced, and framed for decision. Need mechanisms to prevent silent failures in irreversible, institutionally constrained decisions.Method: Proposes a governance architecture that models autonomy as a vector of sovereignty: cognitive autonomy (unconstrained), selection autonomy (bounded), and action autonomy (bounded). Uses mechanical primitives operating outside agent’s optimization space: external candidate generation (CEFL), governed reducer, commit-reveal entropy isolation, rationale validation, and fail-loud circuit breakers.
Result: Evaluated across regulated financial scenarios under adversarial stress targeting variance manipulation, threshold gaming, framing skew, ordering effects, and entropy probing. Results show mechanical selection governance is implementable, auditable, prevents deterministic outcome capture while preserving reasoning capacity. Bounds selection authority relative to conventional scalar pipelines.
Conclusion: Reframes governance as bounded causal power rather than internal intent alignment. Offers foundation for deploying autonomous agents where silent failure is unacceptable. Architecture measurably bounds selection authority while preserving cognitive capabilities.
Abstract: Autonomous agentic systems are increasingly deployed in regulated, high-stakes domains where decisions may be irreversible and institutionally constrained. Existing safety approaches emphasize alignment, interpretability, or action-level filtering. We argue that these mechanisms are necessary but insufficient because they do not directly govern selection power: the authority to determine which options are generated, surfaced, and framed for decision. We propose a governance architecture that separates cognition, selection, and action into distinct domains and models autonomy as a vector of sovereignty. Cognitive autonomy remains unconstrained, while selection and action autonomy are bounded through mechanically enforced primitives operating outside the agent’s optimization space. The architecture integrates external candidate generation (CEFL), a governed reducer, commit-reveal entropy isolation, rationale validation, and fail-loud circuit breakers. We evaluate the system across multiple regulated financial scenarios under adversarial stress targeting variance manipulation, threshold gaming, framing skew, ordering effects, and entropy probing. Metrics quantify selection concentration, narrative diversity, governance activation cost, and failure visibility. Results show that mechanical selection governance is implementable, auditable, and prevents deterministic outcome capture while preserving reasoning capacity. Although probabilistic concentration remains, the architecture measurably bounds selection authority relative to conventional scalar pipelines. This work reframes governance as bounded causal power rather than internal intent alignment, offering a foundation for deploying autonomous agents where silent failure is unacceptable.
[835] ST-EVO: Towards Generative Spatio-Temporal Evolution of Multi-Agent Communication Topologies
Xingjian Wu, Xvyuan Liu, Junkai Lu, Siyuan Wang, Yang Shu, Jilin Hu, Chenjuan Guo, Bin Yang
Main category: cs.MA
TL;DR: ST-EVO is a spatio-temporal evolving multi-agent system that uses flow-matching for dynamic communication scheduling between LLM agents, improving collaborative intelligence through uncertainty-aware self-feedback mechanisms.
Details
Motivation: Current self-evolving multi-agent systems focus only on spatial or temporal evolution separately, limiting LLMs' collaborative potential. The authors aim to create a more comprehensive spatio-temporal evolving system that can dynamically schedule communications and learn from experience.Method: ST-EVO introduces a flow-matching based scheduler for dialogue-wise communication scheduling between LLM agents. It incorporates uncertainty perception of the multi-agent system and self-feedback mechanisms to learn from accumulated experience, enabling precise spatio-temporal scheduling.
Result: Extensive experiments on nine benchmarks show state-of-the-art performance with 5%–25% accuracy improvement compared to existing methods.
Conclusion: ST-EVO demonstrates that spatio-temporal evolving multi-agent systems with flow-matching schedulers and self-feedback mechanisms significantly enhance LLM collaboration capabilities beyond single-dimension evolving approaches.
Abstract: LLM-powered Multi-Agent Systems (MAS) have emerged as an effective approach towards collaborative intelligence, and have attracted wide research interests. Among them, ``self-evolving’’ MAS, treated as a more flexible and powerful technical route, can construct task-adaptive workflows or communication topologies, instead of relying on a predefined static structue template. Current self-evolving MAS mainly focus on Spatial Evolving or Temporal Evolving paradigm, which only considers the single dimension of evolution and does not fully incentivize LLMs’ collaborative capability. In this work, we start from a novel Spatio-Temporal perspective by proposing ST-EVO, which supports dialogue-wise communication scheduling with a compact yet powerful flow-matching based Scheduler. To make precise Spatio-Temporal scheduling, ST-EVO can also perceive the uncertainty of MAS, and possesses self-feedback ability to learn from accumulated experience. Extensive experiments on nine benchmarks demonstrate the state-of-the-art performance of ST-EVO, achieving about 5%–25% accuracy improvement.
[836] ROSA: Roundabout Optimized Speed Advisory with Multi-Agent Trajectory Prediction in Multimodal Traffic
Anna-Lena Schlamp, Jeremias Gerner, Klaus Bogenberger, Werner Huber, Stefanie Schmidtner
Main category: cs.MA
TL;DR: ROSA is a system that combines multi-agent trajectory prediction with coordinated speed guidance for mixed traffic at roundabouts, using Transformer-based prediction to generate proactive speed advisories for safety and efficiency.
Details
Motivation: The paper addresses the challenge of coordinating multimodal, mixed traffic (vehicles and Vulnerable Road Users) at roundabouts, where complex interactions create safety and efficiency issues that traditional systems struggle to handle.Method: Uses a Transformer-based model for joint trajectory prediction of vehicles and VRUs, trained for single-step prediction and deployed autoregressively. Incorporates motion dynamics and route intention data, then generates speed advisories based on predicted conflicts.
Result: Achieves high accuracy (ADE: 1.29m, FDE: 2.99m at 5-second horizon), improves with route intention (ADE: 1.10m, FDE: 2.36m). ROSA significantly improves vehicle efficiency and safety, with positive effects on perceived safety from VRU perspective.
Conclusion: ROSA demonstrates that combining multi-agent trajectory prediction with proactive speed guidance can effectively address roundabout safety and efficiency challenges, showing the value of connected vehicle data for mixed traffic coordination.
Abstract: We present ROSA – Roundabout Optimized Speed Advisory – a system that combines multi-agent trajectory prediction with coordinated speed guidance for multimodal, mixed traffic at roundabouts. Using a Transformer-based model, ROSA jointly predicts the future trajectories of vehicles and Vulnerable Road Users (VRUs) at roundabouts. Trained for single-step prediction and deployed autoregressively, it generates deterministic outputs, enabling actionable speed advisories. Incorporating motion dynamics, the model achieves high accuracy (ADE: 1.29m, FDE: 2.99m at a five-second prediction horizon), surpassing prior work. Adding route intention further improves performance (ADE: 1.10m, FDE: 2.36m), demonstrating the value of connected vehicle data. Based on predicted conflicts with VRUs and circulating vehicles, ROSA provides real-time, proactive speed advisories for approaching and entering the roundabout. Despite prediction uncertainty, ROSA significantly improves vehicle efficiency and safety, with positive effects even on perceived safety from a VRU perspective. The source code of this work is available under: github.com/urbanAIthi/ROSA.
[837] Distributed Quantum Gaussian Processes for Multi-Agent Systems
Meet Gandhi, George P. Kontoudis
Main category: cs.MA
TL;DR: Distributed Quantum Gaussian Process (DQGP) method using quantum computing for enhanced kernel expressivity in multi-agent settings, with distributed Riemannian ADMM for optimization.
Details
Motivation: Classical Gaussian Processes have limited expressivity with traditional kernels, especially for complex, large-scale real-world problems. Quantum computing offers potential to embed data into exponentially large Hilbert spaces to capture complex correlations inaccessible to classical approaches.Method: Proposed Distributed Quantum Gaussian Process (DQGP) method in multi-agent setting. Developed Distributed consensus Riemannian Alternating Direction Method of Multipliers (DR-ADMM) algorithm to aggregate local agent models into global model, addressing non-Euclidean optimization challenges.
Result: Evaluated efficacy through numerical experiments on quantum simulator using classical hardware. Tested on real-world NASA Shuttle Radar Topography Mission elevation datasets and synthetic datasets generated by Quantum Gaussian Processes.
Conclusion: Framework demonstrates modeling advantages and highlights potential computational speedups that quantum hardware may provide, particularly for Gaussian processes and distributed optimization problems.
Abstract: Gaussian Processes (GPs) are a powerful tool for probabilistic modeling, but their performance is often constrained in complex, largescale real-world domains due to the limited expressivity of classical kernels. Quantum computing offers the potential to overcome this limitation by embedding data into exponentially large Hilbert spaces, capturing complex correlations that remain inaccessible to classical computing approaches. In this paper, we propose a Distributed Quantum Gaussian Process (DQGP) method in a multiagent setting to enhance modeling capabilities and scalability. To address the challenging non-Euclidean optimization problem, we develop a Distributed consensus Riemannian Alternating Direction Method of Multipliers (DR-ADMM) algorithm that aggregates local agent models into a global model. We evaluate the efficacy of our method through numerical experiments conducted on a quantum simulator in classical hardware. We use real-world, non-stationary elevation datasets of NASA’s Shuttle Radar Topography Mission and synthetic datasets generated by Quantum Gaussian Processes. Beyond modeling advantages, our framework highlights potential computational speedups that quantum hardware may provide, particularly in Gaussian processes and distributed optimization.
[838] QD-MAPPER: A Quality Diversity Framework to Automatically Evaluate Multi-Agent Path Finding Algorithms in Diverse Maps
Cheng Qian, Yulun Zhang, Varun Bhatt, Matthew Christopher Fontaine, Stefanos Nikolaidis, Jiaoyang Li
Main category: cs.MA
TL;DR: QD-MAPPER uses Quality Diversity algorithms with Neural Cellular Automata to generate diverse maps for evaluating Multi-Agent Path Finding algorithms, addressing overfitting to limited human-designed maps.
Details
Motivation: Current MAPF algorithm evaluation relies on limited human-designed maps, which may not cover all scenarios and can lead to algorithm overfitting. There's a need for systematic evaluation on diverse maps to better understand algorithm performance and enable fair comparisons.Method: Proposes QD-MAPPER framework combining Quality Diversity algorithms with Neural Cellular Automata to automatically generate diverse maps. Uses QD to explore map space and NCA for map generation, enabling comprehensive evaluation of different MAPF algorithm types (search-based, priority-based, rule-based, learning-based).
Result: Enables identification of patterns where each MAPF algorithm excels and detection of runtime/success rate disparities between algorithms through both single-algorithm experiments and algorithm comparisons.
Conclusion: QD-MAPPER provides a general framework for comprehensive MAPF algorithm evaluation, offering insights for algorithm selection and design improvements through systematic map diversity generation.
Abstract: We use the Quality Diversity (QD) algorithm with Neural Cellular Automata (NCA) to automatically evaluate Multi-Agent Path Finding (MAPF) algorithms by generating diverse maps. Previously, researchers typically evaluate MAPF algorithms on a set of specific, human-designed maps at their initial stage of algorithm design. However, such fixed maps may not cover all scenarios, and algorithms may overfit to the small set of maps. To seek further improvements, systematic evaluations on a diverse suite of maps are needed. In this work, we propose Quality-Diversity Multi-Agent Path Finding Performance EvaluatoR (QD-MAPPER), a general framework that takes advantage of the QD algorithm to comprehensively understand the performance of MAPF algorithms by generating maps with patterns, be able to make fair comparisons between two MAPF algorithms, providing further information on the selection between two algorithms and on the design of the algorithms. Empirically, we employ this technique to evaluate and compare the behavior of different types of MAPF algorithms, including search-based, priority-based, rule-based, and learning-based algorithms. Through both single-algorithm experiments and comparisons between algorithms, researchers can identify patterns that each MAPF algorithm excels and detect disparities in runtime or success rates between different algorithms.
[839] Modeling AI-Human Collaboration as a Multi-Agent Adaptation
Prothit Sen, Sai Mihir Jakkaraju
Main category: cs.MA
TL;DR: AI-human collaboration simulation shows task architecture (modular vs sequenced, interdependence) determines complementarity more than industry context, with sequenced tasks benefiting from human-first AI-refinement approach.
Details
Motivation: To understand how AI and humans with different decision heuristics (optimization-based AI vs satisficing-based human adaptation) can effectively collaborate, and to identify design principles for AI-human systems across organizational settings.Method: Agent-based simulation using NK model to examine interactions between AI search and human adaptation across modular and sequenced task structures, varying search breadth, task complexity, and interdependence.
Result: For modular tasks: AI typically substitutes humans, but complementarity emerges with moderate AI search breadth and low human task complexity. For sequenced tasks: human-first AI-refinement maximizes performance (contradicting AI-first design), while AI-first human-following attenuates complementarity with increasing interdependence. Memory-less random AI can help low-capability humans escape local optima.
Conclusion: Effective AI-human collaboration depends primarily on task architecture (division of labor, sequencing, interdependence) rather than industry context. Task decomposition should be the central design principle for strategic decision-making with agentic AI.
Abstract: We formalize AI-human collaboration through an agent-based simulation that distinguishes optimization-based AI search from satisficing-based human adaptation. Using an NK model, we examine how these distinct decision heuristics interact across modular and sequenced task structures. For modular tasks, AI typically substitutes for humans, yet complementarities emerge when AI explores a moderately broad search space and human task complexity remains low. In sequenced tasks, we uncover a counterintuitive result: when a high-performing human initiates search and AI subsequently refines it, joint performance is maximized, contradicting the dominant AI-first design principle. Conversely, when AI leads and human satisficing follows, complementarities attenuate as task interdependence increases. We further show that memory-less random AI, despite lacking structured adaptation, can improve outcomes when augmenting low-capability humans by enabling escape from local optima. Collectively, our findings reveal that effective AI-human collaboration depends less on industry context and more on task architecture: the division of labor, sequencing, and interdependence structure. By elevating task decomposition as the central design principle, we provide a generalizable framework for strategic decision-making involving agentic AI across diverse organizational settings.
[840] HEAS: Hierarchical Evolutionary Agent Simulation Framework for Cross-Scale Modeling and Multi-Objective Search
Ruiyu Zhang, Lin Nie, Xin Zhao
Main category: cs.MA
TL;DR: HEAS is a Python framework that combines agent-based modeling with evolutionary optimization and tournament evaluation in a unified workflow for reproducible multi-level simulations.
Details
Motivation: To create a unified framework that bridges agent-based modeling, evolutionary optimization, and tournament evaluation to enable reproducible cross-scale simulations with explicit couplings between different levels of analysis.Method: HEAS uses hierarchical lightweight processes (“streams”) scheduled in deterministic layers that read/write to a shared context. It provides APIs for simulation, optimization (single/multi-objective evolution), PyTorch policy integration via parameter flattening, and tournament evaluation with custom scoring rules.
Result: The framework standardizes evaluation metrics, persists seeds/logbooks/hall-of-fame archives, provides plotting tools, and reduces glue code while improving comparability across studies. It enables composition of exogenous drivers, endogenous agents, and aggregators without refactoring.
Conclusion: HEAS offers a practical foundation for cross-disciplinary, multi-level inquiry that yields reliable, reproducible results, demonstrated through ecological and enterprise decision-making examples.
Abstract: Hierarchical Evolutionary Agent Simulation (HEAS) is a Python framework that unifies layered agent-based modeling with evolutionary optimization and tournament evaluation in a single, reproducible workflow. HEAS represents models as hierarchies of lightweight processes (“streams”) scheduled in deterministic layers that read and write a shared context, making cross-scale couplings explicit and auditable. A compact API and CLI-simulate, optimize, evaluate-expose single- and multi-objective evolution, PyTorch policy integration via parameter flattening/unflattening, and general tournament tooling with user-defined scoring and voting rules. The framework standardizes evaluation through uniform per-step and episode metrics, persists seeds, logbooks, and hall-of-fame archives, and provides plotting helpers for traces, Pareto fronts, and comparative outcomes, reducing glue code and improving comparability across studies. HEAS emphasizes separation of mechanism from orchestration, allowing exogenous drivers, endogenous agents, and aggregators to be composed and swapped without refactoring, while the same model can be used for forward simulation, optimization, or systematic comparison. We illustrate usage with two compact examples-an ecological system and an enterprise decision-making setting. HEAS offers a practical foundation for cross-disciplinary, multi-level inquiry, yielding reliable, reproducible results.
[841] Heterogeneous RBCs via Deep Multi-Agent Reinforcement Learning
Federico Gabriele, Aldo Glielmo, Marco Taboga
Main category: cs.MA
TL;DR: MARL-BC integrates multi-agent reinforcement learning with real business cycle models to bridge heterogeneous-agent GE models and agent-based models, enabling rich agent heterogeneity while recovering traditional macroeconomic results.
Details
Motivation: Current macroeconomic models face limitations: heterogeneous-agent GE models (HANK/KS) rely on unrealistic rational expectations and are computationally cumbersome, limiting heterogeneity. Agent-based models allow rich heterogeneity but require explicit behavioral rules and trial-and-error development. The paper aims to bridge these paradigms.Method: MARL-BC framework integrates deep multi-agent reinforcement learning (MARL) with real business cycle (RBC) models. The approach uses reinforcement learning agents that learn optimal policies through interaction, rather than requiring explicit behavioral rules or rational expectations assumptions.
Result: The framework can: (1) recover textbook RBC results with single agent; (2) reproduce mean-field Krusell-Smith model results with many identical agents; (3) effectively simulate rich agent heterogeneity, which is challenging for traditional GE approaches. It serves as both an ABM with heterogeneous interacting agents and reproduces GE results in limit cases.
Conclusion: MARL-BC represents a step toward synthesizing heterogeneous-agent GE models and agent-based models, offering a flexible framework that can handle rich agent heterogeneity while recovering traditional macroeconomic results in appropriate limits.
Abstract: Current macroeconomic models with agent heterogeneity can be broadly divided into two main groups. Heterogeneous-agent general equilibrium (GE) models, such as those based on Heterogeneous Agent New Keynesian (HANK) or Krusell-Smith (KS) approaches, rely on GE and ‘rational expectations’, somewhat unrealistic assumptions that make the models very computationally cumbersome, which in turn limits the amount of heterogeneity that can be modelled. In contrast, agent-based models (ABMs) can flexibly encompass a large number of arbitrarily heterogeneous agents, but typically require the specification of explicit behavioural rules, which can lead to a lengthy trial-and-error model-development process. To address these limitations, we introduce MARL-BC, a framework that integrates deep multi-agent reinforcement learning (MARL) with real business cycle (RBC) models. We demonstrate that MARL-BC can: (1) recover textbook RBC results when using a single agent; (2) recover the results of the mean-field KS model using a large number of identical agents; and (3) effectively simulate rich heterogeneity among agents, a hard task for traditional GE approaches. Our framework can be thought of as an ABM if used with a variety of heterogeneous interacting agents, and can reproduce GE results in limit cases. As such, it is a step towards a synthesis of these often opposed modelling paradigms.
[842] Interpreting Emergent Extreme Events in Multi-Agent Systems
Ling Tang, Jilin Mei, Dongrui Liu, Chen Qian, Dawei Cheng, Jing Shao, Xia Hu
Main category: cs.MA
TL;DR: A framework for explaining emergent extreme events in multi-agent systems using Shapley value attribution to identify when events originate, which agents drive them, and what behaviors contribute to them.
Details
Motivation: Multi-agent systems powered by large language models can produce extreme events through complex interactions, but these emergent phenomena remain poorly understood as black-box processes. Understanding the origins of such extreme events is crucial for system safety and reliability.Method: Adapts Shapley value to attribute extreme event occurrence to individual agent actions across time steps, then aggregates attribution scores along time, agent, and behavior dimensions to quantify risk contributions. Designs metrics based on contribution scores to characterize extreme event features.
Result: The framework effectively explains extreme events across diverse multi-agent scenarios (economic, financial, social), providing general insights into the emergence of extreme phenomena through attribution analysis.
Conclusion: Proposes the first systematic framework for interpreting emergent extreme events in multi-agent systems, enabling better understanding of when events originate, which agents drive them, and what behaviors contribute to them, which is critical for system safety.
Abstract: Large language model-powered multi-agent systems have emerged as powerful tools for simulating complex human-like systems. The interactions within these systems often lead to extreme events whose origins remain obscured by the black box of emergence. Interpreting these events is critical for system safety. This paper proposes the first framework for explaining emergent extreme events in multi-agent systems, aiming to answer three fundamental questions: When does the event originate? Who drives it? And what behaviors contribute to it? Specifically, we adapt the Shapley value to faithfully attribute the occurrence of extreme events to each action taken by agents at different time steps, i.e., assigning an attribution score to the action to measure its influence on the event. We then aggregate the attribution scores along the dimensions of time, agent, and behavior to quantify the risk contribution of each dimension. Finally, we design a set of metrics based on these contribution scores to characterize the features of extreme events. Experiments across diverse multi-agent system scenarios (economic, financial, and social) demonstrate the effectiveness of our framework and provide general insights into the emergence of extreme phenomena.
[843] Agent Skills for Large Language Models: Architecture, Acquisition, Security, and the Path Forward
Renjun Xu, Yang Yan
Main category: cs.MA
TL;DR: Survey paper on agent skills - modular, composable packages that enable LLMs to dynamically extend capabilities without retraining, covering architecture, acquisition, deployment, security, and open challenges.
Details
Motivation: The shift from monolithic LLMs to modular, skill-equipped agents represents a fundamental change in deployment. Rather than encoding all knowledge in model weights, agent skills allow dynamic capability extension without retraining, enabling more flexible and scalable AI systems.Method: Comprehensive survey organizing the field along four axes: (1) architectural foundations (SKILL.md specification, progressive context loading, MCP integration), (2) skill acquisition (reinforcement learning with skill libraries, autonomous discovery, compositional synthesis), (3) deployment at scale (computer-use agent stack, GUI grounding, OSWorld/SWE-bench benchmarks), and (4) security analysis and governance framework.
Result: Identifies that 26.1% of community-contributed skills contain vulnerabilities, motivating a proposed Skill Trust and Lifecycle Governance Framework with four-tier permission model. Organizes current landscape and identifies seven open challenges for trustworthy skill ecosystems.
Conclusion: Agent skills represent an emerging abstraction layer for next-generation agentic systems, enabling dynamic capability extension without retraining. The survey provides comprehensive coverage of this rapidly evolving field and proposes research agenda for addressing security, portability, and governance challenges.
Abstract: The transition from monolithic language models to modular, skill-equipped agents marks a defining shift in how large language models (LLMs) are deployed in practice. Rather than encoding all procedural knowledge within model weights, agent skills – composable packages of instructions, code, and resources that agents load on demand – enable dynamic capability extension without retraining. It is formalized in a paradigm of progressive disclosure, portable skill definitions, and integration with the Model Context Protocol (MCP). This survey provides a comprehensive treatment of the agent skills landscape, as it has rapidly evolved during the last few months. We organize the field along four axes: (i) architectural foundations, examining the {SKILL.md} specification, progressive context loading, and the complementary roles of skills and MCP; (ii) skill acquisition, covering reinforcement learning with skill libraries, autonomous skill discovery (SEAgent), and compositional skill synthesis; (iii) deployment at scale, including the computer-use agent (CUA) stack, GUI grounding advances, and benchmark progress on OSWorld and SWE-bench; and (iv) security, where recent empirical analyses reveal that 26.1% of community-contributed skills contain vulnerabilities, motivating our proposed Skill Trust and Lifecycle Governance Framework – a four-tier, gate-based permission model that maps skill provenance to graduated deployment capabilities. We identify seven open challenges – from cross-platform skill portability to capability-based permission models – and propose a research agenda for realizing trustworthy, self-improving skill ecosystems. Unlike prior surveys that broadly cover LLM agents or tool use, this work focuses specifically on the emerging skill abstraction layer and its implications for the next generation of agentic systems. Project repo: https://github.com/scienceaix/agentskills
cs.MM
[844] SRA: Semantic Relation-Aware Flowchart Question Answering
Xinyu Li, Bowei Zou, Yuchong Chen, Yifan Fan, Yu Hong
Main category: cs.MM
TL;DR: SRA FlowchartQA enhances flowchart question answering by using LLMs to detect semantic relations between nodes and implementing interlanguage-controllable reasoning based on question intention.
Details
Motivation: Existing flowchart QA methods convert flowcharts to interlanguages (Graphviz, Mermaid, PlantUML) that capture link relations but miss intricate semantic/logic relationships like Conditional and Causal relations, hindering deep reasoning for complex questions.Method: Proposes Semantic Relation-Aware (SRA) FlowchartQA approach that: 1) Uses LLMs to detect discourse semantic relations between nodes, upgrading link-based interlanguages to semantic relation-based ones; 2) Implements interlanguage-controllable reasoning that analyzes question intention to determine reasoning depth (Shallow/Deep) and selects appropriate interlanguage.
Result: Experiments on FlowVQA benchmark show SRA yields widespread improvements when upgrading different interlanguages like Graphviz, Mermaid and PlantUML.
Conclusion: SRA FlowchartQA effectively addresses limitations of existing methods by incorporating semantic relation detection and adaptive reasoning, improving performance on complex flowchart QA tasks.
Abstract: Flowchart Question Answering (FlowchartQA) is a multi-modal task that automatically answers questions conditioned on graphic flowcharts. Current studies convert flowcharts into interlanguages (e.g., Graphviz) for Question Answering (QA), which effectively bridge modal gaps between questions and flowcharts. More importantly, they reveal the link relations between nodes in the flowchart, facilitating a shallow relation reasoning during tracing answers. However, the existing interlanguages still lose sight of intricate semantic/logic relationships such as Conditional and Causal relations. This hinders the deep reasoning for complex questions. To address the issue, we propose a novel Semantic Relation-Aware (SRA) FlowchartQA approach. It leverages Large Language Model (LLM) to detect the discourse semantic relations between nodes, by which a link-based interlanguage is upgraded to the semantic relation based interlanguage. In addition, we conduct an interlanguage-controllable reasoning process. In this process, the question intention is analyzed with the aim to determine the depth of reasoning (Shallow or Deep reasoning), as well as the well-matched interlanguage. We experiment on the benchmark dataset FlowVQA. The test results show that SRA yields widespread improvements when upgrading different interlanguages like Graphviz, Mermaid and Plantuml
[845] AudioX: A Unified Framework for Anything-to-Audio Generation
Zeyue Tian, Zhaoyang Liu, Yizhu Jin, Ruibin Yuan, Xu Tan, Qifeng Chen, Wei Xue, Yike Guo
Main category: cs.MM
TL;DR: AudioX is a unified framework for multimodal-conditioned audio generation that handles text, video, and audio inputs through a Multimodal Adaptive Fusion module, trained on a large-scale dataset IF-caps with 7M+ samples.
Details
Motivation: The paper addresses two key challenges in multimodal audio generation: 1) lack of a unified modeling framework for diverse multimodal conditions, and 2) scarcity of large-scale, high-quality training data for multimodal-conditioned audio generation.Method: Proposes AudioX framework with a Multimodal Adaptive Fusion module that effectively fuses diverse multimodal inputs (text, video, audio) to enhance cross-modal alignment. Constructs IF-caps dataset with 7M+ samples through structured annotation pipeline for comprehensive supervision.
Result: AudioX achieves superior performance compared to state-of-the-art methods, especially in text-to-audio and text-to-music generation, demonstrating powerful instruction-following potential for multimodal control signals.
Conclusion: AudioX provides an effective unified framework for anything-to-audio generation with multimodal conditions, showing strong performance across various tasks and promising potential for instruction-following audio generation.
Abstract: Audio and music generation based on flexible multimodal control signals is a widely applicable topic, with the following key challenges: 1) a unified multimodal modeling framework, and 2) large-scale, high-quality training data. As such, we propose AudioX, a unified framework for anything-to-audio generation that integrates varied multimodal conditions (i.e., text, video, and audio signals) in this work. The core design in this framework is a Multimodal Adaptive Fusion module, which enables the effective fusion of diverse multimodal inputs, enhancing cross-modal alignment and improving overall generation quality. To train this unified model, we construct a large-scale, high-quality dataset, IF-caps, comprising over 7 million samples curated through a structured data annotation pipeline. This dataset provides comprehensive supervision for multimodal-conditioned audio generation. We benchmark AudioX against state-of-the-art methods across a wide range of tasks, finding that our model achieves superior performance, especially in text-to-audio and text-to-music generation. These results demonstrate our method is capable of audio generation under multimodal control signals, showing powerful instruction-following potential. The code and datasets will be available at https://zeyuet.github.io/AudioX/.
[846] TriniMark: A Robust Generative Speech Watermarking Method for Trinity-Level Traceability
Yue Li, Weizhi Liu, Kaiqing Lin, Dongdong Lin, Kassem Kallas
Main category: cs.MM
TL;DR: TriniMark is a diffusion-based generative speech watermarking framework that provides trinity-level traceability for content, model, and user attribution while maintaining speech quality and robustness.
Details
Motivation: Diffusion-based speech generation has achieved high fidelity, increasing misuse risks. Existing watermarking methods are mainly for GAN-based pipelines, and diffusion-based speech watermarking is underexplored. Prior work focuses on content-level provenance but lacks support for model-level and user-level attribution.Method: Uses lightweight encoder to embed watermark bits into time-domain speech features, temporal-aware gated convolutional decoder for bit recovery, waveform-guided fine-tuning to transfer watermarking capability into diffusion models, and variable-watermark training for scalable user-level traceability.
Result: Maintains speech quality while improving robustness to common signal-processing attacks, supports high-capacity watermarking for large-scale traceability, and enables trinity-level traceability (content, model, user).
Conclusion: TriniMark provides comprehensive traceability for diffusion-based speech generation, addressing security concerns while maintaining generation quality and robustness against attacks.
Abstract: Diffusion-based speech generation has achieved remarkable fidelity, increasing the risk of misuse and unauthorized redistribution. However, most existing generative speech watermarking methods are developed for GAN-based pipelines, and watermarking for diffusion-based speech generation remains comparatively underexplored. In addition, prior work often focuses on content-level provenance, while support for model-level and user-level attribution is less mature. We propose \textbf{TriniMark}, a diffusion-based generative speech watermarking framework that targets trinity-level traceability, i.e., the ability to associate a generated speech sample with (i) the embedded watermark message (content-level provenance), (ii) the source generative model (model-level attribution), and (iii) the end user who requested generation (user-level traceability). TriniMark uses a lightweight encoder to embed watermark bits into time-domain speech features and reconstruct the waveform, and a temporal-aware gated convolutional decoder for reliable bit recovery. We further introduce a waveform-guided fine-tuning strategy to transfer watermarking capability into a diffusion model. Finally, we incorporate variable-watermark training so that a single trained model can embed different watermark messages at inference time, enabling scalable user-level traceability. Experiments on speech datasets indicate that TriniMark maintains speech quality while improving robustness to common single and compound signal-processing attacks, and it supports high-capacity watermarking for large-scale traceability.
eess.AS
[847] ELEAT-SAGA: Early & Late Integration with Evading Alternating Training for Spoof-Robust Speaker Verification
Amro Asali, Yehuda Ben-Shimol, Itshak Lapidot
Main category: eess.AS
TL;DR: Proposes SASV-SAGA: a spoofing-robust speaker verification system using score-aware gated attention to dynamically modulate speaker embeddings based on countermeasure scores, achieving state-of-the-art performance on ASVspoof datasets.
Details
Motivation: Current automatic speaker verification systems are vulnerable to both zero-effort impostor attacks and sophisticated spoofing techniques like voice conversion and text-to-speech. There's a need for systems that can robustly verify speakers while detecting spoofing attempts.Method: Introduces SASV-SAGA architecture with score-aware gated attention (SAGA) that dynamically modulates speaker embeddings based on countermeasure scores. Uses pre-trained ECAPA-TDNN for speaker embeddings and AASIST for countermeasure scores. Explores early, late, and full integration strategies, plus alternating training for multi-module (ATMM) and evading alternating training (EAT).
Result: Achieves SASV-EER of 1.22% and min a-DCF of 0.0304 on ASVspoof 2019 evaluation set, showing significant improvements over baselines. Demonstrates effectiveness on both ASVspoof 2019 LA and Spoofceleb datasets.
Conclusion: Score-aware attention mechanisms and alternating training strategies effectively enhance robustness of spoofing-aware speaker verification systems, providing strong defense against both impostor and spoofing attacks.
Abstract: Spoofing-robust automatic speaker verification (SASV) seeks to build automatic speaker verification systems that are robust against both zero-effort impostor attacks and sophisticated spoofing techniques such as voice conversion (VC) and text-to-speech (TTS). In this work, we propose a novel SASV architecture that introduces score-aware gated attention (SAGA), SASV-SAGA, enabling dynamic modulation of speaker embeddings based on countermeasure (CM) scores. By integrating speaker embeddings and CM scores from pre-trained ECAPA-TDNN and AASIST models respectively, we explore several integration strategies including early, late, and full integration. We further introduce alternating training for multi-module (ATMM) and a refined variant, evading alternating training (EAT). Experimental results on the ASVspoof 2019 Logical Access (LA) and Spoofceleb datasets demonstrate significant improvements over baselines, achieving a spoofing aware speaker verification equal error rate (SASV-EER) of 1.22% and minimum normalized agnostic detection cost function (min a-DCF) of 0.0304 on the ASVspoof 2019 evaluation set. These results confirm the effectiveness of score-aware attention mechanisms and alternating training strategies in enhancing the robustness of SASV systems.
[848] CLAP-Based Automatic Word Naming Recognition in Post-Stroke Aphasia
Yacouba Kaloga, Marina Laganaro, Ina Kodrasi
Main category: eess.AS
TL;DR: CLAP-based approach for word-naming recognition in aphasia patients using audio-text alignment in shared embedding space
Details
Motivation: Conventional word-naming recognition systems fail with post-stroke aphasia patients due to disfluencies and mispronunciations, limiting automated assessment capabilitiesMethod: Uses Contrastive Language-Audio Pretraining (CLAP) to treat word-naming as audio-text matching problem, projecting speech and text into shared embedding space for alignment
Result: Achieves up to 90% accuracy on French post-stroke aphasia datasets, outperforming classification-based and ASR-based baselines
Conclusion: CLAP-based approach effectively addresses challenges in recognizing words from aphasia patients by leveraging multimodal audio-text alignment
Abstract: Conventional automatic word-naming recognition systems struggle to recognize words from post-stroke patients with aphasia because of disfluencies and mispronunciations, limiting reliable automated assessment in this population. In this paper, we propose a Contrastive Language-Audio Pretraining (CLAP) based approach for automatic word-naming recognition to address this challenge by leveraging text-audio alignment. Our approach treats word-naming recognition as an audio-text matching problem, projecting speech signals and textual prompts into a shared embedding space to identify intended words even in challenging recordings. Evaluated on two speech datasets of French post-stroke patients with aphasia, our approach achieves up to 90% accuracy, outperforming existing classification-based and automatic speech recognition-based baselines.
[849] LongAudio-RAG: Event-Grounded Question Answering over Multi-Hour Long Audio
Naveen Vakada, Kartik Hegde, Arvind Krishna Sridhar, Yinyi Guo, Erik Visser
Main category: eess.AS
TL;DR: LA-RAG is a hybrid framework for long-audio question answering that grounds LLM outputs in retrieved acoustic event detections stored in SQL, enabling precise temporal grounding with minimal hallucination.
Details
Motivation: Reviewing multi-hour audio recordings is impractical, motivating systems that can answer natural-language queries about long audio with precise temporal grounding and minimal hallucination. Existing audio-language models struggle with long-audio QA due to context-length limitations.Method: Proposes LongAudio-RAG (LA-RAG), a hybrid framework that converts multi-hour audio streams into structured event records stored in SQL database. At inference, system resolves natural-language time references, classifies intent, retrieves relevant events, and generates answers using constrained evidence. Deployed in hybrid edge-cloud environment with audio grounding model on IoT hardware and LLM on GPU server.
Result: Structured, event-level retrieval significantly improves accuracy compared to vanilla RAG or text-to-SQL approaches. System enables low-latency event extraction at edge and high-quality language reasoning in cloud.
Conclusion: LA-RAG provides effective solution for long-audio question answering by grounding LLM outputs in retrieved acoustic event detections, overcoming context-length limitations through structured event retrieval and hybrid edge-cloud deployment.
Abstract: Long-duration audio is increasingly common in industrial and consumer settings, yet reviewing multi-hour recordings is impractical, motivating systems that answer natural-language queries with precise temporal grounding and minimal hallucination. Existing audio-language models show promise, but long-audio question answering remains difficult due to context-length limits. We introduce LongAudio-RAG (LA-RAG), a hybrid framework that grounds Large Language Model (LLM) outputs in retrieved, timestamped acoustic event detections rather than raw audio. Multi-hour streams are converted into structured event records stored in an SQL database, and at inference time the system resolves natural-language time references, classifies intent, retrieves only the relevant events, and generates answers using this constrained evidence. To evaluate performance, we construct a synthetic long-audio benchmark by concatenating recordings with preserved timestamps and generating template-based question-answer pairs for detection, counting, and summarization tasks. Finally, we demonstrate the practicality of our approach by deploying it in a hybrid edge-cloud environment, where the audio grounding model runs on-device on IoT-class hardware while the LLM is hosted on a GPU-backed server. This architecture enables low-latency event extraction at the edge and high-quality language reasoning in the cloud. Experiments show that structured, event-level retrieval significantly improves accuracy compared to vanilla Retrieval-Augmented Generation (RAG) or text-to-SQL approaches.
[850] Data Augmentation for Pathological Speech Enhancement
Mingchi Hou, Enno Hermann, Ina Kodrasi
Main category: eess.AS
TL;DR: Systematic investigation of data augmentation strategies for speech enhancement models on pathological speech, finding noise augmentation most effective, transformative augmentations moderately helpful, and generative augmentation limited or harmful.
Details
Motivation: Speech enhancement models perform poorly on pathological speech due to atypical acoustic characteristics and limited data availability, creating a need for effective data augmentation strategies to bridge this performance gap.Method: Evaluated three categories of data augmentation (transformative, generative, and noise augmentation) on both predictive and generative speech enhancement models using objective SE metrics on pathological speech data.
Result: Noise augmentation consistently delivered the largest and most robust performance gains, transformative augmentations provided moderate improvements, while generative augmentation yielded limited benefits and could harm performance as synthetic data increased. DA was more beneficial for predictive SE models than generative ones.
Conclusion: While data augmentation improves speech enhancement for pathological speakers, a performance gap between neurotypical and pathological speech persists, highlighting the need for targeted DA strategies specifically designed for pathological speech characteristics.
Abstract: The performance of state-of-the-art speech enhancement (SE) models considerably degrades for pathological speech due to atypical acoustic characteristics and limited data availability. This paper systematically investigates data augmentation (DA) strategies to improve SE performance for pathological speakers, evaluating both predictive and generative SE models. We examine three DA categories, i.e., transformative, generative, and noise augmentation, assessing their impact with objective SE metrics. Experimental results show that noise augmentation consistently delivers the largest and most robust gains, transformative augmentations provide moderate improvements, while generative augmentation yields limited benefits and can harm performance as the amount of synthetic data increases. Furthermore, we show that the effectiveness of DA varies depending on the SE model, with DA being more beneficial for predictive SE models. While our results demonstrate that DA improves SE performance for pathological speakers, a performance gap between neurotypical and pathological speech persists, highlighting the need for future research on targeted DA strategies for pathological speech.
[851] Disentangling Pitch and Creak for Speaker Identity Preservation in Speech Synthesis
Frederik Rautenberg, Jana Wiechmann, Petra Wagner, Reinhold Haeb-Umbach
Main category: eess.AS
TL;DR: A system that modifies vocal creak quality while preserving speaker identity using conditional continuous normalizing flows for pitch-creak disentanglement
Details
Motivation: To develop a system that can modify perceptual voice quality (specifically creak) while maintaining speaker identity, addressing the challenge that creak probability is typically correlated with low pitch but this correlation doesn't always hold across all situationsMethod: Uses a speech synthesis system with a speaker manipulation block based on conditional continuous normalizing flow, augmented training dataset to disentangle pitch from creak
Result: Greatly improved speaker verification performance across a range of creak manipulation strengths, demonstrating successful disentanglement of pitch from creak while preserving speaker identity
Conclusion: The system successfully modifies vocal creak while preserving speaker identity through effective pitch-creak disentanglement using conditional continuous normalizing flows
Abstract: We introduce a system capable of faithfully modifying the perceptual voice quality of creak while preserving the speaker’s perceived identity. While it is well known that high creak probability is typically correlated with low pitch, it is important to note that this is a property observed on a population of speakers but does not necessarily hold across all situations. Disentanglement of pitch from creak is achieved by augmentation of the training dataset of a speech synthesis system with a speaker manipulation block based on conditional continuous normalizing flow. The experiments show greatly improved speaker verification performance over a range of creak manipulation strengths.
[852] SA-SSL-MOS: Self-supervised Learning MOS Prediction with Spectral Augmentation for Generalized Multi-Rate Speech Assessment
Fengyuan Cao, Xinyu Liang, Fredrik Cumlin, Victor Ungureanu, Chandan K. A. Reddy, Christian Schuldt, Saikat Chatterjee
Main category: eess.AS
TL;DR: Proposes a spectrogram-augmented SSL method with parallel-branch architecture to incorporate high-frequency features (up to 48kHz) for multi-rate speech quality assessment, using two-step training for better generalization with limited data.
Details
Motivation: Current SSL models for speech quality assessment are limited because they're pretrained on 16kHz speech, discarding high-frequency information crucial for multi-rate speech with varying sampling frequencies (16-48kHz). There's also limited availability of MOS-labeled multi-rate training datasets.Method: Proposes a spectrogram-augmented SSL method with parallel-branch architecture: one branch uses SSL features from 16kHz speech, another branch extracts high-frequency features from spectrograms of higher sampling rates. Uses two-step training: first pre-trains on large 48kHz dataset, then fine-tunes on smaller multi-rate dataset.
Result: Experimental results show that leveraging high-frequency information overlooked by SSL features is crucial for accurate multi-rate speech quality assessment. The proposed two-step training substantially improves generalization when multi-rate data is limited.
Conclusion: The proposed approach effectively addresses the limitations of SSL models for multi-rate speech quality assessment by incorporating high-frequency features through spectrogram augmentation and two-step training, improving performance especially with limited multi-rate data.
Abstract: Designing a speech quality assessment (SQA) system for estimating mean-opinion-score (MOS) of multi-rate speech with varying sampling frequency (16-48 kHz) is a challenging task. The challenge arises due to the limited availability of a MOS-labeled training dataset comprising multi-rate speech samples. While self-supervised learning (SSL) models have been widely adopted in SQA to boost performance, a key limitation is that they are pretrained on 16 kHz speech and therefore discard high-frequency information present in higher sampling rates. To address this issue, we propose a spectrogram-augmented SSL method that incorporates high-frequency features (up to 48 kHz sampling rate) through a parallel-branch architecture. We further introduce a two-step training scheme: the model is first pre-trained on a large 48 kHz dataset and then fine-tuned on a smaller multi-rate dataset. Experimental results show that leveraging high-frequency information overlooked by SSL features is crucial for accurate multi-rate SQA, and that the proposed two-step training substantially improves generalization when multi-rate data is limited.
[853] RosettaSpeech: Zero-Shot Speech-to-Speech Translation without Parallel Speech
Zhisheng Zheng, Xiaohang Sun, Tuan Dinh, Abhishek Yanamandra, Abhinav Jain, Zhu Liu, Sunil Hadap, Vimal Bhat, Manoj Aggarwal, Gerard Medioni, David Harwath
Main category: eess.AS
TL;DR: RosettaSpeech is a zero-shot speech-to-speech translation framework trained only on monolingual speech-text data with machine translation supervision, achieving SOTA results without parallel speech data.
Details
Motivation: Overcoming the critical data bottleneck in end-to-end speech-to-speech translation systems caused by scarcity of parallel speech-to-speech corpora.Method: Uses text as semantic bridge during training to synthesize translation targets, eliminating need for parallel speech pairs while maintaining end-to-end inference pipeline. Trained exclusively on monolingual speech-text data augmented by machine translation supervision.
Result: Achieves state-of-the-art zero-shot performance on CVSS-C benchmark: ASR-BLEU scores of 25.17 for German-to-English (+27% relative gain) and 29.86 for Spanish-to-English (+14%). Preserves source speaker’s voice without seeing paired speech data.
Conclusion: Offers scalable solution for extending high-quality speech-to-speech translation to “text-rich, speech-poor” languages by eliminating dependency on parallel speech data.
Abstract: End-to-end speech-to-speech translation (S2ST) systems typically struggle with a critical data bottleneck: the scarcity of parallel speech-to-speech corpora. To overcome this, we introduce RosettaSpeech, a novel zero-shot framework trained exclusively on monolingual speech-text data augmented by machine translation supervision. Unlike prior works that rely on complex cascaded pseudo-labeling, our approach strategically utilizes text as a semantic bridge during training to synthesize translation targets, thereby eliminating the need for parallel speech pairs while maintaining a direct, end-to-end inference pipeline. Empirical evaluations on the CVSS-C benchmark demonstrate that RosettaSpeech achieves state-of-the-art zero-shot performance, surpassing leading baselines by significant margins - achieving ASR-BLEU scores of 25.17 for German-to-English (+27% relative gain) and 29.86 for Spanish-to-English (+14%). Crucially, our model effectively preserves the source speaker’s voice without ever seeing paired speech data. We further analyze the impact of data scaling and demonstrate the model’s capability in many-to-one translation, offering a scalable solution for extending high-quality S2ST to “text-rich, speech-poor” languages.
eess.IV
[854] Deep Learning CNN for Pneumonia Detection: Advancing Digital Health in Society 5.0
Hadi Almohab
Main category: eess.IV
TL;DR: CNN-based deep learning model for automated pneumonia detection from chest X-ray images with 91.67% accuracy
Details
Motivation: Pneumonia is a serious global health problem with high morbidity and mortality, especially in areas with limited diagnostic tools and healthcare resources. There's a need for automated, reliable diagnostic aids.Method: Developed a Convolutional Neural Network (CNN) trained on labeled chest X-ray datasets with preprocessing techniques including normalization, data augmentation, and image quality enhancement to improve robustness and generalization.
Result: The optimized model achieves 91.67% accuracy, ROC-AUC of 0.96, and PR-AUC of 0.95, demonstrating strong performance in distinguishing pneumonia from normal chest X-ray images.
Conclusion: The CNN model has significant potential as a fast, consistent, and reliable diagnostic aid, supporting Society 5.0 by integrating artificial intelligence to improve healthcare services and public well-being.
Abstract: Pneumonia is a serious global health problem, contributing to high morbidity and mortality, especially in areas with limited diagnostic tools and healthcare resources. This study develops a Convolutional Neural Network (CNN) based on deep learning to automatically detect pneumonia from chest X-ray images. The method involves training the model on labeled datasets with preprocessing techniques such as normalization, data augmentation, and image quality enhancement to improve robustness and generalization. Testing results show that the optimized model achieves 91.67% accuracy, ROC-AUC of 0.96, and PR-AUC of 0.95, demonstrating strong performance in distinguishing pneumonia from normal images. In conclusion, this CNN model has significant potential as a fast, consistent, and reliable diagnostic aid, supporting Society 5.0 by integrating artificial intelligence to improve healthcare services and public well-being.
[855] Learning to Select Like Humans: Explainable Active Learning for Medical Imaging
Ifrat Ikhtear Uddin, Longwei Wang, Xiao Qin, Yang Zhou, KC Santosh
Main category: eess.IV
TL;DR: Explainability-guided active learning framework for medical imaging that combines classification uncertainty with attention misalignment to select samples that improve both performance and clinical interpretability.
Details
Motivation: Medical image analysis requires expensive expert annotation. Traditional active learning methods rely only on predictive uncertainty and ignore whether models learn clinically meaningful features, which is critical for clinical deployment.Method: Proposes a dual-criterion selection strategy: (1) classification uncertainty to identify informative examples, and (2) attention misalignment between Grad-CAM attention maps and radiologist-defined ROIs measured using Dice similarity. This approach integrates spatial attention alignment into sample acquisition.
Result: Using only 570 strategically selected samples, the approach outperforms random sampling across three medical imaging datasets: 77.22% accuracy on BraTS (MRI brain tumors), 52.37% on VinDr-CXR (chest X-rays), and 52.66% on SIIM-COVID-19 (chest X-rays). Grad-CAM visualizations confirm models focus on diagnostically relevant regions.
Conclusion: Incorporating explanation guidance into active learning sample acquisition yields superior data efficiency while maintaining clinical interpretability, demonstrating that models trained with this dual-criterion selection focus on diagnostically relevant features.
Abstract: Medical image analysis requires substantial labeled data for model training, yet expert annotation is expensive and time-consuming. Active learning (AL) addresses this challenge by strategically selecting the most informative samples for the annotation purpose, but traditional methods solely rely on predictive uncertainty while ignoring whether models learn from clinically meaningful features a critical requirement for clinical deployment. We propose an explainability-guided active learning framework that integrates spatial attention alignment into a sample acquisition process. Our approach advocates for a dual-criterion selection strategy combining: (i) classification uncertainty to identify informative examples, and (ii) attention misalignment with radiologist-defined regions-of-interest (ROIs) to target samples where the model focuses on incorrect features. By measuring misalignment between Grad-CAM attention maps and expert annotations using \emph{Dice similarity}, our acquisition function judiciously identifies samples that enhance both predictive performance and spatial interpretability. We evaluate the framework using three expert-annotated medical imaging datasets, namely, BraTS (MRI brain tumors), VinDr-CXR (chest X-rays), and SIIM-COVID-19 (chest X-rays). Using only 570 strategically selected samples, our explainability-guided approach consistently outperforms random sampling across all the datasets, achieving 77.22% accuracy on BraTS, 52.37% on VinDr-CXR, and 52.66% on SIIM-COVID. Grad-CAM visualizations confirm that the models trained by our dual-criterion selection focus on diagnostically relevant regions, demonstrating that incorporating explanation guidance into sample acquisition yields superior data efficiency while maintaining clinical interpretability.
[856] FUTON: Fourier Tensor Network for Implicit Neural Representations
Pooya Ashtari, Pourya Behmandpoor, Nikos Deligiannis, Aleksandra Pizurica
Main category: eess.IV
TL;DR: FUTON is a novel implicit neural representation using Fourier tensor networks that outperforms MLP-based INRs in speed and performance for signal representation and inverse problems.
Details
Motivation: MLP-based implicit neural representations (INRs) suffer from slow convergence, overfitting to noise, and poor extrapolation. The authors aim to develop a more efficient and effective INR architecture.Method: FUTON models signals as generalized Fourier series with coefficients parameterized by low-rank tensor decomposition. It combines Fourier bases for smoothness/periodicity with low-rank parameterization for spectral structure.
Result: FUTON outperforms state-of-the-art MLP-based INRs on image/volume representation while training 2-5× faster. It also generalizes better and converges faster on inverse problems like denoising and super-resolution.
Conclusion: FUTON provides an efficient alternative to MLP-based INRs with theoretical guarantees and practical advantages for signal representation and inverse problems.
Abstract: Implicit neural representations (INRs) have emerged as powerful tools for encoding signals, yet dominant MLP-based designs often suffer from slow convergence, overfitting to noise, and poor extrapolation. We introduce FUTON (Fourier Tensor Network), which models signals as generalized Fourier series whose coefficients are parameterized by a low-rank tensor decomposition. FUTON implicitly expresses signals as weighted combinations of orthonormal, separable basis functions, combining complementary inductive biases: Fourier bases capture smoothness and periodicity, while the low-rank parameterization enforces low-dimensional spectral structure. We provide theoretical guarantees through a universal approximation theorem and derive an inference algorithm with complexity linear in the spectral resolution and the input dimension. On image and volume representation, FUTON consistently outperforms state-of-the-art MLP-based INRs while training 2–5$\times$ faster. On inverse problems such as image denoising and super-resolution, FUTON generalizes better and converges faster.
[857] A real-time UAS hyperspectral anomaly detection system
Thomas P. Watson, Kevin McKenzie, Joseph Conroy, Eddie L. Jacobs
Main category: eess.IV
TL;DR: Real-time anomaly detection for hyperspectral images on UAVs with wireless transmission and interactive ground station display
Details
Motivation: Current hyperspectral anomaly detection requires post-processing, delaying insights. Need real-time detection and transmission for immediate operator analysis.Method: Deploy anomaly detection algorithm on UAV with push-broom hyperspectral sensor, use fast georectification, transmit concise anomaly data wirelessly to ground station for interactive visualization.
Result: Demonstrated complete end-to-end real-time solution from data capture to ground station interaction using low-cost components.
Conclusion: Successfully implemented real-time hyperspectral anomaly detection system enabling immediate operator insight without post-processing delays.
Abstract: Detecting anomalies in hyperspectral image data, i.e. regions which are spectrally distinct from the image background, is a common task in hyperspectral imaging. Such regions may represent interesting objects to human operators, but obtaining results often requires post-processing of captured data, delaying insight. To address this limitation, we apply an anomaly detection algorithm to a visible and near-infrared (VNIR) push-broom hyperspectral image sensor in real time onboard a small uncrewed aerial system (UAS), exploring how UAS limitations affect the algorithm. As the generated anomaly information is much more concise than the raw hyperspectral data, it can feasibly be transmitted wirelessly. To detection, we couple an innovative and fast georectification algorithm that enables anomalous areas to be interactively investigated and characterized immediately by a human operator receiving the anomaly data at a ground station. Using these elements, we demonstrate a novel and complete end-to-end solution from data capture and preparation, through anomaly detection and transmission, to ground station display and interaction, all in real time and with relatively low cost components.
[858] Frequency-Enhanced Hilbert Scanning Mamba for Short-Term Arctic Sea Ice Concentration Prediction
Feng Gao, Zheng Gong, Wenli Liu, Yanhai Gan, Zhuoran Zheng, Junyu Dong, Qian Du
Main category: eess.IV
TL;DR: FH-Mamba improves Arctic sea ice concentration prediction using 3D Hilbert scanning for spatiotemporal correlation and wavelet transforms for high-frequency detail enhancement.
Details
Motivation: Vanilla Mamba models struggle with temporal correlations and boundary details in Arctic sea ice concentration prediction, requiring better spatiotemporal modeling and detail preservation.Method: Proposes Frequency-enhanced Hilbert scanning Mamba Framework (FH-Mamba) with 3D Hilbert scan mechanism for locality-preserving spatiotemporal traversal, wavelet transform for high-frequency detail amplification, and Hybrid Shuffle Attention module for adaptive feature aggregation.
Result: FH-Mamba achieves superior prediction performance on OSI-450a1 and AMSR2 datasets compared to state-of-the-art baselines, confirming effectiveness in temporal consistency and edge reconstruction.
Conclusion: The proposed Hilbert scanning and frequency-aware attention effectively improve both temporal consistency and edge reconstruction for Arctic SIC forecasting.
Abstract: While Mamba models offer efficient sequence modeling, vanilla versions struggle with temporal correlations and boundary details in Arctic sea ice concentration (SIC) prediction. To address these limitations, we propose Frequency-enhanced Hilbert scanning Mamba Framework (FH-Mamba) for short-term Arctic SIC prediction. Specifically, we introduce a 3D Hilbert scan mechanism that traverses the 3D spatiotemporal grid along a locality-preserving path, ensuring that adjacent indices in the flattened sequence correspond to neighboring voxels in both spatial and temporal dimensions. Additionally, we incorporate wavelet transform to amplify high-frequency details and we also design a Hybrid Shuffle Attention module to adaptively aggregate sequence and frequency features. Experiments conducted on the OSI-450a1 and AMSR2 datasets demonstrate that our FH-Mamba achieves superior prediction performance compared with state-of-the-art baselines. The results confirm the effectiveness of Hilbert scanning and frequency-aware attention in improving both temporal consistency and edge reconstruction for Arctic SIC forecasting. Our codes are publicly available at https://github.com/oucailab/FH-Mamba.
[859] NeuroMambaLLM: Dynamic Graph Learning of fMRI Functional Connectivity in Autistic Brains Using Mamba and Language Model Reasoning
Yasaman Torabi, Parsa Razmara, Hamed Ajorlou, Bardia Baraeinejad
Main category: eess.IV
TL;DR: NeuroMambaLLM integrates dynamic latent graph learning and selective state-space temporal modeling with frozen LLMs for fMRI analysis, enabling both diagnostic classification and language-based clinical report generation.
Details
Motivation: Current fMRI analysis methods rely on static functional connectivity representations that obscure transient neural dynamics critical for neurodevelopmental disorders like autism. While LLMs have strong semantic reasoning capabilities, their integration with graph-based brain connectivity models remains limited, and state-space approaches like Mamba are typically used as standalone feature extractors without high-level reasoning.Method: End-to-end framework that learns functional connectivity dynamically from raw BOLD time series using adaptive latent connectivity instead of fixed correlation graphs. Combines dynamic latent graph learning with selective state-space temporal modeling (Mamba) to suppress motion artifacts and capture long-range dependencies. The resulting dynamic brain representations are projected into the embedding space of a frozen LLM, with lightweight LoRA modules trained for parameter-efficient alignment.
Result: The framework enables LLMs to perform both diagnostic classification and language-based reasoning, allowing analysis of dynamic fMRI patterns and generation of clinically meaningful textual reports. The approach addresses limitations of static connectivity representations and integrates temporal modeling with high-level semantic reasoning.
Conclusion: NeuroMambaLLM successfully bridges the gap between dynamic brain connectivity modeling and LLM-based reasoning, providing a unified framework for neuroimaging analysis that combines temporal dynamics modeling with clinical report generation capabilities.
Abstract: Large Language Models (LLMs) have demonstrated strong semantic reasoning across multimodal domains. However, their integration with graph-based models of brain connectivity remains limited. In addition, most existing fMRI analysis methods rely on static Functional Connectivity (FC) representations, which obscure transient neural dynamics critical for neurodevelopmental disorders such as autism. Recent state-space approaches, including Mamba, model temporal structure efficiently, but are typically used as standalone feature extractors without explicit high-level reasoning. We propose NeuroMambaLLM, an end-to-end framework that integrates dynamic latent graph learning and selective state-space temporal modelling with LLMs. The proposed method learns the functional connectivity dynamically from raw Blood-Oxygen-Level-Dependent (BOLD) time series, replacing fixed correlation graphs with adaptive latent connectivity while suppressing motion-related artifacts and capturing long-range temporal dependencies. The resulting dynamic brain representations are projected into the embedding space of an LLM model, where the base language model remains frozen and lightweight low-rank adaptation (LoRA) modules are trained for parameter-efficient alignment. This design enables the LLM to perform both diagnostic classification and language-based reasoning, allowing it to analyze dynamic fMRI patterns and generate clinically meaningful textual reports.
[860] A Deep Convolutional Network to Extract Real-Time Landmarks for UAV Navigation
Osman Tokluoglu, Mustafa Ozturk
Main category: eess.IV
TL;DR: A deep learning approach using convolutional networks for visual landmark extraction from UAV camera images to enable navigation in GNSS-denied environments.
Details
Motivation: In GNSS-denied environments where satellite signals are unavailable due to interference or attenuation, UAVs need alternative navigation methods. Visual landmark extraction from onboard camera images provides a solution for reliable navigation without GNSS support.Method: Proposes a convolution-based deep learning approach for extracting appropriate visual landmarks from images captured by UAV cameras. The method processes visual data to identify reliable landmarks that can be used for navigation.
Result: The effectiveness of the proposed convolution-based deep learning approach for landmark extraction is examined, though specific performance metrics are not provided in the abstract.
Conclusion: Visual landmark extraction using deep learning offers a viable solution for UAV navigation in GNSS-denied environments, addressing the limitations of traditional satellite-based positioning systems.
Abstract: Recent advances in satellite and communication technologies have significantly improved geographical information and monitoring systems. Global System for Mobile Communications (GSM) and Global Navigation Satellite System (GNSS) technologies, which rely on electromagnetic signals transmitted from satellites and base stations, have long been utilized for geolocation applications. However, signal attenuation due to environmental conditions or intentional interference such as jamming may lead to severe degradation or complete loss of positioning capability. In such GNSS-denied environments, landmark extraction becomes critical for the navigation of unmanned aerial vehicles (UAVs) used in monitoring applications. By processing images captured from onboard UAV cameras, reliable visual landmarks can be identified to enable navigation without GNSS support. In this study, a convolution-based deep learning approach is proposed for the extraction of appropriate landmarks, and its effectiveness is examined.
[861] Scan-Adaptive Dynamic MRI Undersampling Using a Dictionary of Efficiently Learned Patterns
Siddhant Gautam, Angqi Li, Prachi P. Agarwal, Anil K. Attili, Jeffrey A. Fessler, Nicole Seiberlich, Saiprasad Ravishankar
Main category: eess.IV
TL;DR: Learning-based framework designs scan-adaptive Cartesian undersampling masks for dynamic cardiac MRI acceleration, improving reconstruction quality across multiple acceleration factors.
Details
Motivation: Cardiac MRI suffers from long acquisition times causing patient discomfort and motion artifacts; need for efficient acceleration while preserving diagnostic quality.Method: Develop learning-based framework to optimize scan- or slice-adaptive Cartesian undersampling masks using fully sampled training data; at inference, nearest-neighbor search in low-frequency k-space selects optimized mask from learned dictionary.
Result: Learned sampling improves reconstruction quality across multiple acceleration factors: 2-3 dB PSNR gains, reduced NMSE, improved SSIM, and higher radiologist ratings on public and in-house datasets.
Conclusion: Scan-adaptive sampling framework enables faster, higher-quality dynamic cardiac MRI by adapting k-space sampling to individual scans.
Abstract: Cardiac MRI is limited by long acquisition times, which can lead to patient discomfort and motion artifacts. We aim to accelerate Cartesian dynamic cardiac MRI by learning efficient, scan-adaptive undersampling patterns that preserve diagnostic image quality. We develop a learning-based framework for designing scan- or slice-adaptive Cartesian undersampling masks tailored to dynamic cardiac MRI. Undersampling patterns are optimized using fully sampled training dynamic time-series data. At inference time, a nearest-neighbor search in low-frequency $k$-space selects an optimized mask from a dictionary of learned patterns. Our learned sampling approach improves reconstruction quality across multiple acceleration factors on public and in-house cardiac MRI datasets, including PSNR gains of 2-3 dB, reduced NMSE, improved SSIM, and higher radiologist ratings. The proposed scan-adaptive sampling framework enables faster and higher-quality dynamic cardiac MRI by adapting $k$-space sampling to individual scans.
[862] Learnable Multi-level Discrete Wavelet Transforms for 3D Gaussian Splatting Frequency Modulation
Hung Nguyen, An Le, Truong Nguyen
Main category: eess.IV
TL;DR: Multi-level Discrete Wavelet Transform framework for 3D Gaussian Splatting that reduces Gaussian primitives while maintaining rendering quality through progressive frequency modulation.
Details
Motivation: 3D Gaussian Splatting suffers from excessive growth of Gaussian primitives during training, leading to high memory and storage costs. Existing methods like AutoOpti3DGS use 1-level DWT for frequency modulation but have limited depth and suffer from gradient competition issues.Method: Proposes a multi-level DWT framework that recursively decomposes low-frequency subbands to create deeper curriculum learning. Uses progressive coarser supervision during early training with simplified modulation using only a single scaling parameter instead of learning full 2-tap high-pass filters.
Result: Experimental results on standard benchmarks show further reduction in Gaussian counts while maintaining competitive rendering quality compared to existing methods.
Conclusion: Multi-level DWT frequency modulation effectively reduces Gaussian primitives in 3DGS through deeper curriculum learning and simplified parameterization, addressing memory and storage issues without compromising rendering quality.
Abstract: 3D Gaussian Splatting (3DGS) has emerged as a powerful approach for novel view synthesis. However, the number of Gaussian primitives often grows substantially during training as finer scene details are reconstructed, leading to increased memory and storage costs. Recent coarse-to-fine strategies regulate Gaussian growth by modulating the frequency content of the ground-truth images. In particular, AutoOpti3DGS employs the learnable Discrete Wavelet Transform (DWT) to enable data-adaptive frequency modulation. Nevertheless, its modulation depth is limited by the 1-level DWT, and jointly optimizing wavelet regularization with 3D reconstruction introduces gradient competition that promotes excessive Gaussian densification. In this paper, we propose a multi-level DWT-based frequency modulation framework for 3DGS. By recursively decomposing the low-frequency subband, we construct a deeper curriculum that provides progressively coarser supervision during early training, consistently reducing Gaussian counts. Furthermore, we show that the modulation can be performed using only a single scaling parameter, rather than learning the full 2-tap high-pass filter. Experimental results on standard benchmarks demonstrate that our method further reduces Gaussian counts while maintaining competitive rendering quality.
[863] Deep Image Prior for Computed Tomography Reconstruction
Simon Arridge, Riccardo Barbano, Alexander Denker, Zeljko Kereta
Main category: eess.IV
TL;DR: Deep Image Prior (DIP) framework for CT image reconstruction using unsupervised CNN bias without large datasets, with strategies to mitigate overfitting and improve computational efficiency.
Details
Motivation: To address limitations of conventional deep learning methods that require large supervised datasets for CT reconstruction, by leveraging the implicit bias of CNNs in an unsupervised setting with only single measurements.Method: Uses Deep Image Prior framework with convolutional neural networks operating unsupervisedly. Includes strategies like early stopping, explicit regularization, self-guided methods adapting network input, warm-start initialization, and stochastic optimization to reduce reconstruction time.
Result: Methods tested on real μCT measurements, allowing examination of trade-offs among different modifications and extensions for practical CT reconstruction.
Conclusion: DIP provides effective unsupervised CT reconstruction alternative to supervised deep learning methods, with various strategies available to optimize performance and computational efficiency.
Abstract: We present a comprehensive overview of the Deep Image Prior (DIP) framework and its applications to image reconstruction in computed tomography. Unlike conventional deep learning methods that rely on large, supervised datasets, the DIP exploits the implicit bias of convolutional neural networks and operates in a fully unsupervised setting, requiring only a single measurement, even in the presence of noise. We describe the standard DIP formulation, outline key algorithmic design choices, and review several strategies to mitigate overfitting, including early stopping, explicit regularisation, and self-guided methods that adapt the network input. In addition, we examine computational improvements such as warm-start and stochastic optimisation methods to reduce the reconstruction time. The discussed methods are tested on real $μ$CT measurements, which allows examination of trade-offs among the different modifications and extensions.
[864] CellINR: Implicitly Overcoming Photo-induced Artifacts in 4D Live Fluorescence Microscopy
Cunmin Zhao, Ziyuan Luo, Guoye Guan, Zelin Li, Yiming Ma, Zhongying Zhao, Renjie Wan
Main category: eess.IV
TL;DR: CellINR: An implicit neural representation framework for 4D live fluorescence microscopy that reduces photobleaching artifacts through blind convolution and structure amplification, enabling high-accuracy cellular structure reconstruction with minimal illumination.
Details
Motivation: 4D live fluorescence microscopy suffers from prolonged high-intensity illumination causing photobleaching and phototoxic effects, which generate artifacts and impair image continuity and detail recovery, limiting biological research.Method: CellINR uses case-specific optimization with implicit neural representation, employing blind convolution and structure amplification strategies to map 3D spatial coordinates into high frequency domain for precise modeling and artifact removal.
Result: CellINR significantly outperforms existing techniques in artifact removal and restoration of structural continuity, and provides the first paired 4D live cell imaging dataset for evaluating reconstruction performance.
Conclusion: The framework offers a solid foundation for subsequent quantitative analyses and biological research by enabling high-accuracy reconstruction of cellular structures while effectively distinguishing true signals from artifacts.
Abstract: 4D live fluorescence microscopy is often compromised by prolonged high intensity illumination which induces photobleaching and phototoxic effects that generate photo-induced artifacts and severely impair image continuity and detail recovery. To address this challenge, we propose the CellINR framework, a case-specific optimization approach based on implicit neural representation. The method employs blind convolution and structure amplification strategies to map 3D spatial coordinates into the high frequency domain, enabling precise modeling and high-accuracy reconstruction of cellular structures while effectively distinguishing true signals from artifacts. Experimental results demonstrate that CellINR significantly outperforms existing techniques in artifact removal and restoration of structural continuity, and for the first time, a paired 4D live cell imaging dataset is provided for evaluating reconstruction performance, thereby offering a solid foundation for subsequent quantitative analyses and biological research. The code and dataset will be public.